code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" import os import random import sys from . import cryptomath_module as cryptoMath # noqa: N812 from . import rabin_miller as rabinMiller # noqa: N812 def _snake_case ( ): print('Making key files...' ) make_key_files('rsa' , 1024 ) print('Key files generation successful.' ) def _snake_case ( snake_case__ : List[str] ): print('Generating prime p...' ) A = rabinMiller.generate_large_prime(snake_case__ ) print('Generating prime q...' ) A = rabinMiller.generate_large_prime(snake_case__ ) A = p * q print('Generating e that is relatively prime to (p - 1) * (q - 1)...' ) while True: A = random.randrange(2 ** (key_size - 1) , 2 ** (key_size) ) if cryptoMath.gcd(snake_case__ , (p - 1) * (q - 1) ) == 1: break print('Calculating d that is mod inverse of e...' ) A = cryptoMath.find_mod_inverse(snake_case__ , (p - 1) * (q - 1) ) A = (n, e) A = (n, d) return (public_key, private_key) def _snake_case ( snake_case__ : Tuple , snake_case__ : Optional[Any] ): if os.path.exists(F'{name}_pubkey.txt' ) or os.path.exists(F'{name}_privkey.txt' ): print('\nWARNING:' ) print( F'\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n' 'Use a different name or delete these files and re-run this program.' ) sys.exit() A , A = generate_key(snake_case__ ) print(F'\nWriting public key to file {name}_pubkey.txt...' ) with open(F'{name}_pubkey.txt' , 'w' ) as out_file: out_file.write(F'{key_size},{public_key[0]},{public_key[1]}' ) print(F'Writing private key to file {name}_privkey.txt...' ) with open(F'{name}_privkey.txt' , 'w' ) as out_file: out_file.write(F'{key_size},{private_key[0]},{private_key[1]}' ) if __name__ == "__main__": main()
74
"""simple docstring""" from collections import Counter from timeit import timeit def _lowerCamelCase( a = "" , ): return sum(c % 2 for c in Counter(input_str.replace(" " , "" ).lower() ).values() ) < 2 def _lowerCamelCase( a = "" ): if len(a ) == 0: return True __a = input_str.replace(" " , "" ).lower() # character_freq_dict: Stores the frequency of every character in the input string __a = {} for character in lower_case_input_str: __a = character_freq_dict.get(a , 0 ) + 1 __a = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def _lowerCamelCase( a = "" ): print("\nFor string = " , a , ":" ) print( "> can_string_be_rearranged_as_palindrome_counter()" , "\tans =" , can_string_be_rearranged_as_palindrome_counter(a ) , "\ttime =" , timeit( "z.can_string_be_rearranged_as_palindrome_counter(z.check_str)" , setup="import __main__ as z" , ) , "seconds" , ) print( "> can_string_be_rearranged_as_palindrome()" , "\tans =" , can_string_be_rearranged_as_palindrome(a ) , "\ttime =" , timeit( "z.can_string_be_rearranged_as_palindrome(z.check_str)" , setup="import __main__ as z" , ) , "seconds" , ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__:Tuple = input( """Enter string to determine if it can be rearranged as a palindrome or not: """ ).strip() benchmark(check_str) SCREAMING_SNAKE_CASE__:Dict = can_string_be_rearranged_as_palindrome_counter(check_str) print(F'''{check_str} can {'' if status else 'not '}be rearranged as a palindrome''')
261
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices a : str = logging.get_logger(__name__) a : List[str] = { """facebook/convnextv2-tiny-1k-224""": """https://huggingface.co/facebook/convnextv2-tiny-1k-224/resolve/main/config.json""", } class __UpperCAmelCase( snake_case_ , snake_case_ ): """simple docstring""" __lowerCamelCase = "convnextv2" def __init__( self , snake_case__=3 , snake_case__=4 , snake_case__=4 , snake_case__=None , snake_case__=None , snake_case__="gelu" , snake_case__=0.02 , snake_case__=1e-12 , snake_case__=0.0 , snake_case__=224 , snake_case__=None , snake_case__=None , **snake_case__ , ): '''simple docstring''' super().__init__(**snake_case__ ) lowercase__ : Union[str, Any]= num_channels lowercase__ : int= patch_size lowercase__ : int= num_stages lowercase__ : List[str]= [96, 192, 384, 768] if hidden_sizes is None else hidden_sizes lowercase__ : Any= [3, 3, 9, 3] if depths is None else depths lowercase__ : Optional[int]= hidden_act lowercase__ : Dict= initializer_range lowercase__ : Dict= layer_norm_eps lowercase__ : Any= drop_path_rate lowercase__ : Optional[Any]= image_size lowercase__ : Union[str, Any]= ["stem"] + [F'''stage{idx}''' for idx in range(1 , len(self.depths ) + 1 )] lowercase__, lowercase__ : Tuple= get_aligned_output_features_output_indices( out_features=snake_case__ , out_indices=snake_case__ , stage_names=self.stage_names )
353
"""simple docstring""" import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging a : Dict = logging.get_logger(__name__) a : Optional[Any] = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt"""} # See all LED models at https://huggingface.co/models?filter=LED a : List[str] = { """vocab_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json""", }, """merges_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt""", }, """tokenizer_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json""", }, } a : Dict = { """allenai/led-base-16384""": 16384, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def lowercase__() ->List[Any]: """simple docstring""" lowercase__ : Optional[int]= ( list(range(ord("!" ) , ord("~" ) + 1 ) ) + list(range(ord("¡" ) , ord("¬" ) + 1 ) ) + list(range(ord("®" ) , ord("ÿ" ) + 1 ) ) ) lowercase__ : str= bs[:] lowercase__ : List[str]= 0 for b in range(2**8 ): if b not in bs: bs.append(A ) cs.append(2**8 + n ) n += 1 lowercase__ : Union[str, Any]= [chr(A ) for n in cs] return dict(zip(A , A ) ) def lowercase__(A ) ->str: """simple docstring""" lowercase__ : Optional[int]= set() lowercase__ : Optional[int]= word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase__ : Tuple= char return pairs class __UpperCAmelCase( SCREAMING_SNAKE_CASE__ ): """simple docstring""" __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = ["input_ids", "attention_mask"] def __init__( self , snake_case__ , snake_case__ , snake_case__="replace" , snake_case__="<s>" , snake_case__="</s>" , snake_case__="</s>" , snake_case__="<s>" , snake_case__="<unk>" , snake_case__="<pad>" , snake_case__="<mask>" , snake_case__=False , **snake_case__ , ): '''simple docstring''' lowercase__ : Dict= AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else bos_token lowercase__ : Dict= AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else eos_token lowercase__ : str= AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else sep_token lowercase__ : Optional[Any]= AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else cls_token lowercase__ : List[str]= AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else unk_token lowercase__ : Any= AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowercase__ : int= AddedToken(snake_case__ , lstrip=snake_case__ , rstrip=snake_case__ ) if isinstance(snake_case__ , snake_case__ ) else mask_token super().__init__( errors=snake_case__ , bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , sep_token=snake_case__ , cls_token=snake_case__ , pad_token=snake_case__ , mask_token=snake_case__ , add_prefix_space=snake_case__ , **snake_case__ , ) with open(snake_case__ , encoding="utf-8" ) as vocab_handle: lowercase__ : int= json.load(snake_case__ ) lowercase__ : List[str]= {v: k for k, v in self.encoder.items()} lowercase__ : Dict= errors # how to handle errors in decoding lowercase__ : Optional[int]= bytes_to_unicode() lowercase__ : str= {v: k for k, v in self.byte_encoder.items()} with open(snake_case__ , encoding="utf-8" ) as merges_handle: lowercase__ : Tuple= merges_handle.read().split("\n" )[1:-1] lowercase__ : List[str]= [tuple(merge.split() ) for merge in bpe_merges] lowercase__ : int= dict(zip(snake_case__ , range(len(snake_case__ ) ) ) ) lowercase__ : Optional[int]= {} lowercase__ : Tuple= add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowercase__ : Optional[int]= re.compile(r"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def UpperCAmelCase_ ( self ): '''simple docstring''' return len(self.encoder ) def UpperCAmelCase_ ( self ): '''simple docstring''' return dict(self.encoder , **self.added_tokens_encoder ) def UpperCAmelCase_ ( self , snake_case__ ): '''simple docstring''' if token in self.cache: return self.cache[token] lowercase__ : Any= tuple(snake_case__ ) lowercase__ : Union[str, Any]= get_pairs(snake_case__ ) if not pairs: return token while True: lowercase__ : Union[str, Any]= min(snake_case__ , key=lambda snake_case__ : self.bpe_ranks.get(snake_case__ , float("inf" ) ) ) if bigram not in self.bpe_ranks: break lowercase__, lowercase__ : Dict= bigram lowercase__ : Optional[Any]= [] lowercase__ : Optional[int]= 0 while i < len(snake_case__ ): try: lowercase__ : Optional[int]= word.index(snake_case__ , snake_case__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase__ : List[Any]= j if word[i] == first and i < len(snake_case__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase__ : Optional[int]= tuple(snake_case__ ) lowercase__ : List[Any]= new_word if len(snake_case__ ) == 1: break else: lowercase__ : Tuple= get_pairs(snake_case__ ) lowercase__ : Tuple= " ".join(snake_case__ ) lowercase__ : List[str]= word return word def UpperCAmelCase_ ( self , snake_case__ ): '''simple docstring''' lowercase__ : Optional[int]= [] for token in re.findall(self.pat , snake_case__ ): lowercase__ : int= "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(snake_case__ ).split(" " ) ) return bpe_tokens def UpperCAmelCase_ ( self , snake_case__ ): '''simple docstring''' return self.encoder.get(snake_case__ , self.encoder.get(self.unk_token ) ) def UpperCAmelCase_ ( self , snake_case__ ): '''simple docstring''' return self.decoder.get(snake_case__ ) def UpperCAmelCase_ ( self , snake_case__ ): '''simple docstring''' lowercase__ : Union[str, Any]= "".join(snake_case__ ) lowercase__ : Optional[int]= bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" , errors=self.errors ) return text def UpperCAmelCase_ ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if not os.path.isdir(snake_case__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase__ : Dict= os.path.join( snake_case__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) lowercase__ : Tuple= os.path.join( snake_case__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(snake_case__ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=snake_case__ , ensure_ascii=snake_case__ ) + "\n" ) lowercase__ : Any= 0 with open(snake_case__ , "w" , encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda snake_case__ : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' " Please check that the tokenizer is not corrupted!" ) lowercase__ : Optional[Any]= token_index writer.write(" ".join(snake_case__ ) + "\n" ) index += 1 return vocab_file, merge_file def UpperCAmelCase_ ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ : Any= [self.cls_token_id] lowercase__ : Dict= [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCAmelCase_ ( self , snake_case__ , snake_case__ = None , snake_case__ = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case__ , token_ids_a=snake_case__ , already_has_special_tokens=snake_case__ ) if token_ids_a is None: return [1] + ([0] * len(snake_case__ )) + [1] return [1] + ([0] * len(snake_case__ )) + [1, 1] + ([0] * len(snake_case__ )) + [1] def UpperCAmelCase_ ( self , snake_case__ , snake_case__ = None ): '''simple docstring''' lowercase__ : Tuple= [self.sep_token_id] lowercase__ : Optional[int]= [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCAmelCase_ ( self , snake_case__ , snake_case__=False , **snake_case__ ): '''simple docstring''' lowercase__ : str= kwargs.pop("add_prefix_space" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(snake_case__ ) > 0 and not text[0].isspace()): lowercase__ : List[Any]= " " + text return (text, kwargs) def UpperCAmelCase_ ( self , snake_case__ , snake_case__ = None , snake_case__ = PaddingStrategy.DO_NOT_PAD , snake_case__ = None , snake_case__ = None , ): '''simple docstring''' lowercase__ : List[str]= super()._pad( encoded_inputs=snake_case__ , max_length=snake_case__ , padding_strategy=snake_case__ , pad_to_multiple_of=snake_case__ , return_attention_mask=snake_case__ , ) # Load from model defaults if return_attention_mask is None: lowercase__ : Tuple= "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowercase__ : Any= encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowercase__ : int= len(encoded_inputs["global_attention_mask"] ) != len(snake_case__ ) if needs_to_be_padded: lowercase__ : Dict= len(snake_case__ ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowercase__ : Tuple= ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": lowercase__ : Any= [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
150
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __A = logging.get_logger(__name__) __A = { "facebook/vit-mae-base": "https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json", # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowercase_ = "vit_mae" def __init__(self : int , UpperCAmelCase_ : List[Any]=768 , UpperCAmelCase_ : Optional[Any]=12 , UpperCAmelCase_ : Tuple=12 , UpperCAmelCase_ : int=3_072 , UpperCAmelCase_ : Union[str, Any]="gelu" , UpperCAmelCase_ : Union[str, Any]=0.0 , UpperCAmelCase_ : Tuple=0.0 , UpperCAmelCase_ : Optional[Any]=0.02 , UpperCAmelCase_ : List[Any]=1E-1_2 , UpperCAmelCase_ : Any=224 , UpperCAmelCase_ : List[str]=16 , UpperCAmelCase_ : Optional[int]=3 , UpperCAmelCase_ : List[str]=True , UpperCAmelCase_ : Dict=16 , UpperCAmelCase_ : str=512 , UpperCAmelCase_ : List[Any]=8 , UpperCAmelCase_ : List[str]=2_048 , UpperCAmelCase_ : Dict=0.75 , UpperCAmelCase_ : Tuple=False , **UpperCAmelCase_ : str , ) ->List[str]: '''simple docstring''' super().__init__(**UpperCAmelCase_) lowerCamelCase__: Union[str, Any] =hidden_size lowerCamelCase__: int =num_hidden_layers lowerCamelCase__: str =num_attention_heads lowerCamelCase__: int =intermediate_size lowerCamelCase__: str =hidden_act lowerCamelCase__: Optional[int] =hidden_dropout_prob lowerCamelCase__: int =attention_probs_dropout_prob lowerCamelCase__: Dict =initializer_range lowerCamelCase__: int =layer_norm_eps lowerCamelCase__: List[str] =image_size lowerCamelCase__: Optional[Any] =patch_size lowerCamelCase__: Any =num_channels lowerCamelCase__: Optional[Any] =qkv_bias lowerCamelCase__: List[str] =decoder_num_attention_heads lowerCamelCase__: List[Any] =decoder_hidden_size lowerCamelCase__: str =decoder_num_hidden_layers lowerCamelCase__: List[Any] =decoder_intermediate_size lowerCamelCase__: Optional[int] =mask_ratio lowerCamelCase__: Optional[int] =norm_pix_loss
10
"""simple docstring""" def lowercase ( lowerCAmelCase__ : str = "The quick brown fox jumps over the lazy dog" , ) -> bool: __a = set() # Replace all the whitespace in our sentence __a = input_str.replace(''' ''' , '''''' ) for alpha in input_str: if "a" <= alpha.lower() <= "z": frequency.add(alpha.lower() ) return len(lowerCAmelCase__ ) == 26 def lowercase ( lowerCAmelCase__ : str = "The quick brown fox jumps over the lazy dog" , ) -> bool: __a = [False] * 26 for char in input_str: if char.islower(): __a = True elif char.isupper(): __a = True return all(lowerCAmelCase__ ) def lowercase ( lowerCAmelCase__ : str = "The quick brown fox jumps over the lazy dog" , ) -> bool: return len({char for char in input_str.lower() if char.isalpha()} ) == 26 def lowercase ( ) -> None: from timeit import timeit __a = '''from __main__ import is_pangram, is_pangram_faster, is_pangram_fastest''' print(timeit('''is_pangram()''' , setup=lowerCAmelCase__ ) ) print(timeit('''is_pangram_faster()''' , setup=lowerCAmelCase__ ) ) print(timeit('''is_pangram_fastest()''' , setup=lowerCAmelCase__ ) ) # 5.348480500048026, 2.6477354579837993, 1.8470395830227062 # 5.036091582966037, 2.644472333951853, 1.8869528750656173 if __name__ == "__main__": import doctest doctest.testmod() benchmark()
45
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __A : Optional[Any] = {"configuration_swin": ["SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP", "SwinConfig", "SwinOnnxConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : str = [ "SWIN_PRETRAINED_MODEL_ARCHIVE_LIST", "SwinForImageClassification", "SwinForMaskedImageModeling", "SwinModel", "SwinPreTrainedModel", "SwinBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[str] = [ "TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST", "TFSwinForImageClassification", "TFSwinForMaskedImageModeling", "TFSwinModel", "TFSwinPreTrainedModel", ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys __A : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
89
'''simple docstring''' import argparse import os import sys from unittest.mock import patch import pytorch_lightning as pl import timeout_decorator import torch from distillation import SummarizationDistiller, distill_main from finetune import SummarizationModule, main from transformers import MarianMTModel from transformers.file_utils import cached_path from transformers.testing_utils import TestCasePlus, require_torch_gpu, slow from utils import load_json __A : Tuple = "sshleifer/mar_enro_6_3_student" class __snake_case ( _SCREAMING_SNAKE_CASE): """simple docstring""" def __lowercase ( self : List[Any] ) -> Optional[Any]: super().setUp() lowerCAmelCase_ : Any = cached_path( """https://cdn-datasets.huggingface.co/translation/wmt_en_ro-tr40k-va0.5k-te0.5k.tar.gz""" , extract_compressed_file=lowerCamelCase , ) lowerCAmelCase_ : Optional[Any] = F'{data_cached}/wmt_en_ro-tr40k-va0.5k-te0.5k' @slow @require_torch_gpu def __lowercase ( self : str ) -> str: MarianMTModel.from_pretrained(lowerCamelCase ) @slow @require_torch_gpu def __lowercase ( self : List[Any] ) -> Union[str, Any]: lowerCAmelCase_ : str = { """$MAX_LEN""": 64, """$BS""": 64, """$GAS""": 1, """$ENRO_DIR""": self.data_dir, """facebook/mbart-large-cc25""": MARIAN_MODEL, # "val_check_interval=0.25": "val_check_interval=1.0", """--learning_rate=3e-5""": """--learning_rate 3e-4""", """--num_train_epochs 6""": """--num_train_epochs 1""", } # Clean up bash script lowerCAmelCase_ : Dict = (self.test_file_dir / """train_mbart_cc25_enro.sh""").open().read().split("""finetune.py""" )[1].strip() lowerCAmelCase_ : Optional[int] = bash_script.replace("""\\\n""" , """""" ).strip().replace("""\"$@\"""" , """""" ) for k, v in env_vars_to_replace.items(): lowerCAmelCase_ : Optional[int] = bash_script.replace(lowerCamelCase , str(lowerCamelCase ) ) lowerCAmelCase_ : Optional[Any] = self.get_auto_remove_tmp_dir() # bash_script = bash_script.replace("--fp16 ", "") lowerCAmelCase_ : Tuple = F'\n --output_dir {output_dir}\n --tokenizer_name Helsinki-NLP/opus-mt-en-ro\n --sortish_sampler\n --do_predict\n --gpus 1\n --freeze_encoder\n --n_train 40000\n --n_val 500\n --n_test 500\n --fp16_opt_level O1\n --num_sanity_val_steps 0\n --eval_beams 2\n '.split() # XXX: args.gpus > 1 : handle multi_gpu in the future lowerCAmelCase_ : Tuple = ["""finetune.py"""] + bash_script.split() + args with patch.object(lowerCamelCase , """argv""" , lowerCamelCase ): lowerCAmelCase_ : Optional[Any] = argparse.ArgumentParser() lowerCAmelCase_ : Any = pl.Trainer.add_argparse_args(lowerCamelCase ) lowerCAmelCase_ : List[str] = SummarizationModule.add_model_specific_args(lowerCamelCase , os.getcwd() ) lowerCAmelCase_ : Tuple = parser.parse_args() lowerCAmelCase_ : Dict = main(lowerCamelCase ) # Check metrics lowerCAmelCase_ : int = load_json(model.metrics_save_path ) lowerCAmelCase_ : Optional[Any] = metrics["""val"""][0] lowerCAmelCase_ : Tuple = metrics["""val"""][-1] self.assertEqual(len(metrics["""val"""] ) , (args.max_epochs / args.val_check_interval) ) assert isinstance(last_step_stats[F'val_avg_{model.val_metric}'] , lowerCamelCase ) self.assertGreater(last_step_stats["""val_avg_gen_time"""] , 0.01 ) # model hanging on generate. Maybe bad config was saved. (XXX: old comment/assert?) self.assertLessEqual(last_step_stats["""val_avg_gen_time"""] , 1.0 ) # test learning requirements: # 1. BLEU improves over the course of training by more than 2 pts self.assertGreater(last_step_stats["""val_avg_bleu"""] - first_step_stats["""val_avg_bleu"""] , 2 ) # 2. BLEU finishes above 17 self.assertGreater(last_step_stats["""val_avg_bleu"""] , 17 ) # 3. test BLEU and val BLEU within ~1.1 pt. self.assertLess(abs(metrics["""val"""][-1]["""val_avg_bleu"""] - metrics["""test"""][-1]["""test_avg_bleu"""] ) , 1.1 ) # check lightning ckpt can be loaded and has a reasonable statedict lowerCAmelCase_ : Union[str, Any] = os.listdir(lowerCamelCase ) lowerCAmelCase_ : Any = [x for x in contents if x.endswith(""".ckpt""" )][0] lowerCAmelCase_ : Union[str, Any] = os.path.join(args.output_dir , lowerCamelCase ) lowerCAmelCase_ : int = torch.load(lowerCamelCase , map_location="""cpu""" ) lowerCAmelCase_ : List[str] = """model.model.decoder.layers.0.encoder_attn_layer_norm.weight""" assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: lowerCAmelCase_ : List[str] = {os.path.basename(lowerCamelCase ) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics["""test"""] ) == 1 class __snake_case ( _SCREAMING_SNAKE_CASE): """simple docstring""" @timeout_decorator.timeout(6_00 ) @slow @require_torch_gpu def __lowercase ( self : Optional[Any] ) -> Dict: lowerCAmelCase_ : List[str] = F'{self.test_file_dir_str}/test_data/wmt_en_ro' lowerCAmelCase_ : Dict = { """--fp16_opt_level=O1""": """""", """$MAX_LEN""": 1_28, """$BS""": 16, """$GAS""": 1, """$ENRO_DIR""": data_dir, """$m""": """sshleifer/student_marian_en_ro_6_1""", """val_check_interval=0.25""": """val_check_interval=1.0""", } # Clean up bash script lowerCAmelCase_ : int = ( (self.test_file_dir / """distil_marian_no_teacher.sh""").open().read().split("""distillation.py""" )[1].strip() ) lowerCAmelCase_ : str = bash_script.replace("""\\\n""" , """""" ).strip().replace("""\"$@\"""" , """""" ) lowerCAmelCase_ : Tuple = bash_script.replace("""--fp16 """ , """ """ ) for k, v in env_vars_to_replace.items(): lowerCAmelCase_ : Optional[int] = bash_script.replace(lowerCamelCase , str(lowerCamelCase ) ) lowerCAmelCase_ : int = self.get_auto_remove_tmp_dir() lowerCAmelCase_ : Optional[Any] = bash_script.replace("""--fp16""" , """""" ) lowerCAmelCase_ : Dict = 6 lowerCAmelCase_ : List[Any] = ( ["""distillation.py"""] + bash_script.split() + [ F'--output_dir={output_dir}', """--gpus=1""", """--learning_rate=1e-3""", F'--num_train_epochs={epochs}', """--warmup_steps=10""", """--val_check_interval=1.0""", """--do_predict""", ] ) with patch.object(lowerCamelCase , """argv""" , lowerCamelCase ): lowerCAmelCase_ : Dict = argparse.ArgumentParser() lowerCAmelCase_ : int = pl.Trainer.add_argparse_args(lowerCamelCase ) lowerCAmelCase_ : List[str] = SummarizationDistiller.add_model_specific_args(lowerCamelCase , os.getcwd() ) lowerCAmelCase_ : List[Any] = parser.parse_args() # assert args.gpus == gpus THIS BREAKS for multi_gpu lowerCAmelCase_ : str = distill_main(lowerCamelCase ) # Check metrics lowerCAmelCase_ : Union[str, Any] = load_json(model.metrics_save_path ) lowerCAmelCase_ : Union[str, Any] = metrics["""val"""][0] lowerCAmelCase_ : Union[str, Any] = metrics["""val"""][-1] assert len(metrics["""val"""] ) >= (args.max_epochs / args.val_check_interval) # +1 accounts for val_sanity_check assert last_step_stats["val_avg_gen_time"] >= 0.01 assert first_step_stats["val_avg_bleu"] < last_step_stats["val_avg_bleu"] # model learned nothing assert 1.0 >= last_step_stats["val_avg_gen_time"] # model hanging on generate. Maybe bad config was saved. assert isinstance(last_step_stats[F'val_avg_{model.val_metric}'] , lowerCamelCase ) # check lightning ckpt can be loaded and has a reasonable statedict lowerCAmelCase_ : Union[str, Any] = os.listdir(lowerCamelCase ) lowerCAmelCase_ : Dict = [x for x in contents if x.endswith(""".ckpt""" )][0] lowerCAmelCase_ : Optional[int] = os.path.join(args.output_dir , lowerCamelCase ) lowerCAmelCase_ : int = torch.load(lowerCamelCase , map_location="""cpu""" ) lowerCAmelCase_ : Tuple = """model.model.decoder.layers.0.encoder_attn_layer_norm.weight""" assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: lowerCAmelCase_ : Union[str, Any] = {os.path.basename(lowerCamelCase ) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics["""test"""] ) == 1
89
1
import json import logging import os import sys from time import time from unittest.mock import patch from transformers.testing_utils import TestCasePlus, require_torch_tpu logging.basicConfig(level=logging.DEBUG) lowercase__ :Any = logging.getLogger() def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = {} lowercase = os.path.join(lowerCAmelCase__ , '''all_results.json''' ) if os.path.exists(lowerCAmelCase__ ): with open(lowerCAmelCase__ , '''r''' ) as f: lowercase = json.load(lowerCAmelCase__ ) else: raise ValueError(f'can\'t find {path}' ) return results lowercase__ :Optional[int] = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) @require_torch_tpu class lowercase ( SCREAMING_SNAKE_CASE__ ): def A__ ( self): import xla_spawn lowercase = self.get_auto_remove_tmp_dir() lowercase = f'\n ./examples/pytorch/text-classification/run_glue.py\n --num_cores=8\n ./examples/pytorch/text-classification/run_glue.py\n --model_name_or_path distilbert-base-uncased\n --output_dir {tmp_dir}\n --overwrite_output_dir\n --train_file ./tests/fixtures/tests_samples/MRPC/train.csv\n --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv\n --do_train\n --do_eval\n --debug tpu_metrics_debug\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --learning_rate=1e-4\n --max_steps=10\n --warmup_steps=2\n --seed=42\n --max_seq_length=128\n '.split() with patch.object(A__ ,'''argv''' ,A__): lowercase = time() xla_spawn.main() lowercase = time() lowercase = get_results(A__) self.assertGreaterEqual(result['''eval_accuracy'''] ,0.75) # Assert that the script takes less than 500 seconds to make sure it doesn't hang. self.assertLess(end - start ,5_0_0) def A__ ( self): import xla_spawn lowercase = ''' ./tests/test_trainer_tpu.py --num_cores=8 ./tests/test_trainer_tpu.py '''.split() with patch.object(A__ ,'''argv''' ,A__): xla_spawn.main()
101
import random from .binary_exp_mod import bin_exp_mod def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__=1000 ): '''simple docstring''' if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd lowercase = n - 1 lowercase = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) lowercase = 0 while count < prec: lowercase = random.randint(2 , n - 1 ) lowercase = bin_exp_mod(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) if b != 1: lowercase = True for _ in range(lowerCAmelCase__ ): if b == n - 1: lowercase = False break lowercase = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": lowercase__ :Tuple = abs(int(input("Enter bound : ").strip())) print("Here's the list of primes:") print(", ".join(str(i) for i in range(n + 1) if is_prime_big(i)))
101
1
"""simple docstring""" from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar __magic_name__ = TypeVar("KEY") __magic_name__ = TypeVar("VAL") @dataclass(frozen=__a , slots=__a ) class SCREAMING_SNAKE_CASE_ ( Generic[KEY, VAL] ): """simple docstring""" __lowercase : KEY __lowercase : VAL class SCREAMING_SNAKE_CASE_ ( _Item ): """simple docstring""" def __init__( self): super().__init__(lowerCAmelCase__ , lowerCAmelCase__) def __bool__( self): return False __magic_name__ = _DeletedItem() class SCREAMING_SNAKE_CASE_ ( MutableMapping[KEY, VAL] ): """simple docstring""" def __init__( self , lowerCAmelCase__ = 8 , lowerCAmelCase__ = 0.75): __SCREAMING_SNAKE_CASE = initial_block_size __SCREAMING_SNAKE_CASE = [None] * initial_block_size assert 0.0 < capacity_factor < 1.0 __SCREAMING_SNAKE_CASE = capacity_factor __SCREAMING_SNAKE_CASE = 0 def snake_case_ ( self , lowerCAmelCase__): return hash(lowerCAmelCase__) % len(self._buckets) def snake_case_ ( self , lowerCAmelCase__): return (ind + 1) % len(self._buckets) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self._buckets[ind] if not stored: __SCREAMING_SNAKE_CASE = _Item(lowerCAmelCase__ , lowerCAmelCase__) self._len += 1 return True elif stored.key == key: __SCREAMING_SNAKE_CASE = _Item(lowerCAmelCase__ , lowerCAmelCase__) return True else: return False def snake_case_ ( self): __SCREAMING_SNAKE_CASE = len(self._buckets) * self._capacity_factor return len(self) >= int(lowerCAmelCase__) def snake_case_ ( self): if len(self._buckets) <= self._initial_block_size: return False __SCREAMING_SNAKE_CASE = len(self._buckets) * self._capacity_factor / 2 return len(self) < limit def snake_case_ ( self , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self._buckets __SCREAMING_SNAKE_CASE = [None] * new_size __SCREAMING_SNAKE_CASE = 0 for item in old_buckets: if item: self._add_item(item.key , item.val) def snake_case_ ( self): self._resize(len(self._buckets) * 2) def snake_case_ ( self): self._resize(len(self._buckets) // 2) def snake_case_ ( self , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self._get_bucket_index(lowerCAmelCase__) for _ in range(len(self._buckets)): yield ind __SCREAMING_SNAKE_CASE = self._get_next_ind(lowerCAmelCase__) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__): for ind in self._iterate_buckets(lowerCAmelCase__): if self._try_set(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): break def __setitem__( self , lowerCAmelCase__ , lowerCAmelCase__): if self._is_full(): self._size_up() self._add_item(lowerCAmelCase__ , lowerCAmelCase__) def __delitem__( self , lowerCAmelCase__): for ind in self._iterate_buckets(lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self._buckets[ind] if item is None: raise KeyError(lowerCAmelCase__) if item is _deleted: continue if item.key == key: __SCREAMING_SNAKE_CASE = _deleted self._len -= 1 break if self._is_sparse(): self._size_down() def __getitem__( self , lowerCAmelCase__): for ind in self._iterate_buckets(lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self._buckets[ind] if item is None: break if item is _deleted: continue if item.key == key: return item.val raise KeyError(lowerCAmelCase__) def __len__( self): return self._len def __iter__( self): yield from (item.key for item in self._buckets if item) def __repr__( self): __SCREAMING_SNAKE_CASE = """ ,""".join( f"{item.key}: {item.val}" for item in self._buckets if item) return f"HashMap({val_string})"
255
"""simple docstring""" import argparse import os import torch from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) __magic_name__ = { "sample_size": 32, "in_channels": 3, "out_channels": 3, "layers_per_block": 2, "num_class_embeds": 1000, "block_out_channels": [32, 64], "attention_head_dim": 8, "down_block_types": [ "ResnetDownsampleBlock2D", "AttnDownBlock2D", ], "up_block_types": [ "AttnUpBlock2D", "ResnetUpsampleBlock2D", ], "resnet_time_scale_shift": "scale_shift", "upsample_type": "resnet", "downsample_type": "resnet", } __magic_name__ = { "sample_size": 64, "in_channels": 3, "out_channels": 3, "layers_per_block": 3, "num_class_embeds": 1000, "block_out_channels": [192, 192 * 2, 192 * 3, 192 * 4], "attention_head_dim": 64, "down_block_types": [ "ResnetDownsampleBlock2D", "AttnDownBlock2D", "AttnDownBlock2D", "AttnDownBlock2D", ], "up_block_types": [ "AttnUpBlock2D", "AttnUpBlock2D", "AttnUpBlock2D", "ResnetUpsampleBlock2D", ], "resnet_time_scale_shift": "scale_shift", "upsample_type": "resnet", "downsample_type": "resnet", } __magic_name__ = { "sample_size": 256, "in_channels": 3, "out_channels": 3, "layers_per_block": 2, "num_class_embeds": None, "block_out_channels": [256, 256, 256 * 2, 256 * 2, 256 * 4, 256 * 4], "attention_head_dim": 64, "down_block_types": [ "ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D", "AttnDownBlock2D", "AttnDownBlock2D", "AttnDownBlock2D", ], "up_block_types": [ "AttnUpBlock2D", "AttnUpBlock2D", "AttnUpBlock2D", "ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D", ], "resnet_time_scale_shift": "default", "upsample_type": "resnet", "downsample_type": "resnet", } __magic_name__ = { "num_train_timesteps": 40, "sigma_min": 0.002, "sigma_max": 80.0, } __magic_name__ = { "num_train_timesteps": 201, "sigma_min": 0.002, "sigma_max": 80.0, } __magic_name__ = { "num_train_timesteps": 151, "sigma_min": 0.002, "sigma_max": 80.0, } def _lowerCAmelCase ( UpperCamelCase_ ): if isinstance(UpperCamelCase_ , UpperCamelCase_ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise argparse.ArgumentTypeError("""boolean value expected""" ) def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=False ): __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.in_layers.0.weight"] __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.in_layers.0.bias"] __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.in_layers.2.weight"] __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.in_layers.2.bias"] __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.emb_layers.1.weight"] __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.emb_layers.1.bias"] __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.out_layers.0.weight"] __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.out_layers.0.bias"] __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.out_layers.3.weight"] __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.out_layers.3.bias"] if has_skip: __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.skip_connection.weight"] __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.skip_connection.bias"] return new_checkpoint def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=None ): __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.qkv.weight"].chunk(3 , dim=0 ) __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.qkv.bias"].chunk(3 , dim=0 ) __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.norm.weight"] __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.norm.bias"] __SCREAMING_SNAKE_CASE = weight_q.squeeze(-1 ).squeeze(-1 ) __SCREAMING_SNAKE_CASE = bias_q.squeeze(-1 ).squeeze(-1 ) __SCREAMING_SNAKE_CASE = weight_k.squeeze(-1 ).squeeze(-1 ) __SCREAMING_SNAKE_CASE = bias_k.squeeze(-1 ).squeeze(-1 ) __SCREAMING_SNAKE_CASE = weight_v.squeeze(-1 ).squeeze(-1 ) __SCREAMING_SNAKE_CASE = bias_v.squeeze(-1 ).squeeze(-1 ) __SCREAMING_SNAKE_CASE = ( checkpoint[f"{old_prefix}.proj_out.weight"].squeeze(-1 ).squeeze(-1 ) ) __SCREAMING_SNAKE_CASE = checkpoint[f"{old_prefix}.proj_out.bias"].squeeze(-1 ).squeeze(-1 ) return new_checkpoint def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = torch.load(UpperCamelCase_ , map_location="""cpu""" ) __SCREAMING_SNAKE_CASE = {} __SCREAMING_SNAKE_CASE = checkpoint["""time_embed.0.weight"""] __SCREAMING_SNAKE_CASE = checkpoint["""time_embed.0.bias"""] __SCREAMING_SNAKE_CASE = checkpoint["""time_embed.2.weight"""] __SCREAMING_SNAKE_CASE = checkpoint["""time_embed.2.bias"""] if unet_config["num_class_embeds"] is not None: __SCREAMING_SNAKE_CASE = checkpoint["""label_emb.weight"""] __SCREAMING_SNAKE_CASE = checkpoint["""input_blocks.0.0.weight"""] __SCREAMING_SNAKE_CASE = checkpoint["""input_blocks.0.0.bias"""] __SCREAMING_SNAKE_CASE = unet_config["""down_block_types"""] __SCREAMING_SNAKE_CASE = unet_config["""layers_per_block"""] __SCREAMING_SNAKE_CASE = unet_config["""attention_head_dim"""] __SCREAMING_SNAKE_CASE = unet_config["""block_out_channels"""] __SCREAMING_SNAKE_CASE = 1 __SCREAMING_SNAKE_CASE = channels_list[0] for i, layer_type in enumerate(UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = channels_list[i] __SCREAMING_SNAKE_CASE = current_channels != prev_channels if layer_type == "ResnetDownsampleBlock2D": for j in range(UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = f"down_blocks.{i}.resnets.{j}" __SCREAMING_SNAKE_CASE = f"input_blocks.{current_layer}.0" __SCREAMING_SNAKE_CASE = True if j == 0 and downsample_block_has_skip else False __SCREAMING_SNAKE_CASE = convert_resnet(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , has_skip=UpperCamelCase_ ) current_layer += 1 elif layer_type == "AttnDownBlock2D": for j in range(UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = f"down_blocks.{i}.resnets.{j}" __SCREAMING_SNAKE_CASE = f"input_blocks.{current_layer}.0" __SCREAMING_SNAKE_CASE = True if j == 0 and downsample_block_has_skip else False __SCREAMING_SNAKE_CASE = convert_resnet(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , has_skip=UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = f"down_blocks.{i}.attentions.{j}" __SCREAMING_SNAKE_CASE = f"input_blocks.{current_layer}.1" __SCREAMING_SNAKE_CASE = convert_attention( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) current_layer += 1 if i != len(UpperCamelCase_ ) - 1: __SCREAMING_SNAKE_CASE = f"down_blocks.{i}.downsamplers.0" __SCREAMING_SNAKE_CASE = f"input_blocks.{current_layer}.0" __SCREAMING_SNAKE_CASE = convert_resnet(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) current_layer += 1 __SCREAMING_SNAKE_CASE = current_channels # hardcoded the mid-block for now __SCREAMING_SNAKE_CASE = """mid_block.resnets.0""" __SCREAMING_SNAKE_CASE = """middle_block.0""" __SCREAMING_SNAKE_CASE = convert_resnet(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = """mid_block.attentions.0""" __SCREAMING_SNAKE_CASE = """middle_block.1""" __SCREAMING_SNAKE_CASE = convert_attention(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = """mid_block.resnets.1""" __SCREAMING_SNAKE_CASE = """middle_block.2""" __SCREAMING_SNAKE_CASE = convert_resnet(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = unet_config["""up_block_types"""] for i, layer_type in enumerate(UpperCamelCase_ ): if layer_type == "ResnetUpsampleBlock2D": for j in range(layers_per_block + 1 ): __SCREAMING_SNAKE_CASE = f"up_blocks.{i}.resnets.{j}" __SCREAMING_SNAKE_CASE = f"output_blocks.{current_layer}.0" __SCREAMING_SNAKE_CASE = convert_resnet(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , has_skip=UpperCamelCase_ ) current_layer += 1 if i != len(UpperCamelCase_ ) - 1: __SCREAMING_SNAKE_CASE = f"up_blocks.{i}.upsamplers.0" __SCREAMING_SNAKE_CASE = f"output_blocks.{current_layer-1}.1" __SCREAMING_SNAKE_CASE = convert_resnet(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) elif layer_type == "AttnUpBlock2D": for j in range(layers_per_block + 1 ): __SCREAMING_SNAKE_CASE = f"up_blocks.{i}.resnets.{j}" __SCREAMING_SNAKE_CASE = f"output_blocks.{current_layer}.0" __SCREAMING_SNAKE_CASE = convert_resnet(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , has_skip=UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = f"up_blocks.{i}.attentions.{j}" __SCREAMING_SNAKE_CASE = f"output_blocks.{current_layer}.1" __SCREAMING_SNAKE_CASE = convert_attention( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) current_layer += 1 if i != len(UpperCamelCase_ ) - 1: __SCREAMING_SNAKE_CASE = f"up_blocks.{i}.upsamplers.0" __SCREAMING_SNAKE_CASE = f"output_blocks.{current_layer-1}.2" __SCREAMING_SNAKE_CASE = convert_resnet(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = checkpoint["""out.0.weight"""] __SCREAMING_SNAKE_CASE = checkpoint["""out.0.bias"""] __SCREAMING_SNAKE_CASE = checkpoint["""out.2.weight"""] __SCREAMING_SNAKE_CASE = checkpoint["""out.2.bias"""] return new_checkpoint if __name__ == "__main__": __magic_name__ = argparse.ArgumentParser() parser.add_argument("--unet_path", default=None, type=str, required=True, help="Path to the unet.pt to convert.") parser.add_argument( "--dump_path", default=None, type=str, required=True, help="Path to output the converted UNet model." ) parser.add_argument("--class_cond", default=True, type=str, help="Whether the model is class-conditional.") __magic_name__ = parser.parse_args() __magic_name__ = strabool(args.class_cond) __magic_name__ = os.path.basename(args.unet_path) print(F"""Checkpoint: {ckpt_name}""") # Get U-Net config if "imagenet64" in ckpt_name: __magic_name__ = IMAGENET_64_UNET_CONFIG elif "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): __magic_name__ = LSUN_256_UNET_CONFIG elif "test" in ckpt_name: __magic_name__ = TEST_UNET_CONFIG else: raise ValueError(F"""Checkpoint type {ckpt_name} is not currently supported.""") if not args.class_cond: __magic_name__ = None __magic_name__ = con_pt_to_diffuser(args.unet_path, unet_config) __magic_name__ = UNetaDModel(**unet_config) image_unet.load_state_dict(converted_unet_ckpt) # Get scheduler config if "cd" in ckpt_name or "test" in ckpt_name: __magic_name__ = CD_SCHEDULER_CONFIG elif "ct" in ckpt_name and "imagenet64" in ckpt_name: __magic_name__ = CT_IMAGENET_64_SCHEDULER_CONFIG elif "ct" in ckpt_name and "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): __magic_name__ = CT_LSUN_256_SCHEDULER_CONFIG else: raise ValueError(F"""Checkpoint type {ckpt_name} is not currently supported.""") __magic_name__ = CMStochasticIterativeScheduler(**scheduler_config) __magic_name__ = ConsistencyModelPipeline(unet=image_unet, scheduler=cm_scheduler) consistency_model.save_pretrained(args.dump_path)
255
1
"""simple docstring""" import json import multiprocessing import os import re from collections import defaultdict import torch from accelerate import Accelerator from accelerate.utils import set_seed from arguments import HumanEvalArguments from datasets import load_dataset, load_metric from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from tqdm import tqdm import transformers from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, StoppingCriteria, StoppingCriteriaList UpperCAmelCase = ["""\nclass""", """\ndef""", """\n#""", """\n@""", """\nprint""", """\nif"""] class UpperCAmelCase_ ( _lowercase): def __init__( self : Optional[int] , __UpperCamelCase : int , __UpperCamelCase : List[Any] , __UpperCamelCase : Dict=None , __UpperCamelCase : Dict=1 ) -> Dict: _UpperCamelCase = tokenizer _UpperCamelCase = dataset _UpperCamelCase = len(__UpperCamelCase ) if n_tasks is None else n_tasks _UpperCamelCase = n_copies def __iter__( self : Any ) -> int: _UpperCamelCase = [] for task in range(self.n_tasks ): # without strip, the model generate commented codes ... prompts.append(self.tokenizer.eos_token + self.dataset[task]['''prompt'''].strip() ) _UpperCamelCase = self.tokenizer(__UpperCamelCase , padding=__UpperCamelCase , return_tensors='''pt''' ) for task in range(self.n_tasks ): for _ in range(self.n_copies ): yield { "ids": outputs.input_ids[task], "task_id": task, "input_len": outputs.attention_mask[task].sum(), } class UpperCAmelCase_ ( _lowercase): def __init__( self : Optional[int] , __UpperCamelCase : Any , __UpperCamelCase : str , __UpperCamelCase : Any ) -> Optional[int]: _UpperCamelCase = start_length _UpperCamelCase = eof_strings _UpperCamelCase = tokenizer def __call__( self : Dict , __UpperCamelCase : Dict , __UpperCamelCase : Optional[Any] , **__UpperCamelCase : str ) -> Union[str, Any]: _UpperCamelCase = self.tokenizer.batch_decode(input_ids[:, self.start_length :] ) _UpperCamelCase = [] for decoded_generation in decoded_generations: done.append(any(stop_string in decoded_generation for stop_string in self.eof_strings ) ) return all(__UpperCamelCase ) def lowercase ( a__ : Optional[Any] ) -> Union[str, Any]: _UpperCamelCase = re.split('''(%s)''' % '''|'''.join(UpperCAmelCase_ ) , UpperCAmelCase_ ) # last string should be "" return "".join(string_list[:-2] ) def lowercase ( a__ : Dict , a__ : int , a__ : Any , a__ : Dict , a__ : int , a__ : str=20 , **a__ : Optional[Any] ) -> Optional[int]: _UpperCamelCase = defaultdict(UpperCAmelCase_ ) # dict of list of generated tokens for step, batch in tqdm(enumerate(UpperCAmelCase_ ) ): with torch.no_grad(): _UpperCamelCase = batch['''ids'''].shape[-1] _UpperCamelCase = accelerator.unwrap_model(UpperCAmelCase_ ).generate( input_ids=batch['''ids'''][:, : batch['''input_len''']] , num_return_sequences=UpperCAmelCase_ , **UpperCAmelCase_ ) # each task is generated batch_size times _UpperCamelCase = batch['''task_id'''].repeat(UpperCAmelCase_ ) _UpperCamelCase = accelerator.pad_across_processes( UpperCAmelCase_ , dim=1 , pad_index=tokenizer.pad_token_id ) _UpperCamelCase , _UpperCamelCase = accelerator.gather((generated_tokens, generated_tasks) ) _UpperCamelCase = generated_tokens.cpu().numpy() _UpperCamelCase = generated_tasks.cpu().numpy() for task, generated_tokens in zip(UpperCAmelCase_ , UpperCAmelCase_ ): gen_token_dict[task].append(UpperCAmelCase_ ) _UpperCamelCase = [[] for _ in range(UpperCAmelCase_ )] for task, generated_tokens in gen_token_dict.items(): for s in generated_tokens: _UpperCamelCase = tokenizer.decode(UpperCAmelCase_ , skip_special_tokens=UpperCAmelCase_ , clean_up_tokenization_spaces=UpperCAmelCase_ ) code_gens[task].append(remove_last_block(UpperCAmelCase_ ) ) return code_gens def lowercase ( ) -> int: # Setup configuration _UpperCamelCase = HfArgumentParser(UpperCAmelCase_ ) _UpperCamelCase = parser.parse_args() transformers.logging.set_verbosity_error() # enables code execution in code_eval metric _UpperCamelCase = args.HF_ALLOW_CODE_EVAL # make sure tokenizer plays nice with multiprocessing _UpperCamelCase = '''false''' if args.num_workers is None: _UpperCamelCase = multiprocessing.cpu_count() # Use dataset load to feed to accelerate _UpperCamelCase = Accelerator() set_seed(args.seed , device_specific=UpperCAmelCase_ ) # Load model and tokenizer _UpperCamelCase = AutoTokenizer.from_pretrained(args.model_ckpt ) _UpperCamelCase = tokenizer.eos_token _UpperCamelCase = AutoModelForCausalLM.from_pretrained(args.model_ckpt ) # Generation settings _UpperCamelCase = { '''do_sample''': args.do_sample, '''temperature''': args.temperature, '''max_new_tokens''': args.max_new_tokens, '''top_p''': args.top_p, '''top_k''': args.top_k, '''stopping_criteria''': StoppingCriteriaList([EndOfFunctionCriteria(0 , UpperCAmelCase_ , UpperCAmelCase_ )] ), } # Load evaluation dataset and metric _UpperCamelCase = load_dataset('''openai_humaneval''' ) _UpperCamelCase = load_metric('''code_eval''' ) _UpperCamelCase = args.num_tasks if args.num_tasks is not None else len(human_eval['''test'''] ) _UpperCamelCase = args.n_samples // args.batch_size _UpperCamelCase = TokenizedDataset(UpperCAmelCase_ , human_eval['''test'''] , n_copies=UpperCAmelCase_ , n_tasks=UpperCAmelCase_ ) # do not confuse args.batch_size, which is actually the num_return_sequences _UpperCamelCase = DataLoader(UpperCAmelCase_ , batch_size=1 ) # Run a quick test to see if code evaluation is enabled try: _UpperCamelCase = code_eval_metric.compute(references=[''''''] , predictions=[['''''']] ) except ValueError as exception: print( '''Code evaluation not enabled. Read the warning below carefully and then use `--HF_ALLOW_CODE_EVAL=\"1\"`''' ''' flag to enable code evaluation.''' ) raise exception _UpperCamelCase , _UpperCamelCase = accelerator.prepare(UpperCAmelCase_ , UpperCAmelCase_ ) _UpperCamelCase = complete_code( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , n_tasks=UpperCAmelCase_ , batch_size=args.batch_size , **UpperCAmelCase_ , ) if accelerator.is_main_process: _UpperCamelCase = [] for task in tqdm(range(UpperCAmelCase_ ) ): _UpperCamelCase = human_eval['''test'''][task]['''test'''] _UpperCamelCase = F'''check({human_eval["test"][task]["entry_point"]})''' references.append('''\n''' + test_func + '''\n''' + entry_point ) # Evaluate completions with "code_eval" metric _UpperCamelCase , _UpperCamelCase = code_eval_metric.compute( references=UpperCAmelCase_ , predictions=UpperCAmelCase_ , num_workers=args.num_workers ) print(F'''Results: {pass_at_k}''' ) # Save results to json file with open(args.output_file , '''w''' ) as fp: json.dump(UpperCAmelCase_ , UpperCAmelCase_ ) # For some reason the folliwng seems to be necessary sometimes for code_eval to work nice with multiprocessing # https://stackoverflow.com/questions/60804599/python-multiprocessing-keeps-spawning-the-whole-script if __name__ == "__main__": main()
256
'''simple docstring''' from __future__ import annotations import unittest from transformers import EsmConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers.models.esm.modeling_tf_esm import ( TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, TFEsmModel, ) class snake_case : """simple docstring""" def __init__( self , UpperCamelCase , ): """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = 13 lowerCamelCase_ = 7 lowerCamelCase_ = True lowerCamelCase_ = True lowerCamelCase_ = True lowerCamelCase_ = 99 lowerCamelCase_ = 32 lowerCamelCase_ = 2 lowerCamelCase_ = 4 lowerCamelCase_ = 37 lowerCamelCase_ = "gelu" lowerCamelCase_ = 0.1 lowerCamelCase_ = 0.1 lowerCamelCase_ = 512 lowerCamelCase_ = 16 lowerCamelCase_ = 2 lowerCamelCase_ = 0.02 lowerCamelCase_ = 3 lowerCamelCase_ = 4 lowerCamelCase_ = None def snake_case ( self ): """simple docstring""" lowerCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase_ = None if self.use_input_mask: lowerCamelCase_ = random_attention_mask([self.batch_size, self.seq_length] ) lowerCamelCase_ = None lowerCamelCase_ = None lowerCamelCase_ = None if self.use_labels: lowerCamelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCamelCase_ = ids_tensor([self.batch_size] , self.num_choices ) lowerCamelCase_ = EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , pad_token_id=1 , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case ( self ): """simple docstring""" ( ( lowerCamelCase_ ) ,( lowerCamelCase_ ) ,( lowerCamelCase_ ) ,( lowerCamelCase_ ) ,( lowerCamelCase_ ) ,( lowerCamelCase_ ) , ) = self.prepare_config_and_inputs() lowerCamelCase_ = True lowerCamelCase_ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) lowerCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def snake_case ( self , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ): """simple docstring""" lowerCamelCase_ = TFEsmModel(config=UpperCamelCase ) lowerCamelCase_ = {"input_ids": input_ids, "attention_mask": input_mask} lowerCamelCase_ = model(UpperCamelCase ) lowerCamelCase_ = [input_ids, input_mask] lowerCamelCase_ = model(UpperCamelCase ) lowerCamelCase_ = model(UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case ( self , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , ): """simple docstring""" lowerCamelCase_ = True lowerCamelCase_ = TFEsmModel(config=UpperCamelCase ) lowerCamelCase_ = { "input_ids": input_ids, "attention_mask": input_mask, "encoder_hidden_states": encoder_hidden_states, "encoder_attention_mask": encoder_attention_mask, } lowerCamelCase_ = model(UpperCamelCase ) lowerCamelCase_ = [input_ids, input_mask] lowerCamelCase_ = model(UpperCamelCase , encoder_hidden_states=UpperCamelCase ) # Also check the case where encoder outputs are not passed lowerCamelCase_ = model(UpperCamelCase , attention_mask=UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case ( self , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ): """simple docstring""" lowerCamelCase_ = TFEsmForMaskedLM(config=UpperCamelCase ) lowerCamelCase_ = model([input_ids, input_mask] ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case ( self , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ): """simple docstring""" lowerCamelCase_ = self.num_labels lowerCamelCase_ = TFEsmForTokenClassification(config=UpperCamelCase ) lowerCamelCase_ = {"input_ids": input_ids, "attention_mask": input_mask} lowerCamelCase_ = model(UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case ( self ): """simple docstring""" lowerCamelCase_ = self.prepare_config_and_inputs() ( ( lowerCamelCase_ ) ,( lowerCamelCase_ ) ,( lowerCamelCase_ ) ,( lowerCamelCase_ ) ,( lowerCamelCase_ ) ,( lowerCamelCase_ ) , ) = config_and_inputs lowerCamelCase_ = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class snake_case ( lowercase , lowercase , unittest.TestCase ): """simple docstring""" _lowerCamelCase = ( ( TFEsmModel, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, ) if is_tf_available() else () ) _lowerCamelCase = ( { "feature-extraction": TFEsmModel, "fill-mask": TFEsmForMaskedLM, "text-classification": TFEsmForSequenceClassification, "token-classification": TFEsmForTokenClassification, "zero-shot": TFEsmForSequenceClassification, } if is_tf_available() else {} ) _lowerCamelCase = False _lowerCamelCase = False def snake_case ( self ): """simple docstring""" lowerCamelCase_ = TFEsmModelTester(self ) lowerCamelCase_ = ConfigTester(self , config_class=UpperCamelCase , hidden_size=37 ) def snake_case ( self ): """simple docstring""" self.config_tester.run_common_tests() def snake_case ( self ): """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase ) def snake_case ( self ): """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*UpperCamelCase ) def snake_case ( self ): """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCamelCase ) def snake_case ( self ): """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCamelCase ) @slow def snake_case ( self ): """simple docstring""" for model_name in TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase_ = TFEsmModel.from_pretrained(UpperCamelCase ) self.assertIsNotNone(UpperCamelCase ) @unittest.skip("Protein models do not support embedding resizing." ) def snake_case ( self ): """simple docstring""" pass @unittest.skip("Protein models do not support embedding resizing." ) def snake_case ( self ): """simple docstring""" pass def snake_case ( self ): """simple docstring""" lowerCamelCase_ ,lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = model_class(UpperCamelCase ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class is TFEsmForMaskedLM: # Output embedding test differs from the main test because they're a matrix, not a layer lowerCamelCase_ = model.get_bias() assert isinstance(UpperCamelCase , UpperCamelCase ) for k, v in name.items(): assert isinstance(UpperCamelCase , tf.Variable ) else: lowerCamelCase_ = model.get_output_embeddings() assert x is None lowerCamelCase_ = model.get_bias() assert name is None @require_tf class snake_case ( unittest.TestCase ): """simple docstring""" @slow def snake_case ( self ): """simple docstring""" lowerCamelCase_ = TFEsmForMaskedLM.from_pretrained("facebook/esm2_t6_8M_UR50D" ) lowerCamelCase_ = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowerCamelCase_ = model(UpperCamelCase )[0] lowerCamelCase_ = [1, 6, 33] self.assertEqual(list(output.numpy().shape ) , UpperCamelCase ) # compare the actual values for a slice. lowerCamelCase_ = tf.constant( [ [ [8.921_518, -10.589_814, -6.4_671_307], [-6.3_967_156, -13.911_377, -1.1_211_915], [-7.781_247, -13.951_557, -3.740_592], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-2 ) ) @slow def snake_case ( self ): """simple docstring""" lowerCamelCase_ = TFEsmModel.from_pretrained("facebook/esm2_t6_8M_UR50D" ) lowerCamelCase_ = tf.constant([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) lowerCamelCase_ = model(UpperCamelCase )[0] # compare the actual values for a slice. lowerCamelCase_ = tf.constant( [ [ [0.14_443_092, 0.54_125_327, 0.3_247_739], [0.30_340_484, 0.00_526_676, 0.31_077_722], [0.32_278_043, -0.24_987_096, 0.3_414_628], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
55
0
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("4.31.0") lowercase__ = logging.getLogger(__name__) @dataclass class snake_case__ : """simple docstring""" lowerCamelCase = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) lowerCamelCase = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) lowerCamelCase = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) lowerCamelCase = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) lowerCamelCase = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) lowerCamelCase = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) lowerCamelCase = field( default=__SCREAMING_SNAKE_CASE , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) @dataclass class snake_case__ : """simple docstring""" lowerCamelCase = field(default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """The input training data file (a text file)."""} ) lowerCamelCase = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) lowerCamelCase = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) lowerCamelCase = field( default=__SCREAMING_SNAKE_CASE , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) lowerCamelCase = field( default=__SCREAMING_SNAKE_CASE , metadata={ """help""": ( """The maximum total input sequence length after tokenization. If passed, sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) lowerCamelCase = field( default=__SCREAMING_SNAKE_CASE , metadata={ """help""": ( """Whether to pad all samples to the maximum sentence length. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch. More """ """efficient on GPU but very bad for TPU.""" ) } , ) lowerCamelCase = field( default=__SCREAMING_SNAKE_CASE , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) lowerCamelCase = field( default=__SCREAMING_SNAKE_CASE , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) def lowerCAmelCase ( self : int ) -> Optional[int]: """simple docstring""" if self.train_file is not None: snake_case : Optional[Any] = self.train_file.split('''.''' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: snake_case : int = self.validation_file.split('''.''' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class snake_case__ : """simple docstring""" lowerCamelCase = 42 lowerCamelCase = True lowerCamelCase = None lowerCamelCase = None def __call__( self : Union[str, Any] , UpperCamelCase__ : str ) -> Union[str, Any]: """simple docstring""" snake_case : Tuple = '''label''' if '''label''' in features[0].keys() else '''labels''' snake_case : Union[str, Any] = [feature.pop(UpperCamelCase__ ) for feature in features] snake_case : Optional[Any] = len(UpperCamelCase__ ) snake_case : Union[str, Any] = len(features[0]['''input_ids'''] ) snake_case : List[Any] = [ [{k: v[i] for k, v in feature.items()} for i in range(UpperCamelCase__ )] for feature in features ] snake_case : Optional[int] = list(chain(*UpperCamelCase__ ) ) snake_case : int = self.tokenizer.pad( UpperCamelCase__ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' , ) # Un-flatten snake_case : Any = {k: v.view(UpperCamelCase__ , UpperCamelCase__ , -1 ) for k, v in batch.items()} # Add back labels snake_case : str = torch.tensor(UpperCamelCase__ , dtype=torch.intaa ) return batch def _UpperCamelCase ( ) -> Optional[Any]: '''simple docstring''' snake_case : int = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. snake_case ,snake_case ,snake_case : Optional[int] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: snake_case ,snake_case ,snake_case : Optional[Any] = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_swag''' , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() snake_case : Tuple = training_args.get_process_log_level() logger.setLevel(SCREAMING_SNAKE_CASE__ ) datasets.utils.logging.set_verbosity(SCREAMING_SNAKE_CASE__ ) transformers.utils.logging.set_verbosity(SCREAMING_SNAKE_CASE__ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}' + F'distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}' ) logger.info(F'Training/evaluation parameters {training_args}' ) # Detecting last checkpoint. snake_case : Optional[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: snake_case : Union[str, Any] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'Output directory ({training_args.output_dir}) already exists and is not empty. ' '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ' '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: snake_case : Union[str, Any] = {} if data_args.train_file is not None: snake_case : int = data_args.train_file if data_args.validation_file is not None: snake_case : Optional[int] = data_args.validation_file snake_case : List[str] = data_args.train_file.split('''.''' )[-1] snake_case : str = load_dataset( SCREAMING_SNAKE_CASE__ , data_files=SCREAMING_SNAKE_CASE__ , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. snake_case : str = load_dataset( '''swag''' , '''regular''' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case : List[str] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) snake_case : Dict = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) snake_case : Optional[Any] = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=SCREAMING_SNAKE_CASE__ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. snake_case : List[Any] = [F'ending{i}' for i in range(4 )] snake_case : Optional[int] = '''sent1''' snake_case : Union[str, Any] = '''sent2''' if data_args.max_seq_length is None: snake_case : Any = tokenizer.model_max_length if max_seq_length > 1024: logger.warning( '''The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value''' ''' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can''' ''' override this default with `--block_size xxx`.''' ) snake_case : Tuple = 1024 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F'The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the' F'model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.' ) snake_case : Any = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(SCREAMING_SNAKE_CASE__ ): snake_case : Union[str, Any] = [[context] * 4 for context in examples[context_name]] snake_case : Any = examples[question_header_name] snake_case : Any = [ [F'{header} {examples[end][i]}' for end in ending_names] for i, header in enumerate(SCREAMING_SNAKE_CASE__ ) ] # Flatten out snake_case : Optional[int] = list(chain(*SCREAMING_SNAKE_CASE__ ) ) snake_case : str = list(chain(*SCREAMING_SNAKE_CASE__ ) ) # Tokenize snake_case : Optional[Any] = tokenizer( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , truncation=SCREAMING_SNAKE_CASE__ , max_length=SCREAMING_SNAKE_CASE__ , padding='''max_length''' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(SCREAMING_SNAKE_CASE__ ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('''--do_train requires a train dataset''' ) snake_case : Optional[Any] = raw_datasets['''train'''] if data_args.max_train_samples is not None: snake_case : Union[str, Any] = min(len(SCREAMING_SNAKE_CASE__ ) , data_args.max_train_samples ) snake_case : Dict = train_dataset.select(range(SCREAMING_SNAKE_CASE__ ) ) with training_args.main_process_first(desc='''train dataset map pre-processing''' ): snake_case : Optional[int] = train_dataset.map( SCREAMING_SNAKE_CASE__ , batched=SCREAMING_SNAKE_CASE__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('''--do_eval requires a validation dataset''' ) snake_case : List[str] = raw_datasets['''validation'''] if data_args.max_eval_samples is not None: snake_case : str = min(len(SCREAMING_SNAKE_CASE__ ) , data_args.max_eval_samples ) snake_case : Union[str, Any] = eval_dataset.select(range(SCREAMING_SNAKE_CASE__ ) ) with training_args.main_process_first(desc='''validation dataset map pre-processing''' ): snake_case : Tuple = eval_dataset.map( SCREAMING_SNAKE_CASE__ , batched=SCREAMING_SNAKE_CASE__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator snake_case : Optional[int] = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=SCREAMING_SNAKE_CASE__ , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(SCREAMING_SNAKE_CASE__ ): snake_case ,snake_case : int = eval_predictions snake_case : Tuple = np.argmax(SCREAMING_SNAKE_CASE__ , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer snake_case : Any = Trainer( model=SCREAMING_SNAKE_CASE__ , args=SCREAMING_SNAKE_CASE__ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=SCREAMING_SNAKE_CASE__ , data_collator=SCREAMING_SNAKE_CASE__ , compute_metrics=SCREAMING_SNAKE_CASE__ , ) # Training if training_args.do_train: snake_case : List[str] = None if training_args.resume_from_checkpoint is not None: snake_case : int = training_args.resume_from_checkpoint elif last_checkpoint is not None: snake_case : Dict = last_checkpoint snake_case : Union[str, Any] = trainer.train(resume_from_checkpoint=SCREAMING_SNAKE_CASE__ ) trainer.save_model() # Saves the tokenizer too for easy upload snake_case : str = train_result.metrics snake_case : Union[str, Any] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(SCREAMING_SNAKE_CASE__ ) ) snake_case : Optional[int] = min(SCREAMING_SNAKE_CASE__ , len(SCREAMING_SNAKE_CASE__ ) ) trainer.log_metrics('''train''' , SCREAMING_SNAKE_CASE__ ) trainer.save_metrics('''train''' , SCREAMING_SNAKE_CASE__ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('''*** Evaluate ***''' ) snake_case : List[Any] = trainer.evaluate() snake_case : Union[str, Any] = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(SCREAMING_SNAKE_CASE__ ) snake_case : Any = min(SCREAMING_SNAKE_CASE__ , len(SCREAMING_SNAKE_CASE__ ) ) trainer.log_metrics('''eval''' , SCREAMING_SNAKE_CASE__ ) trainer.save_metrics('''eval''' , SCREAMING_SNAKE_CASE__ ) snake_case : int = { '''finetuned_from''': model_args.model_name_or_path, '''tasks''': '''multiple-choice''', '''dataset_tags''': '''swag''', '''dataset_args''': '''regular''', '''dataset''': '''SWAG''', '''language''': '''en''', } if training_args.push_to_hub: trainer.push_to_hub(**SCREAMING_SNAKE_CASE__ ) else: trainer.create_model_card(**SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ ) -> Optional[Any]: '''simple docstring''' main() if __name__ == "__main__": main()
83
'''simple docstring''' def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ ) -> list: '''simple docstring''' snake_case : Any = len(SCREAMING_SNAKE_CASE__ ) for i in range(1 , SCREAMING_SNAKE_CASE__ ): snake_case : Optional[int] = collection[i] snake_case : str = 0 snake_case : List[Any] = i - 1 while low <= high: snake_case : List[Any] = (low + high) // 2 if val < collection[mid]: snake_case : List[str] = mid - 1 else: snake_case : Optional[int] = mid + 1 for j in range(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , -1 ): snake_case : Dict = collection[j - 1] snake_case : Union[str, Any] = val return collection if __name__ == "__main__": lowercase__ = input("Enter numbers separated by a comma:\n").strip() lowercase__ = [int(item) for item in user_input.split(",")] print(binary_insertion_sort(unsorted))
83
1
"""simple docstring""" import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def _snake_case ( lowercase__ : Any , lowercase__ : int , lowercase__ : Tuple ) -> Any: '''simple docstring''' return params[f"""{prefix}/{prefix}/relpos_bias/rel_embedding"""][:, i, :] def _snake_case ( lowercase__ : Union[str, Any] , lowercase__ : List[Any] , lowercase__ : Any , lowercase__ : Tuple="attention" ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ :Optional[Any] = np.ascontiguousarray(params[f"""{prefix}/{prefix}/{layer_name}/key/kernel"""][:, i, :, :] ) lowerCAmelCase_ :str = k_tmp.reshape(k_tmp.shape[0] , k_tmp.shape[1] * k_tmp.shape[2] ) lowerCAmelCase_ :Any = np.ascontiguousarray(params[f"""{prefix}/{prefix}/{layer_name}/out/kernel"""][:, i, :, :] ) lowerCAmelCase_ :Tuple = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] , o_tmp.shape[2] ) lowerCAmelCase_ :str = np.ascontiguousarray(params[f"""{prefix}/{prefix}/{layer_name}/query/kernel"""][:, i, :, :] ) lowerCAmelCase_ :Dict = q_tmp.reshape(q_tmp.shape[0] , q_tmp.shape[1] * q_tmp.shape[2] ) lowerCAmelCase_ :Optional[Any] = np.ascontiguousarray(params[f"""{prefix}/{prefix}/{layer_name}/value/kernel"""][:, i, :, :] ) lowerCAmelCase_ :Optional[Any] = v_tmp.reshape(v_tmp.shape[0] , v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def _snake_case ( lowercase__ : Tuple , lowercase__ : Tuple , lowercase__ : Dict , lowercase__ : Optional[int]=False ) -> List[Any]: '''simple docstring''' if split_mlp_wi: lowerCAmelCase_ :Tuple = params[f"""{prefix}/{prefix}/mlp/wi_0/kernel"""][:, i, :] lowerCAmelCase_ :Optional[Any] = params[f"""{prefix}/{prefix}/mlp/wi_1/kernel"""][:, i, :] lowerCAmelCase_ :List[Any] = (wi_a, wi_a) else: lowerCAmelCase_ :List[str] = params[f"""{prefix}/{prefix}/mlp/wi/kernel"""][:, i, :] lowerCAmelCase_ :Union[str, Any] = params[f"""{prefix}/{prefix}/mlp/wo/kernel"""][:, i, :] return wi, wo def _snake_case ( lowercase__ : int , lowercase__ : Dict , lowercase__ : List[str] , lowercase__ : Optional[int] ) -> List[str]: '''simple docstring''' return params[f"""{prefix}/{prefix}/{layer_name}/scale"""][:, i] def _snake_case ( lowercase__ : dict , *, lowercase__ : int , lowercase__ : bool , lowercase__ : bool = False ) -> List[str]: '''simple docstring''' lowerCAmelCase_ :int = traverse_util.flatten_dict(variables["""target"""] ) lowerCAmelCase_ :Optional[int] = {"""/""".join(lowercase__ ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi lowerCAmelCase_ :Union[str, Any] = """encoder/encoder/mlp/wi_0/kernel""" in old print("""Split MLP:""" , lowercase__ ) lowerCAmelCase_ :List[Any] = collections.OrderedDict() # Shared embeddings. lowerCAmelCase_ :Tuple = old["""token_embedder/embedding"""] # Encoder. for i in range(lowercase__ ): # Block i, layer 0 (Self Attention). lowerCAmelCase_ :List[Any] = tax_layer_norm_lookup(lowercase__ , lowercase__ , """encoder""" , """pre_attention_layer_norm""" ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ :Optional[Any] = tax_attention_lookup(lowercase__ , lowercase__ , """encoder""" , """attention""" ) lowerCAmelCase_ :int = layer_norm lowerCAmelCase_ :Optional[int] = k.T lowerCAmelCase_ :List[Any] = o.T lowerCAmelCase_ :Dict = q.T lowerCAmelCase_ :List[str] = v.T # Block i, layer 1 (MLP). lowerCAmelCase_ :Union[str, Any] = tax_layer_norm_lookup(lowercase__ , lowercase__ , """encoder""" , """pre_mlp_layer_norm""" ) lowerCAmelCase_ , lowerCAmelCase_ :str = tax_mlp_lookup(lowercase__ , lowercase__ , """encoder""" , lowercase__ ) lowerCAmelCase_ :List[str] = layer_norm if split_mlp_wi: lowerCAmelCase_ :Tuple = wi[0].T lowerCAmelCase_ :List[Any] = wi[1].T else: lowerCAmelCase_ :Dict = wi.T lowerCAmelCase_ :str = wo.T if scalable_attention: # convert the rel_embedding of each layer lowerCAmelCase_ :Optional[Any] = tax_relpos_bias_lookup( lowercase__ , lowercase__ , """encoder""" ).T lowerCAmelCase_ :Tuple = old["""encoder/encoder_norm/scale"""] if not scalable_attention: lowerCAmelCase_ :List[str] = tax_relpos_bias_lookup( lowercase__ , 0 , """encoder""" ).T lowerCAmelCase_ :Dict = tax_relpos_bias_lookup( lowercase__ , 0 , """decoder""" ).T if not is_encoder_only: # Decoder. for i in range(lowercase__ ): # Block i, layer 0 (Self Attention). lowerCAmelCase_ :Union[str, Any] = tax_layer_norm_lookup(lowercase__ , lowercase__ , """decoder""" , """pre_self_attention_layer_norm""" ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ :Dict = tax_attention_lookup(lowercase__ , lowercase__ , """decoder""" , """self_attention""" ) lowerCAmelCase_ :Any = layer_norm lowerCAmelCase_ :str = k.T lowerCAmelCase_ :str = o.T lowerCAmelCase_ :str = q.T lowerCAmelCase_ :List[Any] = v.T # Block i, layer 1 (Cross Attention). lowerCAmelCase_ :Tuple = tax_layer_norm_lookup(lowercase__ , lowercase__ , """decoder""" , """pre_cross_attention_layer_norm""" ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ :Union[str, Any] = tax_attention_lookup(lowercase__ , lowercase__ , """decoder""" , """encoder_decoder_attention""" ) lowerCAmelCase_ :str = layer_norm lowerCAmelCase_ :Optional[int] = k.T lowerCAmelCase_ :Optional[int] = o.T lowerCAmelCase_ :Dict = q.T lowerCAmelCase_ :Union[str, Any] = v.T # Block i, layer 2 (MLP). lowerCAmelCase_ :Optional[int] = tax_layer_norm_lookup(lowercase__ , lowercase__ , """decoder""" , """pre_mlp_layer_norm""" ) lowerCAmelCase_ , lowerCAmelCase_ :str = tax_mlp_lookup(lowercase__ , lowercase__ , """decoder""" , lowercase__ ) lowerCAmelCase_ :str = layer_norm if split_mlp_wi: lowerCAmelCase_ :Optional[int] = wi[0].T lowerCAmelCase_ :Tuple = wi[1].T else: lowerCAmelCase_ :str = wi.T lowerCAmelCase_ :Tuple = wo.T if scalable_attention: # convert the rel_embedding of each layer lowerCAmelCase_ :Dict = tax_relpos_bias_lookup(lowercase__ , lowercase__ , """decoder""" ).T lowerCAmelCase_ :int = old["""decoder/decoder_norm/scale"""] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: lowerCAmelCase_ :Dict = old["""decoder/logits_dense/kernel"""].T return new def _snake_case ( lowercase__ : Any , lowercase__ : bool ) -> str: '''simple docstring''' lowerCAmelCase_ :str = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: lowerCAmelCase_ :Any = state_dict["""shared.weight"""] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: lowerCAmelCase_ :Union[str, Any] = state_dict["""shared.weight"""] if "lm_head.weight" not in state_dict: # For old 1.0 models. print("""Using shared word embeddings as lm_head.""" ) lowerCAmelCase_ :Tuple = state_dict["""shared.weight"""] return state_dict def _snake_case ( lowercase__ : Optional[int] , lowercase__ : Union[str, Any] , lowercase__ : str , lowercase__ : int , lowercase__ : Union[str, Any] ) -> Tuple: '''simple docstring''' lowerCAmelCase_ :Tuple = checkpoints.load_tax_checkpoint(lowercase__ ) lowerCAmelCase_ :List[Any] = convert_tax_to_pytorch( lowercase__ , num_layers=config.num_layers , is_encoder_only=lowercase__ , scalable_attention=lowercase__ ) lowerCAmelCase_ :Optional[Any] = make_state_dict(lowercase__ , lowercase__ ) model.load_state_dict(lowercase__ , strict=lowercase__ ) def _snake_case ( lowercase__ : Any , lowercase__ : Dict , lowercase__ : Union[str, Any] , lowercase__ : bool = False , lowercase__ : bool = False , ) -> Tuple: '''simple docstring''' lowerCAmelCase_ :Any = MTaConfig.from_json_file(lowercase__ ) print(f"""Building PyTorch model from configuration: {config}""" ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: lowerCAmelCase_ :Tuple = UMTaEncoderModel(lowercase__ ) else: lowerCAmelCase_ :Union[str, Any] = UMTaForConditionalGeneration(lowercase__ ) # Load weights from tf checkpoint load_tax_weights_in_ta(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) model.save_pretrained(lowercase__ ) # Verify that we can load the checkpoint. model.from_pretrained(lowercase__ ) print("""Done""" ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser(description='Converts a native T5X checkpoint into a PyTorch checkpoint.') # Required parameters parser.add_argument( '--t5x_checkpoint_path', default=None, type=str, required=True, help='Path to the T5X checkpoint.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--is_encoder_only', action='store_true', help='Check if the model is encoder-decoder model', default=False ) parser.add_argument( '--scalable_attention', action='store_true', help='Whether the model uses scaled attention (umt5 model)', default=False, ) __UpperCAmelCase = parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
84
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=__snake_case ) class _snake_case ( __snake_case ): '''simple docstring''' A__ : str = field(default="automatic-speech-recognition" , metadata={"include_in_asdict_even_if_is_default": True} ) A__ : ClassVar[Features] = Features({"audio": Audio()} ) A__ : ClassVar[Features] = Features({"transcription": Value("string" )} ) A__ : str = "audio" A__ : str = "transcription" def A__ ( self: int ,lowerCamelCase_: Union[str, Any] ) -> Optional[Any]: if self.audio_column not in features: raise ValueError(F'''Column {self.audio_column} is not present in features.''' ) if not isinstance(features[self.audio_column] ,lowerCamelCase_ ): raise ValueError(F'''Column {self.audio_column} is not an Audio type.''' ) UpperCAmelCase_ : Any = copy.deepcopy(self ) UpperCAmelCase_ : Union[str, Any] = self.input_schema.copy() UpperCAmelCase_ : Any = features[self.audio_column] UpperCAmelCase_ : Union[str, Any] = input_schema return task_template @property def A__ ( self: List[str] ) -> Dict[str, str]: return {self.audio_column: "audio", self.transcription_column: "transcription"}
345
0
import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets UpperCamelCase__ = '\\n@inproceedings{popovic-2015-chrf,\n title = "chr{F}: character n-gram {F}-score for automatic {MT} evaluation",\n author = "Popovi{\'c}, Maja",\n booktitle = "Proceedings of the Tenth Workshop on Statistical Machine Translation",\n month = sep,\n year = "2015",\n address = "Lisbon, Portugal",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/W15-3049",\n doi = "10.18653/v1/W15-3049",\n pages = "392--395",\n}\n@inproceedings{popovic-2017-chrf,\n title = "chr{F}++: words helping character n-grams",\n author = "Popovi{\'c}, Maja",\n booktitle = "Proceedings of the Second Conference on Machine Translation",\n month = sep,\n year = "2017",\n address = "Copenhagen, Denmark",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/W17-4770",\n doi = "10.18653/v1/W17-4770",\n pages = "612--618",\n}\n@inproceedings{post-2018-call,\n title = "A Call for Clarity in Reporting {BLEU} Scores",\n author = "Post, Matt",\n booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers",\n month = oct,\n year = "2018",\n address = "Belgium, Brussels",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/W18-6319",\n pages = "186--191",\n}\n' UpperCamelCase__ = '\\nChrF and ChrF++ are two MT evaluation metrics. They both use the F-score statistic for character n-gram matches,\nand ChrF++ adds word n-grams as well which correlates more strongly with direct assessment. We use the implementation\nthat is already present in sacrebleu.\n\nThe implementation here is slightly different from sacrebleu in terms of the required input format. The length of\nthe references and hypotheses lists need to be the same, so you may need to transpose your references compared to\nsacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534\n\nSee the README.md file at https://github.com/mjpost/sacreBLEU#chrf--chrf for more information.\n' UpperCamelCase__ = '\nProduces ChrF(++) scores for hypotheses given reference translations.\n\nArgs:\n predictions (list of str): The predicted sentences.\n references (list of list of str): The references. There should be one reference sub-list for each prediction sentence.\n char_order (int): Character n-gram order. Defaults to `6`.\n word_order (int): Word n-gram order. If equals to `2`, the metric is referred to as chrF++. Defaults to `0`.\n beta (int): Determine the importance of recall w.r.t precision. Defaults to `2`.\n lowercase (bool): if `True`, enables case-insensitivity. Defaults to `False`.\n whitespace (bool): If `True`, include whitespaces when extracting character n-grams.\n eps_smoothing (bool): If `True`, applies epsilon smoothing similar\n to reference chrF++.py, NLTK and Moses implementations. If `False`,\n it takes into account effective match order similar to sacreBLEU < 2.0.0. Defaults to `False`.\n\nReturns:\n \'score\' (float): The chrF (chrF++) score,\n \'char_order\' (int): The character n-gram order,\n \'word_order\' (int): The word n-gram order. If equals to 2, the metric is referred to as chrF++,\n \'beta\' (int): Determine the importance of recall w.r.t precision\n\nExamples:\n Example 1--a simple example of calculating chrF:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction, references=reference)\n >>> print(results)\n {\'score\': 84.64214891738334, \'char_order\': 6, \'word_order\': 0, \'beta\': 2}\n\n Example 2--the same example, but with the argument word_order=2, to calculate chrF++ instead of chrF:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2)\n >>> print(results)\n {\'score\': 82.87263732906315, \'char_order\': 6, \'word_order\': 2, \'beta\': 2}\n\n Example 3--the same chrF++ example as above, but with `lowercase=True` to normalize all case:\n >>> prediction = ["The relationship between cats and dogs is not exactly friendly.", "a good bookshop is just a genteel black hole that knows how to read."]\n >>> reference = [["The relationship between dogs and cats is not exactly friendly."], ["A good bookshop is just a genteel Black Hole that knows how to read."]]\n >>> chrf = datasets.load_metric("chrf")\n >>> results = chrf.compute(predictions=prediction,\n ... references=reference,\n ... word_order=2,\n ... lowercase=True)\n >>> print(results)\n {\'score\': 92.12853119829202, \'char_order\': 6, \'word_order\': 2, \'beta\': 2}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A ( datasets.Metric ): def lowercase_ (self : int ) -> List[Any]: """simple docstring""" if version.parse(scb.__version__ ) < version.parse("1.4.12" ): raise ImportWarning( "To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n" "You can install it with `pip install \"sacrebleu>=1.4.12\"`." ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://github.com/mjpost/sacreBLEU#chrf--chrf" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Sequence(datasets.Value("string" , id="sequence" ) , id="references" ), } ) , codebase_urls=["https://github.com/mjpost/sacreBLEU#chrf--chrf"] , reference_urls=[ "https://github.com/m-popovic/chrF", ] , ) def lowercase_ (self : Tuple , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : int = CHRF.CHAR_ORDER , __UpperCAmelCase : int = CHRF.WORD_ORDER , __UpperCAmelCase : int = CHRF.BETA , __UpperCAmelCase : bool = False , __UpperCAmelCase : bool = False , __UpperCAmelCase : bool = False , ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = len(references[0] ) if any(len(__UpperCAmelCase ) != references_per_prediction for refs in references ): raise ValueError("Sacrebleu requires the same number of references for each prediction" ) UpperCAmelCase__ = [[refs[i] for refs in references] for i in range(__UpperCAmelCase )] UpperCAmelCase__ = CHRF(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = sb_chrf.corpus_score(__UpperCAmelCase , __UpperCAmelCase ) return { "score": output.score, "char_order": output.char_order, "word_order": output.word_order, "beta": output.beta, }
356
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__ = logging.get_logger(__name__) def lowerCAmelCase_ ( __A, __A=False ) -> Any: '''simple docstring''' UpperCAmelCase__ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""deit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""deit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""deit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""deit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""deit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""deit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""deit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""deit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""deit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""deit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "deit.embeddings.cls_token"), ("dist_token", "deit.embeddings.distillation_token"), ("patch_embed.proj.weight", "deit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "deit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "deit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" UpperCAmelCase__ = [(pair[0], pair[1][4:]) if pair[1].startswith("deit" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("norm.weight", "deit.layernorm.weight"), ("norm.bias", "deit.layernorm.bias"), ("head.weight", "cls_classifier.weight"), ("head.bias", "cls_classifier.bias"), ("head_dist.weight", "distillation_classifier.weight"), ("head_dist.bias", "distillation_classifier.bias"), ] ) return rename_keys def lowerCAmelCase_ ( __A, __A, __A=False ) -> Tuple: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase__ = "" else: UpperCAmelCase__ = "deit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase__ = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) UpperCAmelCase__ = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase__ = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase__ = in_proj_bias[: config.hidden_size] UpperCAmelCase__ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase__ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase__ = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase__ = in_proj_bias[-config.hidden_size :] def lowerCAmelCase_ ( __A, __A, __A ) -> Dict: '''simple docstring''' UpperCAmelCase__ = dct.pop(__A ) UpperCAmelCase__ = val def lowerCAmelCase_ ( ) -> Dict: '''simple docstring''' UpperCAmelCase__ = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase__ = Image.open(requests.get(__A, stream=__A ).raw ) return im @torch.no_grad() def lowerCAmelCase_ ( __A, __A ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase__ = DeiTConfig() # all deit models have fine-tuned heads UpperCAmelCase__ = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase__ = 1_000 UpperCAmelCase__ = "huggingface/label-files" UpperCAmelCase__ = "imagenet-1k-id2label.json" UpperCAmelCase__ = json.load(open(hf_hub_download(__A, __A, repo_type="dataset" ), "r" ) ) UpperCAmelCase__ = {int(__A ): v for k, v in idalabel.items()} UpperCAmelCase__ = idalabel UpperCAmelCase__ = {v: k for k, v in idalabel.items()} UpperCAmelCase__ = int(deit_name[-6:-4] ) UpperCAmelCase__ = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("tiny" ): UpperCAmelCase__ = 192 UpperCAmelCase__ = 768 UpperCAmelCase__ = 12 UpperCAmelCase__ = 3 elif deit_name[9:].startswith("small" ): UpperCAmelCase__ = 384 UpperCAmelCase__ = 1_536 UpperCAmelCase__ = 12 UpperCAmelCase__ = 6 if deit_name[9:].startswith("base" ): pass elif deit_name[4:].startswith("large" ): UpperCAmelCase__ = 1_024 UpperCAmelCase__ = 4_096 UpperCAmelCase__ = 24 UpperCAmelCase__ = 16 # load original model from timm UpperCAmelCase__ = timm.create_model(__A, pretrained=__A ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCAmelCase__ = timm_model.state_dict() UpperCAmelCase__ = create_rename_keys(__A, __A ) for src, dest in rename_keys: rename_key(__A, __A, __A ) read_in_q_k_v(__A, __A, __A ) # load HuggingFace model UpperCAmelCase__ = DeiTForImageClassificationWithTeacher(__A ).eval() model.load_state_dict(__A ) # Check outputs on an image, prepared by DeiTImageProcessor UpperCAmelCase__ = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 UpperCAmelCase__ = DeiTImageProcessor(size=__A, crop_size=config.image_size ) UpperCAmelCase__ = image_processor(images=prepare_img(), return_tensors="pt" ) UpperCAmelCase__ = encoding["pixel_values"] UpperCAmelCase__ = model(__A ) UpperCAmelCase__ = timm_model(__A ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__A, outputs.logits, atol=1e-3 ) Path(__A ).mkdir(exist_ok=__A ) print(f"""Saving model {deit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__A ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__A ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--deit_name', default='vit_deit_base_distilled_patch16_224', type=str, help='Name of the DeiT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) UpperCamelCase__ = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
143
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy __A = logging.get_logger(__name__) class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ): lowercase__: Union[str, Any] = feature_size lowercase__: Optional[Any] = sampling_rate lowercase__: int = padding_value lowercase__: str = kwargs.pop('''padding_side''' , '''right''' ) lowercase__: int = kwargs.pop('''return_attention_mask''' , _UpperCAmelCase ) super().__init__(**_UpperCAmelCase ) def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase = True , _UpperCAmelCase = None , _UpperCAmelCase = False , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , ): # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(_UpperCAmelCase , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): lowercase__: Union[str, Any] = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( '''You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`''' F""" to this method that includes {self.model_input_names[0]}, but you provided""" F""" {list(processed_features.keys() )}""" ) lowercase__: Optional[Any] = processed_features[self.model_input_names[0]] lowercase__: Union[str, Any] = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(_UpperCAmelCase ) == 0: if return_attention_mask: lowercase__: int = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch lowercase__: Union[str, Any] = required_input[0] if isinstance(_UpperCAmelCase , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. lowercase__: Any = 0 while len(required_input[index] ) == 0: index += 1 if index < len(_UpperCAmelCase ): lowercase__: Union[str, Any] = required_input[index][0] if return_tensors is None: if is_tf_tensor(_UpperCAmelCase ): lowercase__: List[str] = '''tf''' elif is_torch_tensor(_UpperCAmelCase ): lowercase__: Any = '''pt''' elif isinstance(_UpperCAmelCase , (int, float, list, tuple, np.ndarray) ): lowercase__: str = '''np''' else: raise ValueError( F"""type of {first_element} unknown: {type(_UpperCAmelCase )}. """ '''Should be one of a python, numpy, pytorch or tensorflow object.''' ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): lowercase__: Dict = to_numpy(_UpperCAmelCase ) else: lowercase__: Dict = [to_numpy(_UpperCAmelCase ) for v in value] # Convert padding_strategy in PaddingStrategy lowercase__: List[Any] = self._get_padding_strategies(padding=_UpperCAmelCase , max_length=_UpperCAmelCase ) lowercase__: List[Any] = processed_features[self.model_input_names[0]] lowercase__: int = len(_UpperCAmelCase ) if not all(len(_UpperCAmelCase ) == batch_size for v in processed_features.values() ): raise ValueError('''Some items in the output dictionary have a different batch size than others.''' ) lowercase__: Optional[Any] = [] for i in range(_UpperCAmelCase ): lowercase__: Tuple = {k: v[i] for k, v in processed_features.items()} # truncation lowercase__: List[str] = self._truncate( _UpperCAmelCase , max_length=_UpperCAmelCase , pad_to_multiple_of=_UpperCAmelCase , truncation=_UpperCAmelCase , ) truncated_inputs.append(_UpperCAmelCase ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length lowercase__: List[Any] = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) lowercase__: Union[str, Any] = PaddingStrategy.MAX_LENGTH lowercase__: Tuple = {} for i in range(_UpperCAmelCase ): # padding lowercase__: List[str] = self._pad( truncated_inputs[i] , max_length=_UpperCAmelCase , padding_strategy=_UpperCAmelCase , pad_to_multiple_of=_UpperCAmelCase , return_attention_mask=_UpperCAmelCase , ) for key, value in outputs.items(): if key not in batch_outputs: lowercase__: Union[str, Any] = [] if value.dtype is np.dtype(np.floataa ): lowercase__: Dict = value.astype(np.floataa ) batch_outputs[key].append(_UpperCAmelCase ) return BatchFeature(_UpperCAmelCase , tensor_type=_UpperCAmelCase ) def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase = None , _UpperCAmelCase = PaddingStrategy.DO_NOT_PAD , _UpperCAmelCase = None , _UpperCAmelCase = None , ): lowercase__: Tuple = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: lowercase__: Tuple = len(_UpperCAmelCase ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): lowercase__: int = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of lowercase__: Dict = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(_UpperCAmelCase ) < max_length if return_attention_mask and "attention_mask" not in processed_features: lowercase__: Optional[int] = np.ones(len(_UpperCAmelCase ) , dtype=np.intaa ) if needs_to_be_padded: lowercase__: List[Any] = max_length - len(_UpperCAmelCase ) if self.padding_side == "right": if return_attention_mask: lowercase__: Optional[int] = np.pad( processed_features['''attention_mask'''] , (0, difference) ) lowercase__: Optional[int] = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) lowercase__: Union[str, Any] = np.pad( _UpperCAmelCase , _UpperCAmelCase , '''constant''' , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: lowercase__: str = np.pad( processed_features['''attention_mask'''] , (difference, 0) ) lowercase__: str = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) lowercase__: List[Any] = np.pad( _UpperCAmelCase , _UpperCAmelCase , '''constant''' , constant_values=self.padding_value ) else: raise ValueError('''Invalid padding strategy:''' + str(self.padding_side ) ) return processed_features def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , ): if not truncation: return processed_features elif truncation and max_length is None: raise ValueError('''When setting ``truncation=True``, make sure that ``max_length`` is defined.''' ) lowercase__: Any = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): lowercase__: Optional[int] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of lowercase__: int = len(_UpperCAmelCase ) > max_length if needs_to_be_truncated: lowercase__: Any = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: lowercase__: Tuple = processed_features['''attention_mask'''][:max_length] return processed_features def _snake_case ( self , _UpperCAmelCase=False , _UpperCAmelCase=None ): # Get padding strategy if padding is not False: if padding is True: lowercase__: Union[str, Any] = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(_UpperCAmelCase , _UpperCAmelCase ): lowercase__: str = PaddingStrategy(_UpperCAmelCase ) elif isinstance(_UpperCAmelCase , _UpperCAmelCase ): lowercase__: Any = padding else: lowercase__: str = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F"""When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined""" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( '''Asking to pad but the feature_extractor does not have a padding value. Please select a value to use''' ''' as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.''' ) return padding_strategy
177
"""simple docstring""" from jiwer import compute_measures import datasets __A = "\\n@inproceedings{inproceedings,\n author = {Morris, Andrew and Maier, Viktoria and Green, Phil},\n year = {2004},\n month = {01},\n pages = {},\n title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.}\n}\n" __A = "\\nWord error rate (WER) is a common metric of the performance of an automatic speech recognition system.\n\nThe general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort.\n\nThis problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate.\n\nWord error rate can then be computed as:\n\nWER = (S + D + I) / N = (S + D + I) / (S + D + C)\n\nwhere\n\nS is the number of substitutions,\nD is the number of deletions,\nI is the number of insertions,\nC is the number of correct words,\nN is the number of words in the reference (N=S+D+C).\n\nThis value indicates the average number of errors per reference word. The lower the value, the better the\nperformance of the ASR system with a WER of 0 being a perfect score.\n" __A = "\nCompute WER score of transcribed segments against references.\n\nArgs:\n references: List of references for each speech input.\n predictions: List of transcriptions to score.\n concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively.\n\nReturns:\n (float): the word error rate\n\nExamples:\n\n >>> predictions = [\"this is the prediction\", \"there is an other sample\"]\n >>> references = [\"this is the reference\", \"there is another one\"]\n >>> wer = datasets.load_metric(\"wer\")\n >>> wer_score = wer.compute(predictions=predictions, references=references)\n >>> print(wer_score)\n 0.5\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class UpperCAmelCase (datasets.Metric ): """simple docstring""" def _snake_case ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/jitsi/jiwer/'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/Word_error_rate''', ] , ) def _snake_case ( self , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=False ): if concatenate_texts: return compute_measures(_UpperCAmelCase , _UpperCAmelCase )["wer"] else: lowercase__: Dict = 0 lowercase__: Union[str, Any] = 0 for prediction, reference in zip(_UpperCAmelCase , _UpperCAmelCase ): lowercase__: Tuple = compute_measures(_UpperCAmelCase , _UpperCAmelCase ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
177
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __snake_case = { '''configuration_roformer''': ['''ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''RoFormerConfig''', '''RoFormerOnnxConfig'''], '''tokenization_roformer''': ['''RoFormerTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = ['''RoFormerTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = [ '''ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''RoFormerForCausalLM''', '''RoFormerForMaskedLM''', '''RoFormerForMultipleChoice''', '''RoFormerForQuestionAnswering''', '''RoFormerForSequenceClassification''', '''RoFormerForTokenClassification''', '''RoFormerLayer''', '''RoFormerModel''', '''RoFormerPreTrainedModel''', '''load_tf_weights_in_roformer''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = [ '''TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFRoFormerForCausalLM''', '''TFRoFormerForMaskedLM''', '''TFRoFormerForMultipleChoice''', '''TFRoFormerForQuestionAnswering''', '''TFRoFormerForSequenceClassification''', '''TFRoFormerForTokenClassification''', '''TFRoFormerLayer''', '''TFRoFormerModel''', '''TFRoFormerPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = [ '''FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FlaxRoFormerForMaskedLM''', '''FlaxRoFormerForMultipleChoice''', '''FlaxRoFormerForQuestionAnswering''', '''FlaxRoFormerForSequenceClassification''', '''FlaxRoFormerForTokenClassification''', '''FlaxRoFormerModel''', '''FlaxRoFormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_roformer import ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, RoFormerConfig, RoFormerOnnxConfig from .tokenization_roformer import RoFormerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roformer_fast import RoFormerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roformer import ( ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, RoFormerForCausalLM, RoFormerForMaskedLM, RoFormerForMultipleChoice, RoFormerForQuestionAnswering, RoFormerForSequenceClassification, RoFormerForTokenClassification, RoFormerLayer, RoFormerModel, RoFormerPreTrainedModel, load_tf_weights_in_roformer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roformer import ( TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerLayer, TFRoFormerModel, TFRoFormerPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roformer import ( FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, FlaxRoFormerPreTrainedModel, ) else: import sys __snake_case = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
153
"""simple docstring""" from pathlib import Path import fire from tqdm import tqdm def A_ ( _lowerCAmelCase : str="ro", _lowerCAmelCase : Optional[Any]="en", _lowerCAmelCase : Union[str, Any]="wmt16", _lowerCAmelCase : int=None ): """simple docstring""" try: import datasets except (ModuleNotFoundError, ImportError): raise ImportError('''run pip install datasets''' ) _a = f'{src_lang}-{tgt_lang}' print(f'Converting {dataset}-{pair}' ) _a = datasets.load_dataset(_lowerCAmelCase, _lowerCAmelCase ) if save_dir is None: _a = f'{dataset}-{pair}' _a = Path(_lowerCAmelCase ) save_dir.mkdir(exist_ok=_lowerCAmelCase ) for split in ds.keys(): print(f'Splitting {split} with {ds[split].num_rows} records' ) # to save to val.source, val.target like summary datasets _a = '''val''' if split == '''validation''' else split _a = save_dir.joinpath(f'{fn}.source' ) _a = save_dir.joinpath(f'{fn}.target' ) _a = src_path.open('''w+''' ) _a = tgt_path.open('''w+''' ) # reader is the bottleneck so writing one record at a time doesn't slow things down for x in tqdm(ds[split] ): _a = x['''translation'''] src_fp.write(ex[src_lang] + '''\n''' ) tgt_fp.write(ex[tgt_lang] + '''\n''' ) print(f'Saved {dataset} dataset to {save_dir}' ) if __name__ == "__main__": fire.Fire(download_wmt_dataset)
153
1
import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging lowercase__ : int = logging.get_logger(__name__) lowercase__ : str = "▁" lowercase__ : Optional[Any] = { "vocab_file": "vocab.json", "spm_file": "sentencepiece.bpe.model", "tokenizer_config_file": "tokenizer_config.json", } lowercase__ : Tuple = { "vocab_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json", }, "spm_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model", }, "tokenizer_config_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json", }, } lowercase__ : List[str] = { "facebook/m2m100_418M": 1024, } # fmt: off lowercase__ : Optional[Any] = { "m2m100": ["af", "am", "ar", "ast", "az", "ba", "be", "bg", "bn", "br", "bs", "ca", "ceb", "cs", "cy", "da", "de", "el", "en", "es", "et", "fa", "ff", "fi", "fr", "fy", "ga", "gd", "gl", "gu", "ha", "he", "hi", "hr", "ht", "hu", "hy", "id", "ig", "ilo", "is", "it", "ja", "jv", "ka", "kk", "km", "kn", "ko", "lb", "lg", "ln", "lo", "lt", "lv", "mg", "mk", "ml", "mn", "mr", "ms", "my", "ne", "nl", "no", "ns", "oc", "or", "pa", "pl", "ps", "pt", "ro", "ru", "sd", "si", "sk", "sl", "so", "sq", "sr", "ss", "su", "sv", "sw", "ta", "th", "tl", "tn", "tr", "uk", "ur", "uz", "vi", "wo", "xh", "yi", "yo", "zh", "zu"], "wmt21": ["en", "ha", "is", "ja", "cs", "ru", "zh", "de"] } class UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' lowerCAmelCase_ = VOCAB_FILES_NAMES lowerCAmelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ = ['''input_ids''', '''attention_mask'''] lowerCAmelCase_ = [] lowerCAmelCase_ = [] def __init__( self : Optional[int] , __lowercase : List[Any] , __lowercase : List[str] , __lowercase : List[Any]=None , __lowercase : List[str]=None , __lowercase : Any="<s>" , __lowercase : Optional[Any]="</s>" , __lowercase : List[str]="</s>" , __lowercase : Dict="<pad>" , __lowercase : Optional[Any]="<unk>" , __lowercase : Any="m2m100" , __lowercase : Optional[Dict[str, Any]] = None , __lowercase : Tuple=8 , **__lowercase : int , ): """simple docstring""" snake_case_ = {} if sp_model_kwargs is None else sp_model_kwargs snake_case_ = language_codes snake_case_ = FAIRSEQ_LANGUAGE_CODES[language_codes] snake_case_ = {lang_code: f"__{lang_code}__" for lang_code in fairseq_language_code} snake_case_ = kwargs.get("additional_special_tokens" , [] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(__lowercase ) for lang_code in fairseq_language_code if self.get_lang_token(__lowercase ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=__lowercase , tgt_lang=__lowercase , bos_token=__lowercase , eos_token=__lowercase , sep_token=__lowercase , unk_token=__lowercase , pad_token=__lowercase , language_codes=__lowercase , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=__lowercase , **__lowercase , ) snake_case_ = vocab_file snake_case_ = load_json(__lowercase ) snake_case_ = {v: k for k, v in self.encoder.items()} snake_case_ = spm_file snake_case_ = load_spm(__lowercase , self.sp_model_kwargs ) snake_case_ = len(self.encoder ) snake_case_ = { self.get_lang_token(__lowercase ): self.encoder_size + i for i, lang_code in enumerate(__lowercase ) } snake_case_ = {lang_code: self.encoder_size + i for i, lang_code in enumerate(__lowercase )} snake_case_ = {v: k for k, v in self.lang_token_to_id.items()} snake_case_ = src_lang if src_lang is not None else "en" snake_case_ = tgt_lang snake_case_ = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) snake_case_ = num_madeup_words @property def snake_case__ ( self : List[Any] ): """simple docstring""" return len(self.encoder ) + len(self.lang_token_to_id ) @property def snake_case__ ( self : Dict ): """simple docstring""" return self._src_lang @src_lang.setter def snake_case__ ( self : str , __lowercase : str ): """simple docstring""" snake_case_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def snake_case__ ( self : List[str] , __lowercase : str ): """simple docstring""" return self.sp_model.encode(__lowercase , out_type=__lowercase ) def snake_case__ ( self : List[str] , __lowercase : int ): """simple docstring""" if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(__lowercase , self.encoder[self.unk_token] ) def snake_case__ ( self : str , __lowercase : int ): """simple docstring""" if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(__lowercase , self.unk_token ) def snake_case__ ( self : Optional[Any] , __lowercase : Any ): """simple docstring""" snake_case_ = [] snake_case_ = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__lowercase ) + token snake_case_ = [] else: current_sub_tokens.append(__lowercase ) out_string += self.sp_model.decode(__lowercase ) return out_string.strip() def snake_case__ ( self : str , __lowercase : List[int] , __lowercase : Optional[List[int]] = None , __lowercase : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowercase , token_ids_a=__lowercase , already_has_special_tokens=__lowercase ) snake_case_ = [1] * len(self.prefix_tokens ) snake_case_ = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(__lowercase )) + suffix_ones return prefix_ones + ([0] * len(__lowercase )) + ([0] * len(__lowercase )) + suffix_ones def snake_case__ ( self : Dict , __lowercase : List[int] , __lowercase : Optional[List[int]] = None ): """simple docstring""" if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def snake_case__ ( self : Any ): """simple docstring""" snake_case_ = {self.convert_ids_to_tokens(__lowercase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Optional[int] ): """simple docstring""" snake_case_ = self.__dict__.copy() snake_case_ = None return state def __setstate__( self : List[Any] , __lowercase : Dict ): """simple docstring""" snake_case_ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): snake_case_ = {} snake_case_ = load_spm(self.spm_file , self.sp_model_kwargs ) def snake_case__ ( self : str , __lowercase : str , __lowercase : Optional[str] = None ): """simple docstring""" snake_case_ = Path(__lowercase ) if not save_dir.is_dir(): raise OSError(f"{save_directory} should be a directory" ) snake_case_ = save_dir / ( (filename_prefix + "-" if filename_prefix else "") + self.vocab_files_names["vocab_file"] ) snake_case_ = save_dir / ( (filename_prefix + "-" if filename_prefix else "") + self.vocab_files_names["spm_file"] ) save_json(self.encoder , __lowercase ) if os.path.abspath(self.spm_file ) != os.path.abspath(__lowercase ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , __lowercase ) elif not os.path.isfile(self.spm_file ): with open(__lowercase , "wb" ) as fi: snake_case_ = self.sp_model.serialized_model_proto() fi.write(__lowercase ) return (str(__lowercase ), str(__lowercase )) def snake_case__ ( self : List[str] , __lowercase : List[str] , __lowercase : str = "en" , __lowercase : Optional[List[str]] = None , __lowercase : str = "ro" , **__lowercase : List[str] , ): """simple docstring""" snake_case_ = src_lang snake_case_ = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(__lowercase , __lowercase , **__lowercase ) def snake_case__ ( self : List[Any] , __lowercase : Optional[Any] , __lowercase : Optional[str] , __lowercase : Optional[str] , **__lowercase : int ): """simple docstring""" if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) snake_case_ = src_lang snake_case_ = self(__lowercase , add_special_tokens=__lowercase , **__lowercase ) snake_case_ = self.get_lang_id(__lowercase ) snake_case_ = tgt_lang_id return inputs def snake_case__ ( self : List[Any] ): """simple docstring""" self.set_src_lang_special_tokens(self.src_lang ) def snake_case__ ( self : List[str] ): """simple docstring""" self.set_tgt_lang_special_tokens(self.tgt_lang ) def snake_case__ ( self : Tuple , __lowercase : str ): """simple docstring""" snake_case_ = self.get_lang_token(__lowercase ) snake_case_ = self.lang_token_to_id[lang_token] snake_case_ = [self.cur_lang_id] snake_case_ = [self.eos_token_id] def snake_case__ ( self : Any , __lowercase : str ): """simple docstring""" snake_case_ = self.get_lang_token(__lowercase ) snake_case_ = self.lang_token_to_id[lang_token] snake_case_ = [self.cur_lang_id] snake_case_ = [self.eos_token_id] def snake_case__ ( self : Union[str, Any] , __lowercase : str ): """simple docstring""" return self.lang_code_to_token[lang] def snake_case__ ( self : Union[str, Any] , __lowercase : str ): """simple docstring""" snake_case_ = self.get_lang_token(__lowercase ) return self.lang_token_to_id[lang_token] def lowerCamelCase__ ( _A , _A ): '''simple docstring''' snake_case_ = sentencepiece.SentencePieceProcessor(**_A ) spm.Load(str(_A ) ) return spm def lowerCamelCase__ ( _A ): '''simple docstring''' with open(_A , "r" ) as f: return json.load(_A ) def lowerCamelCase__ ( _A , _A ): '''simple docstring''' with open(_A , "w" ) as f: json.dump(_A , _A , indent=2 )
187
class UpperCAmelCase : '''simple docstring''' def __init__( self : Dict ): """simple docstring""" snake_case_ = {} # Mapping from char to TrieNode snake_case_ = False def snake_case__ ( self : Dict , __lowercase : list[str] ): """simple docstring""" for word in words: self.insert(__lowercase ) def snake_case__ ( self : List[str] , __lowercase : str ): """simple docstring""" snake_case_ = self for char in word: if char not in curr.nodes: snake_case_ = TrieNode() snake_case_ = curr.nodes[char] snake_case_ = True def snake_case__ ( self : List[Any] , __lowercase : str ): """simple docstring""" snake_case_ = self for char in word: if char not in curr.nodes: return False snake_case_ = curr.nodes[char] return curr.is_leaf def snake_case__ ( self : Optional[Any] , __lowercase : str ): """simple docstring""" def _delete(__lowercase : TrieNode , __lowercase : str , __lowercase : int ) -> bool: if index == len(__lowercase ): # If word does not exist if not curr.is_leaf: return False snake_case_ = False return len(curr.nodes ) == 0 snake_case_ = word[index] snake_case_ = curr.nodes.get(__lowercase ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted snake_case_ = _delete(__lowercase , __lowercase , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , __lowercase , 0 ) def lowerCamelCase__ ( _A , _A ): '''simple docstring''' if node.is_leaf: print(_A , end=" " ) for key, value in node.nodes.items(): print_words(_A , word + key ) def lowerCamelCase__ ( ): '''simple docstring''' snake_case_ = "banana bananas bandana band apple all beast".split() snake_case_ = TrieNode() root.insert_many(_A ) # print_words(root, "") assert all(root.find(_A ) for word in words ) assert root.find("banana" ) assert not root.find("bandanas" ) assert not root.find("apps" ) assert root.find("apple" ) assert root.find("all" ) root.delete("all" ) assert not root.find("all" ) root.delete("banana" ) assert not root.find("banana" ) assert root.find("bananas" ) return True def lowerCamelCase__ ( _A , _A ): '''simple docstring''' print(str(_A ) , "works!" if passes else "doesn't work :(" ) def lowerCamelCase__ ( ): '''simple docstring''' assert test_trie() def lowerCamelCase__ ( ): '''simple docstring''' print_results("Testing trie functionality" , test_trie() ) if __name__ == "__main__": main()
187
1
def lowerCamelCase__ ( _lowercase ): '''simple docstring''' for i in range(len(_lowercase ) - 1 , 0 , -1 ): UpperCAmelCase_ : Union[str, Any] = False for j in range(_lowercase , 0 , -1 ): if unsorted[j] < unsorted[j - 1]: UpperCAmelCase_ : Optional[int] = unsorted[j - 1], unsorted[j] UpperCAmelCase_ : Tuple = True for j in range(_lowercase ): if unsorted[j] > unsorted[j + 1]: UpperCAmelCase_ : Dict = unsorted[j + 1], unsorted[j] UpperCAmelCase_ : int = True if not swapped: break return unsorted if __name__ == "__main__": import doctest doctest.testmod() __a = input('Enter numbers separated by a comma:\n').strip() __a = [int(item) for item in user_input.split(',')] print(F"""{cocktail_shaker_sort(unsorted) = }""")
368
import argparse import logging import pickle from collections import Counter logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) __a = logging.getLogger(__name__) if __name__ == "__main__": __a = argparse.ArgumentParser( description='Token Counts for smoothing the masking probabilities in MLM (cf XLM/word2vec)' ) parser.add_argument( '--data_file', type=str, default='data/dump.bert-base-uncased.pickle', help='The binarized dataset.' ) parser.add_argument( '--token_counts_dump', type=str, default='data/token_counts.bert-base-uncased.pickle', help='The dump file.' ) parser.add_argument('--vocab_size', default=30_522, type=int) __a = parser.parse_args() logger.info(F"""Loading data from {args.data_file}""") with open(args.data_file, 'rb') as fp: __a = pickle.load(fp) logger.info('Counting occurrences for MLM.') __a = Counter() for tk_ids in data: counter.update(tk_ids) __a = [0] * args.vocab_size for k, v in counter.items(): __a = v logger.info(F"""Dump to {args.token_counts_dump}""") with open(args.token_counts_dump, 'wb') as handle: pickle.dump(counts, handle, protocol=pickle.HIGHEST_PROTOCOL)
235
0
'''simple docstring''' def lowerCamelCase (_SCREAMING_SNAKE_CASE : Optional[Any] , _SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : List[Any] ): def update_area_of_max_square(_SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : List[str] ) -> int: # BASE CASE if row >= rows or col >= cols: return 0 __a : List[Any] = update_area_of_max_square(SCREAMING_SNAKE_CASE_ , col + 1 ) __a : str = update_area_of_max_square(row + 1 , col + 1 ) __a : int = update_area_of_max_square(row + 1 , SCREAMING_SNAKE_CASE_ ) if mat[row][col]: __a : int = 1 + min([right, diagonal, down] ) __a : Any = max(largest_square_area[0] , SCREAMING_SNAKE_CASE_ ) return sub_problem_sol else: return 0 __a : Optional[Any] = [0] update_area_of_max_square(0 , 0 ) return largest_square_area[0] def lowerCamelCase (_SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : List[str] ): def update_area_of_max_square_using_dp_array( _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : List[str] , _SCREAMING_SNAKE_CASE : List[Any] ) -> int: if row >= rows or col >= cols: return 0 if dp_array[row][col] != -1: return dp_array[row][col] __a : Any = update_area_of_max_square_using_dp_array(SCREAMING_SNAKE_CASE_ , col + 1 , SCREAMING_SNAKE_CASE_ ) __a : Dict = update_area_of_max_square_using_dp_array(row + 1 , col + 1 , SCREAMING_SNAKE_CASE_ ) __a : Optional[int] = update_area_of_max_square_using_dp_array(row + 1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if mat[row][col]: __a : Optional[Any] = 1 + min([right, diagonal, down] ) __a : List[str] = max(largest_square_area[0] , SCREAMING_SNAKE_CASE_ ) __a : List[Any] = sub_problem_sol return sub_problem_sol else: return 0 __a : List[str] = [0] __a : List[Any] = [[-1] * cols for _ in range(SCREAMING_SNAKE_CASE_ )] update_area_of_max_square_using_dp_array(0 , 0 , SCREAMING_SNAKE_CASE_ ) return largest_square_area[0] def lowerCamelCase (_SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : Union[str, Any] ): __a : List[str] = [[0] * (cols + 1) for _ in range(rows + 1 )] __a : Any = 0 for row in range(rows - 1 , -1 , -1 ): for col in range(cols - 1 , -1 , -1 ): __a : Optional[int] = dp_array[row][col + 1] __a : Union[str, Any] = dp_array[row + 1][col + 1] __a : Optional[int] = dp_array[row + 1][col] if mat[row][col] == 1: __a : Optional[Any] = 1 + min(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) __a : Tuple = max(dp_array[row][col] , SCREAMING_SNAKE_CASE_ ) else: __a : str = 0 return largest_square_area def lowerCamelCase (_SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : Union[str, Any] ): __a : Tuple = [0] * (cols + 1) __a : Optional[Any] = [0] * (cols + 1) __a : Optional[int] = 0 for row in range(rows - 1 , -1 , -1 ): for col in range(cols - 1 , -1 , -1 ): __a : Optional[Any] = current_row[col + 1] __a : List[Any] = next_row[col + 1] __a : Tuple = next_row[col] if mat[row][col] == 1: __a : Any = 1 + min(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) __a : Optional[int] = max(current_row[col] , SCREAMING_SNAKE_CASE_ ) else: __a : Any = 0 __a : Union[str, Any] = current_row return largest_square_area if __name__ == "__main__": import doctest doctest.testmod() print(largest_square_area_in_matrix_bottom_up(2, 2, [[1, 1], [1, 1]]))
27
import os import tempfile import unittest from transformers import NezhaConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, ) from transformers.models.nezha.modeling_nezha import NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST class A__ : def __init__( self : str , a : Optional[Any] , a : int=13 , a : str=7 , a : str=True , a : List[str]=True , a : Optional[Any]=True , a : int=True , a : List[Any]=99 , a : List[Any]=32 , a : Tuple=5 , a : Any=4 , a : Optional[int]=37 , a : Tuple="gelu" , a : Any=0.1 , a : int=0.1 , a : List[Any]=128 , a : Union[str, Any]=32 , a : Union[str, Any]=16 , a : Dict=2 , a : List[Any]=0.0_2 , a : Optional[Any]=3 , a : List[Any]=4 , a : Optional[int]=None , ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = parent lowerCAmelCase__ : Dict = batch_size lowerCAmelCase__ : Optional[Any] = seq_length lowerCAmelCase__ : Optional[Any] = is_training lowerCAmelCase__ : Union[str, Any] = use_input_mask lowerCAmelCase__ : List[Any] = use_token_type_ids lowerCAmelCase__ : str = use_labels lowerCAmelCase__ : Optional[Any] = vocab_size lowerCAmelCase__ : Union[str, Any] = hidden_size lowerCAmelCase__ : Optional[int] = num_hidden_layers lowerCAmelCase__ : Optional[int] = num_attention_heads lowerCAmelCase__ : List[Any] = intermediate_size lowerCAmelCase__ : List[str] = hidden_act lowerCAmelCase__ : List[Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[int] = attention_probs_dropout_prob lowerCAmelCase__ : Dict = max_position_embeddings lowerCAmelCase__ : Any = type_vocab_size lowerCAmelCase__ : Any = type_sequence_label_size lowerCAmelCase__ : List[Any] = initializer_range lowerCAmelCase__ : Dict = num_labels lowerCAmelCase__ : Any = num_choices lowerCAmelCase__ : Union[str, Any] = scope def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase__ : Tuple = None if self.use_input_mask: lowerCAmelCase__ : Tuple = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase__ : Tuple = None if self.use_token_type_ids: lowerCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase__ : Optional[int] = None lowerCAmelCase__ : Optional[Any] = None lowerCAmelCase__ : Optional[int] = None if self.use_labels: lowerCAmelCase__ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase__ : List[Any] = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase__ : Dict = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' return NezhaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=a , initializer_range=self.initializer_range , ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' ( ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ) : List[Any] = self.prepare_config_and_inputs() lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : Tuple = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) lowerCAmelCase__ : Tuple = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def _lowerCamelCase ( self : Optional[Any] , a : Optional[int] , a : Tuple , a : Optional[int] , a : List[Any] , a : Tuple , a : List[str] , a : Any ): '''simple docstring''' lowerCAmelCase__ : List[str] = NezhaModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Dict = model(a , attention_mask=a , token_type_ids=a ) lowerCAmelCase__ : List[str] = model(a , token_type_ids=a ) lowerCAmelCase__ : Any = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _lowerCamelCase ( self : List[Any] , a : Union[str, Any] , a : Dict , a : List[Any] , a : Optional[Any] , a : int , a : Tuple , a : List[Any] , a : Tuple , a : List[str] , ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = True lowerCAmelCase__ : Optional[int] = NezhaModel(a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model( a , attention_mask=a , token_type_ids=a , encoder_hidden_states=a , encoder_attention_mask=a , ) lowerCAmelCase__ : Dict = model( a , attention_mask=a , token_type_ids=a , encoder_hidden_states=a , ) lowerCAmelCase__ : List[str] = model(a , attention_mask=a , token_type_ids=a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _lowerCamelCase ( self : Tuple , a : Optional[Any] , a : List[Any] , a : str , a : List[str] , a : Tuple , a : List[Any] , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = NezhaForMaskedLM(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Dict = model(a , attention_mask=a , token_type_ids=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self : List[Any] , a : Optional[int] , a : List[Any] , a : int , a : List[str] , a : Union[str, Any] , a : int , a : Any ): '''simple docstring''' lowerCAmelCase__ : List[Any] = NezhaForNextSentencePrediction(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : str = model( a , attention_mask=a , token_type_ids=a , labels=a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def _lowerCamelCase ( self : int , a : Optional[int] , a : str , a : List[str] , a : int , a : Dict , a : Optional[Any] , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = NezhaForPreTraining(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Optional[int] = model( a , attention_mask=a , token_type_ids=a , labels=a , next_sentence_label=a , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def _lowerCamelCase ( self : Union[str, Any] , a : Dict , a : List[str] , a : Any , a : Any , a : Union[str, Any] , a : Tuple , a : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = NezhaForQuestionAnswering(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Tuple = model( a , attention_mask=a , token_type_ids=a , start_positions=a , end_positions=a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self : Tuple , a : str , a : Union[str, Any] , a : Tuple , a : Optional[Any] , a : Dict , a : str , a : int ): '''simple docstring''' lowerCAmelCase__ : Any = self.num_labels lowerCAmelCase__ : Optional[Any] = NezhaForSequenceClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Tuple = model(a , attention_mask=a , token_type_ids=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self : List[str] , a : Dict , a : str , a : Optional[Any] , a : Optional[int] , a : List[str] , a : Dict , a : str ): '''simple docstring''' lowerCAmelCase__ : Dict = self.num_labels lowerCAmelCase__ : str = NezhaForTokenClassification(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a , attention_mask=a , token_type_ids=a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self : int , a : Tuple , a : List[Any] , a : Tuple , a : List[Any] , a : Optional[int] , a : Optional[int] , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.num_choices lowerCAmelCase__ : Any = NezhaForMultipleChoice(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : int = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : str = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : Optional[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase__ : Any = model( a , attention_mask=a , token_type_ids=a , labels=a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self : Tuple ): '''simple docstring''' lowerCAmelCase__ : int = self.prepare_config_and_inputs() ( ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ) : Any = config_and_inputs lowerCAmelCase__ : str = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( NezhaModel, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, ) if is_torch_available() else () ) lowercase = ( { 'feature-extraction': NezhaModel, 'fill-mask': NezhaForMaskedLM, 'question-answering': NezhaForQuestionAnswering, 'text-classification': NezhaForSequenceClassification, 'token-classification': NezhaForTokenClassification, 'zero-shot': NezhaForSequenceClassification, } if is_torch_available() else {} ) lowercase = True def _lowerCamelCase ( self : str , a : Tuple , a : int , a : Dict=False ): '''simple docstring''' lowerCAmelCase__ : int = super()._prepare_for_class(a , a , return_labels=a ) if return_labels: if model_class in get_values(a ): lowerCAmelCase__ : Tuple = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=a ) lowerCAmelCase__ : Optional[Any] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=a ) return inputs_dict def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = NezhaModelTester(self ) lowerCAmelCase__ : Optional[int] = ConfigTester(self , config_class=a , hidden_size=37 ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' ( ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ) : str = self.model_tester.prepare_config_and_inputs_for_decoder() lowerCAmelCase__ : str = None self.model_tester.create_and_check_model_as_decoder( a , a , a , a , a , a , a , a , a , ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*a ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_next_sequence_prediction(*a ) def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*a ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*a ) def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*a ) @slow def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' for model_name in NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Optional[Any] = NezhaModel.from_pretrained(a ) self.assertIsNotNone(a ) @slow @require_torch_gpu def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # NezhaForMultipleChoice behaves incorrectly in JIT environments. if model_class == NezhaForMultipleChoice: return lowerCAmelCase__ : Dict = True lowerCAmelCase__ : Any = model_class(config=a ) lowerCAmelCase__ : Union[str, Any] = self._prepare_for_class(a , a ) lowerCAmelCase__ : int = torch.jit.trace( a , (inputs_dict['input_ids'].to('cpu' ), inputs_dict['attention_mask'].to('cpu' )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(a , os.path.join(a , 'bert.pt' ) ) lowerCAmelCase__ : Any = torch.jit.load(os.path.join(a , 'bert.pt' ) , map_location=a ) loaded(inputs_dict['input_ids'].to(a ) , inputs_dict['attention_mask'].to(a ) ) @require_torch class A__ ( unittest.TestCase ): @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : str = NezhaModel.from_pretrained('sijunhe/nezha-cn-base' ) lowerCAmelCase__ : Any = torch.tensor([[0, 1, 2, 3, 4, 5]] ) lowerCAmelCase__ : Union[str, Any] = torch.tensor([[0, 1, 1, 1, 1, 1]] ) with torch.no_grad(): lowerCAmelCase__ : Optional[int] = model(a , attention_mask=a )[0] lowerCAmelCase__ : Union[str, Any] = torch.Size((1, 6, 768) ) self.assertEqual(output.shape , a ) lowerCAmelCase__ : Optional[int] = torch.tensor([[[0.0_6_8_5, 0.2_4_4_1, 0.1_1_0_2], [0.0_6_0_0, 0.1_9_0_6, 0.1_3_4_9], [0.0_2_2_1, 0.0_8_1_9, 0.0_5_8_6]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , a , atol=1E-4 ) ) @slow def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Any = NezhaForMaskedLM.from_pretrained('sijunhe/nezha-cn-base' ) lowerCAmelCase__ : Optional[int] = torch.tensor([[0, 1, 2, 3, 4, 5]] ) lowerCAmelCase__ : Optional[int] = torch.tensor([[1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): lowerCAmelCase__ : Optional[int] = model(a , attention_mask=a )[0] lowerCAmelCase__ : int = torch.Size((1, 6, 21_128) ) self.assertEqual(output.shape , a ) lowerCAmelCase__ : List[Any] = torch.tensor( [[-2.7_9_3_9, -1.7_9_0_2, -2.2_1_8_9], [-2.8_5_8_5, -1.8_9_0_8, -2.3_7_2_3], [-2.6_4_9_9, -1.7_7_5_0, -2.2_5_5_8]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , a , atol=1E-4 ) )
212
0
def snake_case_ (__A : str , __A : int ) -> str: """simple docstring""" __lowerCAmelCase : list[list[str]] = [[] for _ in range(__A )] __lowerCAmelCase : int = key - 1 if key <= 0: raise ValueError("""Height of grid can\'t be 0 or negative""" ) if key == 1 or len(__A ) <= key: return input_string for position, character in enumerate(__A ): __lowerCAmelCase : Optional[Any] = position % (lowest * 2) # puts it in bounds __lowerCAmelCase : Optional[Any] = min(__A , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(__A ) __lowerCAmelCase : Any = [''''''.join(__A ) for row in temp_grid] __lowerCAmelCase : Union[str, Any] = ''''''.join(__A ) return output_string def snake_case_ (__A : str , __A : int ) -> str: """simple docstring""" __lowerCAmelCase : Tuple = [] __lowerCAmelCase : Any = key - 1 if key <= 0: raise ValueError("""Height of grid can\'t be 0 or negative""" ) if key == 1: return input_string __lowerCAmelCase : list[list[str]] = [[] for _ in range(__A )] # generates template for position in range(len(__A ) ): __lowerCAmelCase : int = position % (lowest * 2) # puts it in bounds __lowerCAmelCase : Tuple = min(__A , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append("""*""" ) __lowerCAmelCase : Any = 0 for row in temp_grid: # fills in the characters __lowerCAmelCase : List[str] = input_string[counter : counter + len(__A )] grid.append(list(__A ) ) counter += len(__A ) __lowerCAmelCase : Optional[Any] = '''''' # reads as zigzag for position in range(len(__A ) ): __lowerCAmelCase : int = position % (lowest * 2) # puts it in bounds __lowerCAmelCase : List[str] = min(__A , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def snake_case_ (__A : str ) -> dict[int, str]: """simple docstring""" __lowerCAmelCase : int = {} for key_guess in range(1 , len(__A ) ): # tries every key __lowerCAmelCase : List[Any] = decrypt(__A , __A ) return results if __name__ == "__main__": import doctest doctest.testmod()
361
import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.esm.modeling_esmfold import EsmForProteinFolding class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self : Any , lowerCAmelCase : str , lowerCAmelCase : List[Any]=13 , lowerCAmelCase : List[str]=7 , lowerCAmelCase : int=False , lowerCAmelCase : List[str]=True , lowerCAmelCase : List[Any]=False , lowerCAmelCase : List[str]=False , lowerCAmelCase : Optional[int]=19 , lowerCAmelCase : Any=32 , lowerCAmelCase : int=5 , lowerCAmelCase : Optional[Any]=4 , lowerCAmelCase : Optional[Any]=37 , lowerCAmelCase : Dict="gelu" , lowerCAmelCase : Dict=0.1 , lowerCAmelCase : Dict=0.1 , lowerCAmelCase : List[str]=5_12 , lowerCAmelCase : Dict=16 , lowerCAmelCase : List[Any]=2 , lowerCAmelCase : Union[str, Any]=0.02 , lowerCAmelCase : Dict=3 , lowerCAmelCase : List[Any]=4 , lowerCAmelCase : Union[str, Any]=None , ) -> List[str]: """simple docstring""" __lowerCAmelCase : List[str] = parent __lowerCAmelCase : Tuple = batch_size __lowerCAmelCase : Union[str, Any] = seq_length __lowerCAmelCase : int = is_training __lowerCAmelCase : Dict = use_input_mask __lowerCAmelCase : Dict = use_token_type_ids __lowerCAmelCase : Optional[int] = use_labels __lowerCAmelCase : str = vocab_size __lowerCAmelCase : Optional[int] = hidden_size __lowerCAmelCase : Union[str, Any] = num_hidden_layers __lowerCAmelCase : Union[str, Any] = num_attention_heads __lowerCAmelCase : Tuple = intermediate_size __lowerCAmelCase : List[str] = hidden_act __lowerCAmelCase : Optional[int] = hidden_dropout_prob __lowerCAmelCase : Any = attention_probs_dropout_prob __lowerCAmelCase : List[str] = max_position_embeddings __lowerCAmelCase : Any = type_vocab_size __lowerCAmelCase : Union[str, Any] = type_sequence_label_size __lowerCAmelCase : List[str] = initializer_range __lowerCAmelCase : int = num_labels __lowerCAmelCase : Tuple = num_choices __lowerCAmelCase : str = scope def SCREAMING_SNAKE_CASE ( self : Dict ) -> Tuple: """simple docstring""" __lowerCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase : int = None if self.use_input_mask: __lowerCAmelCase : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCAmelCase : Any = None __lowerCAmelCase : Optional[Any] = None __lowerCAmelCase : Optional[int] = None if self.use_labels: __lowerCAmelCase : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCAmelCase : List[Any] = ids_tensor([self.batch_size] , self.num_choices ) __lowerCAmelCase : Tuple = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Tuple: """simple docstring""" __lowerCAmelCase : Tuple = EsmConfig( vocab_size=33 , hidden_size=self.hidden_size , pad_token_id=1 , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , is_folding_model=lowerCAmelCase , esmfold_config={"""trunk""": {"""num_blocks""": 2}, """fp16_esm""": False} , ) return config def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase : Any , lowerCAmelCase : Union[str, Any] , lowerCAmelCase : Any , lowerCAmelCase : Dict , lowerCAmelCase : List[Any] , lowerCAmelCase : Dict ) -> Tuple: """simple docstring""" __lowerCAmelCase : List[Any] = EsmForProteinFolding(config=lowerCAmelCase ).float() model.to(lowerCAmelCase ) model.eval() __lowerCAmelCase : Any = model(lowerCAmelCase , attention_mask=lowerCAmelCase ) __lowerCAmelCase : str = model(lowerCAmelCase ) __lowerCAmelCase : Optional[int] = model(lowerCAmelCase ) self.parent.assertEqual(result.positions.shape , (8, self.batch_size, self.seq_length, 14, 3) ) self.parent.assertEqual(result.angles.shape , (8, self.batch_size, self.seq_length, 7, 2) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> str: """simple docstring""" __lowerCAmelCase : Optional[Any] = self.prepare_config_and_inputs() ( ( __lowerCAmelCase ) ,( __lowerCAmelCase ) ,( __lowerCAmelCase ) ,( __lowerCAmelCase ) ,( __lowerCAmelCase ) ,( __lowerCAmelCase ) , ) : Optional[int] = config_and_inputs __lowerCAmelCase : int = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( a_ , a_ , unittest.TestCase ): """simple docstring""" lowerCamelCase : Tuple =False lowerCamelCase : int =(EsmForProteinFolding,) if is_torch_available() else () lowerCamelCase : int =() lowerCamelCase : Dict ={} if is_torch_available() else {} lowerCamelCase : str =False def SCREAMING_SNAKE_CASE ( self : int ) -> Any: """simple docstring""" __lowerCAmelCase : Union[str, Any] = EsmFoldModelTester(self ) __lowerCAmelCase : Optional[Any] = ConfigTester(self , config_class=lowerCAmelCase , hidden_size=37 ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> int: """simple docstring""" self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : str ) -> int: """simple docstring""" __lowerCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase ) @unittest.skip("""Does not support attention outputs""" ) def SCREAMING_SNAKE_CASE ( self : Dict ) -> List[Any]: """simple docstring""" pass @unittest.skip def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" pass @unittest.skip("""Esm does not support embedding resizing""" ) def SCREAMING_SNAKE_CASE ( self : str ) -> Tuple: """simple docstring""" pass @unittest.skip("""Esm does not support embedding resizing""" ) def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Union[str, Any]: """simple docstring""" pass @unittest.skip("""ESMFold does not support passing input embeds!""" ) def SCREAMING_SNAKE_CASE ( self : List[str] ) -> Optional[Any]: """simple docstring""" pass @unittest.skip("""ESMFold does not support head pruning.""" ) def SCREAMING_SNAKE_CASE ( self : Dict ) -> Optional[int]: """simple docstring""" pass @unittest.skip("""ESMFold does not support head pruning.""" ) def SCREAMING_SNAKE_CASE ( self : Tuple ) -> int: """simple docstring""" pass @unittest.skip("""ESMFold does not support head pruning.""" ) def SCREAMING_SNAKE_CASE ( self : Any ) -> Dict: """simple docstring""" pass @unittest.skip("""ESMFold does not support head pruning.""" ) def SCREAMING_SNAKE_CASE ( self : Any ) -> str: """simple docstring""" pass @unittest.skip("""ESMFold does not support head pruning.""" ) def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Dict: """simple docstring""" pass @unittest.skip("""ESMFold does not output hidden states in the normal way.""" ) def SCREAMING_SNAKE_CASE ( self : Tuple ) -> List[Any]: """simple docstring""" pass @unittest.skip("""ESMfold does not output hidden states in the normal way.""" ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> List[str]: """simple docstring""" pass @unittest.skip("""ESMFold only has one output format.""" ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" pass @unittest.skip("""This test doesn't work for ESMFold and doesn't test core functionality""" ) def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Optional[Any]: """simple docstring""" pass @unittest.skip("""ESMFold does not support input chunking.""" ) def SCREAMING_SNAKE_CASE ( self : Any ) -> Any: """simple docstring""" pass @unittest.skip("""ESMFold doesn't respect you and it certainly doesn't respect your initialization arguments.""" ) def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" pass @unittest.skip("""ESMFold doesn't support torchscript compilation.""" ) def SCREAMING_SNAKE_CASE ( self : int ) -> Dict: """simple docstring""" pass @unittest.skip("""ESMFold doesn't support torchscript compilation.""" ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" pass @unittest.skip("""ESMFold doesn't support torchscript compilation.""" ) def SCREAMING_SNAKE_CASE ( self : int ) -> Any: """simple docstring""" pass @unittest.skip("""ESMFold doesn't support data parallel.""" ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> int: """simple docstring""" pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def SCREAMING_SNAKE_CASE ( self : List[Any] ) -> List[Any]: """simple docstring""" pass @require_torch class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" @slow def SCREAMING_SNAKE_CASE ( self : Any ) -> List[str]: """simple docstring""" __lowerCAmelCase : Dict = EsmForProteinFolding.from_pretrained("""facebook/esmfold_v1""" ).float() model.eval() __lowerCAmelCase : Union[str, Any] = torch.tensor([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) __lowerCAmelCase : str = model(lowerCAmelCase )["""positions"""] __lowerCAmelCase : str = torch.tensor([2.5828, 0.7993, -10.9334] , dtype=torch.floataa ) self.assertTrue(torch.allclose(position_outputs[0, 0, 0, 0] , lowerCAmelCase , atol=1e-4 ) )
139
0
'''simple docstring''' from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef import datasets __lowerCAmelCase = '''\ @inproceedings{wang2019glue, title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding}, author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.}, note={In the Proceedings of ICLR.}, year={2019} } ''' __lowerCAmelCase = '''\ GLUE, the General Language Understanding Evaluation benchmark (https://gluebenchmark.com/) is a collection of resources for training, evaluating, and analyzing natural language understanding systems. ''' __lowerCAmelCase = ''' Compute GLUE evaluation metric associated to each GLUE dataset. Args: predictions: list of predictions to score. Each translation should be tokenized into a list of tokens. references: list of lists of references for each translation. Each reference should be tokenized into a list of tokens. Returns: depending on the GLUE subset, one or several of: "accuracy": Accuracy "f1": F1 score "pearson": Pearson Correlation "spearmanr": Spearman Correlation "matthews_correlation": Matthew Correlation Examples: >>> glue_metric = datasets.load_metric(\'glue\', \'sst2\') # \'sst2\' or any of ["mnli", "mnli_mismatched", "mnli_matched", "qnli", "rte", "wnli", "hans"] >>> references = [0, 1] >>> predictions = [0, 1] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0} >>> glue_metric = datasets.load_metric(\'glue\', \'mrpc\') # \'mrpc\' or \'qqp\' >>> references = [0, 1] >>> predictions = [0, 1] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0, \'f1\': 1.0} >>> glue_metric = datasets.load_metric(\'glue\', \'stsb\') >>> references = [0., 1., 2., 3., 4., 5.] >>> predictions = [0., 1., 2., 3., 4., 5.] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print({"pearson": round(results["pearson"], 2), "spearmanr": round(results["spearmanr"], 2)}) {\'pearson\': 1.0, \'spearmanr\': 1.0} >>> glue_metric = datasets.load_metric(\'glue\', \'cola\') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'matthews_correlation\': 1.0} ''' def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> List[Any]: return float((preds == labels).mean() ) def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> Tuple: _a : Optional[Any] = simple_accuracy(lowerCAmelCase_ , lowerCAmelCase_ ) _a : Union[str, Any] = float(fa_score(y_true=lowerCAmelCase_ , y_pred=lowerCAmelCase_ ) ) return { "accuracy": acc, "f1": fa, } def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> List[Any]: _a : Tuple = float(pearsonr(lowerCAmelCase_ , lowerCAmelCase_ )[0] ) _a : str = float(spearmanr(lowerCAmelCase_ , lowerCAmelCase_ )[0] ) return { "pearson": pearson_corr, "spearmanr": spearman_corr, } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __magic_name__ ( datasets.Metric ): def __lowercase ( self : str ): if self.config_name not in [ "sst2", "mnli", "mnli_mismatched", "mnli_matched", "cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans", ]: raise KeyError( 'You should supply a configuration name selected in ' '["sst2", "mnli", "mnli_mismatched", "mnli_matched", ' '"cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans"]' ) return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { 'predictions': datasets.Value('int64' if self.config_name != 'stsb' else 'float32' ), 'references': datasets.Value('int64' if self.config_name != 'stsb' else 'float32' ), } ) ,codebase_urls=[] ,reference_urls=[] ,format='numpy' ,) def __lowercase ( self : int ,_UpperCAmelCase : List[str] ,_UpperCAmelCase : List[Any] ): if self.config_name == "cola": return {"matthews_correlation": matthews_corrcoef(_UpperCAmelCase ,_UpperCAmelCase )} elif self.config_name == "stsb": return pearson_and_spearman(_UpperCAmelCase ,_UpperCAmelCase ) elif self.config_name in ["mrpc", "qqp"]: return acc_and_fa(_UpperCAmelCase ,_UpperCAmelCase ) elif self.config_name in ["sst2", "mnli", "mnli_mismatched", "mnli_matched", "qnli", "rte", "wnli", "hans"]: return {"accuracy": simple_accuracy(_UpperCAmelCase ,_UpperCAmelCase )} else: raise KeyError( 'You should supply a configuration name selected in ' '["sst2", "mnli", "mnli_mismatched", "mnli_matched", ' '"cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans"]' )
89
'''simple docstring''' import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all image processors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...image_processing_utils import ImageProcessingMixin from ...utils import CONFIG_NAME, IMAGE_PROCESSOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = OrderedDict( [ ('''align''', '''EfficientNetImageProcessor'''), ('''beit''', '''BeitImageProcessor'''), ('''bit''', '''BitImageProcessor'''), ('''blip''', '''BlipImageProcessor'''), ('''blip-2''', '''BlipImageProcessor'''), ('''bridgetower''', '''BridgeTowerImageProcessor'''), ('''chinese_clip''', '''ChineseCLIPImageProcessor'''), ('''clip''', '''CLIPImageProcessor'''), ('''clipseg''', '''ViTImageProcessor'''), ('''conditional_detr''', '''ConditionalDetrImageProcessor'''), ('''convnext''', '''ConvNextImageProcessor'''), ('''convnextv2''', '''ConvNextImageProcessor'''), ('''cvt''', '''ConvNextImageProcessor'''), ('''data2vec-vision''', '''BeitImageProcessor'''), ('''deformable_detr''', '''DeformableDetrImageProcessor'''), ('''deit''', '''DeiTImageProcessor'''), ('''deta''', '''DetaImageProcessor'''), ('''detr''', '''DetrImageProcessor'''), ('''dinat''', '''ViTImageProcessor'''), ('''donut-swin''', '''DonutImageProcessor'''), ('''dpt''', '''DPTImageProcessor'''), ('''efficientformer''', '''EfficientFormerImageProcessor'''), ('''efficientnet''', '''EfficientNetImageProcessor'''), ('''flava''', '''FlavaImageProcessor'''), ('''focalnet''', '''BitImageProcessor'''), ('''git''', '''CLIPImageProcessor'''), ('''glpn''', '''GLPNImageProcessor'''), ('''groupvit''', '''CLIPImageProcessor'''), ('''imagegpt''', '''ImageGPTImageProcessor'''), ('''instructblip''', '''BlipImageProcessor'''), ('''layoutlmv2''', '''LayoutLMv2ImageProcessor'''), ('''layoutlmv3''', '''LayoutLMv3ImageProcessor'''), ('''levit''', '''LevitImageProcessor'''), ('''mask2former''', '''Mask2FormerImageProcessor'''), ('''maskformer''', '''MaskFormerImageProcessor'''), ('''mgp-str''', '''ViTImageProcessor'''), ('''mobilenet_v1''', '''MobileNetV1ImageProcessor'''), ('''mobilenet_v2''', '''MobileNetV2ImageProcessor'''), ('''mobilevit''', '''MobileViTImageProcessor'''), ('''mobilevit''', '''MobileViTImageProcessor'''), ('''mobilevitv2''', '''MobileViTImageProcessor'''), ('''nat''', '''ViTImageProcessor'''), ('''oneformer''', '''OneFormerImageProcessor'''), ('''owlvit''', '''OwlViTImageProcessor'''), ('''perceiver''', '''PerceiverImageProcessor'''), ('''pix2struct''', '''Pix2StructImageProcessor'''), ('''poolformer''', '''PoolFormerImageProcessor'''), ('''regnet''', '''ConvNextImageProcessor'''), ('''resnet''', '''ConvNextImageProcessor'''), ('''sam''', '''SamImageProcessor'''), ('''segformer''', '''SegformerImageProcessor'''), ('''swiftformer''', '''ViTImageProcessor'''), ('''swin''', '''ViTImageProcessor'''), ('''swin2sr''', '''Swin2SRImageProcessor'''), ('''swinv2''', '''ViTImageProcessor'''), ('''table-transformer''', '''DetrImageProcessor'''), ('''timesformer''', '''VideoMAEImageProcessor'''), ('''tvlt''', '''TvltImageProcessor'''), ('''upernet''', '''SegformerImageProcessor'''), ('''van''', '''ConvNextImageProcessor'''), ('''videomae''', '''VideoMAEImageProcessor'''), ('''vilt''', '''ViltImageProcessor'''), ('''vit''', '''ViTImageProcessor'''), ('''vit_hybrid''', '''ViTHybridImageProcessor'''), ('''vit_mae''', '''ViTImageProcessor'''), ('''vit_msn''', '''ViTImageProcessor'''), ('''xclip''', '''CLIPImageProcessor'''), ('''yolos''', '''YolosImageProcessor'''), ] ) __lowerCAmelCase = _LazyAutoMapping(CONFIG_MAPPING_NAMES, IMAGE_PROCESSOR_MAPPING_NAMES) def __lowerCamelCase ( lowerCAmelCase_ ) -> Optional[Any]: for module_name, extractors in IMAGE_PROCESSOR_MAPPING_NAMES.items(): if class_name in extractors: _a : List[Any] = model_type_to_module_name(lowerCAmelCase_ ) _a : Optional[Any] = importlib.import_module(f""".{module_name}""" , 'transformers.models' ) try: return getattr(lowerCAmelCase_ , lowerCAmelCase_ ) except AttributeError: continue for _, extractor in IMAGE_PROCESSOR_MAPPING._extra_content.items(): if getattr(lowerCAmelCase_ , '__name__' , lowerCAmelCase_ ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. _a : Dict = importlib.import_module('transformers' ) if hasattr(lowerCAmelCase_ , lowerCAmelCase_ ): return getattr(lowerCAmelCase_ , lowerCAmelCase_ ) return None def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = False , lowerCAmelCase_ = False , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = False , **lowerCAmelCase_ , ) -> Tuple: _a : List[str] = get_file_from_repo( lowerCAmelCase_ , lowerCAmelCase_ , cache_dir=lowerCAmelCase_ , force_download=lowerCAmelCase_ , resume_download=lowerCAmelCase_ , proxies=lowerCAmelCase_ , use_auth_token=lowerCAmelCase_ , revision=lowerCAmelCase_ , local_files_only=lowerCAmelCase_ , ) if resolved_config_file is None: logger.info( 'Could not locate the image processor configuration file, will try to use the model config instead.' ) return {} with open(lowerCAmelCase_ , encoding='utf-8' ) as reader: return json.load(lowerCAmelCase_ ) class __magic_name__ : def __init__( self : List[str] ): raise EnvironmentError( 'AutoImageProcessor is designed to be instantiated ' 'using the `AutoImageProcessor.from_pretrained(pretrained_model_name_or_path)` method.' ) @classmethod @replace_list_option_in_docstrings(_UpperCAmelCase ) def __lowercase ( cls : Dict ,_UpperCAmelCase : Union[str, Any] ,**_UpperCAmelCase : Optional[Any] ): _a : Any = kwargs.pop('config' ,_UpperCAmelCase ) _a : Dict = kwargs.pop('trust_remote_code' ,_UpperCAmelCase ) _a : Any = True _a , _a : Tuple = ImageProcessingMixin.get_image_processor_dict(_UpperCAmelCase ,**_UpperCAmelCase ) _a : List[Any] = config_dict.get('image_processor_type' ,_UpperCAmelCase ) _a : int = None if "AutoImageProcessor" in config_dict.get('auto_map' ,{} ): _a : Any = config_dict['auto_map']['AutoImageProcessor'] # If we still don't have the image processor class, check if we're loading from a previous feature extractor config # and if so, infer the image processor class from there. if image_processor_class is None and image_processor_auto_map is None: _a : List[Any] = config_dict.pop('feature_extractor_type' ,_UpperCAmelCase ) if feature_extractor_class is not None: logger.warning( 'Could not find image processor class in the image processor config or the model config. Loading' ' based on pattern matching with the model\'s feature extractor configuration.' ) _a : Optional[int] = feature_extractor_class.replace('FeatureExtractor' ,'ImageProcessor' ) if "AutoFeatureExtractor" in config_dict.get('auto_map' ,{} ): _a : List[Any] = config_dict['auto_map']['AutoFeatureExtractor'] _a : List[str] = feature_extractor_auto_map.replace('FeatureExtractor' ,'ImageProcessor' ) logger.warning( 'Could not find image processor auto map in the image processor config or the model config.' ' Loading based on pattern matching with the model\'s feature extractor configuration.' ) # If we don't find the image processor class in the image processor config, let's try the model config. if image_processor_class is None and image_processor_auto_map is None: if not isinstance(_UpperCAmelCase ,_UpperCAmelCase ): _a : Dict = AutoConfig.from_pretrained(_UpperCAmelCase ,**_UpperCAmelCase ) # It could be in `config.image_processor_type`` _a : Optional[int] = getattr(_UpperCAmelCase ,'image_processor_type' ,_UpperCAmelCase ) if hasattr(_UpperCAmelCase ,'auto_map' ) and "AutoImageProcessor" in config.auto_map: _a : Union[str, Any] = config.auto_map['AutoImageProcessor'] if image_processor_class is not None: _a : Optional[int] = image_processor_class_from_name(_UpperCAmelCase ) _a : List[str] = image_processor_auto_map is not None _a : Optional[int] = image_processor_class is not None or type(_UpperCAmelCase ) in IMAGE_PROCESSOR_MAPPING _a : Optional[int] = resolve_trust_remote_code( _UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ) if has_remote_code and trust_remote_code: _a : Dict = get_class_from_dynamic_module( _UpperCAmelCase ,_UpperCAmelCase ,**_UpperCAmelCase ) _a : int = kwargs.pop('code_revision' ,_UpperCAmelCase ) if os.path.isdir(_UpperCAmelCase ): image_processor_class.register_for_auto_class() return image_processor_class.from_dict(_UpperCAmelCase ,**_UpperCAmelCase ) elif image_processor_class is not None: return image_processor_class.from_dict(_UpperCAmelCase ,**_UpperCAmelCase ) # Last try: we use the IMAGE_PROCESSOR_MAPPING. elif type(_UpperCAmelCase ) in IMAGE_PROCESSOR_MAPPING: _a : Dict = IMAGE_PROCESSOR_MAPPING[type(_UpperCAmelCase )] return image_processor_class.from_dict(_UpperCAmelCase ,**_UpperCAmelCase ) raise ValueError( F"""Unrecognized image processor in {pretrained_model_name_or_path}. Should have a """ F"""`image_processor_type` key in its {IMAGE_PROCESSOR_NAME} of {CONFIG_NAME}, or one of the following """ F"""`model_type` keys in its {CONFIG_NAME}: {', '.join(c for c in IMAGE_PROCESSOR_MAPPING_NAMES.keys() )}""" ) @staticmethod def __lowercase ( _UpperCAmelCase : Optional[Any] ,_UpperCAmelCase : Dict ): IMAGE_PROCESSOR_MAPPING.register(_UpperCAmelCase ,_UpperCAmelCase )
89
1
class a__ : def __init__( self , A ) -> None: '''simple docstring''' a = set_counts a = max(A ) a = len(A ) a = [1] * num_sets a = list(range(A ) ) def lowerCAmelCase_ ( self , A , A ) -> bool: '''simple docstring''' a = self.get_parent(A ) a = self.get_parent(A ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] a = 0 a = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 a = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] a = 0 a = src_parent a = self.set_counts[src_parent] a = max(self.max_set , A ) return True def lowerCAmelCase_ ( self , A ) -> int: '''simple docstring''' if self.parents[disj_set] == disj_set: return disj_set a = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
180
import math import unittest def SCREAMING_SNAKE_CASE ( __UpperCamelCase) -> bool: assert isinstance(__UpperCamelCase , __UpperCamelCase) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(__UpperCamelCase) + 1) , 6): if number % i == 0 or number % (i + 2) == 0: return False return True class a__ ( unittest.TestCase ): def lowerCAmelCase_ ( self ) -> List[str]: '''simple docstring''' self.assertTrue(is_prime(2 ) ) self.assertTrue(is_prime(3 ) ) self.assertTrue(is_prime(5 ) ) self.assertTrue(is_prime(7 ) ) self.assertTrue(is_prime(11 ) ) self.assertTrue(is_prime(13 ) ) self.assertTrue(is_prime(17 ) ) self.assertTrue(is_prime(19 ) ) self.assertTrue(is_prime(23 ) ) self.assertTrue(is_prime(29 ) ) def lowerCAmelCase_ ( self ) -> Tuple: '''simple docstring''' with self.assertRaises(A ): is_prime(-19 ) self.assertFalse( is_prime(0 ) , "Zero doesn't have any positive factors, primes must have exactly two." , ) self.assertFalse( is_prime(1 ) , "One only has 1 positive factor, primes must have exactly two." , ) self.assertFalse(is_prime(2 * 2 ) ) self.assertFalse(is_prime(2 * 3 ) ) self.assertFalse(is_prime(3 * 3 ) ) self.assertFalse(is_prime(3 * 5 ) ) self.assertFalse(is_prime(3 * 5 * 7 ) ) if __name__ == "__main__": unittest.main()
180
1
"""simple docstring""" class __lowerCAmelCase ( lowercase_ ): '''simple docstring''' pass class __lowerCAmelCase ( lowercase_ ): '''simple docstring''' pass class __lowerCAmelCase : '''simple docstring''' def __init__( self ): __a = [ [], [], [], ] def __UpperCAmelCase ( self , _a , _a ): try: if len(self.queues[priority] ) >= 100: raise OverflowError('''Maximum queue size is 100''' ) self.queues[priority].append(__UpperCamelCase ) except IndexError: raise ValueError('''Valid priorities are 0, 1, and 2''' ) def __UpperCAmelCase ( self ): for queue in self.queues: if queue: return queue.pop(0 ) raise UnderFlowError('''All queues are empty''' ) def __str__( self ): return "\n".join(f'''Priority {i}: {q}''' for i, q in enumerate(self.queues ) ) class __lowerCAmelCase : '''simple docstring''' def __init__( self ): __a = [] def __UpperCAmelCase ( self , _a ): if len(self.queue ) == 100: raise OverFlowError('''Maximum queue size is 100''' ) self.queue.append(__UpperCamelCase ) def __UpperCAmelCase ( self ): if not self.queue: raise UnderFlowError('''The queue is empty''' ) else: __a = min(self.queue ) self.queue.remove(__UpperCamelCase ) return data def __str__( self ): return str(self.queue ) def lowercase ( ) -> Union[str, Any]: __a = FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 100 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 128 ) print(__SCREAMING_SNAKE_CASE ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(__SCREAMING_SNAKE_CASE ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def lowercase ( ) -> Dict: __a = ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(100 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(128 ) print(__SCREAMING_SNAKE_CASE ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(__SCREAMING_SNAKE_CASE ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
45
"""simple docstring""" from __future__ import annotations def lowercase__( __SCREAMING_SNAKE_CASE : list ): if not nums: raise ValueError('List is empty' ) return sum(__SCREAMING_SNAKE_CASE ) / len(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod()
213
0
import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu lowerCAmelCase__ : int =get_tests_dir() + '''/test_data/fsmt/fsmt_val_data.json''' with io.open(filename, '''r''', encoding='''utf-8''') as f: lowerCAmelCase__ : int =json.load(f) @require_torch class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _A ( self , _A ): '''simple docstring''' return FSMTTokenizer.from_pretrained(_A ) def _A ( self , _A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = FSMTForConditionalGeneration.from_pretrained(_A ).to(_A ) if torch_device == "cuda": model.half() return model @parameterized.expand( [ ['en-ru', 2_6.0], ['ru-en', 2_2.0], ['en-de', 2_2.0], ['de-en', 2_9.0], ] ) @slow def _A ( self , _A , _A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = f"""facebook/wmt19-{pair}""" __SCREAMING_SNAKE_CASE = self.get_tokenizer(_A ) __SCREAMING_SNAKE_CASE = self.get_model(_A ) __SCREAMING_SNAKE_CASE = bleu_data[pair]['src'] __SCREAMING_SNAKE_CASE = bleu_data[pair]['tgt'] __SCREAMING_SNAKE_CASE = tokenizer(_A , return_tensors='pt' , truncation=_A , padding='longest' ).to(_A ) __SCREAMING_SNAKE_CASE = model.generate( input_ids=batch.input_ids , num_beams=8 , ) __SCREAMING_SNAKE_CASE = tokenizer.batch_decode( _A , skip_special_tokens=_A , clean_up_tokenization_spaces=_A ) __SCREAMING_SNAKE_CASE = calculate_bleu(_A , _A ) print(_A ) self.assertGreaterEqual(scores['bleu'] , _A )
118
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def __lowercase ( a__ ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = [2, 2, 6, 2] if 'tiny' in model_name else [2, 2, 18, 2] __SCREAMING_SNAKE_CASE = True if 'large' in model_name or 'huge' in model_name else False __SCREAMING_SNAKE_CASE = True if 'large' in model_name or 'huge' in model_name else False __SCREAMING_SNAKE_CASE = True if 'large' in model_name or 'huge' in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: __SCREAMING_SNAKE_CASE = [3, 3, 3, 3] __SCREAMING_SNAKE_CASE = [5, 5, 5, 5] elif "fl4" in model_name: __SCREAMING_SNAKE_CASE = [4, 4, 4, 4] __SCREAMING_SNAKE_CASE = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: __SCREAMING_SNAKE_CASE = [3, 3, 3, 3] if "lrf" in model_name: __SCREAMING_SNAKE_CASE = [3, 3, 3, 3] else: __SCREAMING_SNAKE_CASE = [2, 2, 2, 2] if "tiny" in model_name: __SCREAMING_SNAKE_CASE = 96 elif "small" in model_name: __SCREAMING_SNAKE_CASE = 96 elif "base" in model_name: __SCREAMING_SNAKE_CASE = 1_28 elif "large" in model_name: __SCREAMING_SNAKE_CASE = 1_92 elif "xlarge" in model_name: __SCREAMING_SNAKE_CASE = 2_56 elif "huge" in model_name: __SCREAMING_SNAKE_CASE = 3_52 # set label information __SCREAMING_SNAKE_CASE = 'huggingface/label-files' if "large" in model_name or "huge" in model_name: __SCREAMING_SNAKE_CASE = 'imagenet-22k-id2label.json' else: __SCREAMING_SNAKE_CASE = 'imagenet-1k-id2label.json' __SCREAMING_SNAKE_CASE = json.load(open(hf_hub_download(a__ , a__ , repo_type='dataset' ) , 'r' ) ) __SCREAMING_SNAKE_CASE = {int(a__ ): v for k, v in idalabel.items()} __SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} __SCREAMING_SNAKE_CASE = FocalNetConfig( embed_dim=a__ , depths=a__ , focal_levels=a__ , focal_windows=a__ , use_conv_embed=a__ , idalabel=a__ , labelaid=a__ , use_post_layernorm=a__ , use_layerscale=a__ , ) return config def __lowercase ( a__ ) -> Any: if "patch_embed.proj" in name: __SCREAMING_SNAKE_CASE = name.replace('patch_embed.proj' , 'embeddings.patch_embeddings.projection' ) if "patch_embed.norm" in name: __SCREAMING_SNAKE_CASE = name.replace('patch_embed.norm' , 'embeddings.norm' ) if "layers" in name: __SCREAMING_SNAKE_CASE = 'encoder.' + name if "encoder.layers" in name: __SCREAMING_SNAKE_CASE = name.replace('encoder.layers' , 'encoder.stages' ) if "downsample.proj" in name: __SCREAMING_SNAKE_CASE = name.replace('downsample.proj' , 'downsample.projection' ) if "blocks" in name: __SCREAMING_SNAKE_CASE = name.replace('blocks' , 'layers' ) if "modulation.f.weight" in name or "modulation.f.bias" in name: __SCREAMING_SNAKE_CASE = name.replace('modulation.f' , 'modulation.projection_in' ) if "modulation.h.weight" in name or "modulation.h.bias" in name: __SCREAMING_SNAKE_CASE = name.replace('modulation.h' , 'modulation.projection_context' ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: __SCREAMING_SNAKE_CASE = name.replace('modulation.proj' , 'modulation.projection_out' ) if name == "norm.weight": __SCREAMING_SNAKE_CASE = 'layernorm.weight' if name == "norm.bias": __SCREAMING_SNAKE_CASE = 'layernorm.bias' if "head" in name: __SCREAMING_SNAKE_CASE = name.replace('head' , 'classifier' ) else: __SCREAMING_SNAKE_CASE = 'focalnet.' + name return name def __lowercase ( a__ , a__ , a__=False ) -> Dict: # fmt: off __SCREAMING_SNAKE_CASE = { 'focalnet-tiny': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth', 'focalnet-tiny-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth', 'focalnet-small': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth', 'focalnet-small-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth', 'focalnet-base': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth', 'focalnet-base-lrf': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth', 'focalnet-large-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth', 'focalnet-large-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth', 'focalnet-xlarge-lrf-fl3': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth', 'focalnet-xlarge-lrf-fl4': 'https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth', } # fmt: on __SCREAMING_SNAKE_CASE = model_name_to_url[model_name] print('Checkpoint URL: ' , a__ ) __SCREAMING_SNAKE_CASE = torch.hub.load_state_dict_from_url(a__ , map_location='cpu' )['model'] # rename keys for key in state_dict.copy().keys(): __SCREAMING_SNAKE_CASE = state_dict.pop(a__ ) __SCREAMING_SNAKE_CASE = val __SCREAMING_SNAKE_CASE = get_focalnet_config(a__ ) __SCREAMING_SNAKE_CASE = FocalNetForImageClassification(a__ ) model.eval() # load state dict model.load_state_dict(a__ ) # verify conversion __SCREAMING_SNAKE_CASE = 'http://images.cocodataset.org/val2017/000000039769.jpg' __SCREAMING_SNAKE_CASE = BitImageProcessor( do_resize=a__ , size={'shortest_edge': 2_56} , resample=PILImageResampling.BILINEAR , do_center_crop=a__ , crop_size=2_24 , do_normalize=a__ , image_mean=a__ , image_std=a__ , ) __SCREAMING_SNAKE_CASE = Image.open(requests.get(a__ , stream=a__ ).raw ) __SCREAMING_SNAKE_CASE = processor(images=a__ , return_tensors='pt' ) __SCREAMING_SNAKE_CASE = transforms.Compose( [ transforms.Resize(2_56 ), transforms.CenterCrop(2_24 ), transforms.ToTensor(), transforms.Normalize(mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ), ] ) __SCREAMING_SNAKE_CASE = image_transforms(a__ ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , a__ , atol=1E-4 ) __SCREAMING_SNAKE_CASE = model(**a__ ) __SCREAMING_SNAKE_CASE = outputs.logits.argmax(-1 ).item() print('Predicted class:' , model.config.idalabel[predicted_class_idx] ) print('First values of logits:' , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": __SCREAMING_SNAKE_CASE = torch.tensor([0.2166, -0.4368, 0.2191] ) elif model_name == "focalnet-tiny-lrf": __SCREAMING_SNAKE_CASE = torch.tensor([1.1669, 0.0125, -0.1695] ) elif model_name == "focalnet-small": __SCREAMING_SNAKE_CASE = torch.tensor([0.4917, -0.0430, 0.1341] ) elif model_name == "focalnet-small-lrf": __SCREAMING_SNAKE_CASE = torch.tensor([-0.2588, -0.5342, -0.2331] ) elif model_name == "focalnet-base": __SCREAMING_SNAKE_CASE = torch.tensor([-0.1655, -0.4090, -0.1730] ) elif model_name == "focalnet-base-lrf": __SCREAMING_SNAKE_CASE = torch.tensor([0.5306, -0.0483, -0.3928] ) assert torch.allclose(outputs.logits[0, :3] , a__ , atol=1E-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(f"""Saving model and processor of {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(a__ ) processor.save_pretrained(a__ ) if push_to_hub: print(f"""Pushing model and processor of {model_name} to the hub...""" ) model.push_to_hub(f"""{model_name}""" ) processor.push_to_hub(f"""{model_name}""" ) if __name__ == "__main__": lowerCAmelCase__ : int =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''focalnet-tiny''', type=str, help='''Name of the FocalNet model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model and processor to the hub.''', ) lowerCAmelCase__ : List[Any] =parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
118
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case_ : Optional[Any] = { 'configuration_xlm_roberta_xl': [ 'XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMRobertaXLConfig', 'XLMRobertaXLOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Optional[int] = [ 'XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMRobertaXLForCausalLM', 'XLMRobertaXLForMaskedLM', 'XLMRobertaXLForMultipleChoice', 'XLMRobertaXLForQuestionAnswering', 'XLMRobertaXLForSequenceClassification', 'XLMRobertaXLForTokenClassification', 'XLMRobertaXLModel', 'XLMRobertaXLPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaXLConfig, XLMRobertaXLOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaXLForCausalLM, XLMRobertaXLForMaskedLM, XLMRobertaXLForMultipleChoice, XLMRobertaXLForQuestionAnswering, XLMRobertaXLForSequenceClassification, XLMRobertaXLForTokenClassification, XLMRobertaXLModel, XLMRobertaXLPreTrainedModel, ) else: import sys snake_case_ : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure)
83
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: snake_case_ : List[Any] = None snake_case_ : str = logging.get_logger(__name__) snake_case_ : Dict = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} snake_case_ : List[Any] = { 'vocab_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/spiece.model', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/spiece.model', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/spiece.model', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/spiece.model', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model', }, 'tokenizer_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json', }, } snake_case_ : List[str] = { 'albert-base-v1': 512, 'albert-large-v1': 512, 'albert-xlarge-v1': 512, 'albert-xxlarge-v1': 512, 'albert-base-v2': 512, 'albert-large-v2': 512, 'albert-xlarge-v2': 512, 'albert-xxlarge-v2': 512, } snake_case_ : List[str] = '▁' class lowercase__ ( lowercase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = AlbertTokenizer def __init__( self : Tuple ,lowerCamelCase__ : Optional[int]=None ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : int=True ,lowerCamelCase__ : Any=False ,lowerCamelCase__ : Optional[int]="[CLS]" ,lowerCamelCase__ : Union[str, Any]="[SEP]" ,lowerCamelCase__ : Optional[int]="<unk>" ,lowerCamelCase__ : str="[SEP]" ,lowerCamelCase__ : List[Any]="<pad>" ,lowerCamelCase__ : Dict="[CLS]" ,lowerCamelCase__ : int="[MASK]" ,**lowerCamelCase__ : Any ,): '''simple docstring''' # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. _UpperCamelCase : Dict = ( AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ,normalized=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else mask_token ) super().__init__( lowerCamelCase__ ,tokenizer_file=lowerCamelCase__ ,do_lower_case=lowerCamelCase__ ,remove_space=lowerCamelCase__ ,keep_accents=lowerCamelCase__ ,bos_token=lowerCamelCase__ ,eos_token=lowerCamelCase__ ,unk_token=lowerCamelCase__ ,sep_token=lowerCamelCase__ ,pad_token=lowerCamelCase__ ,cls_token=lowerCamelCase__ ,mask_token=lowerCamelCase__ ,**lowerCamelCase__ ,) _UpperCamelCase : Tuple = do_lower_case _UpperCamelCase : str = remove_space _UpperCamelCase : Optional[Any] = keep_accents _UpperCamelCase : Dict = vocab_file _UpperCamelCase : Dict = False if not self.vocab_file else True def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : List[Any] = [self.sep_token_id] _UpperCamelCase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : int = [self.sep_token_id] _UpperCamelCase : int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(lowerCamelCase__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _UpperCamelCase : Dict = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ): copyfile(self.vocab_file ,lowerCamelCase__ ) return (out_vocab_file,)
83
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import ( DiffusionPipeline, UnCLIPImageVariationPipeline, UnCLIPScheduler, UNetaDConditionModel, UNetaDModel, ) from diffusers.pipelines.unclip.text_proj import UnCLIPTextProjModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, load_image, require_torch_gpu, skip_mps from ..pipeline_params import IMAGE_VARIATION_BATCH_PARAMS, IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _lowerCamelCase ( lowercase__ , unittest.TestCase ): '''simple docstring''' A_ : Union[str, Any] = UnCLIPImageVariationPipeline A_ : str = IMAGE_VARIATION_PARAMS - {"""height""", """width""", """guidance_scale"""} A_ : Optional[Any] = IMAGE_VARIATION_BATCH_PARAMS A_ : Union[str, Any] = [ """generator""", """return_dict""", """decoder_num_inference_steps""", """super_res_num_inference_steps""", ] A_ : Union[str, Any] = False @property def __lowerCAmelCase ( self : int ) -> Any: return 32 @property def __lowerCAmelCase ( self : Union[str, Any] ) -> Tuple: return 32 @property def __lowerCAmelCase ( self : List[Any] ) -> Dict: return self.time_input_dim @property def __lowerCAmelCase ( self : Tuple ) -> Optional[Any]: return self.time_input_dim * 4 @property def __lowerCAmelCase ( self : int ) -> str: return 100 @property def __lowerCAmelCase ( self : List[Any] ) -> Tuple: __magic_name__ : Optional[int] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def __lowerCAmelCase ( self : Dict ) -> List[str]: torch.manual_seed(0 ) __magic_name__ : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(_A ) @property def __lowerCAmelCase ( self : Dict ) -> Tuple: torch.manual_seed(0 ) __magic_name__ : Dict = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , ) return CLIPVisionModelWithProjection(_A ) @property def __lowerCAmelCase ( self : List[Any] ) -> str: torch.manual_seed(0 ) __magic_name__ : Dict = { 'clip_embeddings_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'cross_attention_dim': self.cross_attention_dim, } __magic_name__ : List[str] = UnCLIPTextProjModel(**_A ) return model @property def __lowerCAmelCase ( self : str ) -> Union[str, Any]: torch.manual_seed(0 ) __magic_name__ : str = { 'sample_size': 32, # RGB in channels 'in_channels': 3, # Out channels is double in channels because predicts mean and variance 'out_channels': 6, 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': 'identity', } __magic_name__ : Dict = UNetaDConditionModel(**_A ) return model @property def __lowerCAmelCase ( self : List[str] ) -> Union[str, Any]: return { "sample_size": 64, "layers_per_block": 1, "down_block_types": ("ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D"), "up_block_types": ("ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D"), "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "in_channels": 6, "out_channels": 3, } @property def __lowerCAmelCase ( self : Dict ) -> str: torch.manual_seed(0 ) __magic_name__ : Optional[int] = UNetaDModel(**self.dummy_super_res_kwargs ) return model @property def __lowerCAmelCase ( self : List[str] ) -> Dict: # seeded differently to get different unet than `self.dummy_super_res_first` torch.manual_seed(1 ) __magic_name__ : List[str] = UNetaDModel(**self.dummy_super_res_kwargs ) return model def __lowerCAmelCase ( self : Optional[int] ) -> Dict: __magic_name__ : Tuple = self.dummy_decoder __magic_name__ : Optional[int] = self.dummy_text_proj __magic_name__ : Any = self.dummy_text_encoder __magic_name__ : Tuple = self.dummy_tokenizer __magic_name__ : Any = self.dummy_super_res_first __magic_name__ : Optional[int] = self.dummy_super_res_last __magic_name__ : Union[str, Any] = UnCLIPScheduler( variance_type='learned_range' , prediction_type='epsilon' , num_train_timesteps=1000 , ) __magic_name__ : Optional[Any] = UnCLIPScheduler( variance_type='fixed_small_log' , prediction_type='epsilon' , num_train_timesteps=1000 , ) __magic_name__ : Dict = CLIPImageProcessor(crop_size=32 , size=32 ) __magic_name__ : Tuple = self.dummy_image_encoder return { "decoder": decoder, "text_encoder": text_encoder, "tokenizer": tokenizer, "text_proj": text_proj, "feature_extractor": feature_extractor, "image_encoder": image_encoder, "super_res_first": super_res_first, "super_res_last": super_res_last, "decoder_scheduler": decoder_scheduler, "super_res_scheduler": super_res_scheduler, } def __lowerCAmelCase ( self : List[str] , _A : List[str] , _A : Tuple=0 , _A : str=True ) -> List[Any]: __magic_name__ : Optional[int] = floats_tensor((1, 3, 32, 32) , rng=random.Random(_A ) ).to(_A ) if str(_A ).startswith('mps' ): __magic_name__ : Tuple = torch.manual_seed(_A ) else: __magic_name__ : List[str] = torch.Generator(device=_A ).manual_seed(_A ) if pil_image: __magic_name__ : Tuple = input_image * 0.5 + 0.5 __magic_name__ : Any = input_image.clamp(0 , 1 ) __magic_name__ : Any = input_image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() __magic_name__ : int = DiffusionPipeline.numpy_to_pil(_A )[0] return { "image": input_image, "generator": generator, "decoder_num_inference_steps": 2, "super_res_num_inference_steps": 2, "output_type": "np", } def __lowerCAmelCase ( self : Dict ) -> Any: __magic_name__ : Optional[Any] = 'cpu' __magic_name__ : Dict = self.get_dummy_components() __magic_name__ : List[str] = self.pipeline_class(**_A ) __magic_name__ : str = pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) __magic_name__ : List[str] = self.get_dummy_inputs(_A , pil_image=_A ) __magic_name__ : Any = pipe(**_A ) __magic_name__ : int = output.images __magic_name__ : Union[str, Any] = self.get_dummy_inputs(_A , pil_image=_A ) __magic_name__ : Optional[Any] = pipe( **_A , return_dict=_A , )[0] __magic_name__ : Dict = image[0, -3:, -3:, -1] __magic_name__ : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __magic_name__ : int = np.array( [ 0.9997, 0.0002, 0.9997, 0.9997, 0.9969, 0.0023, 0.9997, 0.9969, 0.9970, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def __lowerCAmelCase ( self : Optional[Any] ) -> List[str]: __magic_name__ : List[str] = 'cpu' __magic_name__ : Union[str, Any] = self.get_dummy_components() __magic_name__ : int = self.pipeline_class(**_A ) __magic_name__ : Dict = pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) __magic_name__ : Dict = self.get_dummy_inputs(_A , pil_image=_A ) __magic_name__ : int = pipe(**_A ) __magic_name__ : List[str] = output.images __magic_name__ : Optional[Any] = self.get_dummy_inputs(_A , pil_image=_A ) __magic_name__ : List[str] = pipe( **_A , return_dict=_A , )[0] __magic_name__ : str = image[0, -3:, -3:, -1] __magic_name__ : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __magic_name__ : Optional[Any] = np.array([0.9997, 0.0003, 0.9997, 0.9997, 0.9970, 0.0024, 0.9997, 0.9971, 0.9971] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def __lowerCAmelCase ( self : Any ) -> Optional[Any]: __magic_name__ : Optional[int] = 'cpu' __magic_name__ : Dict = self.get_dummy_components() __magic_name__ : Any = self.pipeline_class(**_A ) __magic_name__ : List[Any] = pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) __magic_name__ : Optional[Any] = self.get_dummy_inputs(_A , pil_image=_A ) __magic_name__ : Optional[int] = [ pipeline_inputs['image'], pipeline_inputs['image'], ] __magic_name__ : List[str] = pipe(**_A ) __magic_name__ : List[Any] = output.images __magic_name__ : Optional[Any] = self.get_dummy_inputs(_A , pil_image=_A ) __magic_name__ : Optional[Any] = [ tuple_pipeline_inputs['image'], tuple_pipeline_inputs['image'], ] __magic_name__ : Optional[int] = pipe( **_A , return_dict=_A , )[0] __magic_name__ : Dict = image[0, -3:, -3:, -1] __magic_name__ : str = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (2, 64, 64, 3) __magic_name__ : List[str] = np.array( [ 0.9997, 0.9989, 0.0008, 0.0021, 0.9960, 0.0018, 0.0014, 0.0002, 0.9933, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def __lowerCAmelCase ( self : Tuple ) -> Optional[int]: __magic_name__ : Optional[Any] = torch.device('cpu' ) class _lowerCamelCase : '''simple docstring''' A_ : int = 1 __magic_name__ : List[Any] = self.get_dummy_components() __magic_name__ : Optional[Any] = self.pipeline_class(**_A ) __magic_name__ : str = pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) __magic_name__ : int = torch.Generator(device=_A ).manual_seed(0 ) __magic_name__ : Union[str, Any] = pipe.decoder.dtype __magic_name__ : Optional[int] = 1 __magic_name__ : Any = ( batch_size, pipe.decoder.config.in_channels, pipe.decoder.config.sample_size, pipe.decoder.config.sample_size, ) __magic_name__ : Tuple = pipe.prepare_latents( _A , dtype=_A , device=_A , generator=_A , latents=_A , scheduler=DummyScheduler() ) __magic_name__ : List[Any] = ( batch_size, pipe.super_res_first.config.in_channels // 2, pipe.super_res_first.config.sample_size, pipe.super_res_first.config.sample_size, ) __magic_name__ : int = pipe.prepare_latents( _A , dtype=_A , device=_A , generator=_A , latents=_A , scheduler=DummyScheduler() ) __magic_name__ : Any = self.get_dummy_inputs(_A , pil_image=_A ) __magic_name__ : Union[str, Any] = pipe( **_A , decoder_latents=_A , super_res_latents=_A ).images __magic_name__ : Tuple = self.get_dummy_inputs(_A , pil_image=_A ) # Don't pass image, instead pass embedding __magic_name__ : List[Any] = pipeline_inputs.pop('image' ) __magic_name__ : Tuple = pipe.image_encoder(_A ).image_embeds __magic_name__ : Tuple = pipe( **_A , decoder_latents=_A , super_res_latents=_A , image_embeddings=_A , ).images # make sure passing text embeddings manually is identical assert np.abs(img_out_a - img_out_a ).max() < 1E-4 @skip_mps def __lowerCAmelCase ( self : Dict ) -> Optional[int]: __magic_name__ : Tuple = torch_device == 'cpu' # Check is relaxed because there is not a torch 2.0 sliced attention added kv processor __magic_name__ : int = 1E-2 self._test_attention_slicing_forward_pass( test_max_difference=_A , expected_max_diff=_A ) @skip_mps def __lowerCAmelCase ( self : List[str] ) -> int: __magic_name__ : Optional[int] = torch_device == 'cpu' __magic_name__ : Optional[Any] = True __magic_name__ : int = [ 'decoder_num_inference_steps', 'super_res_num_inference_steps', ] self._test_inference_batch_single_identical( test_max_difference=_A , relax_max_difference=_A , additional_params_copy_to_batched_inputs=_A , ) def __lowerCAmelCase ( self : Union[str, Any] ) -> Optional[int]: __magic_name__ : int = [ 'decoder_num_inference_steps', 'super_res_num_inference_steps', ] if torch_device == "mps": # TODO: MPS errors with larger batch sizes __magic_name__ : Optional[Any] = [2, 3] self._test_inference_batch_consistent( batch_sizes=_A , additional_params_copy_to_batched_inputs=_A , ) else: self._test_inference_batch_consistent( additional_params_copy_to_batched_inputs=_A ) @skip_mps def __lowerCAmelCase ( self : str ) -> Tuple: return super().test_dict_tuple_outputs_equivalent() @skip_mps def __lowerCAmelCase ( self : str ) -> List[Any]: return super().test_save_load_local() @skip_mps def __lowerCAmelCase ( self : Optional[Any] ) -> Any: return super().test_save_load_optional_components() @slow @require_torch_gpu class _lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def __lowerCAmelCase ( self : Optional[Any] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self : Union[str, Any] ) -> Dict: __magic_name__ : Dict = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/unclip/cat.png' ) __magic_name__ : Optional[int] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/unclip/karlo_v1_alpha_cat_variation_fp16.npy' ) __magic_name__ : str = UnCLIPImageVariationPipeline.from_pretrained( 'kakaobrain/karlo-v1-alpha-image-variations' , torch_dtype=torch.floataa ) __magic_name__ : Dict = pipeline.to(_A ) pipeline.set_progress_bar_config(disable=_A ) __magic_name__ : Any = torch.Generator(device='cpu' ).manual_seed(0 ) __magic_name__ : Any = pipeline( _A , generator=_A , output_type='np' , ) __magic_name__ : int = output.images[0] assert image.shape == (256, 256, 3) assert_mean_pixel_difference(_A , _A , 15 )
275
'''simple docstring''' def lowerCamelCase ( lowerCAmelCase : str , lowerCAmelCase : Any , lowerCAmelCase : Any=False ): """simple docstring""" if isinstance(lowerCAmelCase , lowerCAmelCase ) and isinstance(lowerCAmelCase , lowerCAmelCase ): __magic_name__ : str = len(set_a.intersection(lowerCAmelCase ) ) if alternative_union: __magic_name__ : List[str] = len(lowerCAmelCase ) + len(lowerCAmelCase ) else: __magic_name__ : Any = len(set_a.union(lowerCAmelCase ) ) return intersection / union if isinstance(lowerCAmelCase , (list, tuple) ) and isinstance(lowerCAmelCase , (list, tuple) ): __magic_name__ : str = [element for element in set_a if element in set_b] if alternative_union: __magic_name__ : Dict = len(lowerCAmelCase ) + len(lowerCAmelCase ) return len(lowerCAmelCase ) / union else: __magic_name__ : Any = set_a + [element for element in set_b if element not in set_a] return len(lowerCAmelCase ) / len(lowerCAmelCase ) return len(lowerCAmelCase ) / len(lowerCAmelCase ) return None if __name__ == "__main__": lowerCAmelCase :Dict = {'''a''', '''b''', '''c''', '''d''', '''e'''} lowerCAmelCase :Tuple = {'''c''', '''d''', '''e''', '''f''', '''h''', '''i'''} print(jaccard_similarity(set_a, set_b))
275
1
def SCREAMING_SNAKE_CASE__ ( ): snake_case_ : Union[str, Any] = 0 for i in range(1 , 10_01 ): total += i**i return str(__a )[-10:] if __name__ == "__main__": print(solution())
327
from collections import namedtuple import requests from lxml import html # type: ignore _SCREAMING_SNAKE_CASE = namedtuple("""covid_data""", """cases deaths recovered""") def SCREAMING_SNAKE_CASE__ ( __a = "https://www.worldometers.info/coronavirus/" ): snake_case_ : Union[str, Any] = '//div[@class = "maincounter-number"]/span/text()' return covid_data(*html.fromstring(requests.get(__a ).content ).xpath(__a ) ) _SCREAMING_SNAKE_CASE = """Total COVID-19 cases in the world: {} Total deaths due to COVID-19 in the world: {} Total COVID-19 patients recovered in the world: {}""" print(fmt.format(*covid_stats()))
327
1
"""simple docstring""" def __SCREAMING_SNAKE_CASE ( lowercase__ , lowercase__ ): """simple docstring""" return base * power(lowercase__ , (exponent - 1) ) if exponent else 1 if __name__ == "__main__": print('Raise base to the power of exponent using recursion...') __A : str = int(input('Enter the base: ').strip()) __A : str = int(input('Enter the exponent: ').strip()) __A : int = power(base, abs(exponent)) if exponent < 0: # power() does not properly deal w/ negative exponents __A : List[Any] = 1 / result print(f"""{base} to the power of {exponent} is {result}""")
357
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() __A : int = logging.get_logger(__name__) __A : Optional[int] = [ ('bert.bert', 'visual_bert'), ('bert.cls', 'cls'), ('bert.classifier', 'cls'), ('token_type_embeddings_visual', 'visual_token_type_embeddings'), ('position_embeddings_visual', 'visual_position_embeddings'), ('projection', 'visual_projection'), ] __A : Union[str, Any] = [ 'nlvr2_coco_pre_trained.th', 'nlvr2_fine_tuned.th', 'nlvr2_pre_trained.th', 'vcr_coco_pre_train.th', 'vcr_fine_tune.th', 'vcr_pre_train.th', 'vqa_coco_pre_trained.th', 'vqa_fine_tuned.th', 'vqa_pre_trained.th', ] def __SCREAMING_SNAKE_CASE ( lowercase__ ): """simple docstring""" A = torch.load(lowercase__ , map_location="cpu" ) return sd def __SCREAMING_SNAKE_CASE ( lowercase__ , lowercase__ , lowercase__=rename_keys_prefix ): """simple docstring""" A = OrderedDict() A = torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue A = key for name_pair in rename_keys_prefix: A = new_key.replace(name_pair[0] , name_pair[1] ) A = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately A = new_d["cls.predictions.bias"] return new_d @torch.no_grad() def __SCREAMING_SNAKE_CASE ( lowercase__ , lowercase__ ): """simple docstring""" assert ( checkpoint_path.split("/" )[-1] in ACCEPTABLE_CHECKPOINTS ), F"""The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.""" # Get Config if "pre" in checkpoint_path: A = "pretraining" if "vcr" in checkpoint_path: A = {"visual_embedding_dim": 512} elif "vqa_advanced" in checkpoint_path: A = {"visual_embedding_dim": 2_048} elif "vqa" in checkpoint_path: A = {"visual_embedding_dim": 2_048} elif "nlvr" in checkpoint_path: A = {"visual_embedding_dim": 1_024} else: raise NotImplementedError(F"""No implementation found for `{checkpoint_path}`.""" ) else: if "vcr" in checkpoint_path: A = {"visual_embedding_dim": 512} A = "multichoice" elif "vqa_advanced" in checkpoint_path: A = {"visual_embedding_dim": 2_048} A = "vqa_advanced" elif "vqa" in checkpoint_path: A = {"visual_embedding_dim": 2_048, "num_labels": 3_129} A = "vqa" elif "nlvr" in checkpoint_path: A = { "visual_embedding_dim": 1_024, "num_labels": 2, } A = "nlvr" A = VisualBertConfig(**lowercase__ ) # Load State Dict A = load_state_dict(lowercase__ ) A = get_new_dict(lowercase__ , lowercase__ ) if model_type == "pretraining": A = VisualBertForPreTraining(lowercase__ ) elif model_type == "vqa": A = VisualBertForQuestionAnswering(lowercase__ ) elif model_type == "nlvr": A = VisualBertForVisualReasoning(lowercase__ ) elif model_type == "multichoice": A = VisualBertForMultipleChoice(lowercase__ ) model.load_state_dict(lowercase__ ) # Save Checkpoints Path(lowercase__ ).mkdir(exist_ok=lowercase__ ) model.save_pretrained(lowercase__ ) if __name__ == "__main__": __A : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument('orig_checkpoint_path', type=str, help='A path to .th on local filesystem.') parser.add_argument('pytorch_dump_folder_path', type=str, help='Path to the output PyTorch model.') __A : Any = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
57
0
"""simple docstring""" from itertools import product def a__ ( lowerCAmelCase , lowerCAmelCase ) -> Optional[Any]: UpperCAmelCase__ : str = sides_number UpperCAmelCase__ : Optional[Any] = max_face_number * dice_number UpperCAmelCase__ : Optional[int] = [0] * (max_total + 1) UpperCAmelCase__ : int = 1 UpperCAmelCase__ : Tuple = range(_UpperCAmelCase , max_face_number + 1 ) for dice_numbers in product(_UpperCAmelCase , repeat=_UpperCAmelCase ): UpperCAmelCase__ : int = sum(_UpperCAmelCase ) totals_frequencies[total] += 1 return totals_frequencies def a__ ( ) -> int: UpperCAmelCase__ : str = total_frequency_distribution( sides_number=4 , dice_number=9 ) UpperCAmelCase__ : Any = total_frequency_distribution( sides_number=6 , dice_number=6 ) UpperCAmelCase__ : List[str] = 0 UpperCAmelCase__ : Tuple = 9 UpperCAmelCase__ : List[Any] = 4 * 9 UpperCAmelCase__ : Dict = 6 for peter_total in range(_UpperCAmelCase , max_peter_total + 1 ): peter_wins_count += peter_totals_frequencies[peter_total] * sum( colin_totals_frequencies[min_colin_total:peter_total] ) UpperCAmelCase__ : Optional[int] = (4**9) * (6**6) UpperCAmelCase__ : Optional[Any] = peter_wins_count / total_games_number UpperCAmelCase__ : str = round(_UpperCAmelCase , ndigits=7 ) return rounded_peter_win_probability if __name__ == "__main__": print(f'''{solution() = }''')
171
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging __A =logging.get_logger(__name__) if is_vision_available(): import PIL class UpperCAmelCase__ ( __UpperCamelCase ): '''simple docstring''' UpperCamelCase = ["""pixel_values"""] def __init__( self : Tuple , a_ : bool = True , a_ : Dict[str, int] = None , a_ : PILImageResampling = PILImageResampling.BICUBIC , a_ : bool = True , a_ : Dict[str, int] = None , a_ : bool = True , a_ : Union[int, float] = 1 / 2_55 , a_ : bool = True , a_ : Optional[Union[float, List[float]]] = None , a_ : Optional[Union[float, List[float]]] = None , a_ : bool = True , **a_ : str , ): '''simple docstring''' super().__init__(**a_ ) __UpperCAmelCase : List[Any] = size if size is not None else {'''shortest_edge''': 2_24} __UpperCAmelCase : List[str] = get_size_dict(a_ , default_to_square=a_ ) __UpperCAmelCase : int = crop_size if crop_size is not None else {'''height''': 2_24, '''width''': 2_24} __UpperCAmelCase : Optional[int] = get_size_dict(a_ , default_to_square=a_ , param_name='''crop_size''' ) __UpperCAmelCase : int = do_resize __UpperCAmelCase : Union[str, Any] = size __UpperCAmelCase : Union[str, Any] = resample __UpperCAmelCase : Any = do_center_crop __UpperCAmelCase : Any = crop_size __UpperCAmelCase : Any = do_rescale __UpperCAmelCase : Dict = rescale_factor __UpperCAmelCase : Union[str, Any] = do_normalize __UpperCAmelCase : Any = image_mean if image_mean is not None else OPENAI_CLIP_MEAN __UpperCAmelCase : int = image_std if image_std is not None else OPENAI_CLIP_STD __UpperCAmelCase : List[str] = do_convert_rgb def snake_case__ ( self : Optional[Any] , a_ : np.ndarray , a_ : Dict[str, int] , a_ : PILImageResampling = PILImageResampling.BICUBIC , a_ : Optional[Union[str, ChannelDimension]] = None , **a_ : Optional[int] , ): '''simple docstring''' __UpperCAmelCase : Dict = get_size_dict(a_ , default_to_square=a_ ) if "shortest_edge" not in size: raise ValueError(F'The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}' ) __UpperCAmelCase : Optional[int] = get_resize_output_image_size(a_ , size=size['''shortest_edge'''] , default_to_square=a_ ) return resize(a_ , size=a_ , resample=a_ , data_format=a_ , **a_ ) def snake_case__ ( self : Union[str, Any] , a_ : np.ndarray , a_ : Dict[str, int] , a_ : Optional[Union[str, ChannelDimension]] = None , **a_ : Dict , ): '''simple docstring''' __UpperCAmelCase : Optional[Any] = get_size_dict(a_ ) if "height" not in size or "width" not in size: raise ValueError(F'The `size` parameter must contain the keys (height, width). Got {size.keys()}' ) return center_crop(a_ , size=(size['''height'''], size['''width''']) , data_format=a_ , **a_ ) def snake_case__ ( self : Union[str, Any] , a_ : np.ndarray , a_ : Union[int, float] , a_ : Optional[Union[str, ChannelDimension]] = None , **a_ : List[str] , ): '''simple docstring''' return rescale(a_ , scale=a_ , data_format=a_ , **a_ ) def snake_case__ ( self : Optional[Any] , a_ : np.ndarray , a_ : Union[float, List[float]] , a_ : Union[float, List[float]] , a_ : Optional[Union[str, ChannelDimension]] = None , **a_ : Dict , ): '''simple docstring''' return normalize(a_ , mean=a_ , std=a_ , data_format=a_ , **a_ ) def snake_case__ ( self : Any , a_ : ImageInput , a_ : bool = None , a_ : Dict[str, int] = None , a_ : PILImageResampling = None , a_ : bool = None , a_ : int = None , a_ : bool = None , a_ : float = None , a_ : bool = None , a_ : Optional[Union[float, List[float]]] = None , a_ : Optional[Union[float, List[float]]] = None , a_ : bool = None , a_ : Optional[Union[str, TensorType]] = None , a_ : Optional[ChannelDimension] = ChannelDimension.FIRST , **a_ : Dict , ): '''simple docstring''' __UpperCAmelCase : Tuple = do_resize if do_resize is not None else self.do_resize __UpperCAmelCase : Optional[Any] = size if size is not None else self.size __UpperCAmelCase : Dict = get_size_dict(a_ , param_name='''size''' , default_to_square=a_ ) __UpperCAmelCase : int = resample if resample is not None else self.resample __UpperCAmelCase : Optional[Any] = do_center_crop if do_center_crop is not None else self.do_center_crop __UpperCAmelCase : Any = crop_size if crop_size is not None else self.crop_size __UpperCAmelCase : Dict = get_size_dict(a_ , param_name='''crop_size''' , default_to_square=a_ ) __UpperCAmelCase : int = do_rescale if do_rescale is not None else self.do_rescale __UpperCAmelCase : Optional[int] = rescale_factor if rescale_factor is not None else self.rescale_factor __UpperCAmelCase : List[str] = do_normalize if do_normalize is not None else self.do_normalize __UpperCAmelCase : Optional[int] = image_mean if image_mean is not None else self.image_mean __UpperCAmelCase : Tuple = image_std if image_std is not None else self.image_std __UpperCAmelCase : List[str] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb __UpperCAmelCase : List[str] = make_list_of_images(a_ ) if not valid_images(a_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # PIL RGBA images are converted to RGB if do_convert_rgb: __UpperCAmelCase : Optional[Any] = [convert_to_rgb(a_ ) for image in images] # All transformations expect numpy arrays. __UpperCAmelCase : int = [to_numpy_array(a_ ) for image in images] if do_resize: __UpperCAmelCase : int = [self.resize(image=a_ , size=a_ , resample=a_ ) for image in images] if do_center_crop: __UpperCAmelCase : List[str] = [self.center_crop(image=a_ , size=a_ ) for image in images] if do_rescale: __UpperCAmelCase : Dict = [self.rescale(image=a_ , scale=a_ ) for image in images] if do_normalize: __UpperCAmelCase : Optional[int] = [self.normalize(image=a_ , mean=a_ , std=a_ ) for image in images] __UpperCAmelCase : Optional[int] = [to_channel_dimension_format(a_ , a_ ) for image in images] __UpperCAmelCase : Union[str, Any] = {'''pixel_values''': images} return BatchFeature(data=a_ , tensor_type=a_ )
226
0
import os import pytest from attr import dataclass lowerCamelCase__ = """us-east-1""" # defaults region @dataclass class A__ : lowercase = 42 lowercase = 'arn:aws:iam::558105141721:role/sagemaker_execution_role' lowercase = { 'task_name': 'mnli', 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'do_train': True, 'do_eval': True, 'do_predict': True, 'output_dir': '/opt/ml/model', 'overwrite_output_dir': True, 'max_steps': 500, 'save_steps': 5500, } lowercase = {**hyperparameters, 'max_steps': 1000} @property def _lowerCamelCase ( self : str ): '''simple docstring''' if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' return f'''{self.framework}-transfromers-test''' @property def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' return f'''./tests/sagemaker/scripts/{self.framework}''' @property def _lowerCamelCase ( self : Dict ): '''simple docstring''' if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='class' ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> Optional[int]: lowerCAmelCase__ : Any = SageMakerTestEnvironment(framework=request.cls.framework )
367
# Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position lowerCamelCase__ = """2.13.1""" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("""3.7"""): raise ImportWarning( """To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition.""" ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( """To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n""" """If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.""" ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip lowerCamelCase__ = concatenate_datasets lowerCamelCase__ = DownloadConfig lowerCamelCase__ = DownloadManager lowerCamelCase__ = DownloadMode lowerCamelCase__ = DownloadConfig lowerCamelCase__ = DownloadMode lowerCamelCase__ = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
307
0
_lowerCAmelCase : Optional[Any] = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(100_000)] def __snake_case ( _lowerCAmelCase : int ) -> int: A_ : Any = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 100000] number //= 100000 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution _lowerCAmelCase : list[bool | None] = [None] * 10_000_000 _lowerCAmelCase : Optional[Any] = True _lowerCAmelCase : int = False def __snake_case ( _lowerCAmelCase : int ) -> bool: if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore A_ : Optional[Any] = chain(next_number(__lowercase ) ) A_ : Dict = number_chain while number < 10000000: A_ : Union[str, Any] = number_chain number *= 10 return number_chain def __snake_case ( _lowerCAmelCase : int = 10000000 ) -> int: for i in range(1 , __lowercase ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(__lowercase ) if __name__ == "__main__": import doctest doctest.testmod() print(F'''{solution() = }''')
300
import os from typing import Any, Callable, Dict, List, Optional, Tuple, Union import torch from torch import nn from ...models.controlnet import ControlNetModel, ControlNetOutput from ...models.modeling_utils import ModelMixin from ...utils import logging _lowerCamelCase : List[str] = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def __init__( self : Dict , lowercase : Union[List[ControlNetModel], Tuple[ControlNetModel]] ): '''simple docstring''' super().__init__() _snake_case = nn.ModuleList(lowercase ) def A ( self : Optional[int] , lowercase : torch.FloatTensor , lowercase : Union[torch.Tensor, float, int] , lowercase : torch.Tensor , lowercase : List[torch.tensor] , lowercase : List[float] , lowercase : Optional[torch.Tensor] = None , lowercase : Optional[torch.Tensor] = None , lowercase : Optional[torch.Tensor] = None , lowercase : Optional[Dict[str, Any]] = None , lowercase : bool = False , lowercase : bool = True , ): '''simple docstring''' for i, (image, scale, controlnet) in enumerate(zip(lowercase , lowercase , self.nets ) ): _snake_case , _snake_case = controlnet( lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) # merge samples if i == 0: _snake_case , _snake_case = down_samples, mid_sample else: _snake_case = [ samples_prev + samples_curr for samples_prev, samples_curr in zip(lowercase , lowercase ) ] mid_block_res_sample += mid_sample return down_block_res_samples, mid_block_res_sample def A ( self : Dict , lowercase : Union[str, os.PathLike] , lowercase : bool = True , lowercase : Callable = None , lowercase : bool = False , lowercase : Optional[str] = None , ): '''simple docstring''' _snake_case = 0 _snake_case = save_directory for controlnet in self.nets: controlnet.save_pretrained( lowercase , is_main_process=lowercase , save_function=lowercase , safe_serialization=lowercase , variant=lowercase , ) idx += 1 _snake_case = model_path_to_save + f'''_{idx}''' @classmethod def A ( cls : Any , lowercase : Optional[Union[str, os.PathLike]] , **lowercase : List[str] ): '''simple docstring''' _snake_case = 0 _snake_case = [] # load controlnet and append to list until no controlnet directory exists anymore # first controlnet has to be saved under `./mydirectory/controlnet` to be compliant with `DiffusionPipeline.from_prertained` # second, third, ... controlnets have to be saved under `./mydirectory/controlnet_1`, `./mydirectory/controlnet_2`, ... _snake_case = pretrained_model_path while os.path.isdir(lowercase ): _snake_case = ControlNetModel.from_pretrained(lowercase , **lowercase ) controlnets.append(lowercase ) idx += 1 _snake_case = pretrained_model_path + f'''_{idx}''' logger.info(f'''{len(lowercase )} controlnets loaded from {pretrained_model_path}.''' ) if len(lowercase ) == 0: raise ValueError( f'''No ControlNets found under {os.path.dirname(lowercase )}. Expected at least {pretrained_model_path + '_0'}.''' ) return cls(lowercase )
282
0
'''simple docstring''' import os from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen, xsplitext from ..table import array_cast from ..utils.py_utils import no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: from .features import FeatureType _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = False, False, False @dataclass class _snake_case : __A : Optional[int] =None __A : bool =True __A : bool =True __A : Optional[str] =None # Automatically constructed __A : ClassVar[str] ="dict" __A : ClassVar[Any] =pa.struct({"bytes": pa.binary(), "path": pa.string()}) __A : str =field(default="Audio" , init=__SCREAMING_SNAKE_CASE , repr=__SCREAMING_SNAKE_CASE) def __call__( self ): return self.pa_type def UpperCamelCase__ ( self ,_snake_case ): try: import soundfile as sf # soundfile is a dependency of librosa, needed to decode audio files. except ImportError as err: raise ImportError("To support encoding audio data, please install 'soundfile'." ) from err if isinstance(_snake_case ,_snake_case ): return {"bytes": None, "path": value} elif isinstance(_snake_case ,_snake_case ): return {"bytes": value, "path": None} elif "array" in value: # convert the audio array to wav bytes UpperCAmelCase_ : Optional[Any] = BytesIO() sf.write(_snake_case ,value["array"] ,value["sampling_rate"] ,format="wav" ) return {"bytes": buffer.getvalue(), "path": None} elif value.get("path" ) is not None and os.path.isfile(value["path"] ): # we set "bytes": None to not duplicate the data if they're already available locally if value["path"].endswith("pcm" ): # "PCM" only has raw audio bytes if value.get("sampling_rate" ) is None: # At least, If you want to convert "PCM-byte" to "WAV-byte", you have to know sampling rate raise KeyError("To use PCM files, please specify a 'sampling_rate' in Audio object" ) if value.get("bytes" ): # If we already had PCM-byte, we don`t have to make "read file, make bytes" (just use it!) UpperCAmelCase_ : Union[str, Any] = np.frombuffer(value["bytes"] ,dtype=np.intaa ).astype(np.floataa ) / 3_27_67 else: UpperCAmelCase_ : int = np.memmap(value["path"] ,dtype="h" ,mode="r" ).astype(np.floataa ) / 3_27_67 UpperCAmelCase_ : Dict = BytesIO(bytes() ) sf.write(_snake_case ,_snake_case ,value["sampling_rate"] ,format="wav" ) return {"bytes": buffer.getvalue(), "path": None} else: return {"bytes": None, "path": value.get("path" )} elif value.get("bytes" ) is not None or value.get("path" ) is not None: # store the audio bytes, and path is used to infer the audio format using the file extension return {"bytes": value.get("bytes" ), "path": value.get("path" )} else: raise ValueError( f'''An audio sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.''' ) def UpperCamelCase__ ( self ,_snake_case ,_snake_case = None ): if not self.decode: raise RuntimeError("Decoding is disabled for this feature. Please use Audio(decode=True) instead." ) UpperCAmelCase_ , UpperCAmelCase_ : str = (value["path"], BytesIO(value["bytes"] )) if value["bytes"] is not None else (value["path"], None) if path is None and file is None: raise ValueError(f'''An audio sample should have one of \'path\' or \'bytes\' but both are None in {value}.''' ) try: import librosa import soundfile as sf except ImportError as err: raise ImportError("To support decoding audio files, please install 'librosa' and 'soundfile'." ) from err UpperCAmelCase_ : Union[str, Any] = xsplitext(_snake_case )[1][1:].lower() if path is not None else None if not config.IS_OPUS_SUPPORTED and audio_format == "opus": raise RuntimeError( "Decoding 'opus' files requires system library 'libsndfile'>=1.0.31, " "You can try to update `soundfile` python library: `pip install \"soundfile>=0.12.1\"`. " ) elif not config.IS_MP3_SUPPORTED and audio_format == "mp3": raise RuntimeError( "Decoding 'mp3' files requires system library 'libsndfile'>=1.1.0, " "You can try to update `soundfile` python library: `pip install \"soundfile>=0.12.1\"`. " ) if file is None: UpperCAmelCase_ : Any = token_per_repo_id or {} UpperCAmelCase_ : int = path.split("::" )[-1] try: UpperCAmelCase_ : Any = string_to_dict(_snake_case ,config.HUB_DATASETS_URL )["repo_id"] UpperCAmelCase_ : Optional[Any] = token_per_repo_id[repo_id] except (ValueError, KeyError): UpperCAmelCase_ : List[Any] = None with xopen(_snake_case ,"rb" ,use_auth_token=_snake_case ) as f: UpperCAmelCase_ , UpperCAmelCase_ : str = sf.read(_snake_case ) else: UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = sf.read(_snake_case ) UpperCAmelCase_ : Optional[Any] = array.T if self.mono: UpperCAmelCase_ : str = librosa.to_mono(_snake_case ) if self.sampling_rate and self.sampling_rate != sampling_rate: UpperCAmelCase_ : Tuple = librosa.resample(_snake_case ,orig_sr=_snake_case ,target_sr=self.sampling_rate ) UpperCAmelCase_ : Optional[int] = self.sampling_rate return {"path": path, "array": array, "sampling_rate": sampling_rate} def UpperCamelCase__ ( self ): from .features import Value if self.decode: raise ValueError("Cannot flatten a decoded Audio feature." ) return { "bytes": Value("binary" ), "path": Value("string" ), } def UpperCamelCase__ ( self ,_snake_case ): if pa.types.is_string(storage.type ): UpperCAmelCase_ : Union[str, Any] = pa.array([None] * len(_snake_case ) ,type=pa.binary() ) UpperCAmelCase_ : List[str] = pa.StructArray.from_arrays([bytes_array, storage] ,["bytes", "path"] ,mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): UpperCAmelCase_ : Tuple = pa.array([None] * len(_snake_case ) ,type=pa.string() ) UpperCAmelCase_ : List[str] = pa.StructArray.from_arrays([storage, path_array] ,["bytes", "path"] ,mask=storage.is_null() ) elif pa.types.is_struct(storage.type ) and storage.type.get_all_field_indices("array" ): UpperCAmelCase_ : List[str] = pa.array([Audio().encode_example(_snake_case ) if x is not None else None for x in storage.to_pylist()] ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("bytes" ) >= 0: UpperCAmelCase_ : Optional[int] = storage.field("bytes" ) else: UpperCAmelCase_ : Dict = pa.array([None] * len(_snake_case ) ,type=pa.binary() ) if storage.type.get_field_index("path" ) >= 0: UpperCAmelCase_ : str = storage.field("path" ) else: UpperCAmelCase_ : List[str] = pa.array([None] * len(_snake_case ) ,type=pa.string() ) UpperCAmelCase_ : List[Any] = pa.StructArray.from_arrays([bytes_array, path_array] ,["bytes", "path"] ,mask=storage.is_null() ) return array_cast(_snake_case ,self.pa_type ) def UpperCamelCase__ ( self ,_snake_case ): @no_op_if_value_is_null def path_to_bytes(_snake_case ): with xopen(_snake_case ,"rb" ) as f: UpperCAmelCase_ : Union[str, Any] = f.read() return bytes_ UpperCAmelCase_ : List[Any] = pa.array( [ (path_to_bytes(x["path"] ) if x["bytes"] is None else x["bytes"]) if x is not None else None for x in storage.to_pylist() ] ,type=pa.binary() ,) UpperCAmelCase_ : List[Any] = pa.array( [os.path.basename(_snake_case ) if path is not None else None for path in storage.field("path" ).to_pylist()] ,type=pa.string() ,) UpperCAmelCase_ : int = pa.StructArray.from_arrays([bytes_array, path_array] ,["bytes", "path"] ,mask=bytes_array.is_null() ) return array_cast(_snake_case ,self.pa_type )
67
'''simple docstring''' from collections import Counter from timeit import timeit def a__ ( _SCREAMING_SNAKE_CASE : str = "" , ) -> bool: """simple docstring""" return sum(c % 2 for c in Counter(input_str.replace(" " , "" ).lower() ).values() ) < 2 def a__ ( _SCREAMING_SNAKE_CASE : str = "" ) -> bool: """simple docstring""" if len(_SCREAMING_SNAKE_CASE ) == 0: return True UpperCAmelCase_ : List[str] = input_str.replace(" " , "" ).lower() # character_freq_dict: Stores the frequency of every character in the input string UpperCAmelCase_ : dict[str, int] = {} for character in lower_case_input_str: UpperCAmelCase_ : Any = character_freq_dict.get(_SCREAMING_SNAKE_CASE , 0 ) + 1 UpperCAmelCase_ : Union[str, Any] = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def a__ ( _SCREAMING_SNAKE_CASE : str = "" ) -> None: """simple docstring""" print("\nFor string = " , _SCREAMING_SNAKE_CASE , ":" ) print( "> can_string_be_rearranged_as_palindrome_counter()" , "\tans =" , can_string_be_rearranged_as_palindrome_counter(_SCREAMING_SNAKE_CASE ) , "\ttime =" , timeit( "z.can_string_be_rearranged_as_palindrome_counter(z.check_str)" , setup="import __main__ as z" , ) , "seconds" , ) print( "> can_string_be_rearranged_as_palindrome()" , "\tans =" , can_string_be_rearranged_as_palindrome(_SCREAMING_SNAKE_CASE ) , "\ttime =" , timeit( "z.can_string_be_rearranged_as_palindrome(z.check_str)" , setup="import __main__ as z" , ) , "seconds" , ) if __name__ == "__main__": _lowerCamelCase = input( """Enter string to determine if it can be rearranged as a palindrome or not: """ ).strip() benchmark(check_str) _lowerCamelCase = can_string_be_rearranged_as_palindrome_counter(check_str) print(f"""{check_str} can {'' if status else 'not '}be rearranged as a palindrome""")
67
1
"""simple docstring""" import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) if is_vision_available(): from PIL import Image else: class snake_case : """simple docstring""" @staticmethod def __lowerCAmelCase ( *lowerCamelCase__ : List[Any] ,**lowerCamelCase__ : Union[str, Any] ): pass def a_ ( lowerCamelCase ): UpperCAmelCase__ = hashlib.mda(image.tobytes() ) return m.hexdigest()[:1_0] def a_ ( lowerCamelCase ): UpperCAmelCase__ = np.array(lowerCamelCase ) UpperCAmelCase__ = npimg.shape return {"hash": hashimage(lowerCamelCase ), "shape": shape} @is_pipeline_test @require_vision @require_torch class snake_case ( unittest.TestCase ): """simple docstring""" snake_case__ = dict( (list(MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if MODEL_FOR_MASK_GENERATION_MAPPING else []) ) snake_case__ = dict( (list(TF_MODEL_FOR_MASK_GENERATION_MAPPING.items() ) if TF_MODEL_FOR_MASK_GENERATION_MAPPING else []) ) def __lowerCAmelCase ( self : List[Any] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : str ): UpperCAmelCase__ = MaskGenerationPipeline(model=lowerCamelCase__ ,image_processor=lowerCamelCase__ ) return image_segmenter, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : str ): pass @require_tf @unittest.skip('Image segmentation not implemented in TF' ) def __lowerCAmelCase ( self : Optional[Any] ): pass @slow @require_torch def __lowerCAmelCase ( self : List[str] ): UpperCAmelCase__ = pipeline('mask-generation' ,model='facebook/sam-vit-huge' ) UpperCAmelCase__ = image_segmenter('http://images.cocodataset.org/val2017/000000039769.jpg' ,points_per_batch=256 ) # Shortening by hashing UpperCAmelCase__ = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(lowerCamelCase__ ), "scores": outputs["scores"][i]}] # fmt: off self.assertEqual( nested_simplify(lowerCamelCase__ ,decimals=4 ) ,[ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.0_4_4_4}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_2_1}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.0_1_6_7}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.0_1_3_2}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.0_0_5_3}, {'mask': {'hash': 'e2d0b7a0b7', 'shape': (480, 640)}, 'scores': 0.9_9_6_7}, {'mask': {'hash': '453c7844bd', 'shape': (480, 640)}, 'scores': 0.9_9_3}, {'mask': {'hash': '3d44f2926d', 'shape': (480, 640)}, 'scores': 0.9_9_0_9}, {'mask': {'hash': '64033ddc3f', 'shape': (480, 640)}, 'scores': 0.9_8_7_9}, {'mask': {'hash': '801064ff79', 'shape': (480, 640)}, 'scores': 0.9_8_3_4}, {'mask': {'hash': '6172f276ef', 'shape': (480, 640)}, 'scores': 0.9_7_1_6}, {'mask': {'hash': 'b49e60e084', 'shape': (480, 640)}, 'scores': 0.9_6_1_2}, {'mask': {'hash': 'a811e775fd', 'shape': (480, 640)}, 'scores': 0.9_5_9_9}, {'mask': {'hash': 'a6a8ebcf4b', 'shape': (480, 640)}, 'scores': 0.9_5_5_2}, {'mask': {'hash': '9d8257e080', 'shape': (480, 640)}, 'scores': 0.9_5_3_2}, {'mask': {'hash': '32de6454a8', 'shape': (480, 640)}, 'scores': 0.9_5_1_6}, {'mask': {'hash': 'af3d4af2c8', 'shape': (480, 640)}, 'scores': 0.9_4_9_9}, {'mask': {'hash': '3c6db475fb', 'shape': (480, 640)}, 'scores': 0.9_4_8_3}, {'mask': {'hash': 'c290813fb9', 'shape': (480, 640)}, 'scores': 0.9_4_6_4}, {'mask': {'hash': 'b6f0b8f606', 'shape': (480, 640)}, 'scores': 0.9_4_3}, {'mask': {'hash': '92ce16bfdf', 'shape': (480, 640)}, 'scores': 0.9_4_3}, {'mask': {'hash': 'c749b25868', 'shape': (480, 640)}, 'scores': 0.9_4_0_8}, {'mask': {'hash': 'efb6cab859', 'shape': (480, 640)}, 'scores': 0.9_3_3_5}, {'mask': {'hash': '1ff2eafb30', 'shape': (480, 640)}, 'scores': 0.9_3_2_6}, {'mask': {'hash': '788b798e24', 'shape': (480, 640)}, 'scores': 0.9_2_6_2}, {'mask': {'hash': 'abea804f0e', 'shape': (480, 640)}, 'scores': 0.8_9_9_9}, {'mask': {'hash': '7b9e8ddb73', 'shape': (480, 640)}, 'scores': 0.8_9_8_6}, {'mask': {'hash': 'cd24047c8a', 'shape': (480, 640)}, 'scores': 0.8_9_8_4}, {'mask': {'hash': '6943e6bcbd', 'shape': (480, 640)}, 'scores': 0.8_8_7_3}, {'mask': {'hash': 'b5f47c9191', 'shape': (480, 640)}, 'scores': 0.8_8_7_1} ] ,) # fmt: on @require_torch @slow def __lowerCAmelCase ( self : Optional[Any] ): UpperCAmelCase__ = 'facebook/sam-vit-huge' UpperCAmelCase__ = pipeline('mask-generation' ,model=lowerCamelCase__ ) UpperCAmelCase__ = image_segmenter( 'http://images.cocodataset.org/val2017/000000039769.jpg' ,pred_iou_thresh=1 ,points_per_batch=256 ) # Shortening by hashing UpperCAmelCase__ = [] for i, o in enumerate(outputs['masks'] ): new_outupt += [{"mask": mask_to_test_readable(lowerCamelCase__ ), "scores": outputs["scores"][i]}] self.assertEqual( nested_simplify(lowerCamelCase__ ,decimals=4 ) ,[ {'mask': {'hash': '115ad19f5f', 'shape': (480, 640)}, 'scores': 1.0_4_4_4}, {'mask': {'hash': '6affa964c6', 'shape': (480, 640)}, 'scores': 1.0_2_1_0}, {'mask': {'hash': 'dfe28a0388', 'shape': (480, 640)}, 'scores': 1.0_1_6_7}, {'mask': {'hash': 'c0a5f4a318', 'shape': (480, 640)}, 'scores': 1.0_1_3_2}, {'mask': {'hash': 'fe8065c197', 'shape': (480, 640)}, 'scores': 1.0_0_5_3}, ] ,)
98
import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() __A = logging.get_logger(__name__) __A = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } __A = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", ] def lowerCamelCase_ ( UpperCamelCase__ : Any , UpperCamelCase__ : Any , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Optional[int] ) -> Tuple: """simple docstring""" for attribute in key.split('.' ): __lowerCamelCase = getattr(UpperCamelCase__ , UpperCamelCase__ ) if weight_type is not None: __lowerCamelCase = getattr(UpperCamelCase__ , UpperCamelCase__ ).shape else: __lowerCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( F"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": __lowerCamelCase = value elif weight_type == "weight_g": __lowerCamelCase = value elif weight_type == "weight_v": __lowerCamelCase = value elif weight_type == "bias": __lowerCamelCase = value else: __lowerCamelCase = value logger.info(F"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def lowerCamelCase_ ( UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Tuple ) -> Optional[Any]: """simple docstring""" __lowerCamelCase = [] __lowerCamelCase = fairseq_model.state_dict() __lowerCamelCase = hf_model.feature_extractor __lowerCamelCase = hf_model.adapter for name, value in fairseq_dict.items(): __lowerCamelCase = False if "conv_layers" in name: load_conv_layer( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , hf_model.config.feat_extract_norm == 'group' , ) __lowerCamelCase = True elif any(x in name for x in ['adaptor', 'w2v_encoder.proj.', 'w2v_proj_ln.'] ): load_adapter(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) __lowerCamelCase = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: __lowerCamelCase = True if "*" in mapped_key: __lowerCamelCase = name.split(UpperCamelCase__ )[0].split('.' )[-2] __lowerCamelCase = mapped_key.replace('*' , UpperCamelCase__ ) if "weight_g" in name: __lowerCamelCase = 'weight_g' elif "weight_v" in name: __lowerCamelCase = 'weight_v' elif "bias" in name: __lowerCamelCase = 'bias' elif "weight" in name: __lowerCamelCase = 'weight' else: __lowerCamelCase = None set_recursively(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) continue if not is_used: unused_weights.append(UpperCamelCase__ ) logger.warning(F"""Unused weights: {unused_weights}""" ) def lowerCamelCase_ ( UpperCamelCase__ : Dict , UpperCamelCase__ : Any , UpperCamelCase__ : str , UpperCamelCase__ : Tuple , UpperCamelCase__ : Tuple ) -> int: """simple docstring""" __lowerCamelCase = full_name.split('conv_layers.' )[-1] __lowerCamelCase = name.split('.' ) __lowerCamelCase = int(items[0] ) __lowerCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __lowerCamelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __lowerCamelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) __lowerCamelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) __lowerCamelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(UpperCamelCase__ ) def lowerCamelCase_ ( UpperCamelCase__ : Tuple , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : int ) -> Union[str, Any]: """simple docstring""" __lowerCamelCase = full_name.split('adaptor.' )[-1] __lowerCamelCase = name.split('.' ) if items[1].isdigit(): __lowerCamelCase = int(items[1] ) else: __lowerCamelCase = None if "adaptor" not in full_name: if "proj_ln" in full_name: # has to be layer norm if "bias" in name: assert ( value.shape == adapter.proj_layer_norm.bias.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj_layer_norm.bias.data.shape} was found.""" __lowerCamelCase = value logger.info(F"""Adapter proj layer norm bias was initialized from {full_name}.""" ) if "weight" in name: assert ( value.shape == adapter.proj_layer_norm.weight.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj_layer_norm.weight.data.shape} was found.""" __lowerCamelCase = value else: # has to be projection layer if "bias" in name: assert ( value.shape == adapter.proj.bias.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj.bias.data.shape} was found.""" __lowerCamelCase = value logger.info(F"""Adapter proj layer bias was initialized from {full_name}.""" ) if "weight" in name: assert ( value.shape == adapter.proj.weight.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj.weight.data.shape} was found.""" __lowerCamelCase = value logger.info(F"""Adapter proj layer weight was initialized from {full_name}.""" ) elif isinstance(UpperCamelCase__ , UpperCamelCase__ ): if "bias" in name: assert ( value.shape == adapter.layers[layer_id].conv.bias.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.bias.data.shape} was found.""" __lowerCamelCase = value logger.info(F"""Adapter layer {layer_id} bias was initialized from {full_name}.""" ) elif "weight" in name: assert ( value.shape == adapter.layers[layer_id].conv.weight.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.weight.data.shape} was found.""" __lowerCamelCase = value logger.info(F"""Adapter layer {layer_id} bias was initialized from {full_name}.""" ) else: unused_weights.append(UpperCamelCase__ ) def lowerCamelCase_ ( UpperCamelCase__ : Tuple ) -> Tuple: """simple docstring""" __lowerCamelCase , __lowerCamelCase = emb.weight.shape __lowerCamelCase = nn.Linear(UpperCamelCase__ , UpperCamelCase__ , bias=UpperCamelCase__ ) __lowerCamelCase = emb.weight.data return lin_layer @torch.no_grad() def lowerCamelCase_ ( UpperCamelCase__ : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Dict , UpperCamelCase__ : List[str] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : str , UpperCamelCase__ : Optional[Any] , ) -> str: """simple docstring""" __lowerCamelCase = WavaVecaConfig.from_pretrained( UpperCamelCase__ , add_adapter=UpperCamelCase__ , adapter_stride=UpperCamelCase__ , adapter_kernel_size=UpperCamelCase__ , use_auth_token=UpperCamelCase__ , output_hidden_size=UpperCamelCase__ , ) __lowerCamelCase = MBartConfig.from_pretrained(UpperCamelCase__ ) # load model __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={ 'config_yaml': config_yaml_path, 'data': '/'.join(dict_path.split('/' )[:-1] ), 'w2v_path': checkpoint_path, 'load_pretrained_decoder_from': None, } , ) __lowerCamelCase = model[0].eval() # load feature extractor __lowerCamelCase = WavaVecaFeatureExtractor.from_pretrained(UpperCamelCase__ , use_auth_token=UpperCamelCase__ ) # set weights for wav2vec2 encoder __lowerCamelCase = WavaVecaModel(UpperCamelCase__ ) recursively_load_weights_wavaveca(model.encoder , UpperCamelCase__ ) # load decoder weights __lowerCamelCase = MBartForCausalLM(UpperCamelCase__ ) __lowerCamelCase , __lowerCamelCase = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=UpperCamelCase__ ) logger.warning(F"""The following keys are missing when loading the decoder weights: {missing_keys}""" ) logger.warning(F"""The following keys are unexpected when loading the decoder weights: {unexpected_keys}""" ) __lowerCamelCase = SpeechEncoderDecoderModel(encoder=UpperCamelCase__ , decoder=UpperCamelCase__ ) __lowerCamelCase = False __lowerCamelCase = MBartaaTokenizer(UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) __lowerCamelCase = hf_wavavec.config.to_dict() __lowerCamelCase = tokenizer.pad_token_id __lowerCamelCase = tokenizer.bos_token_id __lowerCamelCase = tokenizer.eos_token_id __lowerCamelCase = 'mbart50' __lowerCamelCase = 'wav2vec2' __lowerCamelCase = tokenizer.eos_token_id __lowerCamelCase = 25_0004 __lowerCamelCase = tokenizer.eos_token_id __lowerCamelCase = SpeechEncoderDecoderConfig.from_dict(UpperCamelCase__ ) hf_wavavec.save_pretrained(UpperCamelCase__ ) feature_extractor.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": __A = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_yaml_path", default=None, type=str, help="Path to yaml file of fine-tuned model") parser.add_argument( "--encoder_config_path", default="facebook/wav2vec2-xls-r-1b", type=str, help="Path to hf encoder wav2vec2 checkpoint config", ) parser.add_argument( "--decoder_config_path", default="facebook/mbart-large-50-one-to-many-mmt", type=str, help="Path to hf decoder checkpoint config", ) parser.add_argument("--add_adapter", default=True, type=bool, help="whethere to add model adapter layers") parser.add_argument("--adapter_stride", default=2, type=int, help="stride of adapter layers") parser.add_argument("--adapter_kernel_size", default=3, type=int, help="kernel size of adapter layers") parser.add_argument("--encoder_output_dim", default=10_24, type=int, help="encoder output dim") parser.add_argument("--start_token_id", default=25_00_04, type=int, help="`decoder_start_token_id` of model config") __A = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, args.config_yaml_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, add_adapter=args.add_adapter, adapter_kernel_size=args.adapter_kernel_size, adapter_stride=args.adapter_stride, decoder_start_token_id=args.start_token_id, encoder_output_dim=args.encoder_output_dim, )
90
0
'''simple docstring''' def A_( A : int = 50): UpperCamelCase = [1] * (length + 1) for row_length in range(length + 1): for tile_length in range(2 , 5): for tile_start in range(row_length - tile_length + 1): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(f"""{solution() = }""")
251
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase : Optional[int] = logging.get_logger(__name__) lowerCAmelCase : Union[str, Any] = {'openai-gpt': 'https://huggingface.co/openai-gpt/resolve/main/config.json'} class SCREAMING_SNAKE_CASE__ ( snake_case_): lowerCAmelCase_ = """openai-gpt""" lowerCAmelCase_ = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , A_=40478 , A_=512 , A_=768 , A_=12 , A_=12 , A_="gelu" , A_=0.1 , A_=0.1 , A_=0.1 , A_=1e-5 , A_=0.02 , A_="cls_index" , A_=True , A_=None , A_=True , A_=0.1 , **A_ , )-> List[str]: '''simple docstring''' UpperCamelCase = vocab_size UpperCamelCase = n_positions UpperCamelCase = n_embd UpperCamelCase = n_layer UpperCamelCase = n_head UpperCamelCase = afn UpperCamelCase = resid_pdrop UpperCamelCase = embd_pdrop UpperCamelCase = attn_pdrop UpperCamelCase = layer_norm_epsilon UpperCamelCase = initializer_range UpperCamelCase = summary_type UpperCamelCase = summary_use_proj UpperCamelCase = summary_activation UpperCamelCase = summary_first_dropout UpperCamelCase = summary_proj_to_labels super().__init__(**A_ )
251
1
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed _SCREAMING_SNAKE_CASE = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F'''{bindir}/../../examples/pytorch/translation'''): from run_translation import main # noqa set_seed(42) _SCREAMING_SNAKE_CASE = 'sshleifer/student_marian_en_ro_6_1' _SCREAMING_SNAKE_CASE = 'sshleifer/tiny-mbart' @require_torch class a ( __lowerCAmelCase ): """simple docstring""" def UpperCAmelCase ( self , lowerCAmelCase_=False , lowerCAmelCase_=None , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=True , lowerCAmelCase_=True , ) -> int: _A = self.run_trainer( eval_steps=1 , max_len=12 , model_name=lowerCAmelCase_ , num_train_epochs=1 , distributed=lowerCAmelCase_ , extra_args_str=lowerCAmelCase_ , predict_with_generate=lowerCAmelCase_ , do_train=lowerCAmelCase_ , do_eval=lowerCAmelCase_ , do_predict=lowerCAmelCase_ , ) _A = TrainerState.load_from_json(os.path.join(lowerCAmelCase_ , """trainer_state.json""" ) ).log_history if not do_eval: return _A = [log for log in logs if """eval_loss""" in log.keys()] _A = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats _A = eval_metrics[-1] assert isinstance(last_step_stats["""eval_bleu"""] , lowerCAmelCase_ ) assert not math.isnan(float(last_step_stats["""eval_loss"""] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def UpperCAmelCase ( self ) -> Optional[int]: self.run_seqaseq_quick() @require_torch_multi_gpu def UpperCAmelCase ( self ) -> Dict: self.run_seqaseq_quick(distributed=lowerCAmelCase_ ) @require_torch_multi_gpu def UpperCAmelCase ( self ) -> Dict: self.run_seqaseq_quick(distributed=lowerCAmelCase_ ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def UpperCAmelCase ( self ) -> str: self.run_seqaseq_quick(distributed=lowerCAmelCase_ , extra_args_str="""--sharded_ddp simple""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def UpperCAmelCase ( self ) -> Dict: self.run_seqaseq_quick(distributed=lowerCAmelCase_ , extra_args_str="""--sharded_ddp simple --fp16""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def UpperCAmelCase ( self ) -> Optional[Any]: self.run_seqaseq_quick(distributed=lowerCAmelCase_ , extra_args_str="""--sharded_ddp zero_dp_2""" , predict_with_generate=lowerCAmelCase_ ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def UpperCAmelCase ( self ) -> Tuple: self.run_seqaseq_quick( distributed=lowerCAmelCase_ , extra_args_str="""--sharded_ddp zero_dp_2 --fp16""" , predict_with_generate=lowerCAmelCase_ ) @require_apex @require_torch_gpu def UpperCAmelCase ( self ) -> int: # XXX: apex breaks the trainer if it's run twice e.g. run_seq2seq.main() from the same # program and it breaks other tests that run from the same pytest worker, therefore until this is # sorted out it must be run only in an external program, that is distributed=True in this # test and only under one or more gpus - if we want cpu will need to make a special test # # specifically to the problem traced it to self.optimizer.step() - if it's run 2nd time via # 2nd main() call it botches the future eval. # self.run_seqaseq_quick(distributed=lowerCAmelCase_ , extra_args_str="""--fp16 --fp16_backend=apex""" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=lowerCAmelCase_ , extra_args_str="""--fp16 --fp16_backend=apex""" ) @parameterized.expand(["""base""", """low""", """high""", """mixed"""] ) @require_torch_multi_gpu def UpperCAmelCase ( self , lowerCAmelCase_ ) -> int: # as each sub-test is slow-ish split into multiple sub-tests to avoid CI timeout _A = { # test with the default log_level - should be info and thus log info once """base""": {"""extra_args_str""": """""", """n_matches""": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes """low""": {"""extra_args_str""": """--log_level debug --log_level_replica debug""", """n_matches""": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica """high""": {"""extra_args_str""": """--log_level error --log_level_replica debug""", """n_matches""": 1}, # test with high log_level and log_level_replica - should be quiet on all processes """mixed""": {"""extra_args_str""": """--log_level error --log_level_replica error""", """n_matches""": 0}, } _A = experiments[experiment_id] _A = {"""distributed""": True, """predict_with_generate""": False, """do_eval""": False, """do_predict""": False} _A = """Running training""" with CaptureStderr() as cl: self.run_seqaseq_quick(**lowerCAmelCase_ , extra_args_str=data["""extra_args_str"""] ) _A = len(re.findall(lowerCAmelCase_ , cl.err ) ) self.assertEqual(lowerCAmelCase_ , data["""n_matches"""] ) @slow def UpperCAmelCase ( self ) -> Dict: _A = self.run_trainer( eval_steps=2 , max_len=1_28 , model_name=lowerCAmelCase_ , learning_rate=3E-4 , num_train_epochs=10 , distributed=lowerCAmelCase_ , ) # Check metrics _A = TrainerState.load_from_json(os.path.join(lowerCAmelCase_ , """trainer_state.json""" ) ).log_history _A = [log for log in logs if """eval_loss""" in log.keys()] _A = eval_metrics[0] _A = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["""eval_bleu"""] , lowerCAmelCase_ ) # test if do_predict saves generations and metrics _A = os.listdir(lowerCAmelCase_ ) _A = {os.path.basename(lowerCAmelCase_ ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def UpperCAmelCase ( self ) -> Optional[Any]: from transformers.training_args import OptimizerNames def train_and_return_metrics(lowerCAmelCase_ ) -> Tuple[int, float]: _A = """--skip_memory_metrics 0""" _A = self.run_trainer( max_len=1_28 , model_name=lowerCAmelCase_ , learning_rate=3E-4 , num_train_epochs=1 , optim=lowerCAmelCase_ , distributed=lowerCAmelCase_ , extra_args_str=lowerCAmelCase_ , do_eval=lowerCAmelCase_ , do_predict=lowerCAmelCase_ , n_gpus_to_use=1 , ) # Check metrics _A = TrainerState.load_from_json(Path(lowerCAmelCase_ , """trainer_state.json""" ) ).log_history _A = int(logs[0]["""train_mem_gpu_peaked_delta"""] / 2**20 ) _A = int(logs[0]["""train_mem_gpu_alloc_delta"""] / 2**20 ) _A = logs[0]["""train_loss"""] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss _A , _A , _A = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) _A , _A , _A = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) _A = gpu_alloc_mem_orig - gpu_alloc_mem_bnb _A = gpu_peak_mem_orig + gpu_alloc_mem_orig _A = gpu_peak_mem_bnb + gpu_alloc_mem_bnb _A = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings _A = 1_20 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( lowerCAmelCase_ , lowerCAmelCase_ , """should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got""" F''' a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and''' F''' gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB''' , ) self.assertGreater( lowerCAmelCase_ , lowerCAmelCase_ , """should use ~150MB less total gpu memory with BNB, compared to without it for this model but got""" F''' a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and''' F''' gpu_total_mem_bnb={gpu_total_mem_bnb}MB''' , ) self.assertEqual( lowerCAmelCase_ , lowerCAmelCase_ , F'''loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}''' ) def UpperCAmelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = 3E-3 , lowerCAmelCase_ = "adafactor" , lowerCAmelCase_ = False , lowerCAmelCase_ = None , lowerCAmelCase_ = 0 , lowerCAmelCase_ = True , lowerCAmelCase_ = True , lowerCAmelCase_ = True , lowerCAmelCase_ = True , lowerCAmelCase_ = None , ) -> str: _A = self.test_file_dir / """../fixtures/tests_samples/wmt_en_ro""" _A = self.get_auto_remove_tmp_dir() _A = F''' --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(lowerCAmelCase_ )} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(lowerCAmelCase_ )} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX '''.split() _A = F''' --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(lowerCAmelCase_ )} '''.split() _A = """ --do_predict """.split() _A = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F'''--optim {optim}'''.split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: _A = get_gpu_count() _A = get_torch_dist_unique_port() _A = F''' -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py '''.split() _A = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(lowerCAmelCase_ , env=self.get_env() ) else: _A = ["""run_translation.py"""] + args with patch.object(lowerCAmelCase_ , """argv""" , lowerCAmelCase_ ): main() return output_dir
180
import argparse import gc import json import os import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler _SCREAMING_SNAKE_CASE = 16 _SCREAMING_SNAKE_CASE = 32 def snake_case ( snake_case__ :Optional[int]) -> str: return int(x / 2**20) class a : """simple docstring""" def __enter__( self ) -> List[str]: gc.collect() torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() # reset the peak gauge to zero _A = torch.cuda.memory_allocated() return self def __exit__( self , *lowerCAmelCase_ ) -> Optional[int]: gc.collect() torch.cuda.empty_cache() _A = torch.cuda.memory_allocated() _A = torch.cuda.max_memory_allocated() _A = bamb(self.end - self.begin ) _A = bamb(self.peak - self.begin ) # print(f"delta used/peak {self.used:4d}/{self.peaked:4d}") def snake_case ( snake_case__ :Accelerator , snake_case__ :int = 16 , snake_case__ :str = "bert-base-cased" , snake_case__ :int = 320 , snake_case__ :int = 160 , ) -> Dict: _A = AutoTokenizer.from_pretrained(snake_case__) _A = load_dataset( """glue""" , """mrpc""" , split={"""train""": F'''train[:{n_train}]''', """validation""": F'''validation[:{n_val}]'''}) def tokenize_function(snake_case__ :Optional[int]): # max_length=None => use the model max length (it's actually the default) _A = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=snake_case__ , max_length=snake_case__) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset _A = datasets.map( snake_case__ , batched=snake_case__ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , load_from_cache_file=snake_case__) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _A = tokenized_datasets.rename_column("""label""" , """labels""") def collate_fn(snake_case__ :List[str]): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(snake_case__ , padding="""max_length""" , max_length=128 , return_tensors="""pt""") return tokenizer.pad(snake_case__ , padding="""longest""" , return_tensors="""pt""") # Instantiate dataloaders. _A = DataLoader( tokenized_datasets["""train"""] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__) _A = DataLoader( tokenized_datasets["""validation"""] , shuffle=snake_case__ , collate_fn=snake_case__ , batch_size=snake_case__) return train_dataloader, eval_dataloader def snake_case ( snake_case__ :Optional[int] , snake_case__ :Optional[int]) -> Optional[int]: # Initialize accelerator _A = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _A = config["""lr"""] _A = int(config["""num_epochs"""]) _A = int(config["""seed"""]) _A = int(config["""batch_size"""]) _A = args.model_name_or_path set_seed(snake_case__) _A , _A = get_dataloaders(snake_case__ , snake_case__ , snake_case__ , args.n_train , args.n_val) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _A = AutoModelForSequenceClassification.from_pretrained(snake_case__ , return_dict=snake_case__) # Instantiate optimizer _A = ( AdamW if accelerator.state.deepspeed_plugin is None or """optimizer""" not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) _A = optimizer_cls(params=model.parameters() , lr=snake_case__) if accelerator.state.deepspeed_plugin is not None: _A = accelerator.state.deepspeed_plugin.deepspeed_config[ """gradient_accumulation_steps""" ] else: _A = 1 _A = (len(snake_case__) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): _A = get_linear_schedule_with_warmup( optimizer=snake_case__ , num_warmup_steps=0 , num_training_steps=snake_case__ , ) else: _A = DummyScheduler(snake_case__ , total_num_steps=snake_case__ , warmup_num_steps=0) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _A , _A , _A , _A , _A = accelerator.prepare( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__) # We need to keep track of how many total steps we have iterated over _A = 0 # We also need to keep track of the stating epoch so files are named properly _A = 0 # Now we train the model _A = {} for epoch in range(snake_case__ , snake_case__): with TorchTracemalloc() as tracemalloc: model.train() for step, batch in enumerate(snake_case__): _A = model(**snake_case__) _A = outputs.loss _A = loss / gradient_accumulation_steps accelerator.backward(snake_case__) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 # Printing the GPU memory usage details such as allocated memory, peak memory, and total memory usage accelerator.print("""Memory before entering the train : {}""".format(bamb(tracemalloc.begin))) accelerator.print("""Memory consumed at the end of the train (end-begin): {}""".format(tracemalloc.used)) accelerator.print("""Peak Memory consumed during the train (max-begin): {}""".format(tracemalloc.peaked)) accelerator.print( """Total Peak Memory consumed during the train (max): {}""".format( tracemalloc.peaked + bamb(tracemalloc.begin))) _A = tracemalloc.peaked + bamb(tracemalloc.begin) if args.peak_memory_upper_bound is not None: assert ( train_total_peak_memory[F'''epoch-{epoch}'''] <= args.peak_memory_upper_bound ), "Peak memory usage exceeded the upper bound" accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , """peak_memory_utilization.json""") , """w""") as f: json.dump(snake_case__ , snake_case__) def snake_case ( ) -> Optional[int]: _A = argparse.ArgumentParser(description="""Simple example of training script tracking peak GPU memory usage.""") parser.add_argument( """--model_name_or_path""" , type=snake_case__ , default="""bert-base-cased""" , help="""Path to pretrained model or model identifier from huggingface.co/models.""" , required=snake_case__ , ) parser.add_argument( """--output_dir""" , type=snake_case__ , default=""".""" , help="""Optional save directory where all checkpoint folders will be stored. Default is the current working directory.""" , ) parser.add_argument( """--peak_memory_upper_bound""" , type=snake_case__ , default=snake_case__ , help="""The upper bound of peak memory usage in MB. If set, the training will throw an error if the peak memory usage exceeds this value.""" , ) parser.add_argument( """--n_train""" , type=snake_case__ , default=320 , help="""Number of training examples to use.""" , ) parser.add_argument( """--n_val""" , type=snake_case__ , default=160 , help="""Number of validation examples to use.""" , ) parser.add_argument( """--num_epochs""" , type=snake_case__ , default=1 , help="""Number of train epochs.""" , ) _A = parser.parse_args() _A = {"""lr""": 2E-5, """num_epochs""": args.num_epochs, """seed""": 42, """batch_size""": 16} training_function(snake_case__ , snake_case__) if __name__ == "__main__": main()
180
1
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ =logging.get_logger(__name__) UpperCamelCase__ ={ 'facebook/timesformer': 'https://huggingface.co/facebook/timesformer/resolve/main/config.json', } class lowerCAmelCase__( __lowercase ): '''simple docstring''' __snake_case = 'timesformer' def __init__( self , __lowerCamelCase=2_2_4 , __lowerCamelCase=1_6 , __lowerCamelCase=3 , __lowerCamelCase=8 , __lowerCamelCase=7_6_8 , __lowerCamelCase=1_2 , __lowerCamelCase=1_2 , __lowerCamelCase=3_0_7_2 , __lowerCamelCase="gelu" , __lowerCamelCase=0.0 , __lowerCamelCase=0.0 , __lowerCamelCase=0.02 , __lowerCamelCase=1E-6 , __lowerCamelCase=True , __lowerCamelCase="divided_space_time" , __lowerCamelCase=0 , **__lowerCamelCase , ) -> List[str]: super().__init__(**__lowerCamelCase ) _SCREAMING_SNAKE_CASE : int = image_size _SCREAMING_SNAKE_CASE : str = patch_size _SCREAMING_SNAKE_CASE : str = num_channels _SCREAMING_SNAKE_CASE : str = num_frames _SCREAMING_SNAKE_CASE : Dict = hidden_size _SCREAMING_SNAKE_CASE : Any = num_hidden_layers _SCREAMING_SNAKE_CASE : List[Any] = num_attention_heads _SCREAMING_SNAKE_CASE : Optional[Any] = intermediate_size _SCREAMING_SNAKE_CASE : Optional[int] = hidden_act _SCREAMING_SNAKE_CASE : int = hidden_dropout_prob _SCREAMING_SNAKE_CASE : Optional[Any] = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE : int = initializer_range _SCREAMING_SNAKE_CASE : List[str] = layer_norm_eps _SCREAMING_SNAKE_CASE : List[str] = qkv_bias _SCREAMING_SNAKE_CASE : Tuple = attention_type _SCREAMING_SNAKE_CASE : Union[str, Any] = drop_path_rate
325
import math import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from .attention_processor import Attention from .embeddings import get_timestep_embedding from .modeling_utils import ModelMixin class lowerCAmelCase__( __lowercase , __lowercase ): '''simple docstring''' @register_to_config def __init__( self , __lowerCamelCase = 1_2_8 , __lowerCamelCase = 2_5_6 , __lowerCamelCase = 2000.0 , __lowerCamelCase = 7_6_8 , __lowerCamelCase = 1_2 , __lowerCamelCase = 1_2 , __lowerCamelCase = 6_4 , __lowerCamelCase = 2_0_4_8 , __lowerCamelCase = 0.1 , ) -> int: super().__init__() _SCREAMING_SNAKE_CASE : Optional[int] = nn.Sequential( nn.Linear(__lowerCamelCase , d_model * 4 , bias=__lowerCamelCase ) , nn.SiLU() , nn.Linear(d_model * 4 , d_model * 4 , bias=__lowerCamelCase ) , nn.SiLU() , ) _SCREAMING_SNAKE_CASE : str = nn.Embedding(__lowerCamelCase , __lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[Any] = False _SCREAMING_SNAKE_CASE : Optional[Any] = nn.Linear(__lowerCamelCase , __lowerCamelCase , bias=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[Any] = nn.Dropout(p=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[int] = nn.ModuleList() for lyr_num in range(__lowerCamelCase ): # FiLM conditional T5 decoder _SCREAMING_SNAKE_CASE : Optional[int] = DecoderLayer(d_model=__lowerCamelCase , d_kv=__lowerCamelCase , num_heads=__lowerCamelCase , d_ff=__lowerCamelCase , dropout_rate=__lowerCamelCase ) self.decoders.append(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Tuple = TaLayerNorm(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : str = nn.Dropout(p=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : List[Any] = nn.Linear(__lowerCamelCase , __lowerCamelCase , bias=__lowerCamelCase ) def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase ) -> List[Any]: _SCREAMING_SNAKE_CASE : int = torch.mul(query_input.unsqueeze(-1 ) , key_input.unsqueeze(-2 ) ) return mask.unsqueeze(-3 ) def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Dict: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : str = decoder_input_tokens.shape assert decoder_noise_time.shape == (batch,) # decoder_noise_time is in [0, 1), so rescale to expected timing range. _SCREAMING_SNAKE_CASE : Tuple = get_timestep_embedding( decoder_noise_time * self.config.max_decoder_noise_time , embedding_dim=self.config.d_model , max_period=self.config.max_decoder_noise_time , ).to(dtype=self.dtype ) _SCREAMING_SNAKE_CASE : str = self.conditioning_emb(__lowerCamelCase ).unsqueeze(1 ) assert conditioning_emb.shape == (batch, 1, self.config.d_model * 4) _SCREAMING_SNAKE_CASE : Tuple = decoder_input_tokens.shape[1] # If we want to use relative positions for audio context, we can just offset # this sequence by the length of encodings_and_masks. _SCREAMING_SNAKE_CASE : Optional[int] = torch.broadcast_to( torch.arange(__lowerCamelCase , device=decoder_input_tokens.device ) , (batch, seq_length) , ) _SCREAMING_SNAKE_CASE : Union[str, Any] = self.position_encoding(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : str = self.continuous_inputs_projection(__lowerCamelCase ) inputs += position_encodings _SCREAMING_SNAKE_CASE : Any = self.dropout(__lowerCamelCase ) # decoder: No padding present. _SCREAMING_SNAKE_CASE : Any = torch.ones( decoder_input_tokens.shape[:2] , device=decoder_input_tokens.device , dtype=inputs.dtype ) # Translate encoding masks to encoder-decoder masks. _SCREAMING_SNAKE_CASE : List[str] = [(x, self.encoder_decoder_mask(__lowerCamelCase , __lowerCamelCase )) for x, y in encodings_and_masks] # cross attend style: concat encodings _SCREAMING_SNAKE_CASE : Tuple = torch.cat([x[0] for x in encodings_and_encdec_masks] , dim=1 ) _SCREAMING_SNAKE_CASE : Union[str, Any] = torch.cat([x[1] for x in encodings_and_encdec_masks] , dim=-1 ) for lyr in self.decoders: _SCREAMING_SNAKE_CASE : Optional[Any] = lyr( __lowerCamelCase , conditioning_emb=__lowerCamelCase , encoder_hidden_states=__lowerCamelCase , encoder_attention_mask=__lowerCamelCase , )[0] _SCREAMING_SNAKE_CASE : int = self.decoder_norm(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Dict = self.post_dropout(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : int = self.spec_out(__lowerCamelCase ) return spec_out class lowerCAmelCase__( nn.Module ): '''simple docstring''' def __init__( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=1E-6 ) -> Dict: super().__init__() _SCREAMING_SNAKE_CASE : Optional[int] = nn.ModuleList() # cond self attention: layer 0 self.layer.append( TaLayerSelfAttentionCond(d_model=__lowerCamelCase , d_kv=__lowerCamelCase , num_heads=__lowerCamelCase , dropout_rate=__lowerCamelCase ) ) # cross attention: layer 1 self.layer.append( TaLayerCrossAttention( d_model=__lowerCamelCase , d_kv=__lowerCamelCase , num_heads=__lowerCamelCase , dropout_rate=__lowerCamelCase , layer_norm_epsilon=__lowerCamelCase , ) ) # Film Cond MLP + dropout: last layer self.layer.append( TaLayerFFCond(d_model=__lowerCamelCase , d_ff=__lowerCamelCase , dropout_rate=__lowerCamelCase , layer_norm_epsilon=__lowerCamelCase ) ) def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase=None , ) -> Union[str, Any]: _SCREAMING_SNAKE_CASE : int = self.layer[0]( __lowerCamelCase , conditioning_emb=__lowerCamelCase , attention_mask=__lowerCamelCase , ) if encoder_hidden_states is not None: _SCREAMING_SNAKE_CASE : str = torch.where(encoder_attention_mask > 0 , 0 , -1E10 ).to( encoder_hidden_states.dtype ) _SCREAMING_SNAKE_CASE : Tuple = self.layer[1]( __lowerCamelCase , key_value_states=__lowerCamelCase , attention_mask=__lowerCamelCase , ) # Apply Film Conditional Feed Forward layer _SCREAMING_SNAKE_CASE : Optional[Any] = self.layer[-1](__lowerCamelCase , __lowerCamelCase ) return (hidden_states,) class lowerCAmelCase__( nn.Module ): '''simple docstring''' def __init__( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Optional[Any]: super().__init__() _SCREAMING_SNAKE_CASE : List[str] = TaLayerNorm(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Any = TaFiLMLayer(in_features=d_model * 4 , out_features=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : str = Attention(query_dim=__lowerCamelCase , heads=__lowerCamelCase , dim_head=__lowerCamelCase , out_bias=__lowerCamelCase , scale_qk=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[Any] = nn.Dropout(__lowerCamelCase ) def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase=None , __lowerCamelCase=None , ) -> Union[str, Any]: # pre_self_attention_layer_norm _SCREAMING_SNAKE_CASE : int = self.layer_norm(__lowerCamelCase ) if conditioning_emb is not None: _SCREAMING_SNAKE_CASE : Any = self.FiLMLayer(__lowerCamelCase , __lowerCamelCase ) # Self-attention block _SCREAMING_SNAKE_CASE : Optional[int] = self.attention(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Dict = hidden_states + self.dropout(__lowerCamelCase ) return hidden_states class lowerCAmelCase__( nn.Module ): '''simple docstring''' def __init__( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> List[str]: super().__init__() _SCREAMING_SNAKE_CASE : Optional[Any] = Attention(query_dim=__lowerCamelCase , heads=__lowerCamelCase , dim_head=__lowerCamelCase , out_bias=__lowerCamelCase , scale_qk=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : List[str] = TaLayerNorm(__lowerCamelCase , eps=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[Any] = nn.Dropout(__lowerCamelCase ) def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase=None , __lowerCamelCase=None , ) -> List[Any]: _SCREAMING_SNAKE_CASE : Tuple = self.layer_norm(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : str = self.attention( __lowerCamelCase , encoder_hidden_states=__lowerCamelCase , attention_mask=attention_mask.squeeze(1 ) , ) _SCREAMING_SNAKE_CASE : Union[str, Any] = hidden_states + self.dropout(__lowerCamelCase ) return layer_output class lowerCAmelCase__( nn.Module ): '''simple docstring''' def __init__( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> List[Any]: super().__init__() _SCREAMING_SNAKE_CASE : Tuple = TaDenseGatedActDense(d_model=__lowerCamelCase , d_ff=__lowerCamelCase , dropout_rate=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : str = TaFiLMLayer(in_features=d_model * 4 , out_features=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : str = TaLayerNorm(__lowerCamelCase , eps=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[Any] = nn.Dropout(__lowerCamelCase ) def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase=None ) -> List[str]: _SCREAMING_SNAKE_CASE : Optional[int] = self.layer_norm(__lowerCamelCase ) if conditioning_emb is not None: _SCREAMING_SNAKE_CASE : Union[str, Any] = self.film(__lowerCamelCase , __lowerCamelCase ) _SCREAMING_SNAKE_CASE : Any = self.DenseReluDense(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : List[Any] = hidden_states + self.dropout(__lowerCamelCase ) return hidden_states class lowerCAmelCase__( nn.Module ): '''simple docstring''' def __init__( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) -> Optional[Any]: super().__init__() _SCREAMING_SNAKE_CASE : Union[str, Any] = nn.Linear(__lowerCamelCase , __lowerCamelCase , bias=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : List[str] = nn.Linear(__lowerCamelCase , __lowerCamelCase , bias=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Dict = nn.Linear(__lowerCamelCase , __lowerCamelCase , bias=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Union[str, Any] = nn.Dropout(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Union[str, Any] = NewGELUActivation() def UpperCamelCase_ ( self , __lowerCamelCase ) -> Any: _SCREAMING_SNAKE_CASE : Dict = self.act(self.wi_a(__lowerCamelCase ) ) _SCREAMING_SNAKE_CASE : Dict = self.wi_a(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : int = hidden_gelu * hidden_linear _SCREAMING_SNAKE_CASE : Optional[int] = self.dropout(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[int] = self.wo(__lowerCamelCase ) return hidden_states class lowerCAmelCase__( nn.Module ): '''simple docstring''' def __init__( self , __lowerCamelCase , __lowerCamelCase=1E-6 ) -> int: super().__init__() _SCREAMING_SNAKE_CASE : Dict = nn.Parameter(torch.ones(__lowerCamelCase ) ) _SCREAMING_SNAKE_CASE : str = eps def UpperCamelCase_ ( self , __lowerCamelCase ) -> Optional[Any]: # T5 uses a layer_norm which only scales and doesn't shift, which is also known as Root Mean # Square Layer Normalization https://arxiv.org/abs/1910.07467 thus variance is calculated # w/o mean and there is no bias. Additionally we want to make sure that the accumulation for # half-precision inputs is done in fp32 _SCREAMING_SNAKE_CASE : Tuple = hidden_states.to(torch.floataa ).pow(2 ).mean(-1 , keepdim=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : List[Any] = hidden_states * torch.rsqrt(variance + self.variance_epsilon ) # convert into half-precision if necessary if self.weight.dtype in [torch.floataa, torch.bfloataa]: _SCREAMING_SNAKE_CASE : str = hidden_states.to(self.weight.dtype ) return self.weight * hidden_states class lowerCAmelCase__( nn.Module ): '''simple docstring''' def UpperCamelCase_ ( self , __lowerCamelCase ) -> torch.Tensor: return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi ) * (input + 0.04_4715 * torch.pow(__lowerCamelCase , 3.0 )) )) class lowerCAmelCase__( nn.Module ): '''simple docstring''' def __init__( self , __lowerCamelCase , __lowerCamelCase ) -> Optional[Any]: super().__init__() _SCREAMING_SNAKE_CASE : Any = nn.Linear(__lowerCamelCase , out_features * 2 , bias=__lowerCamelCase ) def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase ) -> Dict: _SCREAMING_SNAKE_CASE : List[Any] = self.scale_bias(__lowerCamelCase ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : Dict = torch.chunk(__lowerCamelCase , 2 , -1 ) _SCREAMING_SNAKE_CASE : Optional[int] = x * (1 + scale) + shift return x
325
1
'''simple docstring''' import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=A_ ) class _snake_case ( A_ ): _A : Optional[Any] = field(default='''image-classification''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) _A : List[Any] = Features({'''image''': Image()} ) _A : Any = Features({'''labels''': ClassLabel} ) _A : Optional[int] = '''image''' _A : int = '''labels''' def __UpperCamelCase ( self : Optional[int] ,SCREAMING_SNAKE_CASE__ : str ): if self.label_column not in features: raise ValueError(F'''Column {self.label_column} is not present in features.''' ) if not isinstance(features[self.label_column] ,_lowerCamelCase ): raise ValueError(F'''Column {self.label_column} is not a ClassLabel.''' ) SCREAMING_SNAKE_CASE:List[Any] = copy.deepcopy(self ) SCREAMING_SNAKE_CASE:str = self.label_schema.copy() SCREAMING_SNAKE_CASE:Dict = features[self.label_column] SCREAMING_SNAKE_CASE:Tuple = label_schema return task_template @property def __UpperCamelCase ( self : int ): return { self.image_column: "image", self.label_column: "labels", }
139
"""simple docstring""" from math import sqrt def _UpperCAmelCase ( __lowerCamelCase : int = 1_00_00_00 ) -> int: _snake_case = 0 _snake_case = 0 _snake_case = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(F"{solution() = }")
288
0
"""simple docstring""" import argparse import json import os from collections import OrderedDict import torch from transformers import LukeConfig, LukeForMaskedLM, MLukeTokenizer, XLMRobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def UpperCAmelCase__ ( lowerCAmelCase__ :int , lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Any , lowerCAmelCase__ :int , lowerCAmelCase__ :str ) -> Union[str, Any]: with open(lowerCAmelCase__ ) as metadata_file: lowercase = json.load(lowerCAmelCase__ ) lowercase = LukeConfig(use_entity_aware_attention=lowerCAmelCase__ , **metadata["""model_config"""] ) # Load in the weights from the checkpoint_path lowercase = torch.load(lowerCAmelCase__ , map_location="""cpu""" )["""module"""] # Load the entity vocab file lowercase = load_original_entity_vocab(lowerCAmelCase__ ) # add an entry for [MASK2] lowercase = max(entity_vocab.values() ) + 1 config.entity_vocab_size += 1 lowercase = XLMRobertaTokenizer.from_pretrained(metadata["""model_config"""]["""bert_model_name"""] ) # Add special tokens to the token vocabulary for downstream tasks lowercase = AddedToken("""<ent>""" , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) lowercase = AddedToken("""<ent2>""" , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) tokenizer.add_special_tokens({"""additional_special_tokens""": [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(f'Saving tokenizer to {pytorch_dump_folder_path}' ) tokenizer.save_pretrained(lowerCAmelCase__ ) with open(os.path.join(lowerCAmelCase__ , """tokenizer_config.json""" ) , """r""" ) as f: lowercase = json.load(lowerCAmelCase__ ) lowercase = """MLukeTokenizer""" with open(os.path.join(lowerCAmelCase__ , """tokenizer_config.json""" ) , """w""" ) as f: json.dump(lowerCAmelCase__ , lowerCAmelCase__ ) with open(os.path.join(lowerCAmelCase__ , MLukeTokenizer.vocab_files_names["""entity_vocab_file"""] ) , """w""" ) as f: json.dump(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = MLukeTokenizer.from_pretrained(lowerCAmelCase__ ) # Initialize the embeddings of the special tokens lowercase = tokenizer.convert_tokens_to_ids(["""@"""] )[0] lowercase = tokenizer.convert_tokens_to_ids(["""#"""] )[0] lowercase = state_dict["""embeddings.word_embeddings.weight"""] lowercase = word_emb[ent_init_index].unsqueeze(0 ) lowercase = word_emb[enta_init_index].unsqueeze(0 ) lowercase = torch.cat([word_emb, ent_emb, enta_emb] ) # add special tokens for 'entity_predictions.bias' for bias_name in ["lm_head.decoder.bias", "lm_head.bias"]: lowercase = state_dict[bias_name] lowercase = decoder_bias[ent_init_index].unsqueeze(0 ) lowercase = decoder_bias[enta_init_index].unsqueeze(0 ) lowercase = torch.cat([decoder_bias, ent_decoder_bias, enta_decoder_bias] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: lowercase = f'encoder.layer.{layer_index}.attention.self.' lowercase = state_dict[prefix + matrix_name] lowercase = state_dict[prefix + matrix_name] lowercase = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks lowercase = state_dict["""entity_embeddings.entity_embeddings.weight"""] lowercase = entity_emb[entity_vocab["""[MASK]"""]].unsqueeze(0 ) lowercase = torch.cat([entity_emb, entity_mask_emb] ) # add [MASK2] for 'entity_predictions.bias' lowercase = state_dict["""entity_predictions.bias"""] lowercase = entity_prediction_bias[entity_vocab["""[MASK]"""]].unsqueeze(0 ) lowercase = torch.cat([entity_prediction_bias, entity_mask_bias] ) lowercase = LukeForMaskedLM(config=lowerCAmelCase__ ).eval() state_dict.pop("""entity_predictions.decoder.weight""" ) state_dict.pop("""lm_head.decoder.weight""" ) state_dict.pop("""lm_head.decoder.bias""" ) lowercase = OrderedDict() for key, value in state_dict.items(): if not (key.startswith("""lm_head""" ) or key.startswith("""entity_predictions""" )): lowercase = state_dict[key] else: lowercase = state_dict[key] lowercase , lowercase = model.load_state_dict(lowerCAmelCase__ , strict=lowerCAmelCase__ ) if set(lowerCAmelCase__ ) != {"luke.embeddings.position_ids"}: raise ValueError(f'Unexpected unexpected_keys: {unexpected_keys}' ) if set(lowerCAmelCase__ ) != { "lm_head.decoder.weight", "lm_head.decoder.bias", "entity_predictions.decoder.weight", }: raise ValueError(f'Unexpected missing_keys: {missing_keys}' ) model.tie_weights() assert (model.luke.embeddings.word_embeddings.weight == model.lm_head.decoder.weight).all() assert (model.luke.entity_embeddings.entity_embeddings.weight == model.entity_predictions.decoder.weight).all() # Check outputs lowercase = MLukeTokenizer.from_pretrained(lowerCAmelCase__ , task="""entity_classification""" ) lowercase = """ISO 639-3 uses the code fas for the dialects spoken across Iran and アフガニスタン (Afghanistan).""" lowercase = (0, 9) lowercase = tokenizer(lowerCAmelCase__ , entity_spans=[span] , return_tensors="""pt""" ) lowercase = model(**lowerCAmelCase__ ) # Verify word hidden states if model_size == "large": raise NotImplementedError else: # base lowercase = torch.Size((1, 3_3, 7_6_8) ) lowercase = torch.tensor([[0.0_892, 0.0_596, -0.2_819], [0.0_134, 0.1_199, 0.0_573], [-0.0_169, 0.0_927, 0.0_644]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( f'Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}' ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowerCAmelCase__ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": raise NotImplementedError else: # base lowercase = torch.Size((1, 1, 7_6_8) ) lowercase = torch.tensor([[-0.1_482, 0.0_609, 0.0_322]] ) if not (outputs.entity_last_hidden_state.shape == expected_shape): raise ValueError( f'Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is' f' {expected_shape}' ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , lowerCAmelCase__ , atol=1e-4 ): raise ValueError # Verify masked word/entity prediction lowercase = MLukeTokenizer.from_pretrained(lowerCAmelCase__ ) lowercase = """Tokyo is the capital of <mask>.""" lowercase = (2_4, 3_0) lowercase = tokenizer(lowerCAmelCase__ , entity_spans=[span] , return_tensors="""pt""" ) lowercase = model(**lowerCAmelCase__ ) lowercase = encoding["""input_ids"""][0].tolist() lowercase = input_ids.index(tokenizer.convert_tokens_to_ids("""<mask>""" ) ) lowercase = outputs.logits[0][mask_position_id].argmax(dim=-1 ) assert "Japan" == tokenizer.decode(lowerCAmelCase__ ) lowercase = outputs.entity_logits[0][0].argmax().item() lowercase = [ entity for entity, entity_id in tokenizer.entity_vocab.items() if entity_id == predicted_entity_id ] assert [e for e in multilingual_predicted_entities if e.startswith("""en:""" )][0] == "en:Japan" # Finally, save our PyTorch model and tokenizer print("""Saving PyTorch model to {}""".format(lowerCAmelCase__ ) ) model.save_pretrained(lowerCAmelCase__ ) def UpperCAmelCase__ ( lowerCAmelCase__ :Union[str, Any] ) -> Tuple: lowercase = ["""[MASK]""", """[PAD]""", """[UNK]"""] lowercase = [json.loads(lowerCAmelCase__ ) for line in open(lowerCAmelCase__ )] lowercase = {} for entry in data: lowercase = entry["""id"""] for entity_name, language in entry["entities"]: if entity_name in SPECIAL_TOKENS: lowercase = entity_id break lowercase = f'{language}:{entity_name}' lowercase = entity_id return new_mapping if __name__ == "__main__": __lowerCAmelCase : str =argparse.ArgumentParser() # Required parameters parser.add_argument("""--checkpoint_path""", type=str, help="""Path to a pytorch_model.bin file.""") parser.add_argument( """--metadata_path""", default=None, type=str, help="""Path to a metadata.json file, defining the configuration.""" ) parser.add_argument( """--entity_vocab_path""", default=None, type=str, help="""Path to an entity_vocab.tsv file, containing the entity vocabulary.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to where to dump the output PyTorch model.""" ) parser.add_argument( """--model_size""", default="""base""", type=str, choices=["""base""", """large"""], help="""Size of the model to be converted.""" ) __lowerCAmelCase : Any =parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
360
"""simple docstring""" class _A : def __init__( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): """simple docstring""" lowercase = None lowercase = None lowercase = graph self._normalize_graph(__lowerCAmelCase , __lowerCAmelCase ) lowercase = len(__lowerCAmelCase ) lowercase = None def A__ ( self , __lowerCAmelCase , __lowerCAmelCase ): """simple docstring""" if sources is int: lowercase = [sources] if sinks is int: lowercase = [sinks] if len(__lowerCAmelCase ) == 0 or len(__lowerCAmelCase ) == 0: return lowercase = sources[0] lowercase = sinks[0] # make fake vertex if there are more # than one source or sink if len(__lowerCAmelCase ) > 1 or len(__lowerCAmelCase ) > 1: lowercase = 0 for i in sources: max_input_flow += sum(self.graph[i] ) lowercase = len(self.graph ) + 1 for room in self.graph: room.insert(0 , 0 ) self.graph.insert(0 , [0] * size ) for i in sources: lowercase = max_input_flow lowercase = 0 lowercase = len(self.graph ) + 1 for room in self.graph: room.append(0 ) self.graph.append([0] * size ) for i in sinks: lowercase = max_input_flow lowercase = size - 1 def A__ ( self ): """simple docstring""" if self.maximum_flow_algorithm is None: raise Exception("""You need to set maximum flow algorithm before.""" ) if self.source_index is None or self.sink_index is None: return 0 self.maximum_flow_algorithm.execute() return self.maximum_flow_algorithm.getMaximumFlow() def A__ ( self , __lowerCAmelCase ): """simple docstring""" lowercase = algorithm(self ) class _A : def __init__( self , __lowerCAmelCase ): """simple docstring""" lowercase = flow_network lowercase = flow_network.verticesCount lowercase = flow_network.sourceIndex lowercase = flow_network.sinkIndex # it's just a reference, so you shouldn't change # it in your algorithms, use deep copy before doing that lowercase = flow_network.graph lowercase = False def A__ ( self ): """simple docstring""" if not self.executed: self._algorithm() lowercase = True def A__ ( self ): """simple docstring""" pass class _A ( lowerCAmelCase ): def __init__( self , __lowerCAmelCase ): """simple docstring""" super().__init__(__lowerCAmelCase ) # use this to save your result lowercase = -1 def A__ ( self ): """simple docstring""" if not self.executed: raise Exception("""You should execute algorithm before using its result!""" ) return self.maximum_flow class _A ( lowerCAmelCase ): def __init__( self , __lowerCAmelCase ): """simple docstring""" super().__init__(__lowerCAmelCase ) lowercase = [[0] * self.verticies_count for i in range(self.verticies_count )] lowercase = [0] * self.verticies_count lowercase = [0] * self.verticies_count def A__ ( self ): """simple docstring""" lowercase = self.verticies_count # push some substance to graph for nextvertex_index, bandwidth in enumerate(self.graph[self.source_index] ): self.preflow[self.source_index][nextvertex_index] += bandwidth self.preflow[nextvertex_index][self.source_index] -= bandwidth self.excesses[nextvertex_index] += bandwidth # Relabel-to-front selection rule lowercase = [ i for i in range(self.verticies_count ) if i != self.source_index and i != self.sink_index ] # move through list lowercase = 0 while i < len(__lowerCAmelCase ): lowercase = vertices_list[i] lowercase = self.heights[vertex_index] self.process_vertex(__lowerCAmelCase ) if self.heights[vertex_index] > previous_height: # if it was relabeled, swap elements # and start from 0 index vertices_list.insert(0 , vertices_list.pop(__lowerCAmelCase ) ) lowercase = 0 else: i += 1 lowercase = sum(self.preflow[self.source_index] ) def A__ ( self , __lowerCAmelCase ): """simple docstring""" while self.excesses[vertex_index] > 0: for neighbour_index in range(self.verticies_count ): # if it's neighbour and current vertex is higher if ( self.graph[vertex_index][neighbour_index] - self.preflow[vertex_index][neighbour_index] > 0 and self.heights[vertex_index] > self.heights[neighbour_index] ): self.push(__lowerCAmelCase , __lowerCAmelCase ) self.relabel(__lowerCAmelCase ) def A__ ( self , __lowerCAmelCase , __lowerCAmelCase ): """simple docstring""" lowercase = min( self.excesses[from_index] , self.graph[from_index][to_index] - self.preflow[from_index][to_index] , ) self.preflow[from_index][to_index] += preflow_delta self.preflow[to_index][from_index] -= preflow_delta self.excesses[from_index] -= preflow_delta self.excesses[to_index] += preflow_delta def A__ ( self , __lowerCAmelCase ): """simple docstring""" lowercase = None for to_index in range(self.verticies_count ): if ( self.graph[vertex_index][to_index] - self.preflow[vertex_index][to_index] > 0 ) and (min_height is None or self.heights[to_index] < min_height): lowercase = self.heights[to_index] if min_height is not None: lowercase = min_height + 1 if __name__ == "__main__": __lowerCAmelCase : int =[0] __lowerCAmelCase : List[Any] =[3] # graph = [ # [0, 0, 4, 6, 0, 0], # [0, 0, 5, 2, 0, 0], # [0, 0, 0, 0, 4, 4], # [0, 0, 0, 0, 6, 6], # [0, 0, 0, 0, 0, 0], # [0, 0, 0, 0, 0, 0], # ] __lowerCAmelCase : Optional[int] =[[0, 7, 0, 0], [0, 0, 6, 0], [0, 0, 0, 8], [9, 0, 0, 0]] # prepare our network __lowerCAmelCase : Tuple =FlowNetwork(graph, entrances, exits) # set algorithm flow_network.set_maximum_flow_algorithm(PushRelabelExecutor) # and calculate __lowerCAmelCase : Optional[int] =flow_network.find_maximum_flow() print(F"""maximum flow is {maximum_flow}""")
32
0
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _UpperCamelCase = logging.get_logger(__name__) class __lowercase (_UpperCAmelCase ): _UpperCamelCase = ["""pixel_values"""] def __init__( self , A_ = True , A_ = 1 / 255 , A_ = True , A_ = 8 , **A_ , ) ->None: '''simple docstring''' super().__init__(**A_ ) __lowerCAmelCase : Tuple = do_rescale __lowerCAmelCase : int = rescale_factor __lowerCAmelCase : List[Any] = do_pad __lowerCAmelCase : Any = pad_size def UpperCamelCase__ ( self , A_ , A_ , A_ = None , **A_ ) ->np.ndarray: '''simple docstring''' return rescale(A_ , scale=A_ , data_format=A_ , **A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ = None ) ->int: '''simple docstring''' __lowerCAmelCase, __lowerCAmelCase : str = get_image_size(A_ ) __lowerCAmelCase : Tuple = (old_height // size + 1) * size - old_height __lowerCAmelCase : Tuple = (old_width // size + 1) * size - old_width return pad(A_ , ((0, pad_height), (0, pad_width)) , mode='''symmetric''' , data_format=A_ ) def UpperCamelCase__ ( self , A_ , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = ChannelDimension.FIRST , **A_ , ) ->Any: '''simple docstring''' __lowerCAmelCase : Tuple = do_rescale if do_rescale is not None else self.do_rescale __lowerCAmelCase : Tuple = rescale_factor if rescale_factor is not None else self.rescale_factor __lowerCAmelCase : Dict = do_pad if do_pad is not None else self.do_pad __lowerCAmelCase : str = pad_size if pad_size is not None else self.pad_size __lowerCAmelCase : Any = make_list_of_images(A_ ) if not valid_images(A_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) # All transformations expect numpy arrays. __lowerCAmelCase : Union[str, Any] = [to_numpy_array(A_ ) for image in images] if do_rescale: __lowerCAmelCase : Optional[int] = [self.rescale(image=A_ , scale=A_ ) for image in images] if do_pad: __lowerCAmelCase : List[str] = [self.pad(A_ , size=A_ ) for image in images] __lowerCAmelCase : Optional[int] = [to_channel_dimension_format(A_ , A_ ) for image in images] __lowerCAmelCase : Any = {'''pixel_values''': images} return BatchFeature(data=A_ , tensor_type=A_ )
275
from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class __lowercase (_UpperCAmelCase ): _UpperCamelCase = 42 _UpperCamelCase = 42 _UpperCamelCase = None class __lowercase (_UpperCAmelCase , _UpperCAmelCase ): _UpperCamelCase = 2 @register_to_config def __init__( self , A_ = 0.02 , A_ = 100 , A_ = 1.007 , A_ = 80 , A_ = 0.05 , A_ = 50 , ) ->int: '''simple docstring''' __lowerCAmelCase : Optional[int] = sigma_max # setable values __lowerCAmelCase : int = None __lowerCAmelCase : np.IntTensor = None __lowerCAmelCase : torch.FloatTensor = None # sigma(t_i) def UpperCamelCase__ ( self , A_ , A_ = None ) ->torch.FloatTensor: '''simple docstring''' return sample def UpperCamelCase__ ( self , A_ , A_ = None ) ->List[str]: '''simple docstring''' __lowerCAmelCase : str = num_inference_steps __lowerCAmelCase : Dict = np.arange(0 , self.num_inference_steps )[::-1].copy() __lowerCAmelCase : Optional[Any] = torch.from_numpy(A_ ).to(A_ ) __lowerCAmelCase : Tuple = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] __lowerCAmelCase : Optional[int] = torch.tensor(A_ , dtype=torch.floataa , device=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ = None ) ->Tuple[torch.FloatTensor, float]: '''simple docstring''' if self.config.s_min <= sigma <= self.config.s_max: __lowerCAmelCase : List[str] = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: __lowerCAmelCase : List[str] = 0 # sample eps ~ N(0, S_noise^2 * I) __lowerCAmelCase : int = self.config.s_noise * randn_tensor(sample.shape , generator=A_ ).to(sample.device ) __lowerCAmelCase : str = sigma + gamma * sigma __lowerCAmelCase : Any = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ = True , ) ->Union[KarrasVeOutput, Tuple]: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = sample_hat + sigma_hat * model_output __lowerCAmelCase : int = (sample_hat - pred_original_sample) / sigma_hat __lowerCAmelCase : Tuple = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=A_ , derivative=A_ , pred_original_sample=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ = True , ) ->Union[KarrasVeOutput, Tuple]: '''simple docstring''' __lowerCAmelCase : str = sample_prev + sigma_prev * model_output __lowerCAmelCase : List[Any] = (sample_prev - pred_original_sample) / sigma_prev __lowerCAmelCase : Any = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=A_ , derivative=A_ , pred_original_sample=A_ ) def UpperCamelCase__ ( self , A_ , A_ , A_ ) ->Any: '''simple docstring''' raise NotImplementedError()
275
1
from math import pow, sqrt def A ( *_UpperCAmelCase : float ) -> bool: '''simple docstring''' _UpperCAmelCase = len(_UpperCAmelCase ) > 0 and all(value > 0.0 for value in values ) return result def A ( _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float | ValueError: '''simple docstring''' return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_UpperCAmelCase , _UpperCAmelCase ) else ValueError('Input Error: Molar mass values must greater than 0.' ) ) def A ( _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float | ValueError: '''simple docstring''' return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def A ( _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float | ValueError: '''simple docstring''' return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def A ( _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float | ValueError: '''simple docstring''' return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) ) def A ( _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float | ValueError: '''simple docstring''' return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) else ValueError( 'Input Error: Molar mass and effusion rate values must greater than 0.' ) )
290
UpperCAmelCase__ = { "A": ["B", "C", "E"], "B": ["A", "D", "E"], "C": ["A", "F", "G"], "D": ["B"], "E": ["A", "B", "D"], "F": ["C"], "G": ["C"], } def A ( _UpperCAmelCase : dict , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : List[Any] ) -> list[str]: '''simple docstring''' _UpperCAmelCase = set() # keep track of all the paths to be checked _UpperCAmelCase = [[start]] # return path if start is goal if start == goal: return [start] # keeps looping until all possible paths have been checked while queue: # pop the first path from the queue _UpperCAmelCase = queue.pop(0 ) # get the last node from the path _UpperCAmelCase = path[-1] if node not in explored: _UpperCAmelCase = graph[node] # go through all neighbour nodes, construct a new path and # push it into the queue for neighbour in neighbours: _UpperCAmelCase = list(_UpperCAmelCase ) new_path.append(_UpperCAmelCase ) queue.append(_UpperCAmelCase ) # return path if neighbour is goal if neighbour == goal: return new_path # mark node as explored explored.add(_UpperCAmelCase ) # in case there's no path between the 2 nodes return [] def A ( _UpperCAmelCase : dict , _UpperCAmelCase : Any , _UpperCAmelCase : List[str] ) -> int: '''simple docstring''' if not graph or start not in graph or target not in graph: return -1 if start == target: return 0 _UpperCAmelCase = [start] _UpperCAmelCase = set(_UpperCAmelCase ) # Keep tab on distances from `start` node. _UpperCAmelCase = {start: 0, target: -1} while queue: _UpperCAmelCase = queue.pop(0 ) if node == target: _UpperCAmelCase = ( dist[node] if dist[target] == -1 else min(dist[target] , dist[node] ) ) for adjacent in graph[node]: if adjacent not in visited: visited.add(_UpperCAmelCase ) queue.append(_UpperCAmelCase ) _UpperCAmelCase = dist[node] + 1 return dist[target] if __name__ == "__main__": print(bfs_shortest_path(demo_graph, "G", "D")) # returns ['G', 'C', 'A', 'B', 'D'] print(bfs_shortest_path_distance(demo_graph, "G", "D")) # returns 4
290
1
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): _lowerCamelCase : int = 0 def A_ ( self ): _lowerCamelCase : Dict = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32' ) self.assertIsInstance(__a , __a ) def A_ ( self ): with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : Dict = Path(__a ) / 'preprocessor_config.json' _lowerCamelCase : int = Path(__a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(__a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(__a , 'w' ) ) _lowerCamelCase : List[str] = AutoImageProcessor.from_pretrained(__a ) self.assertIsInstance(__a , __a ) def A_ ( self ): # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : Tuple = Path(__a ) / 'preprocessor_config.json' _lowerCamelCase : List[Any] = Path(__a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(__a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(__a , 'w' ) ) _lowerCamelCase : int = AutoImageProcessor.from_pretrained(__a ) self.assertIsInstance(__a , __a ) def A_ ( self ): with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : Optional[Any] = CLIPConfig() # Create a dummy config file with image_proceesor_type _lowerCamelCase : List[Any] = Path(__a ) / 'preprocessor_config.json' _lowerCamelCase : List[str] = Path(__a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(__a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(__a , 'w' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally _lowerCamelCase : Dict = AutoImageProcessor.from_pretrained(__a ).to_dict() config_dict.pop('image_processor_type' ) _lowerCamelCase : Union[str, Any] = CLIPImageProcessor(**__a ) # save in new folder model_config.save_pretrained(__a ) config.save_pretrained(__a ) _lowerCamelCase : Any = AutoImageProcessor.from_pretrained(__a ) # make sure private variable is not incorrectly saved _lowerCamelCase : Optional[Any] = json.loads(config.to_json_string() ) self.assertTrue('_processor_class' not in dict_as_saved ) self.assertIsInstance(__a , __a ) def A_ ( self ): with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : List[Any] = Path(__a ) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(__a , 'w' ) , ) _lowerCamelCase : Optional[Any] = AutoImageProcessor.from_pretrained(__a ) self.assertIsInstance(__a , __a ) def A_ ( self ): with self.assertRaisesRegex( __a , 'clip-base is not a local folder and is not a valid model identifier' ): _lowerCamelCase : Union[str, Any] = AutoImageProcessor.from_pretrained('clip-base' ) def A_ ( self ): with self.assertRaisesRegex( __a , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): _lowerCamelCase : str = AutoImageProcessor.from_pretrained(__a , revision='aaaaaa' ) def A_ ( self ): with self.assertRaisesRegex( __a , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): _lowerCamelCase : Optional[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model' ) def A_ ( self ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(__a ): _lowerCamelCase : Optional[int] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(__a ): _lowerCamelCase : Optional[Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=__a ) _lowerCamelCase : Dict = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=__a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(__a ) _lowerCamelCase : List[str] = AutoImageProcessor.from_pretrained(__a , trust_remote_code=__a ) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor' ) def A_ ( self ): try: AutoConfig.register('custom' , __a ) AutoImageProcessor.register(__a , __a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__a ): AutoImageProcessor.register(__a , __a ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCamelCase : Optional[int] = Path(__a ) / 'preprocessor_config.json' _lowerCamelCase : List[str] = Path(__a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(__a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(__a , 'w' ) ) _lowerCamelCase : List[Any] = CustomImageProcessor.from_pretrained(__a ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(__a ) _lowerCamelCase : Union[str, Any] = AutoImageProcessor.from_pretrained(__a ) self.assertIsInstance(__a , __a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def A_ ( self ): class lowerCAmelCase__ ( lowerCAmelCase__ ): '''simple docstring''' lowerCamelCase__ = True try: AutoConfig.register('custom' , __a ) AutoImageProcessor.register(__a , __a ) # If remote code is not set, the default is to use local _lowerCamelCase : Optional[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. _lowerCamelCase : str = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=__a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub _lowerCamelCase : Any = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=__a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(not hasattr(__a , 'is_local' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
96
"""simple docstring""" class _UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' pass class _UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' pass class _UpperCamelCase : '''simple docstring''' def __init__( self ): __lowerCAmelCase = [ [], [], [], ] def snake_case ( self , __a , __a ): try: if len(self.queues[priority] ) >= 1_00: raise OverflowError("Maximum queue size is 100" ) self.queues[priority].append(__a ) except IndexError: raise ValueError("Valid priorities are 0, 1, and 2" ) def snake_case ( self ): for queue in self.queues: if queue: return queue.pop(0 ) raise UnderFlowError("All queues are empty" ) def __str__( self ): return "\n".join(f"Priority {i}: {q}" for i, q in enumerate(self.queues ) ) class _UpperCamelCase : '''simple docstring''' def __init__( self ): __lowerCAmelCase = [] def snake_case ( self , __a ): if len(self.queue ) == 1_00: raise OverFlowError("Maximum queue size is 100" ) self.queue.append(__a ) def snake_case ( self ): if not self.queue: raise UnderFlowError("The queue is empty" ) else: __lowerCAmelCase = min(self.queue ) self.queue.remove(__a ) return data def __str__( self ): return str(self.queue ) def _lowerCamelCase ( ): '''simple docstring''' __lowerCAmelCase = FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 100 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 128 ) print(_UpperCamelCase ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(_UpperCamelCase ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def _lowerCamelCase ( ): '''simple docstring''' __lowerCAmelCase = ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(100 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(128 ) print(_UpperCamelCase ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(_UpperCamelCase ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
57
0
'''simple docstring''' from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class SCREAMING_SNAKE_CASE (a__ ): lowerCAmelCase = ['''image_processor''', '''tokenizer'''] lowerCAmelCase = '''BlipImageProcessor''' lowerCAmelCase = ('''BertTokenizer''', '''BertTokenizerFast''') def __init__( self , _UpperCAmelCase , _UpperCAmelCase): '''simple docstring''' __A : Union[str, Any] = False super().__init__(_UpperCAmelCase , _UpperCAmelCase) __A : Optional[int] = self.image_processor def __call__( self , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = True , _UpperCAmelCase = False , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = 0 , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = False , _UpperCAmelCase = False , _UpperCAmelCase = False , _UpperCAmelCase = False , _UpperCAmelCase = False , _UpperCAmelCase = True , _UpperCAmelCase = None , **_UpperCAmelCase , ): '''simple docstring''' if images is None and text is None: raise ValueError('You have to specify either images or text.') # Get only text if images is None: __A : str = self.tokenizer __A : List[Any] = self.tokenizer( text=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , max_length=_UpperCAmelCase , stride=_UpperCAmelCase , pad_to_multiple_of=_UpperCAmelCase , return_attention_mask=_UpperCAmelCase , return_overflowing_tokens=_UpperCAmelCase , return_special_tokens_mask=_UpperCAmelCase , return_offsets_mapping=_UpperCAmelCase , return_token_type_ids=_UpperCAmelCase , return_length=_UpperCAmelCase , verbose=_UpperCAmelCase , return_tensors=_UpperCAmelCase , **_UpperCAmelCase , ) return text_encoding # add pixel_values __A : Union[str, Any] = self.image_processor(_UpperCAmelCase , return_tensors=_UpperCAmelCase) if text is not None: __A : str = self.tokenizer( text=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , max_length=_UpperCAmelCase , stride=_UpperCAmelCase , pad_to_multiple_of=_UpperCAmelCase , return_attention_mask=_UpperCAmelCase , return_overflowing_tokens=_UpperCAmelCase , return_special_tokens_mask=_UpperCAmelCase , return_offsets_mapping=_UpperCAmelCase , return_token_type_ids=_UpperCAmelCase , return_length=_UpperCAmelCase , verbose=_UpperCAmelCase , return_tensors=_UpperCAmelCase , **_UpperCAmelCase , ) else: __A : Optional[Any] = None if text_encoding is not None: encoding_image_processor.update(_UpperCAmelCase) return encoding_image_processor def SCREAMING_SNAKE_CASE ( self , *_UpperCAmelCase , **_UpperCAmelCase): '''simple docstring''' return self.tokenizer.batch_decode(*_UpperCAmelCase , **_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self , *_UpperCAmelCase , **_UpperCAmelCase): '''simple docstring''' return self.tokenizer.decode(*_UpperCAmelCase , **_UpperCAmelCase) @property def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Tuple = self.tokenizer.model_input_names __A : Optional[Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names))
354
'''simple docstring''' import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse('''9.1.0'''): lowercase__ : Dict = { '''linear''': PIL.Image.Resampling.BILINEAR, '''bilinear''': PIL.Image.Resampling.BILINEAR, '''bicubic''': PIL.Image.Resampling.BICUBIC, '''lanczos''': PIL.Image.Resampling.LANCZOS, '''nearest''': PIL.Image.Resampling.NEAREST, } else: lowercase__ : Any = { '''linear''': PIL.Image.LINEAR, '''bilinear''': PIL.Image.BILINEAR, '''bicubic''': PIL.Image.BICUBIC, '''lanczos''': PIL.Image.LANCZOS, '''nearest''': PIL.Image.NEAREST, } def _lowerCAmelCase ( __snake_case : Any ) -> Optional[Any]: __A : Dict = (images / 2 + 0.5).clamp(0 , 1 ) __A : str = images.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() __A : Dict = numpy_to_pil(__snake_case ) return images def _lowerCAmelCase ( __snake_case : List[Any] ) -> Optional[Any]: if images.ndim == 3: __A : List[Any] = images[None, ...] __A : List[str] = (images * 2_55).round().astype('uint8' ) if images.shape[-1] == 1: # special case for grayscale (single channel) images __A : str = [Image.fromarray(image.squeeze() , mode='L' ) for image in images] else: __A : str = [Image.fromarray(__snake_case ) for image in images] return pil_images
190
0
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model _lowerCamelCase ="""0.12""" # assumed parallelism: 8 if is_torch_available(): import torch def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase=None ): if rng is None: lowerCamelCase : Tuple = random.Random() lowerCamelCase : Tuple = 1 for dim in shape: total_dims *= dim lowerCamelCase : List[str] = [] for _ in range(lowerCamelCase ): values.append(rng.randint(0, vocab_size - 1 ) ) lowerCamelCase : Dict = np.array(lowerCamelCase, dtype=jnp.intaa ).reshape(lowerCamelCase ) return output def _a ( lowerCamelCase, lowerCamelCase=None ): lowerCamelCase : Tuple = ids_tensor(lowerCamelCase, vocab_size=2, rng=lowerCamelCase ) # make sure that at least one token is attended to for each batch lowerCamelCase : str = 1 return attn_mask @require_flax class A__ : _UpperCAmelCase : Dict = None _UpperCAmelCase : List[str] = () def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 lowerCamelCase : List[str] = 2 lowerCamelCase : Optional[int] = inputs["""input_ids"""].shape[-1] // 2 lowerCamelCase : Any = inputs["""input_ids"""][:max_batch_size, :sequence_length] lowerCamelCase : str = jnp.ones_like(__magic_name__ ) lowerCamelCase : str = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens lowerCamelCase : Tuple = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` lowerCamelCase : Dict = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : List[Any] = self._get_input_ids_and_config() lowerCamelCase : List[Any] = False lowerCamelCase : List[str] = max_length lowerCamelCase : str = 0 for model_class in self.all_generative_model_classes: lowerCamelCase : Optional[int] = model_class(__magic_name__ ) lowerCamelCase : Optional[Any] = model_class.__name__[4:] # Skip the "Flax" at the beginning lowerCamelCase : int = getattr(__magic_name__ , __magic_name__ ) lowerCamelCase : List[Any] = pt_model_class(__magic_name__ ).eval() lowerCamelCase : List[str] = load_flax_weights_in_pytorch_model(__magic_name__ , flax_model.params ) lowerCamelCase : Union[str, Any] = flax_model.generate(__magic_name__ ).sequences lowerCamelCase : Union[str, Any] = pt_model.generate(torch.tensor(__magic_name__ , dtype=torch.long ) ) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: lowerCamelCase : int = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() , flax_generation_outputs.tolist() ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Union[str, Any] = self._get_input_ids_and_config() lowerCamelCase : Tuple = False lowerCamelCase : Tuple = max_length for model_class in self.all_generative_model_classes: lowerCamelCase : List[Any] = model_class(__magic_name__ ) lowerCamelCase : Dict = model.generate(__magic_name__ ).sequences self.assertEqual(generation_outputs.shape[-1] , __magic_name__ ) lowerCamelCase : str = jit(model.generate ) lowerCamelCase : Any = jit_generate(__magic_name__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Any = self._get_input_ids_and_config() lowerCamelCase : int = True lowerCamelCase : int = max_length for model_class in self.all_generative_model_classes: lowerCamelCase : int = model_class(__magic_name__ ) lowerCamelCase : List[str] = model.generate(__magic_name__ ).sequences self.assertEqual(generation_outputs.shape[-1] , __magic_name__ ) lowerCamelCase : List[str] = jit(model.generate ) lowerCamelCase : Tuple = jit_generate(__magic_name__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : List[str] = self._get_input_ids_and_config() lowerCamelCase : Any = False lowerCamelCase : int = max_length lowerCamelCase : Union[str, Any] = 2 for model_class in self.all_generative_model_classes: lowerCamelCase : Dict = model_class(__magic_name__ ) lowerCamelCase : Dict = model.generate(__magic_name__ ).sequences self.assertEqual(generation_outputs.shape[-1] , __magic_name__ ) lowerCamelCase : Any = jit(model.generate ) lowerCamelCase : Union[str, Any] = jit_generate(__magic_name__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : List[Any] = self._get_input_ids_and_config() lowerCamelCase : Optional[int] = False lowerCamelCase : Tuple = max_length lowerCamelCase : List[Any] = 2 lowerCamelCase : Optional[int] = 2 for model_class in self.all_generative_model_classes: lowerCamelCase : str = model_class(__magic_name__ ) lowerCamelCase : Any = model.generate(__magic_name__ ).sequences self.assertEqual(generation_outputs.shape[0] , input_ids.shape[0] * config.num_return_sequences ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Tuple = self._get_input_ids_and_config() lowerCamelCase : List[str] = True lowerCamelCase : Union[str, Any] = max_length lowerCamelCase : Dict = 0.8 lowerCamelCase : List[Any] = 1_0 lowerCamelCase : Any = 0.3 lowerCamelCase : Dict = 1 lowerCamelCase : List[str] = 8 lowerCamelCase : Tuple = 9 for model_class in self.all_generative_model_classes: lowerCamelCase : Optional[int] = model_class(__magic_name__ ) lowerCamelCase : Dict = model.generate(__magic_name__ ).sequences self.assertEqual(generation_outputs.shape[-1] , __magic_name__ ) lowerCamelCase : Dict = jit(model.generate ) lowerCamelCase : List[Any] = jit_generate(__magic_name__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[int] = self._get_input_ids_and_config() lowerCamelCase : Union[str, Any] = max_length lowerCamelCase : str = 1 lowerCamelCase : Optional[int] = 8 lowerCamelCase : List[str] = 9 for model_class in self.all_generative_model_classes: lowerCamelCase : Union[str, Any] = model_class(__magic_name__ ) lowerCamelCase : List[str] = model.generate(__magic_name__ ).sequences self.assertEqual(generation_outputs.shape[-1] , __magic_name__ ) lowerCamelCase : Dict = jit(model.generate ) lowerCamelCase : Dict = jit_generate(__magic_name__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[Any] = self._get_input_ids_and_config() lowerCamelCase : str = max_length lowerCamelCase : str = 2 lowerCamelCase : Optional[Any] = 1 lowerCamelCase : List[str] = 8 lowerCamelCase : List[str] = 9 for model_class in self.all_generative_model_classes: lowerCamelCase : Union[str, Any] = model_class(__magic_name__ ) lowerCamelCase : Optional[int] = model.generate(__magic_name__ ).sequences self.assertEqual(generation_outputs.shape[-1] , __magic_name__ ) lowerCamelCase : Optional[int] = jit(model.generate ) lowerCamelCase : Optional[Any] = jit_generate(__magic_name__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Optional[Any] = self._get_input_ids_and_config() # pad attention mask on the left lowerCamelCase : Union[str, Any] = attention_mask.at[(0, 0)].set(0 ) lowerCamelCase : Dict = False lowerCamelCase : int = max_length for model_class in self.all_generative_model_classes: lowerCamelCase : List[str] = model_class(__magic_name__ ) lowerCamelCase : Any = model.generate(__magic_name__ , attention_mask=__magic_name__ ).sequences self.assertEqual(generation_outputs.shape[-1] , __magic_name__ ) lowerCamelCase : List[str] = jit(model.generate ) lowerCamelCase : List[str] = jit_generate(__magic_name__ , attention_mask=__magic_name__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Any = self._get_input_ids_and_config() # pad attention mask on the left lowerCamelCase : Optional[int] = attention_mask.at[(0, 0)].set(0 ) lowerCamelCase : Optional[Any] = True lowerCamelCase : List[str] = max_length for model_class in self.all_generative_model_classes: lowerCamelCase : List[str] = model_class(__magic_name__ ) lowerCamelCase : Any = model.generate(__magic_name__ , attention_mask=__magic_name__ ).sequences self.assertEqual(generation_outputs.shape[-1] , __magic_name__ ) lowerCamelCase : Optional[int] = jit(model.generate ) lowerCamelCase : int = jit_generate(__magic_name__ , attention_mask=__magic_name__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def UpperCamelCase__ ( self ): lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Union[str, Any] = self._get_input_ids_and_config() # pad attention mask on the left lowerCamelCase : Tuple = attention_mask.at[(0, 0)].set(0 ) lowerCamelCase : Any = 2 lowerCamelCase : Any = max_length for model_class in self.all_generative_model_classes: lowerCamelCase : Union[str, Any] = model_class(__magic_name__ ) lowerCamelCase : Any = model.generate(__magic_name__ , attention_mask=__magic_name__ ).sequences self.assertEqual(generation_outputs.shape[-1] , __magic_name__ ) lowerCamelCase : str = jit(model.generate ) lowerCamelCase : int = jit_generate(__magic_name__ , attention_mask=__magic_name__ ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) @require_flax class A__ ( unittest.TestCase): def UpperCamelCase__ ( self ): lowerCamelCase : List[str] = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-bert""" ) lowerCamelCase : List[Any] = FlaxAutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-bert-flax-only""" ) lowerCamelCase : Tuple = """Hello world""" lowerCamelCase : Dict = tokenizer(__magic_name__ , return_tensors="""np""" ).input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(__magic_name__ , """do_samples""" ): model.generate(__magic_name__ , do_samples=__magic_name__ ) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(__magic_name__ , """foo""" ): lowerCamelCase : Any = {"""foo""": """bar"""} model.generate(__magic_name__ , **__magic_name__ )
287
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL _lowerCamelCase =logging.get_logger(__name__) class A__ ( __SCREAMING_SNAKE_CASE): _UpperCAmelCase : Union[str, Any] = ["""pixel_values"""] def __init__( self , __magic_name__ = True , __magic_name__ = None , __magic_name__ = None , __magic_name__ = PILImageResampling.BILINEAR , __magic_name__ = True , __magic_name__ = 1 / 2_5_5 , __magic_name__ = True , __magic_name__ = None , __magic_name__ = None , **__magic_name__ , ): super().__init__(**__magic_name__ ) lowerCamelCase : Dict = size if size is not None else {"""shortest_edge""": 3_8_4} lowerCamelCase : Tuple = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : Dict = do_resize lowerCamelCase : List[Any] = size # Default value set here for backwards compatibility where the value in config is None lowerCamelCase : Any = crop_pct if crop_pct is not None else 2_2_4 / 2_5_6 lowerCamelCase : Union[str, Any] = resample lowerCamelCase : str = do_rescale lowerCamelCase : Union[str, Any] = rescale_factor lowerCamelCase : Tuple = do_normalize lowerCamelCase : Optional[Any] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCamelCase : int = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = PILImageResampling.BICUBIC , __magic_name__ = None , **__magic_name__ , ): lowerCamelCase : Union[str, Any] = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) if "shortest_edge" not in size: raise ValueError(F'''Size dictionary must contain \'shortest_edge\' key. Got {size.keys()}''' ) lowerCamelCase : str = size["""shortest_edge"""] if shortest_edge < 3_8_4: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct lowerCamelCase : List[str] = int(shortest_edge / crop_pct ) lowerCamelCase : Optional[Any] = get_resize_output_image_size(__magic_name__ , size=__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : Optional[int] = resize(image=__magic_name__ , size=__magic_name__ , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=__magic_name__ , size=(shortest_edge, shortest_edge) , data_format=__magic_name__ , **__magic_name__ ) else: # warping (no cropping) when evaluated at 384 or larger return resize( __magic_name__ , size=(shortest_edge, shortest_edge) , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ , ): return rescale(__magic_name__ , scale=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ = None , **__magic_name__ , ): return normalize(__magic_name__ , mean=__magic_name__ , std=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = None , __magic_name__ = ChannelDimension.FIRST , **__magic_name__ , ): lowerCamelCase : str = do_resize if do_resize is not None else self.do_resize lowerCamelCase : Optional[Any] = crop_pct if crop_pct is not None else self.crop_pct lowerCamelCase : Optional[int] = resample if resample is not None else self.resample lowerCamelCase : Optional[int] = do_rescale if do_rescale is not None else self.do_rescale lowerCamelCase : List[str] = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCamelCase : Tuple = do_normalize if do_normalize is not None else self.do_normalize lowerCamelCase : Optional[Any] = image_mean if image_mean is not None else self.image_mean lowerCamelCase : Tuple = image_std if image_std is not None else self.image_std lowerCamelCase : Dict = size if size is not None else self.size lowerCamelCase : str = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) lowerCamelCase : List[str] = make_list_of_images(__magic_name__ ) if not valid_images(__magic_name__ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_resize and size["shortest_edge"] < 3_8_4 and crop_pct is None: raise ValueError("""crop_pct must be specified if size < 384.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. lowerCamelCase : Optional[Any] = [to_numpy_array(__magic_name__ ) for image in images] if do_resize: lowerCamelCase : List[Any] = [self.resize(image=__magic_name__ , size=__magic_name__ , crop_pct=__magic_name__ , resample=__magic_name__ ) for image in images] if do_rescale: lowerCamelCase : Union[str, Any] = [self.rescale(image=__magic_name__ , scale=__magic_name__ ) for image in images] if do_normalize: lowerCamelCase : List[Any] = [self.normalize(image=__magic_name__ , mean=__magic_name__ , std=__magic_name__ ) for image in images] lowerCamelCase : Optional[int] = [to_channel_dimension_format(__magic_name__ , __magic_name__ ) for image in images] lowerCamelCase : List[str] = {"""pixel_values""": images} return BatchFeature(data=__magic_name__ , tensor_type=__magic_name__ )
287
1
from __future__ import annotations def __UpperCamelCase (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> list: lowercase__ = [] lowercase__ , lowercase__ = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0 ) ) lowercase__ = result + left + right return input_list def __UpperCamelCase (_SCREAMING_SNAKE_CASE ) -> list: if len(_SCREAMING_SNAKE_CASE ) <= 1: return input_list lowercase__ = list(_SCREAMING_SNAKE_CASE ) # iteration for two-way merging lowercase__ = 2 while p <= len(_SCREAMING_SNAKE_CASE ): # getting low, high and middle value for merge-sort of single list for i in range(0 , len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ): lowercase__ = i lowercase__ = i + p - 1 lowercase__ = (low + high + 1) // 2 lowercase__ = merge(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # final merge of last two parts if p * 2 >= len(_SCREAMING_SNAKE_CASE ): lowercase__ = i lowercase__ = merge(_SCREAMING_SNAKE_CASE , 0 , _SCREAMING_SNAKE_CASE , len(_SCREAMING_SNAKE_CASE ) - 1 ) break p *= 2 return input_list if __name__ == "__main__": lowercase_ = input("""Enter numbers separated by a comma:\n""").strip() if user_input == "": lowercase_ = [] else: lowercase_ = [int(item.strip()) for item in user_input.split(""",""")] print(iter_merge_sort(unsorted))
352
def __UpperCamelCase (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> int: if exponent == 1: return base if exponent % 2 == 0: lowercase__ = _modexpt(_SCREAMING_SNAKE_CASE , exponent // 2 , _SCREAMING_SNAKE_CASE ) % modulo_value return (x * x) % modulo_value else: return (base * _modexpt(_SCREAMING_SNAKE_CASE , exponent - 1 , _SCREAMING_SNAKE_CASE )) % modulo_value def __UpperCamelCase (_SCREAMING_SNAKE_CASE = 1777 , _SCREAMING_SNAKE_CASE = 1855 , _SCREAMING_SNAKE_CASE = 8 ) -> int: lowercase__ = base for _ in range(1 , _SCREAMING_SNAKE_CASE ): lowercase__ = _modexpt(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , 10**digits ) return result if __name__ == "__main__": print(f'''{solution() = }''')
269
0
'''simple docstring''' import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class a__ ( UpperCAmelCase__ ): lowerCamelCase : Union[str, Any] =["image_processor", "tokenizer"] lowerCamelCase : int ="BlipImageProcessor" lowerCamelCase : int ="AutoTokenizer" def __init__( self : Dict , a : Any , a : Union[str, Any] , a : str ): """simple docstring""" super().__init__(a , a ) # add QFormer tokenizer __lowerCamelCase = qformer_tokenizer def __call__( self : Union[str, Any] , a : ImageInput = None , a : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , a : bool = True , a : Union[bool, str, PaddingStrategy] = False , a : Union[bool, str, TruncationStrategy] = None , a : Optional[int] = None , a : int = 0 , a : Optional[int] = None , a : Optional[bool] = None , a : bool = False , a : bool = False , a : bool = False , a : bool = False , a : bool = False , a : bool = True , a : Optional[Union[str, TensorType]] = None , **a : List[Any] , ): """simple docstring""" if images is None and text is None: raise ValueError('''You have to specify at least images or text.''' ) __lowerCamelCase = BatchFeature() if text is not None: __lowerCamelCase = self.tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) encoding.update(a ) __lowerCamelCase = self.qformer_tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) __lowerCamelCase = qformer_text_encoding.pop('''input_ids''' ) __lowerCamelCase = qformer_text_encoding.pop('''attention_mask''' ) if images is not None: __lowerCamelCase = self.image_processor(a , return_tensors=a ) encoding.update(a ) return encoding def SCREAMING_SNAKE_CASE__ ( self : Dict , *a : Dict , **a : List[str] ): """simple docstring""" return self.tokenizer.batch_decode(*a , **a ) def SCREAMING_SNAKE_CASE__ ( self : Any , *a : List[str] , **a : Tuple ): """simple docstring""" return self.tokenizer.decode(*a , **a ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ): """simple docstring""" __lowerCamelCase = self.tokenizer.model_input_names __lowerCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def SCREAMING_SNAKE_CASE__ ( self : Dict , a : str , **a : List[Any] ): """simple docstring""" if os.path.isfile(a ): raise ValueError(f"""Provided path ({save_directory}) should be a directory, not a file""" ) os.makedirs(a , exist_ok=a ) __lowerCamelCase = os.path.join(a , '''qformer_tokenizer''' ) self.qformer_tokenizer.save_pretrained(a ) return super().save_pretrained(a , **a ) @classmethod def SCREAMING_SNAKE_CASE__ ( cls : str , a : List[str] , **a : Dict ): """simple docstring""" __lowerCamelCase = AutoTokenizer.from_pretrained(a , subfolder='''qformer_tokenizer''' ) __lowerCamelCase = cls._get_arguments_from_pretrained(a , **a ) args.append(a ) return cls(*a )
67
'''simple docstring''' from typing import List, Union import numpy as np from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_DEPTH_ESTIMATION_MAPPING __UpperCAmelCase =logging.get_logger(__name__) @add_end_docstrings(UpperCAmelCase__ ) class a__ ( UpperCAmelCase__ ): def __init__( self : List[str] , *a : Union[str, Any] , **a : Optional[Any] ): """simple docstring""" super().__init__(*a , **a ) requires_backends(self , '''vision''' ) self.check_model_type(a ) def __call__( self : Any , a : Union[str, List[str], "Image.Image", List["Image.Image"]] , **a : Optional[int] ): """simple docstring""" return super().__call__(a , **a ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , **a : Any ): """simple docstring""" return {}, {}, {} def SCREAMING_SNAKE_CASE__ ( self : List[Any] , a : List[str] ): """simple docstring""" __lowerCamelCase = load_image(a ) __lowerCamelCase = image.size __lowerCamelCase = self.image_processor(images=a , return_tensors=self.framework ) return model_inputs def SCREAMING_SNAKE_CASE__ ( self : int , a : Optional[Any] ): """simple docstring""" __lowerCamelCase = self.model(**a ) return model_outputs def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , a : Any ): """simple docstring""" __lowerCamelCase = model_outputs.predicted_depth __lowerCamelCase = torch.nn.functional.interpolate( predicted_depth.unsqueeze(1 ) , size=self.image_size[::-1] , mode='''bicubic''' , align_corners=a ) __lowerCamelCase = prediction.squeeze().cpu().numpy() __lowerCamelCase = (output * 2_55 / np.max(a )).astype('''uint8''' ) __lowerCamelCase = Image.fromarray(a ) __lowerCamelCase = {} __lowerCamelCase = predicted_depth __lowerCamelCase = depth return output_dict
67
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_barthez import BarthezTokenizer else: __UpperCAmelCase = None __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = {"vocab_file": "sentencepiece.bpe.model", "tokenizer_file": "tokenizer.json"} __UpperCAmelCase = { "vocab_file": { "moussaKam/mbarthez": "https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez": "https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez-orangesum-title": ( "https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model" ), }, "tokenizer_file": { "moussaKam/mbarthez": "https://huggingface.co/moussaKam/mbarthez/resolve/main/tokenizer.json", "moussaKam/barthez": "https://huggingface.co/moussaKam/barthez/resolve/main/tokenizer.json", "moussaKam/barthez-orangesum-title": ( "https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/tokenizer.json" ), }, } __UpperCAmelCase = { "moussaKam/mbarthez": 10_24, "moussaKam/barthez": 10_24, "moussaKam/barthez-orangesum-title": 10_24, } __UpperCAmelCase = "▁" class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCAmelCase_ =VOCAB_FILES_NAMES UpperCAmelCase_ =PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase_ =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase_ =["input_ids", "attention_mask"] UpperCAmelCase_ =BarthezTokenizer def __init__( self , _A=None , _A=None , _A="<s>" , _A="</s>" , _A="</s>" , _A="<s>" , _A="<unk>" , _A="<pad>" , _A="<mask>" , **_A , ) -> List[str]: # Mask token behave like a normal word, i.e. include the space before it SCREAMING_SNAKE_CASE_ = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else mask_token super().__init__( _A , tokenizer_file=_A , bos_token=_A , eos_token=_A , unk_token=_A , sep_token=_A , cls_token=_A , pad_token=_A , mask_token=_A , **_A , ) SCREAMING_SNAKE_CASE_ = vocab_file SCREAMING_SNAKE_CASE_ = False if not self.vocab_file else True def _UpperCamelCase ( self , _A , _A = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] SCREAMING_SNAKE_CASE_ = [self.cls_token_id] SCREAMING_SNAKE_CASE_ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _UpperCamelCase ( self , _A , _A = None ) -> List[int]: SCREAMING_SNAKE_CASE_ = [self.sep_token_id] SCREAMING_SNAKE_CASE_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _UpperCamelCase ( self , _A , _A = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(_A ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return SCREAMING_SNAKE_CASE_ = os.path.join( _A , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_A ): copyfile(self.vocab_file , _A ) return (out_vocab_file,)
257
import argparse from transformers import BigBirdConfig, BigBirdForPreTraining, BigBirdForQuestionAnswering, load_tf_weights_in_big_bird from transformers.utils import logging logging.set_verbosity_info() def A__ ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # Initialise PyTorch model SCREAMING_SNAKE_CASE_ = BigBirdConfig.from_json_file(__lowerCamelCase ) print(F'''Building PyTorch model from configuration: {config}''' ) if is_trivia_qa: SCREAMING_SNAKE_CASE_ = BigBirdForQuestionAnswering(__lowerCamelCase ) else: SCREAMING_SNAKE_CASE_ = BigBirdForPreTraining(__lowerCamelCase ) # Load weights from tf checkpoint load_tf_weights_in_big_bird(__lowerCamelCase, __lowerCamelCase, is_trivia_qa=__lowerCamelCase ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--big_bird_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained BERT model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--is_trivia_qa", action="store_true", help="Whether to convert a model with a trivia_qa head." ) __UpperCAmelCase = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.big_bird_config_file, args.pytorch_dump_path, args.is_trivia_qa )
257
1
'''simple docstring''' from math import log from scipy.constants import Boltzmann, physical_constants A_ : Union[str, Any] = 3_0_0 # TEMPERATURE (unit = K) def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , )-> float: '''simple docstring''' if donor_conc <= 0: raise ValueError("""Donor concentration should be positive""" ) elif acceptor_conc <= 0: raise ValueError("""Acceptor concentration should be positive""" ) elif intrinsic_conc <= 0: raise ValueError("""Intrinsic concentration should be positive""" ) elif donor_conc <= intrinsic_conc: raise ValueError( """Donor concentration should be greater than intrinsic concentration""" ) elif acceptor_conc <= intrinsic_conc: raise ValueError( """Acceptor concentration should be greater than intrinsic concentration""" ) else: return ( Boltzmann * T * log((donor_conc * acceptor_conc) / intrinsic_conc**2 ) / physical_constants["electron volt"][0] ) if __name__ == "__main__": import doctest doctest.testmod()
215
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class lowercase ( unittest.TestCase ): """simple docstring""" @slow def _snake_case ( self ) -> Any: _UpperCAmelCase : Dict = TFCamembertModel.from_pretrained("""jplu/tf-camembert-base""" ) _UpperCAmelCase : Optional[int] = tf.convert_to_tensor( [[5, 121, 11, 660, 16, 730, 25_543, 110, 83, 6]] ,dtype=tf.intaa ,) # J'aime le camembert !" _UpperCAmelCase : Dict = model(a_ )["""last_hidden_state"""] _UpperCAmelCase : Dict = tf.TensorShape((1, 10, 768) ) self.assertEqual(output.shape ,a_ ) # compare the actual values for a slice. _UpperCAmelCase : Tuple = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] ,dtype=tf.floataa ,) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() ,expected_slice.numpy() ,atol=1E-4 ) )
215
1
'''simple docstring''' import cmath import math def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): UpperCAmelCase__ : Optional[int] = math.radians(UpperCamelCase__ ) UpperCAmelCase__ : List[Any] = math.radians(UpperCamelCase__ ) # Convert voltage and current to rectangular form UpperCAmelCase__ : Tuple = cmath.rect(UpperCamelCase__ , UpperCamelCase__ ) UpperCAmelCase__ : List[str] = cmath.rect(UpperCamelCase__ , UpperCamelCase__ ) # Calculate apparent power return voltage_rect * current_rect if __name__ == "__main__": import doctest doctest.testmod()
283
'''simple docstring''' import os from typing import Optional import fsspec from fsspec.archive import AbstractArchiveFileSystem from fsspec.utils import DEFAULT_BLOCK_SIZE class _snake_case ( a__ ): lowerCAmelCase :Optional[int] = '''''' lowerCAmelCase :str = ( None # protocol passed in prefix to the url. ex: "gzip", for gzip://file.txt::http://foo.bar/file.txt.gz ) lowerCAmelCase :str = None # compression type in fsspec. ex: "gzip" lowerCAmelCase :str = None # extension of the filename to strip. ex: "".gz" to get file.txt from file.txt.gz def __init__( self , _lowerCamelCase = "" , _lowerCamelCase = None , _lowerCamelCase = None , **_lowerCamelCase): super().__init__(self , **_lowerCamelCase) # always open as "rb" since fsspec can then use the TextIOWrapper to make it work for "r" mode UpperCAmelCase__ : Optional[Any] = fsspec.open( _lowerCamelCase , mode="""rb""" , protocol=_lowerCamelCase , compression=self.compression , client_kwargs={ """requote_redirect_url""": False, # see https://github.com/huggingface/datasets/pull/5459 """trust_env""": True, # Enable reading proxy env variables. **(target_options or {}).pop("""client_kwargs""" , {}), # To avoid issues if it was already passed. } , **(target_options or {}) , ) UpperCAmelCase__ : List[Any] = os.path.basename(self.file.path.split("""::""")[0]) UpperCAmelCase__ : Dict = ( self.compressed_name[: self.compressed_name.rindex(""".""")] if """.""" in self.compressed_name else self.compressed_name ) UpperCAmelCase__ : Tuple = None @classmethod def snake_case__ ( cls , _lowerCamelCase): # compressed file paths are always relative to the archive root return super()._strip_protocol(_lowerCamelCase).lstrip("""/""") def snake_case__ ( self): if self.dir_cache is None: UpperCAmelCase__ : Optional[Any] = {**self.file.fs.info(self.file.path), """name""": self.uncompressed_name} UpperCAmelCase__ : Union[str, Any] = {f["""name"""]: f} def snake_case__ ( self , _lowerCamelCase): return self.file.open().read() def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase = "rb" , _lowerCamelCase=None , _lowerCamelCase=True , _lowerCamelCase=None , **_lowerCamelCase , ): UpperCAmelCase__ : List[str] = self._strip_protocol(_lowerCamelCase) if mode != "rb": raise ValueError(f'''Tried to read with mode {mode} on file {self.file.path} opened with mode \'rb\'''') return self.file.open() class _snake_case ( a__ ): lowerCAmelCase :Dict = '''bz2''' lowerCAmelCase :List[str] = '''bz2''' lowerCAmelCase :Dict = '''.bz2''' class _snake_case ( a__ ): lowerCAmelCase :int = '''gzip''' lowerCAmelCase :Tuple = '''gzip''' lowerCAmelCase :str = '''.gz''' class _snake_case ( a__ ): lowerCAmelCase :List[str] = '''lz4''' lowerCAmelCase :Any = '''lz4''' lowerCAmelCase :int = '''.lz4''' class _snake_case ( a__ ): lowerCAmelCase :Union[str, Any] = '''xz''' lowerCAmelCase :int = '''xz''' lowerCAmelCase :List[Any] = '''.xz''' class _snake_case ( a__ ): lowerCAmelCase :Tuple = '''zstd''' lowerCAmelCase :List[str] = '''zstd''' lowerCAmelCase :Union[str, Any] = '''.zst''' def __init__( self , _lowerCamelCase , _lowerCamelCase = "rb" , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = DEFAULT_BLOCK_SIZE , **_lowerCamelCase , ): super().__init__( fo=_lowerCamelCase , mode=_lowerCamelCase , target_protocol=_lowerCamelCase , target_options=_lowerCamelCase , block_size=_lowerCamelCase , **_lowerCamelCase , ) # We need to wrap the zstd decompressor to avoid this error in fsspec==2021.7.0 and zstandard==0.15.2: # # File "/Users/user/.virtualenvs/hf-datasets/lib/python3.7/site-packages/fsspec/core.py", line 145, in open # out.close = close # AttributeError: 'zstd.ZstdDecompressionReader' object attribute 'close' is read-only # # see https://github.com/intake/filesystem_spec/issues/725 UpperCAmelCase__ : Dict = self.file.__enter__ class _snake_case : def __init__( self , _lowerCamelCase): UpperCAmelCase__ : Optional[int] = file_ def __enter__( self): self._file.__enter__() return self def __exit__( self , *_lowerCamelCase , **_lowerCamelCase): self._file.__exit__(*_lowerCamelCase , **_lowerCamelCase) def __iter__( self): return iter(self._file) def snake_case__ ( self): return next(self._file) def __getattr__( self , _lowerCamelCase): return getattr(self._file , _lowerCamelCase) def fixed_enter(*_lowerCamelCase , **_lowerCamelCase): return WrappedFile(_enter(*_lowerCamelCase , **_lowerCamelCase)) UpperCAmelCase__ : List[Any] = fixed_enter
283
1
"""simple docstring""" from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class SCREAMING_SNAKE_CASE_ ( __a ): """simple docstring""" __lowercase : int = ['''image_processor''', '''tokenizer'''] __lowercase : List[Any] = '''BlipImageProcessor''' __lowercase : Dict = ('''BertTokenizer''', '''BertTokenizerFast''') def __init__( self , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = False super().__init__(lowerCAmelCase__ , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = self.image_processor def __call__( self , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = True , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = 0 , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = False , lowerCAmelCase__ = True , lowerCAmelCase__ = None , **lowerCAmelCase__ , ): if images is None and text is None: raise ValueError("""You have to specify either images or text.""") # Get only text if images is None: __SCREAMING_SNAKE_CASE = self.tokenizer __SCREAMING_SNAKE_CASE = self.tokenizer( text=lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , padding=lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=lowerCAmelCase__ , stride=lowerCAmelCase__ , pad_to_multiple_of=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , return_overflowing_tokens=lowerCAmelCase__ , return_special_tokens_mask=lowerCAmelCase__ , return_offsets_mapping=lowerCAmelCase__ , return_token_type_ids=lowerCAmelCase__ , return_length=lowerCAmelCase__ , verbose=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ , ) return text_encoding # add pixel_values __SCREAMING_SNAKE_CASE = self.image_processor(lowerCAmelCase__ , return_tensors=lowerCAmelCase__) if text is not None: __SCREAMING_SNAKE_CASE = self.tokenizer( text=lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , padding=lowerCAmelCase__ , truncation=lowerCAmelCase__ , max_length=lowerCAmelCase__ , stride=lowerCAmelCase__ , pad_to_multiple_of=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , return_overflowing_tokens=lowerCAmelCase__ , return_special_tokens_mask=lowerCAmelCase__ , return_offsets_mapping=lowerCAmelCase__ , return_token_type_ids=lowerCAmelCase__ , return_length=lowerCAmelCase__ , verbose=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ , ) else: __SCREAMING_SNAKE_CASE = None if text_encoding is not None: encoding_image_processor.update(lowerCAmelCase__) return encoding_image_processor def snake_case_ ( self , *lowerCAmelCase__ , **lowerCAmelCase__): return self.tokenizer.batch_decode(*lowerCAmelCase__ , **lowerCAmelCase__) def snake_case_ ( self , *lowerCAmelCase__ , **lowerCAmelCase__): return self.tokenizer.decode(*lowerCAmelCase__ , **lowerCAmelCase__) @property def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.tokenizer.model_input_names __SCREAMING_SNAKE_CASE = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names))
100
from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS A__ : Tuple = logging.get_logger(__name__) A__ : int = { 'linear': get_linear_schedule_with_warmup, 'cosine': get_cosine_schedule_with_warmup, 'cosine_w_restarts': get_cosine_with_hard_restarts_schedule_with_warmup, 'polynomial': get_polynomial_decay_schedule_with_warmup, 'constant': get_constant_schedule, 'constant_w_warmup': get_constant_schedule_with_warmup, } class _UpperCAmelCase ( A__ ): """simple docstring""" def __init__( self : Optional[int], lowerCamelCase : int=None, lowerCamelCase : int=None, *lowerCamelCase : List[Any], **lowerCamelCase : Any ): '''simple docstring''' super().__init__(*lowerCamelCase, **lowerCamelCase ) if config is None: assert isinstance(self.model, lowerCamelCase ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F""" {self.model.__class__}""" ) lowercase__ = self.model.config else: lowercase__ = config lowercase__ = data_args lowercase__ = self.config.tgt_vocab_size if isinstance(self.config, lowerCamelCase ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ''' padding..''' ) if self.args.label_smoothing == 0: lowercase__ = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss lowercase__ = label_smoothed_nll_loss def lowercase__ ( self : List[Any], lowerCamelCase : int ): '''simple docstring''' if self.optimizer is None: lowercase__ = ['''bias''', '''LayerNorm.weight'''] lowercase__ = [ { '''params''': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], '''weight_decay''': self.args.weight_decay, }, { '''params''': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0, }, ] lowercase__ = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: lowercase__ = Adafactor lowercase__ = {'''scale_parameter''': False, '''relative_step''': False} else: lowercase__ = AdamW lowercase__ = { '''betas''': (self.args.adam_betaa, self.args.adam_betaa), '''eps''': self.args.adam_epsilon, } lowercase__ = self.args.learning_rate if self.sharded_ddp: lowercase__ = OSS( params=lowerCamelCase, optim=lowerCamelCase, **lowerCamelCase, ) else: lowercase__ = optimizer_cls(lowerCamelCase, **lowerCamelCase ) if self.lr_scheduler is None: lowercase__ = self._get_lr_scheduler(lowerCamelCase ) else: # ignoring --lr_scheduler logger.warning('''scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.''' ) def lowercase__ ( self : List[str], lowerCamelCase : Optional[int] ): '''simple docstring''' lowercase__ = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": lowercase__ = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": lowercase__ = schedule_func(self.optimizer, num_warmup_steps=self.args.warmup_steps ) else: lowercase__ = schedule_func( self.optimizer, num_warmup_steps=self.args.warmup_steps, num_training_steps=lowerCamelCase ) return scheduler def lowercase__ ( self : List[Any] ): '''simple docstring''' if isinstance(self.train_dataset, torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size, distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED), ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def lowercase__ ( self : Any, lowerCamelCase : List[Any], lowerCamelCase : str, lowerCamelCase : Union[str, Any] ): '''simple docstring''' if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token lowercase__ = model(**lowerCamelCase, use_cache=lowerCamelCase )[0] lowercase__ = self.loss_fn(logits.view(-1, logits.shape[-1] ), labels.view(-1 ) ) else: # compute usual loss via models lowercase__ , lowercase__ = model(**lowerCamelCase, labels=lowerCamelCase, use_cache=lowerCamelCase )[:2] else: # compute label smoothed loss lowercase__ = model(**lowerCamelCase, use_cache=lowerCamelCase )[0] lowercase__ = torch.nn.functional.log_softmax(lowerCamelCase, dim=-1 ) lowercase__ , lowercase__ = self.loss_fn(lowerCamelCase, lowerCamelCase, self.args.label_smoothing, ignore_index=self.config.pad_token_id ) return loss, logits def lowercase__ ( self : List[str], lowerCamelCase : Optional[Any], lowerCamelCase : Tuple ): '''simple docstring''' lowercase__ = inputs.pop('''labels''' ) lowercase__ , lowercase__ = self._compute_loss(lowerCamelCase, lowerCamelCase, lowerCamelCase ) return loss def lowercase__ ( self : str, lowerCamelCase : nn.Module, lowerCamelCase : Dict[str, Union[torch.Tensor, Any]], lowerCamelCase : bool, lowerCamelCase : Optional[List[str]] = None, ): '''simple docstring''' lowercase__ = self._prepare_inputs(lowerCamelCase ) lowercase__ = { '''max_length''': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, '''num_beams''': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: lowercase__ = self.model.generate( inputs['''input_ids'''], attention_mask=inputs['''attention_mask'''], **lowerCamelCase, ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: lowercase__ = self._pad_tensors_to_max_len(lowerCamelCase, gen_kwargs['''max_length'''] ) lowercase__ = inputs.pop('''labels''' ) with torch.no_grad(): # compute loss on predict data lowercase__ , lowercase__ = self._compute_loss(lowerCamelCase, lowerCamelCase, lowerCamelCase ) lowercase__ = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) lowercase__ = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: lowercase__ = self._pad_tensors_to_max_len(lowerCamelCase, gen_kwargs['''max_length'''] ) return (loss, logits, labels) def lowercase__ ( self : List[Any], lowerCamelCase : str, lowerCamelCase : Any ): '''simple docstring''' # If PAD token is not defined at least EOS token has to be defined lowercase__ = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( '''Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be''' F""" padded to `max_length`={max_length}""" ) lowercase__ = pad_token_id * torch.ones( (tensor.shape[0], max_length), dtype=tensor.dtype, device=tensor.device ) lowercase__ = tensor return padded_tensor
207
0
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class UpperCAmelCase ( SCREAMING_SNAKE_CASE__): _lowerCamelCase : torch.FloatTensor class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__): @register_to_config def __init__( self : str, a_ : int = 6_5536, a_ : Optional[int] = None, a_ : int = 2, a_ : int = 2, a_ : int = 0, a_ : str = "fourier", a_ : bool = True, a_ : bool = False, a_ : float = 0.0, a_ : Tuple[str] = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D"), a_ : Tuple[str] = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip"), a_ : Tuple[str] = "UNetMidBlock1D", a_ : str = None, a_ : Tuple[int] = (32, 32, 64), a_ : str = None, a_ : int = 8, a_ : int = 1, a_ : bool = False, ): """simple docstring""" super().__init__() UpperCamelCase__ = sample_size # time if time_embedding_type == "fourier": UpperCamelCase__ = GaussianFourierProjection( embedding_size=8, set_W_to_weight=a_, log=a_, flip_sin_to_cos=a_ ) UpperCamelCase__ = 2 * block_out_channels[0] elif time_embedding_type == "positional": UpperCamelCase__ = Timesteps( block_out_channels[0], flip_sin_to_cos=a_, downscale_freq_shift=a_ ) UpperCamelCase__ = block_out_channels[0] if use_timestep_embedding: UpperCamelCase__ = block_out_channels[0] * 4 UpperCamelCase__ = TimestepEmbedding( in_channels=a_, time_embed_dim=a_, act_fn=a_, out_dim=block_out_channels[0], ) UpperCamelCase__ = nn.ModuleList([] ) UpperCamelCase__ = None UpperCamelCase__ = nn.ModuleList([] ) UpperCamelCase__ = None # down UpperCamelCase__ = in_channels for i, down_block_type in enumerate(a_ ): UpperCamelCase__ = output_channel UpperCamelCase__ = block_out_channels[i] if i == 0: input_channel += extra_in_channels UpperCamelCase__ = i == len(a_ ) - 1 UpperCamelCase__ = get_down_block( a_, num_layers=a_, in_channels=a_, out_channels=a_, temb_channels=block_out_channels[0], add_downsample=not is_final_block or downsample_each_block, ) self.down_blocks.append(a_ ) # mid UpperCamelCase__ = get_mid_block( a_, in_channels=block_out_channels[-1], mid_channels=block_out_channels[-1], out_channels=block_out_channels[-1], embed_dim=block_out_channels[0], num_layers=a_, add_downsample=a_, ) # up UpperCamelCase__ = list(reversed(a_ ) ) UpperCamelCase__ = reversed_block_out_channels[0] if out_block_type is None: UpperCamelCase__ = out_channels else: UpperCamelCase__ = block_out_channels[0] for i, up_block_type in enumerate(a_ ): UpperCamelCase__ = output_channel UpperCamelCase__ = ( reversed_block_out_channels[i + 1] if i < len(a_ ) - 1 else final_upsample_channels ) UpperCamelCase__ = i == len(a_ ) - 1 UpperCamelCase__ = get_up_block( a_, num_layers=a_, in_channels=a_, out_channels=a_, temb_channels=block_out_channels[0], add_upsample=not is_final_block, ) self.up_blocks.append(a_ ) UpperCamelCase__ = output_channel # out UpperCamelCase__ = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4, 32 ) UpperCamelCase__ = get_out_block( out_block_type=a_, num_groups_out=a_, embed_dim=block_out_channels[0], out_channels=a_, act_fn=a_, fc_dim=block_out_channels[-1] // 4, ) def lowercase_ ( self : int, a_ : torch.FloatTensor, a_ : Union[torch.Tensor, float, int], a_ : bool = True, ): """simple docstring""" UpperCamelCase__ = timestep if not torch.is_tensor(a_ ): UpperCamelCase__ = torch.tensor([timesteps], dtype=torch.long, device=sample.device ) elif torch.is_tensor(a_ ) and len(timesteps.shape ) == 0: UpperCamelCase__ = timesteps[None].to(sample.device ) UpperCamelCase__ = self.time_proj(a_ ) if self.config.use_timestep_embedding: UpperCamelCase__ = self.time_mlp(a_ ) else: UpperCamelCase__ = timestep_embed[..., None] UpperCamelCase__ = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) UpperCamelCase__ = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down UpperCamelCase__ = () for downsample_block in self.down_blocks: UpperCamelCase__ , UpperCamelCase__ = downsample_block(hidden_states=a_, temb=a_ ) down_block_res_samples += res_samples # 3. mid if self.mid_block: UpperCamelCase__ = self.mid_block(a_, a_ ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): UpperCamelCase__ = down_block_res_samples[-1:] UpperCamelCase__ = down_block_res_samples[:-1] UpperCamelCase__ = upsample_block(a_, res_hidden_states_tuple=a_, temb=a_ ) # 5. post-process if self.out_block: UpperCamelCase__ = self.out_block(a_, a_ ) if not return_dict: return (sample,) return UNetaDOutput(sample=a_ )
31
'''simple docstring''' from __future__ import annotations def SCREAMING_SNAKE_CASE__( _UpperCamelCase : int | float | str , _UpperCamelCase : int | float | str ) -> list[str]: '''simple docstring''' if nth_term == "": return [""] UpperCamelCase__ = int(_UpperCamelCase ) UpperCamelCase__ = int(_UpperCamelCase ) UpperCamelCase__ = [] for temp in range(int(_UpperCamelCase ) ): series.append(F'1 / {pow(temp + 1 , int(_UpperCamelCase ) )}' if series else "1" ) return series if __name__ == "__main__": import doctest doctest.testmod() __lowercase: Dict = int(input("Enter the last number (nth term) of the P-Series")) __lowercase: Optional[int] = int(input("Enter the power for P-Series")) print("Formula of P-Series => 1+1/2^p+1/3^p ..... 1/n^p") print(p_series(nth_term, power))
31
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool A : Tuple = { 'Acehnese Arabic': 'ace_Arab', 'Acehnese Latin': 'ace_Latn', 'Mesopotamian Arabic': 'acm_Arab', 'Ta\'izzi-Adeni Arabic': 'acq_Arab', 'Tunisian Arabic': 'aeb_Arab', 'Afrikaans': 'afr_Latn', 'South Levantine Arabic': 'ajp_Arab', 'Akan': 'aka_Latn', 'Amharic': 'amh_Ethi', 'North Levantine Arabic': 'apc_Arab', 'Modern Standard Arabic': 'arb_Arab', 'Modern Standard Arabic Romanized': 'arb_Latn', 'Najdi Arabic': 'ars_Arab', 'Moroccan Arabic': 'ary_Arab', 'Egyptian Arabic': 'arz_Arab', 'Assamese': 'asm_Beng', 'Asturian': 'ast_Latn', 'Awadhi': 'awa_Deva', 'Central Aymara': 'ayr_Latn', 'South Azerbaijani': 'azb_Arab', 'North Azerbaijani': 'azj_Latn', 'Bashkir': 'bak_Cyrl', 'Bambara': 'bam_Latn', 'Balinese': 'ban_Latn', 'Belarusian': 'bel_Cyrl', 'Bemba': 'bem_Latn', 'Bengali': 'ben_Beng', 'Bhojpuri': 'bho_Deva', 'Banjar Arabic': 'bjn_Arab', 'Banjar Latin': 'bjn_Latn', 'Standard Tibetan': 'bod_Tibt', 'Bosnian': 'bos_Latn', 'Buginese': 'bug_Latn', 'Bulgarian': 'bul_Cyrl', 'Catalan': 'cat_Latn', 'Cebuano': 'ceb_Latn', 'Czech': 'ces_Latn', 'Chokwe': 'cjk_Latn', 'Central Kurdish': 'ckb_Arab', 'Crimean Tatar': 'crh_Latn', 'Welsh': 'cym_Latn', 'Danish': 'dan_Latn', 'German': 'deu_Latn', 'Southwestern Dinka': 'dik_Latn', 'Dyula': 'dyu_Latn', 'Dzongkha': 'dzo_Tibt', 'Greek': 'ell_Grek', 'English': 'eng_Latn', 'Esperanto': 'epo_Latn', 'Estonian': 'est_Latn', 'Basque': 'eus_Latn', 'Ewe': 'ewe_Latn', 'Faroese': 'fao_Latn', 'Fijian': 'fij_Latn', 'Finnish': 'fin_Latn', 'Fon': 'fon_Latn', 'French': 'fra_Latn', 'Friulian': 'fur_Latn', 'Nigerian Fulfulde': 'fuv_Latn', 'Scottish Gaelic': 'gla_Latn', 'Irish': 'gle_Latn', 'Galician': 'glg_Latn', 'Guarani': 'grn_Latn', 'Gujarati': 'guj_Gujr', 'Haitian Creole': 'hat_Latn', 'Hausa': 'hau_Latn', 'Hebrew': 'heb_Hebr', 'Hindi': 'hin_Deva', 'Chhattisgarhi': 'hne_Deva', 'Croatian': 'hrv_Latn', 'Hungarian': 'hun_Latn', 'Armenian': 'hye_Armn', 'Igbo': 'ibo_Latn', 'Ilocano': 'ilo_Latn', 'Indonesian': 'ind_Latn', 'Icelandic': 'isl_Latn', 'Italian': 'ita_Latn', 'Javanese': 'jav_Latn', 'Japanese': 'jpn_Jpan', 'Kabyle': 'kab_Latn', 'Jingpho': 'kac_Latn', 'Kamba': 'kam_Latn', 'Kannada': 'kan_Knda', 'Kashmiri Arabic': 'kas_Arab', 'Kashmiri Devanagari': 'kas_Deva', 'Georgian': 'kat_Geor', 'Central Kanuri Arabic': 'knc_Arab', 'Central Kanuri Latin': 'knc_Latn', 'Kazakh': 'kaz_Cyrl', 'Kabiyè': 'kbp_Latn', 'Kabuverdianu': 'kea_Latn', 'Khmer': 'khm_Khmr', 'Kikuyu': 'kik_Latn', 'Kinyarwanda': 'kin_Latn', 'Kyrgyz': 'kir_Cyrl', 'Kimbundu': 'kmb_Latn', 'Northern Kurdish': 'kmr_Latn', 'Kikongo': 'kon_Latn', 'Korean': 'kor_Hang', 'Lao': 'lao_Laoo', 'Ligurian': 'lij_Latn', 'Limburgish': 'lim_Latn', 'Lingala': 'lin_Latn', 'Lithuanian': 'lit_Latn', 'Lombard': 'lmo_Latn', 'Latgalian': 'ltg_Latn', 'Luxembourgish': 'ltz_Latn', 'Luba-Kasai': 'lua_Latn', 'Ganda': 'lug_Latn', 'Luo': 'luo_Latn', 'Mizo': 'lus_Latn', 'Standard Latvian': 'lvs_Latn', 'Magahi': 'mag_Deva', 'Maithili': 'mai_Deva', 'Malayalam': 'mal_Mlym', 'Marathi': 'mar_Deva', 'Minangkabau Arabic ': 'min_Arab', 'Minangkabau Latin': 'min_Latn', 'Macedonian': 'mkd_Cyrl', 'Plateau Malagasy': 'plt_Latn', 'Maltese': 'mlt_Latn', 'Meitei Bengali': 'mni_Beng', 'Halh Mongolian': 'khk_Cyrl', 'Mossi': 'mos_Latn', 'Maori': 'mri_Latn', 'Burmese': 'mya_Mymr', 'Dutch': 'nld_Latn', 'Norwegian Nynorsk': 'nno_Latn', 'Norwegian Bokmål': 'nob_Latn', 'Nepali': 'npi_Deva', 'Northern Sotho': 'nso_Latn', 'Nuer': 'nus_Latn', 'Nyanja': 'nya_Latn', 'Occitan': 'oci_Latn', 'West Central Oromo': 'gaz_Latn', 'Odia': 'ory_Orya', 'Pangasinan': 'pag_Latn', 'Eastern Panjabi': 'pan_Guru', 'Papiamento': 'pap_Latn', 'Western Persian': 'pes_Arab', 'Polish': 'pol_Latn', 'Portuguese': 'por_Latn', 'Dari': 'prs_Arab', 'Southern Pashto': 'pbt_Arab', 'Ayacucho Quechua': 'quy_Latn', 'Romanian': 'ron_Latn', 'Rundi': 'run_Latn', 'Russian': 'rus_Cyrl', 'Sango': 'sag_Latn', 'Sanskrit': 'san_Deva', 'Santali': 'sat_Olck', 'Sicilian': 'scn_Latn', 'Shan': 'shn_Mymr', 'Sinhala': 'sin_Sinh', 'Slovak': 'slk_Latn', 'Slovenian': 'slv_Latn', 'Samoan': 'smo_Latn', 'Shona': 'sna_Latn', 'Sindhi': 'snd_Arab', 'Somali': 'som_Latn', 'Southern Sotho': 'sot_Latn', 'Spanish': 'spa_Latn', 'Tosk Albanian': 'als_Latn', 'Sardinian': 'srd_Latn', 'Serbian': 'srp_Cyrl', 'Swati': 'ssw_Latn', 'Sundanese': 'sun_Latn', 'Swedish': 'swe_Latn', 'Swahili': 'swh_Latn', 'Silesian': 'szl_Latn', 'Tamil': 'tam_Taml', 'Tatar': 'tat_Cyrl', 'Telugu': 'tel_Telu', 'Tajik': 'tgk_Cyrl', 'Tagalog': 'tgl_Latn', 'Thai': 'tha_Thai', 'Tigrinya': 'tir_Ethi', 'Tamasheq Latin': 'taq_Latn', 'Tamasheq Tifinagh': 'taq_Tfng', 'Tok Pisin': 'tpi_Latn', 'Tswana': 'tsn_Latn', 'Tsonga': 'tso_Latn', 'Turkmen': 'tuk_Latn', 'Tumbuka': 'tum_Latn', 'Turkish': 'tur_Latn', 'Twi': 'twi_Latn', 'Central Atlas Tamazight': 'tzm_Tfng', 'Uyghur': 'uig_Arab', 'Ukrainian': 'ukr_Cyrl', 'Umbundu': 'umb_Latn', 'Urdu': 'urd_Arab', 'Northern Uzbek': 'uzn_Latn', 'Venetian': 'vec_Latn', 'Vietnamese': 'vie_Latn', 'Waray': 'war_Latn', 'Wolof': 'wol_Latn', 'Xhosa': 'xho_Latn', 'Eastern Yiddish': 'ydd_Hebr', 'Yoruba': 'yor_Latn', 'Yue Chinese': 'yue_Hant', 'Chinese Simplified': 'zho_Hans', 'Chinese Traditional': 'zho_Hant', 'Standard Malay': 'zsm_Latn', 'Zulu': 'zul_Latn', } class __A( lowercase__ ): snake_case_ = '''facebook/nllb-200-distilled-600M''' snake_case_ = ( '''This is a tool that translates text from a language to another. It takes three inputs: `text`, which should ''' '''be the text to translate, `src_lang`, which should be the language of the text to translate and `tgt_lang`, ''' '''which should be the language for the desired ouput language. Both `src_lang` and `tgt_lang` are written in ''' '''plain English, such as \'Romanian\', or \'Albanian\'. It returns the text translated in `tgt_lang`.''' ) snake_case_ = '''translator''' snake_case_ = AutoTokenizer snake_case_ = AutoModelForSeqaSeqLM snake_case_ = LANGUAGE_CODES snake_case_ = ['''text''', '''text''', '''text'''] snake_case_ = ['''text'''] def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case ) -> Tuple: '''simple docstring''' if src_lang not in self.lang_to_code: raise ValueError(F"""{src_lang} is not a supported language.""" ) if tgt_lang not in self.lang_to_code: raise ValueError(F"""{tgt_lang} is not a supported language.""" ) __a = self.lang_to_code[src_lang] __a = self.lang_to_code[tgt_lang] return self.pre_processor._build_translation_inputs( SCREAMING_SNAKE_CASE__ , return_tensors='''pt''' , src_lang=SCREAMING_SNAKE_CASE__ , tgt_lang=SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> Any: '''simple docstring''' return self.model.generate(**SCREAMING_SNAKE_CASE__ ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> Dict: '''simple docstring''' return self.post_processor.decode(outputs[0].tolist() , skip_special_tokens=SCREAMING_SNAKE_CASE__ )
6
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES UpperCAmelCase_ : Union[str, Any] = logging.get_logger(__name__) UpperCAmelCase_ : List[str] = OrderedDict( [ # Base model mapping ('albert', 'FlaxAlbertModel'), ('bart', 'FlaxBartModel'), ('beit', 'FlaxBeitModel'), ('bert', 'FlaxBertModel'), ('big_bird', 'FlaxBigBirdModel'), ('blenderbot', 'FlaxBlenderbotModel'), ('blenderbot-small', 'FlaxBlenderbotSmallModel'), ('clip', 'FlaxCLIPModel'), ('distilbert', 'FlaxDistilBertModel'), ('electra', 'FlaxElectraModel'), ('gpt-sw3', 'FlaxGPT2Model'), ('gpt2', 'FlaxGPT2Model'), ('gpt_neo', 'FlaxGPTNeoModel'), ('gptj', 'FlaxGPTJModel'), ('longt5', 'FlaxLongT5Model'), ('marian', 'FlaxMarianModel'), ('mbart', 'FlaxMBartModel'), ('mt5', 'FlaxMT5Model'), ('opt', 'FlaxOPTModel'), ('pegasus', 'FlaxPegasusModel'), ('regnet', 'FlaxRegNetModel'), ('resnet', 'FlaxResNetModel'), ('roberta', 'FlaxRobertaModel'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormModel'), ('roformer', 'FlaxRoFormerModel'), ('t5', 'FlaxT5Model'), ('vision-text-dual-encoder', 'FlaxVisionTextDualEncoderModel'), ('vit', 'FlaxViTModel'), ('wav2vec2', 'FlaxWav2Vec2Model'), ('whisper', 'FlaxWhisperModel'), ('xglm', 'FlaxXGLMModel'), ('xlm-roberta', 'FlaxXLMRobertaModel'), ] ) UpperCAmelCase_ : str = OrderedDict( [ # Model for pre-training mapping ('albert', 'FlaxAlbertForPreTraining'), ('bart', 'FlaxBartForConditionalGeneration'), ('bert', 'FlaxBertForPreTraining'), ('big_bird', 'FlaxBigBirdForPreTraining'), ('electra', 'FlaxElectraForPreTraining'), ('longt5', 'FlaxLongT5ForConditionalGeneration'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('mt5', 'FlaxMT5ForConditionalGeneration'), ('roberta', 'FlaxRobertaForMaskedLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMaskedLM'), ('roformer', 'FlaxRoFormerForMaskedLM'), ('t5', 'FlaxT5ForConditionalGeneration'), ('wav2vec2', 'FlaxWav2Vec2ForPreTraining'), ('whisper', 'FlaxWhisperForConditionalGeneration'), ('xlm-roberta', 'FlaxXLMRobertaForMaskedLM'), ] ) UpperCAmelCase_ : Dict = OrderedDict( [ # Model for Masked LM mapping ('albert', 'FlaxAlbertForMaskedLM'), ('bart', 'FlaxBartForConditionalGeneration'), ('bert', 'FlaxBertForMaskedLM'), ('big_bird', 'FlaxBigBirdForMaskedLM'), ('distilbert', 'FlaxDistilBertForMaskedLM'), ('electra', 'FlaxElectraForMaskedLM'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('roberta', 'FlaxRobertaForMaskedLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMaskedLM'), ('roformer', 'FlaxRoFormerForMaskedLM'), ('xlm-roberta', 'FlaxXLMRobertaForMaskedLM'), ] ) UpperCAmelCase_ : Optional[Any] = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ('bart', 'FlaxBartForConditionalGeneration'), ('blenderbot', 'FlaxBlenderbotForConditionalGeneration'), ('blenderbot-small', 'FlaxBlenderbotSmallForConditionalGeneration'), ('encoder-decoder', 'FlaxEncoderDecoderModel'), ('longt5', 'FlaxLongT5ForConditionalGeneration'), ('marian', 'FlaxMarianMTModel'), ('mbart', 'FlaxMBartForConditionalGeneration'), ('mt5', 'FlaxMT5ForConditionalGeneration'), ('pegasus', 'FlaxPegasusForConditionalGeneration'), ('t5', 'FlaxT5ForConditionalGeneration'), ] ) UpperCAmelCase_ : List[str] = OrderedDict( [ # Model for Image-classsification ('beit', 'FlaxBeitForImageClassification'), ('regnet', 'FlaxRegNetForImageClassification'), ('resnet', 'FlaxResNetForImageClassification'), ('vit', 'FlaxViTForImageClassification'), ] ) UpperCAmelCase_ : int = OrderedDict( [ ('vision-encoder-decoder', 'FlaxVisionEncoderDecoderModel'), ] ) UpperCAmelCase_ : List[str] = OrderedDict( [ # Model for Causal LM mapping ('bart', 'FlaxBartForCausalLM'), ('bert', 'FlaxBertForCausalLM'), ('big_bird', 'FlaxBigBirdForCausalLM'), ('electra', 'FlaxElectraForCausalLM'), ('gpt-sw3', 'FlaxGPT2LMHeadModel'), ('gpt2', 'FlaxGPT2LMHeadModel'), ('gpt_neo', 'FlaxGPTNeoForCausalLM'), ('gptj', 'FlaxGPTJForCausalLM'), ('opt', 'FlaxOPTForCausalLM'), ('roberta', 'FlaxRobertaForCausalLM'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForCausalLM'), ('xglm', 'FlaxXGLMForCausalLM'), ('xlm-roberta', 'FlaxXLMRobertaForCausalLM'), ] ) UpperCAmelCase_ : List[str] = OrderedDict( [ # Model for Sequence Classification mapping ('albert', 'FlaxAlbertForSequenceClassification'), ('bart', 'FlaxBartForSequenceClassification'), ('bert', 'FlaxBertForSequenceClassification'), ('big_bird', 'FlaxBigBirdForSequenceClassification'), ('distilbert', 'FlaxDistilBertForSequenceClassification'), ('electra', 'FlaxElectraForSequenceClassification'), ('mbart', 'FlaxMBartForSequenceClassification'), ('roberta', 'FlaxRobertaForSequenceClassification'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForSequenceClassification'), ('roformer', 'FlaxRoFormerForSequenceClassification'), ('xlm-roberta', 'FlaxXLMRobertaForSequenceClassification'), ] ) UpperCAmelCase_ : List[str] = OrderedDict( [ # Model for Question Answering mapping ('albert', 'FlaxAlbertForQuestionAnswering'), ('bart', 'FlaxBartForQuestionAnswering'), ('bert', 'FlaxBertForQuestionAnswering'), ('big_bird', 'FlaxBigBirdForQuestionAnswering'), ('distilbert', 'FlaxDistilBertForQuestionAnswering'), ('electra', 'FlaxElectraForQuestionAnswering'), ('mbart', 'FlaxMBartForQuestionAnswering'), ('roberta', 'FlaxRobertaForQuestionAnswering'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForQuestionAnswering'), ('roformer', 'FlaxRoFormerForQuestionAnswering'), ('xlm-roberta', 'FlaxXLMRobertaForQuestionAnswering'), ] ) UpperCAmelCase_ : Union[str, Any] = OrderedDict( [ # Model for Token Classification mapping ('albert', 'FlaxAlbertForTokenClassification'), ('bert', 'FlaxBertForTokenClassification'), ('big_bird', 'FlaxBigBirdForTokenClassification'), ('distilbert', 'FlaxDistilBertForTokenClassification'), ('electra', 'FlaxElectraForTokenClassification'), ('roberta', 'FlaxRobertaForTokenClassification'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForTokenClassification'), ('roformer', 'FlaxRoFormerForTokenClassification'), ('xlm-roberta', 'FlaxXLMRobertaForTokenClassification'), ] ) UpperCAmelCase_ : Dict = OrderedDict( [ # Model for Multiple Choice mapping ('albert', 'FlaxAlbertForMultipleChoice'), ('bert', 'FlaxBertForMultipleChoice'), ('big_bird', 'FlaxBigBirdForMultipleChoice'), ('distilbert', 'FlaxDistilBertForMultipleChoice'), ('electra', 'FlaxElectraForMultipleChoice'), ('roberta', 'FlaxRobertaForMultipleChoice'), ('roberta-prelayernorm', 'FlaxRobertaPreLayerNormForMultipleChoice'), ('roformer', 'FlaxRoFormerForMultipleChoice'), ('xlm-roberta', 'FlaxXLMRobertaForMultipleChoice'), ] ) UpperCAmelCase_ : List[str] = OrderedDict( [ ('bert', 'FlaxBertForNextSentencePrediction'), ] ) UpperCAmelCase_ : Dict = OrderedDict( [ ('speech-encoder-decoder', 'FlaxSpeechEncoderDecoderModel'), ('whisper', 'FlaxWhisperForConditionalGeneration'), ] ) UpperCAmelCase_ : Union[str, Any] = OrderedDict( [ ('whisper', 'FlaxWhisperForAudioClassification'), ] ) UpperCAmelCase_ : Any = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) UpperCAmelCase_ : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) UpperCAmelCase_ : Dict = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) UpperCAmelCase_ : List[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) UpperCAmelCase_ : int = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) UpperCAmelCase_ : Any = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) UpperCAmelCase_ : int = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) UpperCAmelCase_ : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) UpperCAmelCase_ : Optional[int] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) UpperCAmelCase_ : List[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) UpperCAmelCase_ : int = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) UpperCAmelCase_ : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) UpperCAmelCase_ : Tuple = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) UpperCAmelCase_ : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : List[Any] = FLAX_MODEL_MAPPING UpperCAmelCase_ : Tuple = auto_class_update(FlaxAutoModel) class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Any = FLAX_MODEL_FOR_PRETRAINING_MAPPING UpperCAmelCase_ : Optional[Any] = auto_class_update(FlaxAutoModelForPreTraining, head_doc='pretraining') class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : List[Any] = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING UpperCAmelCase_ : Optional[Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc='causal language modeling') class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Optional[Any] = FLAX_MODEL_FOR_MASKED_LM_MAPPING UpperCAmelCase_ : Union[str, Any] = auto_class_update(FlaxAutoModelForMaskedLM, head_doc='masked language modeling') class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Tuple = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING UpperCAmelCase_ : Optional[int] = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc='sequence-to-sequence language modeling', checkpoint_for_example='t5-base' ) class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Tuple = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING UpperCAmelCase_ : Optional[Any] = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc='sequence classification' ) class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Tuple = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING UpperCAmelCase_ : str = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc='question answering') class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : List[str] = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING UpperCAmelCase_ : Tuple = auto_class_update( FlaxAutoModelForTokenClassification, head_doc='token classification' ) class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Dict = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING UpperCAmelCase_ : int = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc='multiple choice') class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Optional[int] = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING UpperCAmelCase_ : Dict = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc='next sentence prediction' ) class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Dict = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING UpperCAmelCase_ : str = auto_class_update( FlaxAutoModelForImageClassification, head_doc='image classification' ) class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Optional[Any] = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING UpperCAmelCase_ : Tuple = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc='vision-to-text modeling') class SCREAMING_SNAKE_CASE__ ( _BaseAutoModelClass ): snake_case__ : Optional[int] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING UpperCAmelCase_ : Union[str, Any] = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc='sequence-to-sequence speech-to-text modeling' )
32
0
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation UpperCAmelCase : int = logging.get_logger(__name__) UpperCAmelCase : List[Any] = {"""vocab_file""": """spiece.model"""} UpperCAmelCase : int = { """vocab_file""": { """AI-Sweden/gpt-sw3-126m""": """https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-350m""": """https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-1.6b""": """https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-6.7b""": """https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-20b""": """https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model""", } } UpperCAmelCase : str = { """AI-Sweden/gpt-sw3-126m""": 2048, """AI-Sweden/gpt-sw3-350m""": 2048, """AI-Sweden/gpt-sw3-1.6b""": 2048, """AI-Sweden/gpt-sw3-6.7b""": 2048, """AI-Sweden/gpt-sw3-20b""": 2048, } class __lowerCAmelCase ( UpperCamelCase__): _lowercase : int = VOCAB_FILES_NAMES _lowercase : Optional[int] = PRETRAINED_VOCAB_FILES_MAP _lowercase : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowercase : List[str] = ["""input_ids""", """attention_mask"""] def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=False , lowerCAmelCase__=False , lowerCAmelCase__=False , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__ = None , **lowerCAmelCase__ , ) -> None: '''simple docstring''' a__ : str ={} if sp_model_kwargs is None else sp_model_kwargs a__ : Any =kwargs.get("name_or_path" ) if name_or_path is None: logger.warning( "name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b," " you are testing the model, this can safely be ignored" ) a__ : int ="None" # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing a__ : Dict ="<|endoftext|>" if eos_token is None else eos_token a__ : Any ="<unk>" if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: a__ : Optional[int] =unk_token if pad_token is None else pad_token a__ : List[str] =eos_token if bos_token is None else bos_token else: a__ : Optional[int] ="<pad>" if pad_token is None else pad_token a__ : Any ="<s>" if bos_token is None else bos_token super().__init__( do_lower_case=lowerCAmelCase__ , remove_space=lowerCAmelCase__ , keep_accents=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , **lowerCAmelCase__ , ) a__ : List[str] =do_lower_case a__ : Union[str, Any] =remove_space a__ : Union[str, Any] =keep_accents a__ : int =vocab_file a__ : Any =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCAmelCase__ ) # Used for whitespace normalization in input texts # fmt : off a__ : List[Any] ={" ", " ", " ", " ", " ", " ", " ", " ", " ", " ", "", "„"} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing a__ : Optional[Any] =re.compile( F'''[{"".join(map(lowerCAmelCase__ , list(range(0 , 9 ) ) + list(range(1_1 , 3_2 ) ) + list(range(1_2_7 , 1_6_0 ) ) + [1_6_0, 1_7_3, 8_2_0_3] ) )}]''' ) def __getstate__( self ) -> str: '''simple docstring''' a__ : Dict =self.__dict__.copy() a__ : Optional[int] =None return state def __setstate__( self , lowerCAmelCase__ ) -> Optional[Any]: '''simple docstring''' a__ : List[Any] =d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): a__ : str ={} a__ : Any =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def _lowercase ( self ) -> int: '''simple docstring''' return len(self.sp_model ) def _lowercase ( self , lowerCAmelCase__ ) -> str: '''simple docstring''' a__ : List[str] =self.non_printing_characters_re.sub("" , lowerCAmelCase__ ) # Normalize whitespaces a__ : Optional[int] ="".join([char if char not in self.whitespaces else " " for char in text] ) # NFC Unicode normalization a__ : Optional[Any] =unicodedata.normalize("NFC" , lowerCAmelCase__ ) return text def _lowercase ( self , lowerCAmelCase__ , **lowerCAmelCase__ ) -> List[str]: '''simple docstring''' a__ : Dict =self.preprocess_text(lowerCAmelCase__ ) return self.sp_model.encode(lowerCAmelCase__ , out_type=lowerCAmelCase__ ) def _lowercase ( self , lowerCAmelCase__ ) -> int: '''simple docstring''' return self.sp_model.PieceToId(lowerCAmelCase__ ) def _lowercase ( self , lowerCAmelCase__ ) -> str: '''simple docstring''' return self.sp_model.IdToPiece(lowerCAmelCase__ ) @staticmethod def _lowercase ( lowerCAmelCase__ ) -> str: '''simple docstring''' return out_string def _lowercase ( self , lowerCAmelCase__ ) -> str: '''simple docstring''' a__ : int =[] a__ : List[Any] ="" a__ : Tuple =False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowerCAmelCase__ ) + token a__ : Tuple =True a__ : Any =[] else: current_sub_tokens.append(lowerCAmelCase__ ) a__ : List[str] =False out_string += self.sp_model.decode(lowerCAmelCase__ ) return out_string def _lowercase ( self ) -> Dict[str, int]: '''simple docstring''' a__ : List[str] ={self.convert_ids_to_tokens(lowerCAmelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(lowerCAmelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return a__ : List[Any] =os.path.join( lowerCAmelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCAmelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCAmelCase__ , "wb" ) as fi: a__ : List[str] =self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase__ ) return (out_vocab_file,) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ = False ) -> Union[List[int], List[List[int]], "torch.Tensor"]: '''simple docstring''' if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): a__ : Tuple =self.preprocess_text(lowerCAmelCase__ ) a__ : str =self.sp_model.encode(lowerCAmelCase__ ) else: a__ : List[str] =[self.preprocess_text(lowerCAmelCase__ ) for t in text] a__ : int =self.sp_model.encode(lowerCAmelCase__ ) if return_tensors is True or return_tensors == "pt": a__ : Optional[int] =torch.tensor(lowerCAmelCase__ ) return token_ids def _lowercase ( self , lowerCAmelCase__ ) -> str: '''simple docstring''' return self.sp_model.decode(lowerCAmelCase__ ) def _lowercase ( self , lowerCAmelCase__ ) -> List[int]: '''simple docstring''' a__ : Tuple =[F'''User: {text}''' if is_user else F'''Bot: {text}''' for is_user, text in conversation.iter_texts()] a__ : Tuple =( F'''{self.eos_token}{self.bos_token}''' + F'''{self.bos_token}'''.join(lowerCAmelCase__ ) + F'''{self.bos_token}Bot:''' ) return self.encode(text=lowerCAmelCase__ )
148
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features UpperCAmelCase : Tuple = logging.get_logger(__name__) UpperCAmelCase : List[Any] = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) UpperCAmelCase : Dict = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class __lowerCAmelCase : _lowercase : str = field( default=UpperCamelCase__ , metadata={"""help""": """Model type selected in the list: """ + """, """.join(UpperCamelCase__)}) _lowercase : str = field( default=UpperCamelCase__ , metadata={"""help""": """The input data dir. Should contain the .json files for the SQuAD task."""}) _lowercase : int = field( default=128 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) _lowercase : int = field( default=128 , metadata={"""help""": """When splitting up a long document into chunks, how much stride to take between chunks."""} , ) _lowercase : int = field( default=64 , metadata={ """help""": ( """The maximum number of tokens for the question. Questions longer than this will """ """be truncated to this length.""" ) } , ) _lowercase : int = field( default=30 , metadata={ """help""": ( """The maximum length of an answer that can be generated. This is needed because the start """ """and end predictions are not conditioned on one another.""" ) } , ) _lowercase : bool = field( default=UpperCamelCase__ , metadata={"""help""": """Overwrite the cached training and evaluation sets"""}) _lowercase : bool = field( default=UpperCamelCase__ , metadata={"""help""": """If true, the SQuAD examples contain some that do not have an answer."""}) _lowercase : float = field( default=0.0 , metadata={"""help""": """If null_score - best_non_null is greater than the threshold predict null."""}) _lowercase : int = field( default=20 , metadata={"""help""": """If null_score - best_non_null is greater than the threshold predict null."""}) _lowercase : int = field( default=0 , metadata={ """help""": ( """language id of input for language-specific xlm models (see""" """ tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)""" ) } , ) _lowercase : int = field(default=1 , metadata={"""help""": """multiple threads for converting example to features"""}) class __lowerCAmelCase ( UpperCamelCase__): _lowercase : List[Any] = """train""" _lowercase : Any = """dev""" class __lowerCAmelCase ( UpperCamelCase__): _lowercase : SquadDataTrainingArguments _lowercase : List[SquadFeatures] _lowercase : Split _lowercase : bool def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = Split.train , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = "pt" , ) -> str: '''simple docstring''' a__ : List[Any] =args a__ : int =is_language_sensitive a__ : List[str] =SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): try: a__ : str =Split[mode] except KeyError: raise KeyError("mode is not a valid split name" ) a__ : Any =mode # Load data features from cache or dataset file a__ : str ="v2" if args.version_2_with_negative else "v1" a__ : str =os.path.join( cache_dir if cache_dir is not None else args.data_dir , F'''cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}''' , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. a__ : Dict =cached_features_file + ".lock" with FileLock(lowerCAmelCase__ ): if os.path.exists(lowerCAmelCase__ ) and not args.overwrite_cache: a__ : Any =time.time() a__ : List[Any] =torch.load(lowerCAmelCase__ ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. a__ : str =self.old_features["features"] a__ : str =self.old_features.get("dataset" , lowerCAmelCase__ ) a__ : Optional[int] =self.old_features.get("examples" , lowerCAmelCase__ ) logger.info( F'''Loading features from cached file {cached_features_file} [took %.3f s]''' , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( F'''Deleting cached file {cached_features_file} will allow dataset and examples to be cached in''' " future run" ) else: if mode == Split.dev: a__ : Dict =self.processor.get_dev_examples(args.data_dir ) else: a__ : str =self.processor.get_train_examples(args.data_dir ) a__ , a__ : Union[str, Any] =squad_convert_examples_to_features( examples=self.examples , tokenizer=lowerCAmelCase__ , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=lowerCAmelCase__ , ) a__ : Any =time.time() torch.save( {"features": self.features, "dataset": self.dataset, "examples": self.examples} , lowerCAmelCase__ , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( F'''Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]''' ) def __len__( self ) -> Tuple: '''simple docstring''' return len(self.features ) def __getitem__( self , lowerCAmelCase__ ) -> Dict[str, torch.Tensor]: '''simple docstring''' a__ : str =self.features[i] a__ : Optional[Any] =torch.tensor(feature.input_ids , dtype=torch.long ) a__ : List[Any] =torch.tensor(feature.attention_mask , dtype=torch.long ) a__ : int =torch.tensor(feature.token_type_ids , dtype=torch.long ) a__ : List[str] =torch.tensor(feature.cls_index , dtype=torch.long ) a__ : int =torch.tensor(feature.p_mask , dtype=torch.float ) a__ : Tuple =torch.tensor(feature.is_impossible , dtype=torch.float ) a__ : Tuple ={ "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"cls_index": cls_index, "p_mask": p_mask} ) if self.args.version_2_with_negative: inputs.update({"is_impossible": is_impossible} ) if self.is_language_sensitive: inputs.update({"langs": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: a__ : int =torch.tensor(feature.start_position , dtype=torch.long ) a__ : Any =torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"start_positions": start_positions, "end_positions": end_positions} ) return inputs
148
1
'''simple docstring''' import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor UpperCamelCase__ : Optional[int] = logging.get_logger(__name__) class _UpperCamelCase ( lowerCamelCase__ ): '''simple docstring''' def __init__( self : Dict , *lowerCAmelCase__ : Any , **lowerCAmelCase__ : Any ): """simple docstring""" warnings.warn( """The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use ImageGPTImageProcessor instead.""" , lowerCAmelCase__ , ) super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__ )
112
'''simple docstring''' def lowerCAmelCase_ ( _lowerCamelCase: int ): __SCREAMING_SNAKE_CASE : str = int(_lowerCamelCase ) if n_element < 1: __SCREAMING_SNAKE_CASE : List[str] = ValueError("""a should be a positive number""" ) raise my_error __SCREAMING_SNAKE_CASE : List[Any] = [1] __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Optional[int] = (0, 0, 0) __SCREAMING_SNAKE_CASE : List[str] = 1 while index < n_element: while hamming_list[i] * 2 <= hamming_list[-1]: i += 1 while hamming_list[j] * 3 <= hamming_list[-1]: j += 1 while hamming_list[k] * 5 <= hamming_list[-1]: k += 1 hamming_list.append( min(hamming_list[i] * 2 , hamming_list[j] * 3 , hamming_list[k] * 5 ) ) index += 1 return hamming_list if __name__ == "__main__": UpperCamelCase__ : Optional[Any] = input('''Enter the last number (nth term) of the Hamming Number Series: ''') print('''Formula of Hamming Number Series => 2^i * 3^j * 5^k''') UpperCamelCase__ : List[str] = hamming(int(n)) print('''-----------------------------------------------------''') print(f"The list with nth numbers is: {hamming_numbers}") print('''-----------------------------------------------------''')
112
1
import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis from lavis.models import load_model_and_preprocess from PIL import Image from transformers import ( AutoTokenizer, BlipaConfig, BlipaForConditionalGeneration, BlipaProcessor, BlipaVisionConfig, BlipImageProcessor, OPTConfig, TaConfig, ) from transformers.utils.constants import OPENAI_CLIP_MEAN, OPENAI_CLIP_STD def UpperCamelCase ( ) -> Tuple: UpperCamelCase : Any = 'https://storage.googleapis.com/sfr-vision-language-research/LAVIS/assets/merlion.png' UpperCamelCase : List[str] = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ).convert('RGB' ) return image def UpperCamelCase ( snake_case__ : int ) -> List[Any]: UpperCamelCase : Optional[int] = [] # fmt: off # vision encoder rename_keys.append(('visual_encoder.cls_token', 'vision_model.embeddings.class_embedding') ) rename_keys.append(('visual_encoder.pos_embed', 'vision_model.embeddings.position_embedding') ) rename_keys.append(('visual_encoder.patch_embed.proj.weight', 'vision_model.embeddings.patch_embedding.weight') ) rename_keys.append(('visual_encoder.patch_embed.proj.bias', 'vision_model.embeddings.patch_embedding.bias') ) rename_keys.append(('ln_vision.weight', 'vision_model.post_layernorm.weight') ) rename_keys.append(('ln_vision.bias', 'vision_model.post_layernorm.bias') ) for i in range(config.vision_config.num_hidden_layers ): rename_keys.append((F"""visual_encoder.blocks.{i}.norm1.weight""", F"""vision_model.encoder.layers.{i}.layer_norm1.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.norm1.bias""", F"""vision_model.encoder.layers.{i}.layer_norm1.bias""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.norm2.weight""", F"""vision_model.encoder.layers.{i}.layer_norm2.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.norm2.bias""", F"""vision_model.encoder.layers.{i}.layer_norm2.bias""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.attn.qkv.weight""", F"""vision_model.encoder.layers.{i}.self_attn.qkv.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.attn.proj.weight""", F"""vision_model.encoder.layers.{i}.self_attn.projection.weight""",) ) rename_keys.append((F"""visual_encoder.blocks.{i}.attn.proj.bias""", F"""vision_model.encoder.layers.{i}.self_attn.projection.bias""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.mlp.fc1.weight""", F"""vision_model.encoder.layers.{i}.mlp.fc1.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.mlp.fc1.bias""", F"""vision_model.encoder.layers.{i}.mlp.fc1.bias""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.mlp.fc2.weight""", F"""vision_model.encoder.layers.{i}.mlp.fc2.weight""") ) rename_keys.append((F"""visual_encoder.blocks.{i}.mlp.fc2.bias""", F"""vision_model.encoder.layers.{i}.mlp.fc2.bias""") ) # QFormer rename_keys.append(('Qformer.bert.embeddings.LayerNorm.weight', 'qformer.layernorm.weight') ) rename_keys.append(('Qformer.bert.embeddings.LayerNorm.bias', 'qformer.layernorm.bias') ) # fmt: on return rename_keys def UpperCamelCase ( snake_case__ : Optional[Any] , snake_case__ : Dict , snake_case__ : Optional[int] ) -> Optional[int]: UpperCamelCase : Dict = dct.pop(snake_case__ ) UpperCamelCase : str = val def UpperCamelCase ( snake_case__ : str , snake_case__ : Union[str, Any] ) -> Optional[int]: for i in range(config.vision_config.num_hidden_layers ): # read in original q and v biases UpperCamelCase : Optional[Any] = state_dict.pop(F"""visual_encoder.blocks.{i}.attn.q_bias""" ) UpperCamelCase : int = state_dict.pop(F"""visual_encoder.blocks.{i}.attn.v_bias""" ) # next, set bias in the state dict UpperCamelCase : int = torch.cat((q_bias, torch.zeros_like(snake_case__ , requires_grad=snake_case__ ), v_bias) ) UpperCamelCase : Tuple = qkv_bias def UpperCamelCase ( snake_case__ : Tuple , snake_case__ : Optional[Any] ) -> Dict: UpperCamelCase : str = 364 if 'coco' in model_name else 224 UpperCamelCase : Union[str, Any] = BlipaVisionConfig(image_size=snake_case__ ).to_dict() # make sure the models have proper bos_token_id and eos_token_id set (important for generation) # seems like flan-T5 models don't have bos_token_id properly set? if "opt-2.7b" in model_name: UpperCamelCase : List[Any] = OPTConfig.from_pretrained('facebook/opt-2.7b' , eos_token_id=snake_case__ ).to_dict() elif "opt-6.7b" in model_name: UpperCamelCase : int = OPTConfig.from_pretrained('facebook/opt-6.7b' , eos_token_id=snake_case__ ).to_dict() elif "t5-xl" in model_name: UpperCamelCase : List[str] = TaConfig.from_pretrained('google/flan-t5-xl' , dense_act_fn='gelu' , bos_token_id=1 ).to_dict() elif "t5-xxl" in model_name: UpperCamelCase : int = TaConfig.from_pretrained('google/flan-t5-xxl' , dense_act_fn='gelu' , bos_token_id=1 ).to_dict() UpperCamelCase : Any = BlipaConfig(vision_config=snake_case__ , text_config=snake_case__ ) return config, image_size @torch.no_grad() def UpperCamelCase ( snake_case__ : int , snake_case__ : Dict=None , snake_case__ : int=False ) -> List[Any]: UpperCamelCase : str = ( AutoTokenizer.from_pretrained('facebook/opt-2.7b' ) if 'opt' in model_name else AutoTokenizer.from_pretrained('google/flan-t5-xl' ) ) UpperCamelCase : int = tokenizer('\n' , add_special_tokens=snake_case__ ).input_ids[0] UpperCamelCase , UpperCamelCase : Union[str, Any] = get_blipa_config(snake_case__ , eos_token_id=snake_case__ ) UpperCamelCase : Dict = BlipaForConditionalGeneration(snake_case__ ).eval() UpperCamelCase : Optional[Any] = { 'blip2-opt-2.7b': ('blip2_opt', 'pretrain_opt2.7b'), 'blip2-opt-6.7b': ('blip2_opt', 'pretrain_opt6.7b'), 'blip2-opt-2.7b-coco': ('blip2_opt', 'caption_coco_opt2.7b'), 'blip2-opt-6.7b-coco': ('blip2_opt', 'caption_coco_opt6.7b'), 'blip2-flan-t5-xl': ('blip2_t5', 'pretrain_flant5xl'), 'blip2-flan-t5-xl-coco': ('blip2_t5', 'caption_coco_flant5xl'), 'blip2-flan-t5-xxl': ('blip2_t5', 'pretrain_flant5xxl'), } UpperCamelCase , UpperCamelCase : Optional[Any] = model_name_to_original[model_name] # load original model print('Loading original model...' ) UpperCamelCase : List[str] = 'cuda' if torch.cuda.is_available() else 'cpu' UpperCamelCase , UpperCamelCase , UpperCamelCase : Tuple = load_model_and_preprocess( name=snake_case__ , model_type=snake_case__ , is_eval=snake_case__ , device=snake_case__ ) original_model.eval() print('Done!' ) # update state dict keys UpperCamelCase : List[Any] = original_model.state_dict() UpperCamelCase : Tuple = create_rename_keys(snake_case__ ) for src, dest in rename_keys: rename_key(snake_case__ , snake_case__ , snake_case__ ) # some keys can be renamed efficiently for key, val in state_dict.copy().items(): UpperCamelCase : Optional[Any] = state_dict.pop(snake_case__ ) if key.startswith('Qformer.bert' ): UpperCamelCase : List[str] = key.replace('Qformer.bert' , 'qformer' ) if "attention.self" in key: UpperCamelCase : Tuple = key.replace('self' , 'attention' ) if "opt_proj" in key: UpperCamelCase : Union[str, Any] = key.replace('opt_proj' , 'language_projection' ) if "t5_proj" in key: UpperCamelCase : Optional[Any] = key.replace('t5_proj' , 'language_projection' ) if key.startswith('opt' ): UpperCamelCase : Dict = key.replace('opt' , 'language' ) if key.startswith('t5' ): UpperCamelCase : Dict = key.replace('t5' , 'language' ) UpperCamelCase : Optional[int] = val # read in qv biases read_in_q_v_bias(snake_case__ , snake_case__ ) UpperCamelCase , UpperCamelCase : Any = hf_model.load_state_dict(snake_case__ , strict=snake_case__ ) assert len(snake_case__ ) == 0 assert unexpected_keys == ["qformer.embeddings.position_ids"] UpperCamelCase : List[str] = load_demo_image() UpperCamelCase : str = vis_processors['eval'](snake_case__ ).unsqueeze(0 ).to(snake_case__ ) UpperCamelCase : Any = tokenizer(['\n'] , return_tensors='pt' ).input_ids.to(snake_case__ ) # create processor UpperCamelCase : Optional[Any] = BlipImageProcessor( size={'height': image_size, 'width': image_size} , image_mean=snake_case__ , image_std=snake_case__ ) UpperCamelCase : Any = BlipaProcessor(image_processor=snake_case__ , tokenizer=snake_case__ ) UpperCamelCase : Optional[int] = processor(images=snake_case__ , return_tensors='pt' ).pixel_values.to(snake_case__ ) # make sure processor creates exact same pixel values assert torch.allclose(snake_case__ , snake_case__ ) original_model.to(snake_case__ ) hf_model.to(snake_case__ ) with torch.no_grad(): if "opt" in model_name: UpperCamelCase : Tuple = original_model({'image': original_pixel_values, 'text_input': ['']} ).logits UpperCamelCase : str = hf_model(snake_case__ , snake_case__ ).logits else: UpperCamelCase : Tuple = original_model( {'image': original_pixel_values, 'text_input': ['\n'], 'text_output': ['\n']} ).logits UpperCamelCase : List[Any] = input_ids.masked_fill(input_ids == tokenizer.pad_token_id , -100 ) UpperCamelCase : Optional[int] = hf_model(snake_case__ , snake_case__ , labels=snake_case__ ).logits assert original_logits.shape == logits.shape print('First values of original logits:' , original_logits[0, :3, :3] ) print('First values of HF logits:' , logits[0, :3, :3] ) # assert values if model_name == "blip2-flan-t5-xl": UpperCamelCase : List[str] = torch.tensor( [[-41.5850, -4.4440, -8.9922], [-47.4322, -5.9143, -1.7340]] , device=snake_case__ ) assert torch.allclose(logits[0, :3, :3] , snake_case__ , atol=1E-4 ) elif model_name == "blip2-flan-t5-xl-coco": UpperCamelCase : Union[str, Any] = torch.tensor( [[-57.0109, -9.8967, -12.6280], [-68.6578, -12.7191, -10.5065]] , device=snake_case__ ) else: # cast to same type UpperCamelCase : Optional[int] = logits.dtype assert torch.allclose(original_logits.to(snake_case__ ) , snake_case__ , atol=1E-2 ) print('Looks ok!' ) print('Generating a caption...' ) UpperCamelCase : Optional[int] = '' UpperCamelCase : Union[str, Any] = tokenizer(snake_case__ , return_tensors='pt' ).input_ids.to(snake_case__ ) UpperCamelCase : str = original_model.generate({'image': original_pixel_values} ) UpperCamelCase : str = hf_model.generate( snake_case__ , snake_case__ , do_sample=snake_case__ , num_beams=5 , max_length=30 , min_length=1 , top_p=0.9 , repetition_penalty=1.0 , length_penalty=1.0 , temperature=1 , ) print('Original generation:' , snake_case__ ) UpperCamelCase : Optional[int] = input_ids.shape[1] UpperCamelCase : Union[str, Any] = processor.batch_decode(outputs[:, prompt_length:] , skip_special_tokens=snake_case__ ) UpperCamelCase : Dict = [text.strip() for text in output_text] print('HF generation:' , snake_case__ ) if pytorch_dump_folder_path is not None: processor.save_pretrained(snake_case__ ) hf_model.save_pretrained(snake_case__ ) if push_to_hub: processor.push_to_hub(F"""nielsr/{model_name}""" ) hf_model.push_to_hub(F"""nielsr/{model_name}""" ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() __UpperCAmelCase = [ '''blip2-opt-2.7b''', '''blip2-opt-6.7b''', '''blip2-opt-2.7b-coco''', '''blip2-opt-6.7b-coco''', '''blip2-flan-t5-xl''', '''blip2-flan-t5-xl-coco''', '''blip2-flan-t5-xxl''', ] parser.add_argument( '''--model_name''', default='''blip2-opt-2.7b''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model and processor to the hub after converting''', ) __UpperCAmelCase = parser.parse_args() convert_blipa_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
103
import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def UpperCamelCase ( snake_case__ : List[Any] , snake_case__ : Union[str, Any] , snake_case__ : Tuple , snake_case__ : Union[str, Any] , snake_case__ : List[Any] ) -> Tuple: # load base model UpperCamelCase : Tuple = StableDiffusionPipeline.from_pretrained(snake_case__ , torch_dtype=torch.floataa ) # load LoRA weight from .safetensors UpperCamelCase : Union[str, Any] = load_file(snake_case__ ) UpperCamelCase : int = [] # directly update weight in diffusers model for key in state_dict: # it is suggested to print out the key, it usually will be something like below # "lora_te_text_model_encoder_layers_0_self_attn_k_proj.lora_down.weight" # as we have set the alpha beforehand, so just skip if ".alpha" in key or key in visited: continue if "text" in key: UpperCamelCase : Optional[Any] = key.split('.' )[0].split(LORA_PREFIX_TEXT_ENCODER + '_' )[-1].split('_' ) UpperCamelCase : Optional[Any] = pipeline.text_encoder else: UpperCamelCase : Tuple = key.split('.' )[0].split(LORA_PREFIX_UNET + '_' )[-1].split('_' ) UpperCamelCase : List[str] = pipeline.unet # find the target layer UpperCamelCase : Optional[Any] = layer_infos.pop(0 ) while len(snake_case__ ) > -1: try: UpperCamelCase : Dict = curr_layer.__getattr__(snake_case__ ) if len(snake_case__ ) > 0: UpperCamelCase : Dict = layer_infos.pop(0 ) elif len(snake_case__ ) == 0: break except Exception: if len(snake_case__ ) > 0: temp_name += "_" + layer_infos.pop(0 ) else: UpperCamelCase : Tuple = layer_infos.pop(0 ) UpperCamelCase : List[Any] = [] if "lora_down" in key: pair_keys.append(key.replace('lora_down' , 'lora_up' ) ) pair_keys.append(snake_case__ ) else: pair_keys.append(snake_case__ ) pair_keys.append(key.replace('lora_up' , 'lora_down' ) ) # update weight if len(state_dict[pair_keys[0]].shape ) == 4: UpperCamelCase : Dict = state_dict[pair_keys[0]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) UpperCamelCase : Union[str, Any] = state_dict[pair_keys[1]].squeeze(3 ).squeeze(2 ).to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(snake_case__ , snake_case__ ).unsqueeze(2 ).unsqueeze(3 ) else: UpperCamelCase : Dict = state_dict[pair_keys[0]].to(torch.floataa ) UpperCamelCase : List[Any] = state_dict[pair_keys[1]].to(torch.floataa ) curr_layer.weight.data += alpha * torch.mm(snake_case__ , snake_case__ ) # update visited list for item in pair_keys: visited.append(snake_case__ ) return pipeline if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() parser.add_argument( '''--base_model_path''', default=None, type=str, required=True, help='''Path to the base model in diffusers format.''' ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the checkpoint to convert.''' ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument( '''--lora_prefix_unet''', default='''lora_unet''', type=str, help='''The prefix of UNet weight in safetensors''' ) parser.add_argument( '''--lora_prefix_text_encoder''', default='''lora_te''', type=str, help='''The prefix of text encoder weight in safetensors''', ) parser.add_argument('''--alpha''', default=0.75, type=float, help='''The merging ratio in W = W0 + alpha * deltaW''') parser.add_argument( '''--to_safetensors''', action='''store_true''', help='''Whether to store pipeline in safetensors format or not.''' ) parser.add_argument('''--device''', type=str, help='''Device to use (e.g. cpu, cuda:0, cuda:1, etc.)''') __UpperCAmelCase = parser.parse_args() __UpperCAmelCase = args.base_model_path __UpperCAmelCase = args.checkpoint_path __UpperCAmelCase = args.dump_path __UpperCAmelCase = args.lora_prefix_unet __UpperCAmelCase = args.lora_prefix_text_encoder __UpperCAmelCase = args.alpha __UpperCAmelCase = convert(base_model_path, checkpoint_path, lora_prefix_unet, lora_prefix_text_encoder, alpha) __UpperCAmelCase = pipe.to(args.device) pipe.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
103
1
import argparse import dataclasses import json import logging import os import shutil from typing import List, Optional import datasets from accelerate import Accelerator from datasets import load_dataset from finetuning import finetune from tqdm.auto import tqdm import transformers from transformers import AutoConfig, set_seed from transformers.trainer_utils import IntervalStrategy __lowerCamelCase = logging.getLogger(__name__) __lowerCamelCase = """pytorch_model.bin""" @dataclasses.dataclass class UpperCAmelCase : A__ : str = dataclasses.field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models."} ) A__ : Optional[str] = dataclasses.field( default=A_ ,metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co."} ,) @dataclasses.dataclass class UpperCAmelCase : A__ : str = dataclasses.field(metadata={"help": "A csv or a json file containing the training data."} ) A__ : str = dataclasses.field(metadata={"help": "A csv or a json file containing the data to predict on."} ) A__ : Optional[str] = dataclasses.field( default=A_ ,metadata={"help": "A csv or a json file containing the validation data."} ) A__ : Optional[str] = dataclasses.field( default=A_ ,metadata={"help": "The name of the task to train on."} ,) A__ : Optional[List[str]] = dataclasses.field( default=A_ ,metadata={"help": "The list of labels for the task."} ) @dataclasses.dataclass class UpperCAmelCase : A__ : str = dataclasses.field( metadata={"help": "The output directory where the model predictions and checkpoints will be written."} ) A__ : Optional[str] = dataclasses.field( default="accuracy" ,metadata={"help": "The evaluation metric used for the task."} ) A__ : Optional[str] = dataclasses.field( default="no" ,metadata={ "help": "The evaluation strategy to adopt during training. Possible values are: [\"no\", \"step\", \"epoch]" } ,) A__ : Optional[int] = dataclasses.field( default=10 ,metadata={"help": "Number of evaluation calls with no improvement after which training will be stopped."} ,) A__ : Optional[float] = dataclasses.field( default=0.0 ,metadata={ "help": "How much the specified evaluation metric must improve to satisfy early stopping conditions." } ,) A__ : Optional[bool] = dataclasses.field( default=A_ ,metadata={"help": "Whether to filter the pseudo-labeled data based on the confidence score."} ,) A__ : Optional[bool] = dataclasses.field( default=A_ ,metadata={"help": "Whether to filter the pseudo-labeled data based on the validation performance."} ,) A__ : Optional[bool] = dataclasses.field( default=A_ ,metadata={"help": "Whether to fine-tune on labeled data after pseudo training."} ,) A__ : Optional[float] = dataclasses.field( default=0.0 ,metadata={"help": "Confidence threshold for pseudo-labeled data filtering."} ,) A__ : Optional[int] = dataclasses.field( default=1_00 ,metadata={"help": "Number of evaluation calls with no improvement after which training will be stopped."} ,) A__ : Optional[int] = dataclasses.field( default=A_ ,metadata={"help": "Random seed for initialization."} ,) def UpperCamelCase ( __lowerCamelCase : int , __lowerCamelCase : Tuple , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Tuple , __lowerCamelCase : Dict , __lowerCamelCase : Optional[int] ): snake_case : Tuple = datasets.concatenate_datasets([infer_input, infer_output] , axis=1 ) if args.do_filter_by_confidence: snake_case : Optional[int] = dataset.filter(lambda __lowerCamelCase : example["probability"] > args.confidence_threshold ) if args.do_filter_by_val_performance: assert eval_result >= 0.0 and eval_result <= 1.0 snake_case : int = int(eval_result * len(__lowerCamelCase ) ) print(__lowerCamelCase ) snake_case : List[str] = dataset.sort("probability" , reverse=__lowerCamelCase ) snake_case : Tuple = dataset.select(range(__lowerCamelCase ) ) snake_case : List[Any] = dataset.remove_columns(["label", "probability"] ) snake_case : Any = dataset.rename_column("prediction" , "label" ) snake_case : str = dataset.map(lambda __lowerCamelCase : {"label": idalabel[example["label"]]} ) snake_case : List[str] = dataset.shuffle(seed=args.seed ) snake_case : int = os.path.join(__lowerCamelCase , f"""train_pseudo.{args.data_file_extension}""" ) if args.data_file_extension == "csv": dataset.to_csv(__lowerCamelCase , index=__lowerCamelCase ) else: dataset.to_json(__lowerCamelCase ) def UpperCamelCase ( __lowerCamelCase : Optional[int] , __lowerCamelCase : str , __lowerCamelCase : Optional[int] , __lowerCamelCase : Tuple , **__lowerCamelCase : List[Any] ): snake_case : int = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , level=logging.INFO , ) logger.info(accelerator.state ) # Setup logging, we only want one process per machine to log things on the # screen. accelerator.is_local_main_process is only True for one process per # machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() snake_case : Dict = STModelArguments(model_name_or_path=__lowerCamelCase ) snake_case : Tuple = STDataArguments(train_file=__lowerCamelCase , infer_file=__lowerCamelCase ) snake_case : str = STTrainingArguments(output_dir=__lowerCamelCase ) snake_case : int = argparse.Namespace() for arg_class in (model_args, data_args, training_args): for key, value in vars(__lowerCamelCase ).items(): setattr(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) for key, value in kwargs.items(): if hasattr(__lowerCamelCase , __lowerCamelCase ): setattr(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Sanity checks snake_case : List[str] = {} snake_case : Optional[int] = None # You need to provide the training data and the data to predict on assert args.train_file is not None assert args.infer_file is not None snake_case : str = args.train_file snake_case : Tuple = args.infer_file if args.evaluation_strategy != IntervalStrategy.NO.value: assert args.eval_file is not None snake_case : Tuple = args.eval_file for key in data_files: snake_case : List[Any] = data_files[key].split("." )[-1] assert extension in ["csv", "json"], f"""`{key}_file` should be a csv or a json file.""" if args.data_file_extension is None: snake_case : Union[str, Any] = extension else: assert extension == args.data_file_extension, f"""`{key}_file` should be a {args.data_file_extension} file`.""" assert ( args.eval_metric in datasets.list_metrics() ), f"""{args.eval_metric} not in the list of supported metrics {datasets.list_metrics()}.""" # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed ) logger.info("Creating the initial data directory for self-training..." ) snake_case : List[Any] = f"""{args.output_dir}/self-train_iter-{{}}""".format snake_case : Optional[int] = data_dir_format(0 ) if accelerator.is_main_process: if args.output_dir is not None: os.makedirs(args.output_dir , exist_ok=__lowerCamelCase ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) accelerator.wait_for_everyone() snake_case : Dict = None snake_case : Union[str, Any] = None snake_case : Tuple = 0 snake_case : List[Any] = False # Show the progress bar snake_case : List[Any] = tqdm(range(args.max_selftrain_iterations ) , disable=not accelerator.is_local_main_process ) # Self-train for iteration in range(0 , int(args.max_selftrain_iterations ) ): snake_case : str = data_dir_format(__lowerCamelCase ) assert os.path.exists(__lowerCamelCase ) # Stage 1: initial fine-tuning for iteration = 0 or pseudo-training for # iteration > 0 snake_case : Dict = os.path.join(__lowerCamelCase , "stage-1" ) snake_case : Optional[Any] = { "accelerator": accelerator, "model_name_or_path": args.model_name_or_path, "cache_dir": args.cache_dir, "do_train": True, "train_file": data_files["train"] if iteration == 0 else data_files["train_pseudo"], "do_eval": True if args.eval_file is not None else False, "eval_file": data_files["eval"], "do_predict": True, "infer_file": data_files["infer"], "task_name": args.task_name, "label_list": args.label_list, "output_dir": current_output_dir, "eval_metric": args.eval_metric, "evaluation_strategy": args.evaluation_strategy, "early_stopping_patience": args.early_stopping_patience, "early_stopping_threshold": args.early_stopping_threshold, "seed": args.seed, } # Add additional training arguments for key, value in kwargs.items(): if key not in arguments_dict and not hasattr(__lowerCamelCase , __lowerCamelCase ): arguments_dict.update({key: value} ) snake_case : int = os.path.join(__lowerCamelCase , "best-checkpoint" , __lowerCamelCase ) if os.path.exists(__lowerCamelCase ): logger.info( "Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 1." , __lowerCamelCase , __lowerCamelCase , ) else: logger.info("***** Running self-training: iteration: %d, stage: 1 *****" , __lowerCamelCase ) finetune(**__lowerCamelCase ) accelerator.wait_for_everyone() assert os.path.exists(__lowerCamelCase ) logger.info("Self-training job completed: iteration: %d, stage: 1." , __lowerCamelCase ) if iteration > 0 and args.finetune_on_labeled_data: # Stage 2 (optional): fine-tuning on the original labeled data snake_case : str = os.path.join(__lowerCamelCase , "best-checkpoint" ) snake_case : Dict = os.path.join(__lowerCamelCase , "stage-2" ) # Update arguments_dict snake_case : List[str] = model_path snake_case : Optional[Any] = data_files["train"] snake_case : Optional[Any] = current_output_dir snake_case : Union[str, Any] = os.path.join(__lowerCamelCase , "best-checkpoint" , __lowerCamelCase ) if os.path.exists(__lowerCamelCase ): logger.info( "Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 2." , __lowerCamelCase , __lowerCamelCase , ) else: logger.info("***** Running self-training: iteration: %d, stage: 2 *****" , __lowerCamelCase ) finetune(**__lowerCamelCase ) accelerator.wait_for_everyone() assert os.path.exists(__lowerCamelCase ) logger.info("Self-training job completed: iteration: %d, stage: 2." , __lowerCamelCase ) snake_case : int = iteration snake_case : Tuple = data_dir_format(iteration + 1 ) snake_case : Tuple = AutoConfig.from_pretrained(os.path.join(__lowerCamelCase , "best-checkpoint" ) ) snake_case : Optional[int] = config.idalabel snake_case : List[Any] = os.path.join(__lowerCamelCase , "eval_results_best-checkpoint.json" ) snake_case : Union[str, Any] = os.path.join(__lowerCamelCase , "test_results_best-checkpoint.json" ) assert os.path.exists(__lowerCamelCase ) with open(__lowerCamelCase , "r" ) as f: snake_case : Dict = float(json.load(__lowerCamelCase )[args.eval_metric] ) snake_case : Optional[int] = os.path.join(__lowerCamelCase , "infer_output_best-checkpoint.csv" ) assert os.path.exists(__lowerCamelCase ) # Loading the dataset from local csv or json files. snake_case : Optional[Any] = load_dataset(args.data_file_extension , data_files={"data": data_files["infer"]} )["data"] snake_case : Dict = load_dataset("csv" , data_files={"data": infer_output_file} )["data"] if accelerator.is_main_process: os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) shutil.copy(__lowerCamelCase , os.path.join(__lowerCamelCase , f"""eval_results_iter-{iteration}.json""" ) ) if os.path.exists(__lowerCamelCase ): shutil.copy(__lowerCamelCase , os.path.join(__lowerCamelCase , f"""test_results_iter-{iteration}.json""" ) ) create_pseudo_labeled_data(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) accelerator.wait_for_everyone() snake_case : str = os.path.join(__lowerCamelCase , f"""train_pseudo.{args.data_file_extension}""" ) if args.evaluation_strategy != IntervalStrategy.NO.value: snake_case : List[Any] = eval_result if best_iteration is None: snake_case : List[Any] = new_iteration snake_case : int = new_eval_result else: if new_eval_result - best_eval_result > args.early_stopping_threshold: snake_case : int = new_iteration snake_case : Union[str, Any] = new_eval_result snake_case : str = 0 else: if new_eval_result == best_eval_result: snake_case : Any = new_iteration snake_case : Union[str, Any] = new_eval_result early_stopping_patience_counter += 1 if early_stopping_patience_counter >= args.early_stopping_patience: snake_case : Tuple = True progress_bar.update(1 ) if should_training_stop: break if best_iteration is not None: # Save the best iteration logger.info("Best iteration: %d" , __lowerCamelCase ) logger.info("Best evaluation result: %s = %f" , args.eval_metric , __lowerCamelCase ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(__lowerCamelCase , f"""eval_results_iter-{iteration}.json""" ) , os.path.join(__lowerCamelCase , "eval_results_best-iteration.json" ) , ) else: # Assume that the last iteration is the best logger.info("Best iteration: %d" , args.max_selftrain_iterations - 1 ) logger.info("Best evaluation result: %s = %f" , args.eval_metric , __lowerCamelCase ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(__lowerCamelCase , f"""eval_results_iter-{args.max_selftrain_iterations - 1}.json""" ) , os.path.join(__lowerCamelCase , "eval_results_best-iteration.json" ) , )
59
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import platform import numpy as np import psutil import torch from accelerate import __version__ as version from accelerate.commands.config import default_config_file, load_config_from_file from ..utils import is_npu_available, is_xpu_available def lowercase__ ( snake_case_ :Union[str, Any]=None ): if subparsers is not None: __UpperCAmelCase = subparsers.add_parser('''env''' ) else: __UpperCAmelCase = argparse.ArgumentParser('''Accelerate env command''' ) parser.add_argument( '''--config_file''' , default=snake_case_ , help='''The config file to use for the default values in the launching script.''' ) if subparsers is not None: parser.set_defaults(func=snake_case_ ) return parser def lowercase__ ( snake_case_ :List[Any] ): __UpperCAmelCase = torch.__version__ __UpperCAmelCase = torch.cuda.is_available() __UpperCAmelCase = is_xpu_available() __UpperCAmelCase = is_npu_available() __UpperCAmelCase = '''Not found''' # Get the default from the config file. if args.config_file is not None or os.path.isfile(snake_case_ ): __UpperCAmelCase = load_config_from_file(args.config_file ).to_dict() __UpperCAmelCase = { '''`Accelerate` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''Numpy version''': np.__version__, '''PyTorch version (GPU?)''': F'''{pt_version} ({pt_cuda_available})''', '''PyTorch XPU available''': str(snake_case_ ), '''PyTorch NPU available''': str(snake_case_ ), '''System RAM''': F'''{psutil.virtual_memory().total / 1_024 ** 3:.2f} GB''', } if pt_cuda_available: __UpperCAmelCase = torch.cuda.get_device_name() print('''\nCopy-and-paste the text below in your GitHub issue\n''' ) print('''\n'''.join([F'''- {prop}: {val}''' for prop, val in info.items()] ) ) print('''- `Accelerate` default config:''' if args.config_file is None else '''- `Accelerate` config passed:''' ) __UpperCAmelCase = ( '''\n'''.join([F'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(snake_case_ , snake_case_ ) else F'''\t{accelerate_config}''' ) print(snake_case_ ) __UpperCAmelCase = accelerate_config return info def lowercase__ ( ): __UpperCAmelCase = env_command_parser() __UpperCAmelCase = parser.parse_args() env_command(snake_case_ ) return 0 if __name__ == "__main__": raise SystemExit(main())
332
0
'''simple docstring''' # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ : List[str] = { '''configuration_xmod''': [ '''XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XmodConfig''', '''XmodOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ : List[str] = [ '''XMOD_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XmodForCausalLM''', '''XmodForMaskedLM''', '''XmodForMultipleChoice''', '''XmodForQuestionAnswering''', '''XmodForSequenceClassification''', '''XmodForTokenClassification''', '''XmodModel''', '''XmodPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys lowercase__ : Dict = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
190
'''simple docstring''' import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel lowercase__ : Optional[int] = HfApi() lowercase__ : Dict = {} # fmt: off lowercase__ : List[str] = torch.tensor([ -0.7515, -1.6883, 0.2420, 0.0300, 0.6347, 1.3433, -1.1743, -3.7467, 1.2342, -2.2485, 0.4636, 0.8076, -0.7991, 0.3969, 0.8498, 0.9189, -1.8887, -3.3522, 0.7639, 0.2040, 0.6271, -2.7148, -1.6316, 3.0839, 0.3186, 0.2721, -0.9759, -1.2461, 2.6257, 1.3557 ]) lowercase__ : Tuple = torch.tensor([ -2.3639, -2.5344, 0.0054, -0.6674, 1.5990, 1.0158, 0.3124, -2.1436, 1.8795, -2.5429, -0.1566, -0.3973, 1.2490, 2.6447, 1.2283, -0.5208, -2.8154, -3.5119, 2.3838, 1.2033, 1.7201, -2.1256, -1.4576, 2.7948, 2.4204, -0.9752, -1.2546, 0.8027, 3.2758, 3.1365 ]) lowercase__ : Optional[Any] = torch.tensor([ -0.6531, -0.6891, -0.3172, -0.5375, -0.9140, -0.5367, -0.1175, -0.7869, -0.3808, -0.4513, -0.2098, -0.0083, 0.3183, 0.5140, 0.2247, -0.1304, -0.1302, -0.2802, -0.2084, -0.2025, -0.4967, -0.4873, -0.0861, 0.6925, 0.0250, 0.1290, -0.1543, 0.6316, 1.0460, 1.4943 ]) lowercase__ : List[Any] = torch.tensor([ 0.0911, 0.1107, 0.0182, 0.0435, -0.0805, -0.0608, 0.0381, 0.2172, -0.0280, 0.1327, -0.0299, -0.0255, -0.0050, -0.1170, -0.1046, 0.0309, 0.1367, 0.1728, -0.0533, -0.0748, -0.0534, 0.1624, 0.0384, -0.1805, -0.0707, 0.0642, 0.0220, -0.0134, -0.1333, -0.1505 ]) lowercase__ : Dict = torch.tensor([ 0.1321, 0.1337, 0.0440, 0.0622, -0.0591, -0.0370, 0.0503, 0.2133, -0.0177, 0.1415, -0.0116, -0.0112, 0.0044, -0.0980, -0.0789, 0.0395, 0.1502, 0.1785, -0.0488, -0.0514, -0.0404, 0.1539, 0.0454, -0.1559, -0.0665, 0.0659, 0.0383, -0.0005, -0.1266, -0.1386 ]) lowercase__ : Optional[int] = torch.tensor([ 0.1154, 0.1218, 0.0307, 0.0526, -0.0711, -0.0541, 0.0366, 0.2078, -0.0267, 0.1317, -0.0226, -0.0193, -0.0014, -0.1055, -0.0902, 0.0330, 0.1391, 0.1709, -0.0562, -0.0693, -0.0560, 0.1482, 0.0381, -0.1683, -0.0681, 0.0661, 0.0331, -0.0046, -0.1268, -0.1431 ]) lowercase__ : List[Any] = torch.tensor([ 0.1192, 0.1240, 0.0414, 0.0606, -0.0557, -0.0412, 0.0430, 0.2042, -0.0200, 0.1385, -0.0115, -0.0132, 0.0017, -0.0965, -0.0802, 0.0398, 0.1433, 0.1747, -0.0458, -0.0533, -0.0407, 0.1545, 0.0419, -0.1574, -0.0645, 0.0626, 0.0341, -0.0010, -0.1199, -0.1390 ]) lowercase__ : List[str] = torch.tensor([ 0.1075, 0.1074, 0.0205, 0.0431, -0.0774, -0.0607, 0.0298, 0.2042, -0.0320, 0.1267, -0.0281, -0.0250, -0.0064, -0.1091, -0.0946, 0.0290, 0.1328, 0.1650, -0.0580, -0.0738, -0.0586, 0.1440, 0.0337, -0.1746, -0.0712, 0.0605, 0.0250, -0.0099, -0.1316, -0.1473 ]) lowercase__ : Dict = torch.tensor([ -1.4572, -2.0481, -0.0414, -0.6005, 1.4136, 0.5848, 0.4028, -2.7330, 1.2212, -2.1228, 0.2155, 0.4039, 0.7662, 2.0535, 0.7477, -0.3243, -2.1758, -2.7648, 1.6947, 0.7026, 1.2338, -1.6078, -0.8682, 2.2810, 1.8574, -0.5718, -0.5586, -0.0186, 2.3415, 2.1251]) lowercase__ : Optional[int] = torch.tensor([ -1.3690, -1.9720, -0.4090, -0.6966, 1.4660, 0.9938, -0.1385, -2.7324, 0.7736, -1.8917, 0.2923, 0.4293, 0.1693, 1.4112, 1.1887, -0.3181, -2.2160, -2.6381, 1.3170, 0.8163, 0.9240, -1.6544, -0.6099, 2.5259, 1.6430, -0.9090, -0.9392, -0.0126, 2.4268, 2.3266 ]) lowercase__ : List[str] = torch.tensor([ -1.3525, -1.9628, -0.3956, -0.6860, 1.4664, 1.0014, -0.1259, -2.7212, 0.7772, -1.8811, 0.2996, 0.4388, 0.1704, 1.4029, 1.1701, -0.3027, -2.2053, -2.6287, 1.3350, 0.8131, 0.9274, -1.6292, -0.6098, 2.5131, 1.6505, -0.8958, -0.9298, -0.0151, 2.4257, 2.3355 ]) lowercase__ : Optional[int] = torch.tensor([ -2.0585, -2.7897, -0.2850, -0.8940, 1.9052, 0.5702, 0.6345, -3.8959, 1.5932, -3.2319, 0.1974, 0.0287, 1.7566, 2.6543, 0.8387, -0.5351, -3.2736, -4.3375, 2.9029, 1.6390, 1.4640, -2.1701, -1.9013, 2.9341, 3.4981, -0.6255, -1.1644, -0.1591, 3.7097, 3.2066 ]) lowercase__ : int = torch.tensor([ -2.3139, -2.5594, -0.0197, -0.6785, 1.7001, 1.1606, 0.3075, -2.1740, 1.8071, -2.5630, -0.0926, -0.3811, 1.2116, 2.6246, 1.2731, -0.5398, -2.8153, -3.6140, 2.3893, 1.3262, 1.6258, -2.1856, -1.3267, 2.8395, 2.3779, -1.0623, -1.2468, 0.8959, 3.3367, 3.2243 ]) lowercase__ : int = torch.tensor([ -2.0628, -2.7667, -0.2089, -0.8263, 2.0539, 0.5992, 0.6495, -3.8336, 1.6025, -3.2817, 0.1721, -0.0633, 1.7516, 2.7039, 0.8100, -0.5908, -3.2113, -4.4343, 2.9257, 1.3632, 1.5562, -2.1489, -1.9894, 3.0560, 3.3396, -0.7328, -1.0417, 0.0383, 3.7093, 3.2343 ]) lowercase__ : List[Any] = torch.tensor([ -1.4574, -2.0569, -0.0473, -0.6117, 1.4018, 0.5769, 0.4129, -2.7344, 1.2241, -2.1397, 0.2000, 0.3937, 0.7616, 2.0453, 0.7324, -0.3391, -2.1746, -2.7744, 1.6963, 0.6921, 1.2187, -1.6172, -0.8877, 2.2439, 1.8471, -0.5839, -0.5605, -0.0464, 2.3250, 2.1219 ]) # fmt: on lowercase__ : str = api.list_models(filter='''diffusers''') for mod in models: if "google" in mod.author or mod.modelId == "CompVis/ldm-celebahq-256": lowercase__ : int = '''/home/patrick/google_checkpoints/''' + mod.modelId.split('''/''')[-1] print(f"""Started running {mod.modelId}!!!""") if mod.modelId.startswith('''CompVis'''): lowercase__ : Optional[Any] = UNetaDModel.from_pretrained(local_checkpoint, subfolder='''unet''') else: lowercase__ : Tuple = UNetaDModel.from_pretrained(local_checkpoint) torch.manual_seed(0) random.seed(0) lowercase__ : List[str] = torch.randn(1, model.config.in_channels, model.config.sample_size, model.config.sample_size) lowercase__ : int = torch.tensor([10] * noise.shape[0]) with torch.no_grad(): lowercase__ : Tuple = model(noise, time_step).sample assert torch.allclose( logits[0, 0, 0, :30], results['''_'''.join('''_'''.join(mod.modelId.split('''/''')).split('''-'''))], atol=1e-3 ) print(f"""{mod.modelId} has passed successfully!!!""")
190
1
def __lowercase ( _SCREAMING_SNAKE_CASE = 1_00 ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = 0 for i in range(1 , n + 1 ): sum_of_squares += i**2 sum_of_ints += i return sum_of_ints**2 - sum_of_squares if __name__ == "__main__": print(F'''{solution() = }''')
296
import dataclasses import json import warnings from dataclasses import dataclass, field from time import time from typing import List from ..utils import logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) def __lowercase ( _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None ) -> List[str]: '''simple docstring''' return field(default_factory=lambda: default , metadata=_SCREAMING_SNAKE_CASE ) @dataclass class UpperCamelCase__ : '''simple docstring''' __snake_case : List[str] = list_field( default=[] , metadata={ "help": ( "Model checkpoints to be provided to the AutoModel classes. Leave blank to benchmark the base version" " of all available models" ) } , ) __snake_case : List[int] = list_field( default=[8] , metadata={"help": "List of batch sizes for which memory and time performance will be evaluated"} ) __snake_case : List[int] = list_field( default=[8, 32, 128, 512] , metadata={"help": "List of sequence lengths for which memory and time performance will be evaluated"} , ) __snake_case : bool = field( default=lowerCAmelCase_ , metadata={"help": "Whether to benchmark inference of model. Inference can be disabled via --no-inference."} , ) __snake_case : bool = field( default=lowerCAmelCase_ , metadata={"help": "Whether to run on available cuda devices. Cuda can be disabled via --no-cuda."} , ) __snake_case : bool = field( default=lowerCAmelCase_ , metadata={"help": "Whether to run on available tpu devices. TPU can be disabled via --no-tpu."} ) __snake_case : bool = field(default=lowerCAmelCase_ , metadata={"help": "Use FP16 to accelerate inference."} ) __snake_case : bool = field(default=lowerCAmelCase_ , metadata={"help": "Benchmark training of model"} ) __snake_case : bool = field(default=lowerCAmelCase_ , metadata={"help": "Verbose memory tracing"} ) __snake_case : bool = field( default=lowerCAmelCase_ , metadata={"help": "Whether to perform speed measurements. Speed measurements can be disabled via --no-speed."} , ) __snake_case : bool = field( default=lowerCAmelCase_ , metadata={ "help": "Whether to perform memory measurements. Memory measurements can be disabled via --no-memory" } , ) __snake_case : bool = field(default=lowerCAmelCase_ , metadata={"help": "Trace memory line by line"} ) __snake_case : bool = field(default=lowerCAmelCase_ , metadata={"help": "Save result to a CSV file"} ) __snake_case : bool = field(default=lowerCAmelCase_ , metadata={"help": "Save all print statements in a log file"} ) __snake_case : bool = field(default=lowerCAmelCase_ , metadata={"help": "Whether to print environment information"} ) __snake_case : bool = field( default=lowerCAmelCase_ , metadata={ "help": ( "Whether to use multiprocessing for memory and speed measurement. It is highly recommended to use" " multiprocessing for accurate CPU and GPU memory measurements. This option should only be disabled" " for debugging / testing and on TPU." ) } , ) __snake_case : str = field( default=F"inference_time_{round(time() )}.csv" , metadata={"help": "CSV filename used if saving time results to csv."} , ) __snake_case : str = field( default=F"inference_memory_{round(time() )}.csv" , metadata={"help": "CSV filename used if saving memory results to csv."} , ) __snake_case : str = field( default=F"train_time_{round(time() )}.csv" , metadata={"help": "CSV filename used if saving time results to csv for training."} , ) __snake_case : str = field( default=F"train_memory_{round(time() )}.csv" , metadata={"help": "CSV filename used if saving memory results to csv for training."} , ) __snake_case : str = field( default=F"env_info_{round(time() )}.csv" , metadata={"help": "CSV filename used if saving environment information."} , ) __snake_case : str = field( default=F"log_{round(time() )}.csv" , metadata={"help": "Log filename used if print statements are saved in log."} , ) __snake_case : int = field(default=3 , metadata={"help": "Times an experiment will be run."} ) __snake_case : bool = field( default=lowerCAmelCase_ , metadata={ "help": ( "Instead of loading the model as defined in `config.architectures` if exists, just load the pretrain" " model weights." ) } , ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' warnings.warn( F"""The class {self.__class__} is deprecated. Hugging Face Benchmarking utils""" """ are deprecated in general and it is advised to use external Benchmarking libraries """ """ to benchmark Transformer models.""" ,lowerCamelCase__ ,) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ) -> Any: '''simple docstring''' return json.dumps(dataclasses.asdict(self ) ,indent=2 ) @property def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> List[str]: '''simple docstring''' if len(self.models ) <= 0: raise ValueError( """Please make sure you provide at least one model name / model identifier, *e.g.* `--models""" """ bert-base-cased` or `args.models = ['bert-base-cased'].""" ) return self.models @property def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ) -> int: '''simple docstring''' if not self.multi_process: return False elif self.is_tpu: logger.info("""Multiprocessing is currently not possible on TPU.""" ) return False else: return True
296
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _A = { """configuration_chinese_clip""": [ """CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ChineseCLIPConfig""", """ChineseCLIPOnnxConfig""", """ChineseCLIPTextConfig""", """ChineseCLIPVisionConfig""", ], """processing_chinese_clip""": ["""ChineseCLIPProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = ["""ChineseCLIPFeatureExtractor"""] _A = ["""ChineseCLIPImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ """CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """ChineseCLIPModel""", """ChineseCLIPPreTrainedModel""", """ChineseCLIPTextModel""", """ChineseCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys _A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
212
"""simple docstring""" import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class _lowerCamelCase ( a_ ): # to overwrite at feature extractactor specific tests _lowerCamelCase :Optional[int] = None _lowerCamelCase :List[Any] = None @property def _lowerCAmelCase ( self : int ) -> Tuple: """simple docstring""" return self.feat_extract_tester.prepare_feat_extract_dict() def _lowerCAmelCase ( self : Tuple ) -> List[Any]: """simple docstring""" lowerCAmelCase__ : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(UpperCamelCase , """feature_size""" ) ) self.assertTrue(hasattr(UpperCamelCase , """sampling_rate""" ) ) self.assertTrue(hasattr(UpperCamelCase , """padding_value""" ) ) def _lowerCAmelCase ( self : Any ) -> int: """simple docstring""" lowerCAmelCase__ : Dict = self.feat_extract_tester.prepare_inputs_for_common() lowerCAmelCase__ : List[Any] = self.feature_extraction_class(**self.feat_extract_dict ) lowerCAmelCase__ : Union[str, Any] = feat_extract.model_input_names[0] lowerCAmelCase__ : Any = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(UpperCamelCase ) == len(UpperCamelCase ) for x, y in zip(UpperCamelCase , processed_features[input_name] ) ) ) lowerCAmelCase__ : List[str] = self.feat_extract_tester.prepare_inputs_for_common(equal_length=UpperCamelCase ) lowerCAmelCase__ : List[Any] = BatchFeature({input_name: speech_inputs} , tensor_type="""np""" ) lowerCAmelCase__ : Dict = processed_features[input_name] if len(batch_features_input.shape ) < 3: lowerCAmelCase__ : int = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def _lowerCAmelCase ( self : Tuple ) -> Optional[Any]: """simple docstring""" lowerCAmelCase__ : int = self.feat_extract_tester.prepare_inputs_for_common(equal_length=UpperCamelCase ) lowerCAmelCase__ : str = self.feature_extraction_class(**self.feat_extract_dict ) lowerCAmelCase__ : Optional[int] = feat_extract.model_input_names[0] lowerCAmelCase__ : Union[str, Any] = BatchFeature({input_name: speech_inputs} , tensor_type="""pt""" ) lowerCAmelCase__ : str = processed_features[input_name] if len(batch_features_input.shape ) < 3: lowerCAmelCase__ : int = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def _lowerCAmelCase ( self : Optional[Any] ) -> int: """simple docstring""" lowerCAmelCase__ : List[str] = self.feat_extract_tester.prepare_inputs_for_common(equal_length=UpperCamelCase ) lowerCAmelCase__ : str = self.feature_extraction_class(**self.feat_extract_dict ) lowerCAmelCase__ : Dict = feat_extract.model_input_names[0] lowerCAmelCase__ : Tuple = BatchFeature({input_name: speech_inputs} , tensor_type="""tf""" ) lowerCAmelCase__ : Union[str, Any] = processed_features[input_name] if len(batch_features_input.shape ) < 3: lowerCAmelCase__ : Optional[Any] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def _lowerCAmelCase ( self : List[Any] , UpperCamelCase : Union[str, Any]=False ) -> int: """simple docstring""" def _inputs_have_equal_length(UpperCamelCase : int ): lowerCAmelCase__ : Optional[Any] = len(input[0] ) for input_slice in input[1:]: if len(UpperCamelCase ) != length: return False return True def _inputs_are_equal(UpperCamelCase : int , UpperCamelCase : Dict ): if len(UpperCamelCase ) != len(UpperCamelCase ): return False for input_slice_a, input_slice_a in zip(UpperCamelCase , UpperCamelCase ): if not np.allclose(np.asarray(UpperCamelCase ) , np.asarray(UpperCamelCase ) , atol=1E-3 ): return False return True lowerCAmelCase__ : str = self.feature_extraction_class(**self.feat_extract_dict ) lowerCAmelCase__ : Dict = self.feat_extract_tester.prepare_inputs_for_common(numpify=UpperCamelCase ) lowerCAmelCase__ : str = feat_extract.model_input_names[0] lowerCAmelCase__ : str = BatchFeature({input_name: speech_inputs} ) lowerCAmelCase__ : Any = self.feat_extract_tester.seq_length_diff lowerCAmelCase__ : List[str] = self.feat_extract_tester.max_seq_length + pad_diff lowerCAmelCase__ : Optional[int] = self.feat_extract_tester.min_seq_length lowerCAmelCase__ : Optional[int] = self.feat_extract_tester.batch_size lowerCAmelCase__ : str = self.feat_extract_tester.feature_size # test padding for List[int] + numpy lowerCAmelCase__ : List[Any] = feat_extract.pad(UpperCamelCase , padding=UpperCamelCase ) lowerCAmelCase__ : int = input_a[input_name] lowerCAmelCase__ : Tuple = feat_extract.pad(UpperCamelCase , padding="""longest""" ) lowerCAmelCase__ : Any = input_a[input_name] lowerCAmelCase__ : List[str] = feat_extract.pad(UpperCamelCase , padding="""max_length""" , max_length=len(speech_inputs[-1] ) ) lowerCAmelCase__ : Dict = input_a[input_name] lowerCAmelCase__ : str = feat_extract.pad(UpperCamelCase , padding="""longest""" , return_tensors="""np""" ) lowerCAmelCase__ : Optional[Any] = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(UpperCamelCase ): feat_extract.pad(UpperCamelCase , padding="""max_length""" )[input_name] lowerCAmelCase__ : Optional[int] = feat_extract.pad( UpperCamelCase , padding="""max_length""" , max_length=UpperCamelCase , return_tensors="""np""" ) lowerCAmelCase__ : int = input_a[input_name] self.assertFalse(_inputs_have_equal_length(UpperCamelCase ) ) self.assertTrue(_inputs_have_equal_length(UpperCamelCase ) ) self.assertTrue(_inputs_have_equal_length(UpperCamelCase ) ) self.assertTrue(_inputs_are_equal(UpperCamelCase , UpperCamelCase ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy lowerCAmelCase__ : Any = feat_extract.pad(UpperCamelCase , pad_to_multiple_of=10 ) lowerCAmelCase__ : int = input_a[input_name] lowerCAmelCase__ : Dict = feat_extract.pad(UpperCamelCase , padding="""longest""" , pad_to_multiple_of=10 ) lowerCAmelCase__ : Union[str, Any] = input_a[input_name] lowerCAmelCase__ : Optional[int] = feat_extract.pad( UpperCamelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=UpperCamelCase ) lowerCAmelCase__ : str = input_a[input_name] lowerCAmelCase__ : Union[str, Any] = feat_extract.pad( UpperCamelCase , padding="""max_length""" , pad_to_multiple_of=10 , max_length=UpperCamelCase , return_tensors="""np""" , ) lowerCAmelCase__ : List[Any] = input_a[input_name] self.assertTrue(all(len(UpperCamelCase ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(UpperCamelCase , UpperCamelCase ) ) lowerCAmelCase__ : Optional[int] = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(UpperCamelCase ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] , (batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct lowerCAmelCase__ : str = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1E-3 ) def _lowerCAmelCase ( self : Optional[Any] , UpperCamelCase : int=False ) -> List[str]: """simple docstring""" def _inputs_have_equal_length(UpperCamelCase : List[Any] ): lowerCAmelCase__ : List[str] = len(input[0] ) for input_slice in input[1:]: if len(UpperCamelCase ) != length: return False return True def _inputs_are_equal(UpperCamelCase : Union[str, Any] , UpperCamelCase : str ): if len(UpperCamelCase ) != len(UpperCamelCase ): return False for input_slice_a, input_slice_a in zip(UpperCamelCase , UpperCamelCase ): if not np.allclose(np.asarray(UpperCamelCase ) , np.asarray(UpperCamelCase ) , atol=1E-3 ): return False return True lowerCAmelCase__ : Dict = self.feature_extraction_class(**self.feat_extract_dict ) lowerCAmelCase__ : Any = self.feat_extract_tester.prepare_inputs_for_common(numpify=UpperCamelCase ) lowerCAmelCase__ : str = feat_extract.model_input_names[0] lowerCAmelCase__ : Optional[Any] = BatchFeature({input_name: speech_inputs} ) # truncate to smallest lowerCAmelCase__ : str = feat_extract.pad( UpperCamelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , truncation=UpperCamelCase ) lowerCAmelCase__ : Dict = input_a[input_name] lowerCAmelCase__ : str = feat_extract.pad(UpperCamelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) ) lowerCAmelCase__ : Optional[int] = input_a[input_name] self.assertTrue(_inputs_have_equal_length(UpperCamelCase ) ) self.assertFalse(_inputs_have_equal_length(UpperCamelCase ) ) # truncate to smallest with np lowerCAmelCase__ : Tuple = feat_extract.pad( UpperCamelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" , truncation=UpperCamelCase , ) lowerCAmelCase__ : List[Any] = input_a[input_name] lowerCAmelCase__ : List[Any] = feat_extract.pad( UpperCamelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , return_tensors="""np""" ) lowerCAmelCase__ : Union[str, Any] = input_a[input_name] self.assertTrue(_inputs_have_equal_length(UpperCamelCase ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(UpperCamelCase ) ) # truncate to middle lowerCAmelCase__ : List[Any] = feat_extract.pad( UpperCamelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=UpperCamelCase , return_tensors="""np""" , ) lowerCAmelCase__ : List[Any] = input_a[input_name] lowerCAmelCase__ : List[Any] = feat_extract.pad( UpperCamelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , truncation=UpperCamelCase ) lowerCAmelCase__ : List[str] = input_a[input_name] lowerCAmelCase__ : Any = feat_extract.pad( UpperCamelCase , padding="""max_length""" , max_length=len(speech_inputs[1] ) , return_tensors="""np""" ) lowerCAmelCase__ : Dict = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(UpperCamelCase ) ) self.assertTrue(_inputs_have_equal_length(UpperCamelCase ) ) self.assertTrue(_inputs_are_equal(UpperCamelCase , UpperCamelCase ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(UpperCamelCase ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(UpperCamelCase ): feat_extract.pad(UpperCamelCase , truncation=UpperCamelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(UpperCamelCase ): feat_extract.pad(UpperCamelCase , padding="""longest""" , truncation=UpperCamelCase )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(UpperCamelCase ): feat_extract.pad(UpperCamelCase , padding="""longest""" , truncation=UpperCamelCase )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(UpperCamelCase ): feat_extract.pad(UpperCamelCase , padding="""max_length""" , truncation=UpperCamelCase )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy lowerCAmelCase__ : Any = 12 lowerCAmelCase__ : Optional[int] = feat_extract.pad( UpperCamelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=UpperCamelCase , truncation=UpperCamelCase , ) lowerCAmelCase__ : List[Any] = input_a[input_name] lowerCAmelCase__ : Any = feat_extract.pad( UpperCamelCase , padding="""max_length""" , max_length=len(speech_inputs[0] ) , pad_to_multiple_of=UpperCamelCase , ) lowerCAmelCase__ : Optional[Any] = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of lowerCAmelCase__ : Union[str, Any] = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: lowerCAmelCase__ : Any = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(UpperCamelCase ) ) self.assertFalse(_inputs_have_equal_length(UpperCamelCase ) ) def _lowerCAmelCase ( self : List[str] ) -> Optional[int]: """simple docstring""" self._check_padding(numpify=UpperCamelCase ) def _lowerCAmelCase ( self : List[Any] ) -> List[Any]: """simple docstring""" self._check_padding(numpify=UpperCamelCase ) def _lowerCAmelCase ( self : Dict ) -> Optional[int]: """simple docstring""" self._check_truncation(numpify=UpperCamelCase ) def _lowerCAmelCase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" self._check_truncation(numpify=UpperCamelCase ) @require_torch def _lowerCAmelCase ( self : Optional[int] ) -> str: """simple docstring""" lowerCAmelCase__ : int = self.feature_extraction_class(**self.feat_extract_dict ) lowerCAmelCase__ : str = self.feat_extract_tester.prepare_inputs_for_common() lowerCAmelCase__ : str = feat_extract.model_input_names[0] lowerCAmelCase__ : Optional[int] = BatchFeature({input_name: speech_inputs} ) lowerCAmelCase__ : Dict = feat_extract.pad(UpperCamelCase , padding="""longest""" , return_tensors="""np""" )[input_name] lowerCAmelCase__ : str = feat_extract.pad(UpperCamelCase , padding="""longest""" , return_tensors="""pt""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) @require_tf def _lowerCAmelCase ( self : Optional[Any] ) -> str: """simple docstring""" lowerCAmelCase__ : Tuple = self.feature_extraction_class(**self.feat_extract_dict ) lowerCAmelCase__ : Tuple = self.feat_extract_tester.prepare_inputs_for_common() lowerCAmelCase__ : List[Any] = feat_extract.model_input_names[0] lowerCAmelCase__ : int = BatchFeature({input_name: speech_inputs} ) lowerCAmelCase__ : int = feat_extract.pad(UpperCamelCase , padding="""longest""" , return_tensors="""np""" )[input_name] lowerCAmelCase__ : Any = feat_extract.pad(UpperCamelCase , padding="""longest""" , return_tensors="""tf""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def _lowerCAmelCase ( self : List[Any] ) -> int: """simple docstring""" lowerCAmelCase__ : Any = self.feat_extract_dict lowerCAmelCase__ : int = True lowerCAmelCase__ : str = self.feature_extraction_class(**UpperCamelCase ) lowerCAmelCase__ : int = self.feat_extract_tester.prepare_inputs_for_common() lowerCAmelCase__ : List[Any] = [len(UpperCamelCase ) for x in speech_inputs] lowerCAmelCase__ : Optional[Any] = feat_extract.model_input_names[0] lowerCAmelCase__ : List[str] = BatchFeature({input_name: speech_inputs} ) lowerCAmelCase__ : Tuple = feat_extract.pad(UpperCamelCase , padding="""longest""" , return_tensors="""np""" ) self.assertIn("""attention_mask""" , UpperCamelCase ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , UpperCamelCase ) def _lowerCAmelCase ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" lowerCAmelCase__ : Union[str, Any] = self.feat_extract_dict lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : str = self.feature_extraction_class(**UpperCamelCase ) lowerCAmelCase__ : int = self.feat_extract_tester.prepare_inputs_for_common() lowerCAmelCase__ : Any = [len(UpperCamelCase ) for x in speech_inputs] lowerCAmelCase__ : int = feat_extract.model_input_names[0] lowerCAmelCase__ : Optional[Any] = BatchFeature({input_name: speech_inputs} ) lowerCAmelCase__ : Any = min(UpperCamelCase ) lowerCAmelCase__ : Tuple = feat_extract.pad( UpperCamelCase , padding="""max_length""" , max_length=UpperCamelCase , truncation=UpperCamelCase , return_tensors="""np""" ) self.assertIn("""attention_mask""" , UpperCamelCase ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] )
212
1
"""simple docstring""" from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
132
"""simple docstring""" import itertools import string from collections.abc import Generator, Iterable def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> Generator[tuple[str, ...], None, None]: SCREAMING_SNAKE_CASE__ : List[Any] = iter(__lowerCAmelCase ) while True: SCREAMING_SNAKE_CASE__ : Optional[int] = tuple(itertools.islice(__lowerCAmelCase , __lowerCAmelCase ) ) if not chunk: return yield chunk def _lowercase ( __lowerCAmelCase ) -> str: SCREAMING_SNAKE_CASE__ : List[Any] = """""".join([c.upper() for c in dirty if c in string.ascii_letters] ) SCREAMING_SNAKE_CASE__ : Tuple = """""" if len(__lowerCAmelCase ) < 2: return dirty for i in range(len(__lowerCAmelCase ) - 1 ): clean += dirty[i] if dirty[i] == dirty[i + 1]: clean += "X" clean += dirty[-1] if len(__lowerCAmelCase ) & 1: clean += "X" return clean def _lowercase ( __lowerCAmelCase ) -> list[str]: # I and J are used interchangeably to allow # us to use a 5x5 table (25 letters) SCREAMING_SNAKE_CASE__ : str = """ABCDEFGHIKLMNOPQRSTUVWXYZ""" # we're using a list instead of a '2d' array because it makes the math # for setting up the table and doing the actual encoding/decoding simpler SCREAMING_SNAKE_CASE__ : Optional[int] = [] # copy key chars into the table if they are in `alphabet` ignoring duplicates for char in key.upper(): if char not in table and char in alphabet: table.append(__lowerCAmelCase ) # fill the rest of the table in with the remaining alphabet chars for char in alphabet: if char not in table: table.append(__lowerCAmelCase ) return table def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> str: SCREAMING_SNAKE_CASE__ : Tuple = generate_table(__lowerCAmelCase ) SCREAMING_SNAKE_CASE__ : List[Any] = prepare_input(__lowerCAmelCase ) SCREAMING_SNAKE_CASE__ : Tuple = """""" # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(__lowerCAmelCase , 2 ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[Any] = divmod(table.index(__lowerCAmelCase ) , 5 ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[str] = divmod(table.index(__lowerCAmelCase ) , 5 ) if rowa == rowa: ciphertext += table[rowa * 5 + (cola + 1) % 5] ciphertext += table[rowa * 5 + (cola + 1) % 5] elif cola == cola: ciphertext += table[((rowa + 1) % 5) * 5 + cola] ciphertext += table[((rowa + 1) % 5) * 5 + cola] else: # rectangle ciphertext += table[rowa * 5 + cola] ciphertext += table[rowa * 5 + cola] return ciphertext def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> str: SCREAMING_SNAKE_CASE__ : str = generate_table(__lowerCAmelCase ) SCREAMING_SNAKE_CASE__ : Dict = """""" # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(__lowerCAmelCase , 2 ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Dict = divmod(table.index(__lowerCAmelCase ) , 5 ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Union[str, Any] = divmod(table.index(__lowerCAmelCase ) , 5 ) if rowa == rowa: plaintext += table[rowa * 5 + (cola - 1) % 5] plaintext += table[rowa * 5 + (cola - 1) % 5] elif cola == cola: plaintext += table[((rowa - 1) % 5) * 5 + cola] plaintext += table[((rowa - 1) % 5) * 5 + cola] else: # rectangle plaintext += table[rowa * 5 + cola] plaintext += table[rowa * 5 + cola] return plaintext
132
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase__ : str =logging.get_logger(__name__) def _lowercase ( _UpperCAmelCase , _UpperCAmelCase=False ) -> str: lowerCamelCase =[] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """vit.embeddings.cls_token"""), ("""patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """vit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" lowerCamelCase =[(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def _lowercase ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=False ) -> Optional[Any]: for i in range(config.num_hidden_layers ): if base_model: lowerCamelCase ="""""" else: lowerCamelCase ="""vit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCamelCase =state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) lowerCamelCase =state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase =in_proj_weight[ : config.hidden_size, : ] lowerCamelCase =in_proj_bias[: config.hidden_size] lowerCamelCase =in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCamelCase =in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCamelCase =in_proj_weight[ -config.hidden_size :, : ] lowerCamelCase =in_proj_bias[-config.hidden_size :] def _lowercase ( _UpperCAmelCase ) -> Optional[Any]: lowerCamelCase =["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) def _lowercase ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[Any]: lowerCamelCase =dct.pop(_UpperCAmelCase ) lowerCamelCase =val def _lowercase ( ) -> Optional[Any]: lowerCamelCase ="""http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCamelCase =Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return im @torch.no_grad() def _lowercase ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=True ) -> List[str]: lowerCamelCase =ViTConfig() # patch_size if model_name[-1] == "8": lowerCamelCase =8 # set labels if required if not base_model: lowerCamelCase =10_00 lowerCamelCase ="""huggingface/label-files""" lowerCamelCase ="""imagenet-1k-id2label.json""" lowerCamelCase =json.load(open(hf_hub_download(_UpperCAmelCase , _UpperCAmelCase , repo_type="""dataset""" ) , """r""" ) ) lowerCamelCase ={int(_UpperCAmelCase ): v for k, v in idalabel.items()} lowerCamelCase =idalabel lowerCamelCase ={v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: lowerCamelCase =3_84 lowerCamelCase =15_36 lowerCamelCase =12 lowerCamelCase =6 # load original model from torch hub lowerCamelCase =torch.hub.load("""facebookresearch/dino:main""" , _UpperCAmelCase ) original_model.eval() # load state_dict of original model, remove and rename some keys lowerCamelCase =original_model.state_dict() if base_model: remove_classification_head_(_UpperCAmelCase ) lowerCamelCase =create_rename_keys(_UpperCAmelCase , base_model=_UpperCAmelCase ) for src, dest in rename_keys: rename_key(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) read_in_q_k_v(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # load HuggingFace model if base_model: lowerCamelCase =ViTModel(_UpperCAmelCase , add_pooling_layer=_UpperCAmelCase ).eval() else: lowerCamelCase =ViTForImageClassification(_UpperCAmelCase ).eval() model.load_state_dict(_UpperCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor lowerCamelCase =ViTImageProcessor() lowerCamelCase =image_processor(images=prepare_img() , return_tensors="""pt""" ) lowerCamelCase =encoding["""pixel_values"""] lowerCamelCase =model(_UpperCAmelCase ) if base_model: lowerCamelCase =original_model(_UpperCAmelCase ) assert torch.allclose(_UpperCAmelCase , outputs.last_hidden_state[:, 0, :] , atol=1e-1 ) else: lowerCamelCase =original_model(_UpperCAmelCase ) assert logits.shape == outputs.logits.shape assert torch.allclose(_UpperCAmelCase , outputs.logits , atol=1e-3 ) Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCAmelCase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": UpperCAmelCase__ : Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''dino_vitb16''', type=str, help='''Name of the model trained with DINO you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--base_model''', action='''store_true''', help='''Whether to only convert the base model (no projection head weights).''', ) parser.set_defaults(base_model=True) UpperCAmelCase__ : Tuple =parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
361
from .glue import glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels from .squad import SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features from .utils import DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor from .xnli import xnli_output_modes, xnli_processors, xnli_tasks_num_labels
262
0
import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class __snake_case : @staticmethod def __a ( *_lowercase : Union[str, Any] , **_lowercase : Union[str, Any] ): """simple docstring""" pass def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Image ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class __snake_case ( unittest.TestCase ): lowerCAmelCase_ = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def __a ( self : Tuple , _lowercase : Union[str, Any] , _lowercase : Any , _lowercase : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = DepthEstimationPipeline(model=_lowercase , image_processor=_lowercase ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def __a ( self : int , _lowercase : List[str] , _lowercase : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = depth_estimator("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) self.assertEqual({"""predicted_depth""": ANY(torch.Tensor ), """depth""": ANY(Image.Image )} , _lowercase ) import datasets SCREAMING_SNAKE_CASE__ = datasets.load_dataset("""hf-internal-testing/fixtures_image_utils""" , """image""" , split="""test""" ) SCREAMING_SNAKE_CASE__ = depth_estimator( [ Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ), """http://images.cocodataset.org/val2017/000000039769.jpg""", # RGBA dataset[0]["""file"""], # LA dataset[1]["""file"""], # L dataset[2]["""file"""], ] ) self.assertEqual( [ {"""predicted_depth""": ANY(torch.Tensor ), """depth""": ANY(Image.Image )}, {"""predicted_depth""": ANY(torch.Tensor ), """depth""": ANY(Image.Image )}, {"""predicted_depth""": ANY(torch.Tensor ), """depth""": ANY(Image.Image )}, {"""predicted_depth""": ANY(torch.Tensor ), """depth""": ANY(Image.Image )}, {"""predicted_depth""": ANY(torch.Tensor ), """depth""": ANY(Image.Image )}, ] , _lowercase , ) @require_tf @unittest.skip("""Depth estimation is not implemented in TF""" ) def __a ( self : Union[str, Any] ): """simple docstring""" pass @slow @require_torch def __a ( self : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = """Intel/dpt-large""" SCREAMING_SNAKE_CASE__ = pipeline("""depth-estimation""" , model=_lowercase ) SCREAMING_SNAKE_CASE__ = depth_estimator("""http://images.cocodataset.org/val2017/000000039769.jpg""" ) SCREAMING_SNAKE_CASE__ = hashimage(outputs["""depth"""] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs["""predicted_depth"""].max().item() ) , 29.3_04 ) self.assertEqual(nested_simplify(outputs["""predicted_depth"""].min().item() ) , 2.6_62 ) @require_torch def __a ( self : Any ): """simple docstring""" self.skipTest("""There is not hf-internal-testing tiny model for either GLPN nor DPT""" )
219
import doctest from collections import deque import numpy as np class __snake_case : def __init__( self : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE__ = [2, 1, 2, -1] SCREAMING_SNAKE_CASE__ = [1, 2, 3, 4] def __a ( self : int ): """simple docstring""" SCREAMING_SNAKE_CASE__ = len(self.first_signal ) SCREAMING_SNAKE_CASE__ = len(self.second_signal ) SCREAMING_SNAKE_CASE__ = max(_lowercase , _lowercase ) # create a zero matrix of max_length x max_length SCREAMING_SNAKE_CASE__ = [[0] * max_length for i in range(_lowercase )] # fills the smaller signal with zeros to make both signals of same length if length_first_signal < length_second_signal: self.first_signal += [0] * (max_length - length_first_signal) elif length_first_signal > length_second_signal: self.second_signal += [0] * (max_length - length_second_signal) for i in range(_lowercase ): SCREAMING_SNAKE_CASE__ = deque(self.second_signal ) rotated_signal.rotate(_lowercase ) for j, item in enumerate(_lowercase ): matrix[i][j] += item # multiply the matrix with the first signal SCREAMING_SNAKE_CASE__ = np.matmul(np.transpose(_lowercase ) , np.transpose(self.first_signal ) ) # rounding-off to two decimal places return [round(_lowercase , 2 ) for i in final_signal] if __name__ == "__main__": doctest.testmod()
219
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase : Optional[int] = logging.get_logger(__name__) def _A ( SCREAMING_SNAKE_CASE : Any , SCREAMING_SNAKE_CASE : Any=False ): """simple docstring""" a__ : int =[] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''blocks.{i}.norm1.weight''', f'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''blocks.{i}.norm1.bias''', f'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((f'''blocks.{i}.attn.proj.weight''', f'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.attn.proj.bias''', f'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''blocks.{i}.norm2.weight''', f'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''blocks.{i}.norm2.bias''', f'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.weight''', f'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.bias''', f'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.weight''', f'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.bias''', f'''vit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "vit.embeddings.cls_token"), ("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "vit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" a__ : Union[str, Any] =[(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def _A ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : List[Any]=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: a__ : Any ="" else: a__ : int ="vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) a__ : int =state_dict.pop(f'''blocks.{i}.attn.qkv.weight''' ) a__ : Tuple =state_dict.pop(f'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict a__ : int =in_proj_weight[ : config.hidden_size, : ] a__ : str =in_proj_bias[: config.hidden_size] a__ : Dict =in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] a__ : List[str] =in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] a__ : Tuple =in_proj_weight[ -config.hidden_size :, : ] a__ : Union[str, Any] =in_proj_bias[-config.hidden_size :] def _A ( SCREAMING_SNAKE_CASE : Optional[Any] ): """simple docstring""" a__ : List[Any] =["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def _A ( SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : Any ): """simple docstring""" a__ : Optional[Any] =dct.pop(SCREAMING_SNAKE_CASE ) a__ : List[Any] =val def _A ( ): """simple docstring""" a__ : List[Any] ="http://images.cocodataset.org/val2017/000000039769.jpg" a__ : Optional[int] =Image.open(requests.get(SCREAMING_SNAKE_CASE , stream=SCREAMING_SNAKE_CASE ).raw ) return im @torch.no_grad() def _A ( SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : Optional[Any] , SCREAMING_SNAKE_CASE : Union[str, Any]=True ): """simple docstring""" a__ : Optional[int] =ViTConfig() # patch_size if model_name[-1] == "8": a__ : List[str] =8 # set labels if required if not base_model: a__ : str =1_000 a__ : Union[str, Any] ="huggingface/label-files" a__ : Dict ="imagenet-1k-id2label.json" a__ : Union[str, Any] =json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , repo_type="dataset" ) , "r" ) ) a__ : Any ={int(SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} a__ : List[Any] =idalabel a__ : Dict ={v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: a__ : Optional[Any] =384 a__ : List[str] =1_536 a__ : Optional[Any] =12 a__ : Optional[int] =6 # load original model from torch hub a__ : Tuple =torch.hub.load("facebookresearch/dino:main" , SCREAMING_SNAKE_CASE ) original_model.eval() # load state_dict of original model, remove and rename some keys a__ : int =original_model.state_dict() if base_model: remove_classification_head_(SCREAMING_SNAKE_CASE ) a__ : List[Any] =create_rename_keys(SCREAMING_SNAKE_CASE , base_model=SCREAMING_SNAKE_CASE ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) read_in_q_k_v(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # load HuggingFace model if base_model: a__ : Optional[int] =ViTModel(SCREAMING_SNAKE_CASE , add_pooling_layer=SCREAMING_SNAKE_CASE ).eval() else: a__ : Union[str, Any] =ViTForImageClassification(SCREAMING_SNAKE_CASE ).eval() model.load_state_dict(SCREAMING_SNAKE_CASE ) # Check outputs on an image, prepared by ViTImageProcessor a__ : List[Any] =ViTImageProcessor() a__ : int =image_processor(images=prepare_img() , return_tensors="pt" ) a__ : Tuple =encoding["pixel_values"] a__ : Tuple =model(SCREAMING_SNAKE_CASE ) if base_model: a__ : Any =original_model(SCREAMING_SNAKE_CASE ) assert torch.allclose(SCREAMING_SNAKE_CASE , outputs.last_hidden_state[:, 0, :] , atol=1e-1 ) else: a__ : Any =original_model(SCREAMING_SNAKE_CASE ) assert logits.shape == outputs.logits.shape assert torch.allclose(SCREAMING_SNAKE_CASE , outputs.logits , atol=1e-3 ) Path(SCREAMING_SNAKE_CASE ).mkdir(exist_ok=SCREAMING_SNAKE_CASE ) print(f'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": UpperCAmelCase : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""dino_vitb16""", type=str, help="""Name of the model trained with DINO you\'d like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--base_model""", action="""store_true""", help="""Whether to only convert the base model (no projection head weights).""", ) parser.set_defaults(base_model=True) UpperCAmelCase : List[str] = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
370
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def _A ( SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : Any , SCREAMING_SNAKE_CASE : Tuple ): """simple docstring""" a__ : List[Any] ={ "en": "Machine learning is great, isn't it?", "ru": "Машинное обучение - это здорово, не так ли?", "de": "Maschinelles Lernen ist großartig, oder?", } # BLUE scores as follows: # "pair": [fairseq, transformers] a__ : List[Any] ={ "ru-en": ["[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)", "39.20"], "en-ru": ["[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)", "33.47"], "en-de": ["[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)", "42.83"], "de-en": ["[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)", "41.35"], } a__ : Optional[int] =f'''{src_lang}-{tgt_lang}''' a__ : Any =f''' --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "facebook/wmt19-{src_lang}-{tgt_lang}" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = "{texts[src_lang]}" input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR\'s WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) ''' os.makedirs(SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) a__ : Tuple =os.path.join(SCREAMING_SNAKE_CASE , "README.md" ) print(f'''Generating {path}''' ) with open(SCREAMING_SNAKE_CASE , "w" , encoding="utf-8" ) as f: f.write(SCREAMING_SNAKE_CASE ) # make sure we are under the root of the project UpperCAmelCase : str = Path(__file__).resolve().parent.parent.parent UpperCAmelCase : Dict = repo_dir / """model_cards""" for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : List[str] = model_name.split("""-""") UpperCAmelCase : Tuple = model_cards_dir / """facebook""" / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
148
0
'''simple docstring''' import shutil import tempfile import unittest from transformers import ( SPIECE_UNDERLINE, AddedToken, BatchEncoding, NllbTokenizer, NllbTokenizerFast, is_torch_available, ) from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Union[str, Any] = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right __SCREAMING_SNAKE_CASE : Optional[int] = 256_047 __SCREAMING_SNAKE_CASE : Optional[int] = 256_145 @require_sentencepiece @require_tokenizers class lowerCamelCase_ (snake_case__ , unittest.TestCase ): '''simple docstring''' __UpperCamelCase: int = NllbTokenizer __UpperCamelCase: Tuple = NllbTokenizerFast __UpperCamelCase: Union[str, Any] = True __UpperCamelCase: Dict = True __UpperCamelCase: Optional[Any] = {} def _A ( self : Union[str, Any] ): super().setUp() # We have a SentencePiece fixture for testing _UpperCAmelCase : Tuple = NllbTokenizer(A , keep_accents=A ) tokenizer.save_pretrained(self.tmpdirname ) def _A ( self : Dict ): _UpperCAmelCase : Tuple = NllbTokenizer(A , keep_accents=A ) _UpperCAmelCase : Optional[Any] = tokenizer.tokenize("This is a test" ) self.assertListEqual(A , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(A ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) _UpperCAmelCase : List[str] = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( A , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) _UpperCAmelCase : Optional[Any] = tokenizer.convert_tokens_to_ids(A ) self.assertListEqual( A , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) _UpperCAmelCase : Union[str, Any] = tokenizer.convert_ids_to_tokens(A ) self.assertListEqual( A , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def _A ( self : List[Any] ): _UpperCAmelCase : Any = (self.rust_tokenizer_class, "hf-internal-testing/tiny-random-nllb", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _UpperCAmelCase : Dict = self.rust_tokenizer_class.from_pretrained(A , **A ) _UpperCAmelCase : str = self.tokenizer_class.from_pretrained(A , **A ) _UpperCAmelCase : Optional[int] = tempfile.mkdtemp() _UpperCAmelCase : Dict = tokenizer_r.save_pretrained(A ) _UpperCAmelCase : Dict = tokenizer_p.save_pretrained(A ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) _UpperCAmelCase : Optional[int] = tuple(f for f in tokenizer_r_files if "tokenizer.json" not in f ) self.assertSequenceEqual(A , A ) # Checks everything loads correctly in the same way _UpperCAmelCase : List[Any] = tokenizer_r.from_pretrained(A ) _UpperCAmelCase : List[str] = tokenizer_p.from_pretrained(A ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A , A ) ) shutil.rmtree(A ) # Save tokenizer rust, legacy_format=True _UpperCAmelCase : Optional[Any] = tempfile.mkdtemp() _UpperCAmelCase : str = tokenizer_r.save_pretrained(A , legacy_format=A ) _UpperCAmelCase : str = tokenizer_p.save_pretrained(A ) # Checks it save with the same files self.assertSequenceEqual(A , A ) # Checks everything loads correctly in the same way _UpperCAmelCase : Optional[int] = tokenizer_r.from_pretrained(A ) _UpperCAmelCase : Dict = tokenizer_p.from_pretrained(A ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A , A ) ) shutil.rmtree(A ) # Save tokenizer rust, legacy_format=False _UpperCAmelCase : Optional[int] = tempfile.mkdtemp() _UpperCAmelCase : Optional[int] = tokenizer_r.save_pretrained(A , legacy_format=A ) _UpperCAmelCase : Dict = tokenizer_p.save_pretrained(A ) # Checks it saved the tokenizer.json file self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _UpperCAmelCase : List[Any] = tokenizer_r.from_pretrained(A ) _UpperCAmelCase : Optional[int] = tokenizer_p.from_pretrained(A ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(A , A ) ) shutil.rmtree(A ) @require_torch def _A ( self : Tuple ): if not self.test_seqaseq: return _UpperCAmelCase : Union[str, Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): # Longer text that will definitely require truncation. _UpperCAmelCase : Optional[Any] = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for" " Syria is that 'there is no military solution' to the nearly five-year conflict and more weapons" " will only worsen the violence and misery for millions of people.", ] _UpperCAmelCase : Optional[Any] = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al" " Rusiei pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi" " că noi arme nu vor face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] try: _UpperCAmelCase : Optional[int] = tokenizer.prepare_seqaseq_batch( src_texts=A , tgt_texts=A , max_length=3 , max_target_length=10 , return_tensors="pt" , src_lang="eng_Latn" , tgt_lang="ron_Latn" , ) except NotImplementedError: return self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.labels.shape[1] , 10 ) # max_target_length will default to max_length if not specified _UpperCAmelCase : Tuple = tokenizer.prepare_seqaseq_batch( A , tgt_texts=A , max_length=3 , return_tensors="pt" ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.labels.shape[1] , 3 ) _UpperCAmelCase : Union[str, Any] = tokenizer.prepare_seqaseq_batch( src_texts=A , max_length=3 , max_target_length=10 , return_tensors="pt" ) self.assertEqual(batch_encoder_only.input_ids.shape[1] , 3 ) self.assertEqual(batch_encoder_only.attention_mask.shape[1] , 3 ) self.assertNotIn("decoder_input_ids" , A ) @unittest.skip("Unfortunately way too slow to build a BPE with SentencePiece." ) def _A ( self : List[Any] ): pass def _A ( self : Union[str, Any] ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _UpperCAmelCase : Any = [AddedToken("<special>" , lstrip=A )] _UpperCAmelCase : List[Any] = self.rust_tokenizer_class.from_pretrained( A , additional_special_tokens=A , **A ) _UpperCAmelCase : Dict = tokenizer_r.encode("Hey this is a <special> token" ) _UpperCAmelCase : Any = tokenizer_r.encode("<special>" , add_special_tokens=A )[0] self.assertTrue(special_token_id in r_output ) if self.test_slow_tokenizer: _UpperCAmelCase : Dict = self.rust_tokenizer_class.from_pretrained( A , additional_special_tokens=A , **A , ) _UpperCAmelCase : Optional[int] = self.tokenizer_class.from_pretrained( A , additional_special_tokens=A , **A ) _UpperCAmelCase : Union[str, Any] = tokenizer_p.encode("Hey this is a <special> token" ) _UpperCAmelCase : Any = tokenizer_cr.encode("Hey this is a <special> token" ) self.assertEqual(A , A ) self.assertEqual(A , A ) self.assertTrue(special_token_id in p_output ) self.assertTrue(special_token_id in cr_output ) @require_torch @require_sentencepiece @require_tokenizers class lowerCamelCase_ (unittest.TestCase ): '''simple docstring''' __UpperCamelCase: Dict = "facebook/nllb-200-distilled-600M" __UpperCamelCase: Optional[int] = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] __UpperCamelCase: str = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] __UpperCamelCase: str = [ 2_5_6_0_4_7, 1_6_2_9_7, 1_3_4_4_0_8, 8_1_6_5, 2_4_8_0_6_6, 1_4_7_3_4, 9_5_0, 1_1_3_5, 1_0_5_7_2_1, 3_5_7_3, 8_3, 2_7_3_5_2, 1_0_8, 4_9_4_8_6, 2, ] @classmethod def _A ( cls : int ): _UpperCAmelCase : NllbTokenizer = NllbTokenizer.from_pretrained( cls.checkpoint_name , src_lang="eng_Latn" , tgt_lang="ron_Latn" ) _UpperCAmelCase : Union[str, Any] = 1 return cls def _A ( self : Any ): self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ace_Arab"] , 256001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ace_Latn"] , 256002 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["fra_Latn"] , 256057 ) def _A ( self : Union[str, Any] ): _UpperCAmelCase : Optional[int] = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , A ) def _A ( self : Tuple ): self.assertIn(A , self.tokenizer.all_special_ids ) # fmt: off _UpperCAmelCase : List[Any] = [RO_CODE, 4254, 98068, 112923, 39072, 3909, 713, 102767, 26, 17314, 35642, 14683, 33118, 2022, 66987, 2, 256047] # fmt: on _UpperCAmelCase : Tuple = self.tokenizer.decode(A , skip_special_tokens=A ) _UpperCAmelCase : Optional[Any] = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=A ) self.assertEqual(A , A ) self.assertNotIn(self.tokenizer.eos_token , A ) def _A ( self : Optional[int] ): _UpperCAmelCase : List[Any] = ["this is gunna be a long sentence " * 20] assert isinstance(src_text[0] , A ) _UpperCAmelCase : Dict = 10 _UpperCAmelCase : Tuple = self.tokenizer(A , max_length=A , truncation=A ).input_ids[0] self.assertEqual(ids[-1] , 2 ) self.assertEqual(ids[0] , A ) self.assertEqual(len(A ) , A ) def _A ( self : Dict ): self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["<mask>", "ar_AR"] ) , [256203, 3] ) def _A ( self : Optional[Any] ): _UpperCAmelCase : Dict = tempfile.mkdtemp() _UpperCAmelCase : str = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(A ) _UpperCAmelCase : Tuple = NllbTokenizer.from_pretrained(A ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , A ) @require_torch def _A ( self : Dict ): _UpperCAmelCase : List[str] = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=A , truncation=A , max_length=len(self.expected_src_tokens ) , return_tensors="pt" , ) _UpperCAmelCase : Tuple = shift_tokens_right( batch["labels"] , self.tokenizer.pad_token_id , self.tokenizer.lang_code_to_id["ron_Latn"] ) self.assertIsInstance(A , A ) self.assertEqual((2, 15) , batch.input_ids.shape ) self.assertEqual((2, 15) , batch.attention_mask.shape ) _UpperCAmelCase : Dict = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , A ) self.assertEqual(A , batch.decoder_input_ids[0, 0] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) def _A ( self : str ): _UpperCAmelCase : Optional[Any] = self.tokenizer(self.src_text , padding=A , truncation=A , max_length=3 , return_tensors="pt" ) _UpperCAmelCase : Dict = self.tokenizer( text_target=self.tgt_text , padding=A , truncation=A , max_length=10 , return_tensors="pt" ) _UpperCAmelCase : List[Any] = targets["input_ids"] _UpperCAmelCase : Union[str, Any] = shift_tokens_right( A , self.tokenizer.pad_token_id , decoder_start_token_id=self.tokenizer.lang_code_to_id[self.tokenizer.tgt_lang] , ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def _A ( self : List[Any] ): _UpperCAmelCase : str = self.tokenizer._build_translation_inputs( "A test" , return_tensors="pt" , src_lang="eng_Latn" , tgt_lang="fra_Latn" ) self.assertEqual( nested_simplify(A ) , { # A, test, EOS, en_XX "input_ids": [[256047, 70, 7356, 2]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 256057, } , ) @require_torch def _A ( self : Any ): _UpperCAmelCase : Dict = True _UpperCAmelCase : Any = self.tokenizer( "UN Chief says there is no military solution in Syria" , src_lang="eng_Latn" , tgt_lang="fra_Latn" ) self.assertEqual( inputs.input_ids , [16297, 134408, 25653, 6370, 248, 254, 103929, 94995, 108, 49486, 2, 256047] ) _UpperCAmelCase : Optional[int] = False _UpperCAmelCase : str = self.tokenizer( "UN Chief says there is no military solution in Syria" , src_lang="eng_Latn" , tgt_lang="fra_Latn" ) self.assertEqual( inputs.input_ids , [256047, 16297, 134408, 25653, 6370, 248, 254, 103929, 94995, 108, 49486, 2] )
31
'''simple docstring''' import numpy as np from matplotlib import pyplot as plt from sklearn.datasets import load_iris from sklearn.metrics import ConfusionMatrixDisplay from sklearn.model_selection import train_test_split from xgboost import XGBClassifier def UpperCamelCase_ ( _UpperCAmelCase : dict ) -> tuple: """simple docstring""" return (data["data"], data["target"]) def UpperCamelCase_ ( _UpperCAmelCase : np.ndarray , _UpperCAmelCase : np.ndarray ) -> XGBClassifier: """simple docstring""" _UpperCAmelCase : Any = XGBClassifier() classifier.fit(_UpperCAmelCase , _UpperCAmelCase ) return classifier def UpperCamelCase_ ( ) -> None: """simple docstring""" _UpperCAmelCase : List[str] = load_iris() _UpperCAmelCase , _UpperCAmelCase : Dict = data_handling(_UpperCAmelCase ) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : str = train_test_split( _UpperCAmelCase , _UpperCAmelCase , test_size=0.2_5 ) _UpperCAmelCase : Optional[Any] = iris["target_names"] # Create an XGBoost Classifier from the training data _UpperCAmelCase : Tuple = xgboost(_UpperCAmelCase , _UpperCAmelCase ) # Display the confusion matrix of the classifier with both training and test sets ConfusionMatrixDisplay.from_estimator( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , display_labels=_UpperCAmelCase , cmap="Blues" , normalize="true" , ) plt.title("Normalized Confusion Matrix - IRIS Dataset" ) plt.show() if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
31
1
def lowerCAmelCase__ ( ) -> int: return [ a * b * (1_000 - a - b) for a in range(1 , 999 ) for b in range(SCREAMING_SNAKE_CASE_ , 999 ) if (a * a + b * b == (1_000 - a - b) ** 2) ][0] if __name__ == "__main__": print(F"""{solution() = }""")
364
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> str: stooge(SCREAMING_SNAKE_CASE_ , 0 , len(SCREAMING_SNAKE_CASE_ ) - 1 ) return arr def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Dict: if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: lowerCAmelCase__ , lowerCAmelCase__ : Tuple = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: lowerCAmelCase__ : Union[str, Any] = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , (h - t) ) # Recursively sort last 2/3 elements stooge(SCREAMING_SNAKE_CASE_ , i + t , (SCREAMING_SNAKE_CASE_) ) # Recursively sort first 2/3 elements stooge(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , (h - t) ) if __name__ == "__main__": lowerCamelCase__ = input("""Enter numbers separated by a comma:\n""").strip() lowerCamelCase__ = [int(item) for item in user_input.split(""",""")] print(stooge_sort(unsorted))
307
0
"""simple docstring""" def UpperCamelCase__ ( lowercase__ : Tuple , lowercase__ : str ): snake_case : Dict = [0 for i in range(r + 1 )] # nc0 = 1 snake_case : Any = 1 for i in range(1 , n + 1 ): # to compute current row from previous row. snake_case : int = min(lowercase__ , lowercase__ ) while j > 0: c[j] += c[j - 1] j -= 1 return c[r] print(binomial_coefficient(n=10, r=5))
148
"""simple docstring""" import importlib import inspect import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __A = "src/transformers" # This is to make sure the transformers module imported is the one in the repo. __A = importlib.util.spec_from_file_location( "transformers", os.path.join(PATH_TO_TRANSFORMERS, "__init__.py"), submodule_search_locations=[PATH_TO_TRANSFORMERS], ) __A = spec.loader.load_module() __A = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __A = re.compile("\[(.+?)\]\((https://huggingface\.co/.+?)\)") __A = { "CLIPConfigMixin", "DecisionTransformerConfigMixin", "EncoderDecoderConfigMixin", "RagConfigMixin", "SpeechEncoderDecoderConfigMixin", "VisionEncoderDecoderConfigMixin", "VisionTextDualEncoderConfigMixin", } def UpperCamelCase__ ( ): snake_case : Dict = [] for config_class in list(CONFIG_MAPPING.values() ): snake_case : Tuple = False # source code of `config_class` snake_case : Tuple = inspect.getsource(lowercase__ ) snake_case : Optional[int] = _re_checkpoint.findall(lowercase__ ) for checkpoint in checkpoints: # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` snake_case , snake_case : str = checkpoint # verify the checkpoint name corresponds to the checkpoint link snake_case : Optional[int] = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: snake_case : Any = True break snake_case : Optional[Any] = config_class.__name__ if not checkpoint_found and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(lowercase__ ) if len(lowercase__ ) > 0: snake_case : Optional[Any] = "\n".join(sorted(lowercase__ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
148
1
import itertools import os import random import tempfile import unittest import numpy as np from transformers import TvltFeatureExtractor, is_datasets_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch if is_datasets_available(): from datasets import load_dataset __magic_name__: Tuple = random.Random() def UpperCamelCase ( _A, _A=1.0, _A=None, _A=None ): """simple docstring""" if rng is None: __magic_name__ : Optional[Any] = global_rng __magic_name__ : Union[str, Any] = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class snake_case__ ( unittest.TestCase ): def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7 , lowerCAmelCase__=4_00 , lowerCAmelCase__=20_00 , lowerCAmelCase__=20_48 , lowerCAmelCase__=1_28 , lowerCAmelCase__=1 , lowerCAmelCase__=5_12 , lowerCAmelCase__=30 , lowerCAmelCase__=4_41_00 , ) -> List[Any]: __magic_name__ : List[str] = parent __magic_name__ : List[Any] = batch_size __magic_name__ : Optional[Any] = min_seq_length __magic_name__ : Tuple = max_seq_length __magic_name__ : Union[str, Any] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __magic_name__ : List[str] = spectrogram_length __magic_name__ : List[Any] = feature_size __magic_name__ : Tuple = num_audio_channels __magic_name__ : str = hop_length __magic_name__ : Optional[int] = chunk_length __magic_name__ : str = sampling_rate def __magic_name__ ( self ) -> List[Any]: return { "spectrogram_length": self.spectrogram_length, "feature_size": self.feature_size, "num_audio_channels": self.num_audio_channels, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "sampling_rate": self.sampling_rate, } def __magic_name__ ( self , lowerCAmelCase__=False , lowerCAmelCase__=False ) -> Dict: def _flatten(lowerCAmelCase__ ): return list(itertools.chain(*lowerCAmelCase__ ) ) if equal_length: __magic_name__ : Dict = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __magic_name__ : Optional[Any] = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __magic_name__ : Tuple = [np.asarray(lowerCAmelCase__ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class snake_case__ ( _lowerCAmelCase , unittest.TestCase ): lowercase__ : Optional[Any] = TvltFeatureExtractor def __magic_name__ ( self ) -> Optional[Any]: __magic_name__ : List[Any] = TvltFeatureExtractionTester(self ) def __magic_name__ ( self ) -> Optional[Any]: __magic_name__ : str = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(lowerCAmelCase__ , """spectrogram_length""" ) ) self.assertTrue(hasattr(lowerCAmelCase__ , """feature_size""" ) ) self.assertTrue(hasattr(lowerCAmelCase__ , """num_audio_channels""" ) ) self.assertTrue(hasattr(lowerCAmelCase__ , """hop_length""" ) ) self.assertTrue(hasattr(lowerCAmelCase__ , """chunk_length""" ) ) self.assertTrue(hasattr(lowerCAmelCase__ , """sampling_rate""" ) ) def __magic_name__ ( self ) -> Optional[int]: __magic_name__ : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __magic_name__ : Dict = feat_extract_first.save_pretrained(lowerCAmelCase__ )[0] check_json_file_has_correct_format(lowerCAmelCase__ ) __magic_name__ : Any = self.feature_extraction_class.from_pretrained(lowerCAmelCase__ ) __magic_name__ : int = feat_extract_first.to_dict() __magic_name__ : Dict = feat_extract_second.to_dict() __magic_name__ : str = dict_first.pop("""mel_filters""" ) __magic_name__ : str = dict_second.pop("""mel_filters""" ) self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ ) ) self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ ) def __magic_name__ ( self ) -> Any: __magic_name__ : int = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __magic_name__ : Union[str, Any] = os.path.join(lowerCAmelCase__ , """feat_extract.json""" ) feat_extract_first.to_json_file(lowerCAmelCase__ ) __magic_name__ : int = self.feature_extraction_class.from_json_file(lowerCAmelCase__ ) __magic_name__ : str = feat_extract_first.to_dict() __magic_name__ : Optional[Any] = feat_extract_second.to_dict() __magic_name__ : Dict = dict_first.pop("""mel_filters""" ) __magic_name__ : Optional[Any] = dict_second.pop("""mel_filters""" ) self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ ) ) self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__ ) def __magic_name__ ( self ) -> Optional[int]: # Initialize feature_extractor __magic_name__ : Optional[int] = self.feature_extraction_class(**self.feat_extract_dict ) # create three inputs of length 800, 1000, and 1200 __magic_name__ : Tuple = [floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __magic_name__ : Union[str, Any] = [np.asarray(lowerCAmelCase__ ) for speech_input in speech_inputs] # Test not batched input __magic_name__ : Union[str, Any] = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" , sampling_rate=4_41_00 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test batched __magic_name__ : str = feature_extractor(lowerCAmelCase__ , return_tensors="""np""" , sampling_rate=4_41_00 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test audio masking __magic_name__ : Dict = feature_extractor( lowerCAmelCase__ , return_tensors="""np""" , sampling_rate=4_41_00 , mask_audio=lowerCAmelCase__ ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test 2-D numpy arrays are batched. __magic_name__ : int = [floats_list((1, x) )[0] for x in (8_00, 8_00, 8_00)] __magic_name__ : Optional[Any] = np.asarray(lowerCAmelCase__ ) __magic_name__ : Optional[int] = feature_extractor(lowerCAmelCase__ , return_tensors="""np""" , sampling_rate=4_41_00 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) def __magic_name__ ( self , lowerCAmelCase__ ) -> Any: __magic_name__ : Optional[Any] = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech __magic_name__ : int = ds.sort("""id""" ).select(range(lowerCAmelCase__ ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def __magic_name__ ( self ) -> Optional[Any]: __magic_name__ : List[str] = self._load_datasamples(1 ) __magic_name__ : List[str] = TvltFeatureExtractor() __magic_name__ : Tuple = feature_extractor(lowerCAmelCase__ , return_tensors="""pt""" ).audio_values self.assertEquals(audio_values.shape , (1, 1, 1_92, 1_28) ) __magic_name__ : List[str] = torch.tensor([[-0.3_0_3_2, -0.2_7_0_8], [-0.4_4_3_4, -0.4_0_0_7]] ) self.assertTrue(torch.allclose(audio_values[0, 0, :2, :2] , lowerCAmelCase__ , atol=1e-4 ) )
138
import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class snake_case__ ( unittest.TestCase ): @slow def __magic_name__ ( self ) -> Optional[Any]: __magic_name__ : Any = FlaxMTaForConditionalGeneration.from_pretrained("""google/mt5-small""" ) __magic_name__ : Dict = AutoTokenizer.from_pretrained("""google/mt5-small""" ) __magic_name__ : Tuple = tokenizer("""Hello there""" , return_tensors="""np""" ).input_ids __magic_name__ : List[Any] = tokenizer("""Hi I am""" , return_tensors="""np""" ).input_ids __magic_name__ : Any = shift_tokens_right(lowerCAmelCase__ , model.config.pad_token_id , model.config.decoder_start_token_id ) __magic_name__ : List[Any] = model(lowerCAmelCase__ , decoder_input_ids=lowerCAmelCase__ ).logits __magic_name__ : Tuple = optax.softmax_cross_entropy(lowerCAmelCase__ , onehot(lowerCAmelCase__ , logits.shape[-1] ) ).mean() __magic_name__ : List[Any] = -(labels.shape[-1] * loss.item()) __magic_name__ : List[Any] = -8_4.9_1_2_7 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1e-4 )
138
1
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING A__ : Tuple = logging.get_logger(__name__) A__ : List[str] = { '''salesforce/blip2-opt-2.7b''': '''https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json''', } class __snake_case ( UpperCamelCase_ ): _a = '''blip_2_vision_model''' def __init__( self : Tuple , A_ : Optional[int]=1_4_0_8 , A_ : Tuple=6_1_4_4 , A_ : Tuple=3_9 , A_ : Dict=1_6 , A_ : Tuple=2_2_4 , A_ : Any=1_4 , A_ : Dict="gelu" , A_ : Any=0.0_0001 , A_ : int=0.0 , A_ : Dict=1e-10 , A_ : Optional[Any]=True , **A_ : Tuple , ): super().__init__(**A_) lowerCAmelCase_ : Tuple = hidden_size lowerCAmelCase_ : Dict = intermediate_size lowerCAmelCase_ : List[str] = num_hidden_layers lowerCAmelCase_ : Dict = num_attention_heads lowerCAmelCase_ : Dict = patch_size lowerCAmelCase_ : Dict = image_size lowerCAmelCase_ : str = initializer_range lowerCAmelCase_ : List[str] = attention_dropout lowerCAmelCase_ : Optional[int] = layer_norm_eps lowerCAmelCase_ : Union[str, Any] = hidden_act lowerCAmelCase_ : Tuple = qkv_bias @classmethod def UpperCAmelCase__ ( cls : str , A_ : Union[str, os.PathLike] , **A_ : List[str]): cls._set_token_in_kwargs(A_) lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = cls.get_config_dict(A_ , **A_) # get the vision config dict if we are loading from Blip2Config if config_dict.get('''model_type''') == "blip-2": lowerCAmelCase_ : Optional[Any] = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''') and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""") return cls.from_dict(A_ , **A_) class __snake_case ( UpperCamelCase_ ): _a = '''blip_2_qformer''' def __init__( self : List[str] , A_ : List[str]=3_0_5_2_2 , A_ : Union[str, Any]=7_6_8 , A_ : Optional[int]=1_2 , A_ : Optional[Any]=1_2 , A_ : str=3_0_7_2 , A_ : Tuple="gelu" , A_ : Union[str, Any]=0.1 , A_ : Optional[Any]=0.1 , A_ : Any=5_1_2 , A_ : Optional[Any]=0.02 , A_ : List[Any]=1e-12 , A_ : Optional[int]=0 , A_ : Dict="absolute" , A_ : List[Any]=2 , A_ : Union[str, Any]=1_4_0_8 , **A_ : Optional[int] , ): super().__init__(pad_token_id=A_ , **A_) lowerCAmelCase_ : Dict = vocab_size lowerCAmelCase_ : str = hidden_size lowerCAmelCase_ : Union[str, Any] = num_hidden_layers lowerCAmelCase_ : Tuple = num_attention_heads lowerCAmelCase_ : List[str] = hidden_act lowerCAmelCase_ : Optional[int] = intermediate_size lowerCAmelCase_ : Any = hidden_dropout_prob lowerCAmelCase_ : int = attention_probs_dropout_prob lowerCAmelCase_ : str = max_position_embeddings lowerCAmelCase_ : Union[str, Any] = initializer_range lowerCAmelCase_ : Union[str, Any] = layer_norm_eps lowerCAmelCase_ : List[Any] = position_embedding_type lowerCAmelCase_ : Optional[Any] = cross_attention_frequency lowerCAmelCase_ : Dict = encoder_hidden_size @classmethod def UpperCAmelCase__ ( cls : Dict , A_ : Union[str, os.PathLike] , **A_ : Tuple): cls._set_token_in_kwargs(A_) lowerCAmelCase_ , lowerCAmelCase_ : str = cls.get_config_dict(A_ , **A_) # get the qformer config dict if we are loading from Blip2Config if config_dict.get('''model_type''') == "blip-2": lowerCAmelCase_ : List[Any] = config_dict['''qformer_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''') and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""") return cls.from_dict(A_ , **A_) class __snake_case ( UpperCamelCase_ ): _a = '''blip-2''' _a = True def __init__( self : str , A_ : str=None , A_ : Dict=None , A_ : List[str]=None , A_ : str=3_2 , **A_ : Any): super().__init__(**A_) if vision_config is None: lowerCAmelCase_ : Any = {} logger.info('''vision_config is None. initializing the Blip2VisionConfig with default values.''') if qformer_config is None: lowerCAmelCase_ : List[str] = {} logger.info('''qformer_config is None. Initializing the Blip2QFormerConfig with default values.''') if text_config is None: lowerCAmelCase_ : Any = {} logger.info('''text_config is None. Initializing the text config with default values (`OPTConfig`).''') lowerCAmelCase_ : Union[str, Any] = BlipaVisionConfig(**A_) lowerCAmelCase_ : int = BlipaQFormerConfig(**A_) lowerCAmelCase_ : Any = text_config['''model_type'''] if '''model_type''' in text_config else '''opt''' lowerCAmelCase_ : Dict = CONFIG_MAPPING[text_model_type](**A_) lowerCAmelCase_ : List[str] = self.text_config.tie_word_embeddings lowerCAmelCase_ : Optional[Any] = self.text_config.is_encoder_decoder lowerCAmelCase_ : Tuple = num_query_tokens lowerCAmelCase_ : Tuple = self.vision_config.hidden_size lowerCAmelCase_ : str = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES lowerCAmelCase_ : List[Any] = 1.0 lowerCAmelCase_ : List[Any] = 0.02 @classmethod def UpperCAmelCase__ ( cls : Dict , A_ : BlipaVisionConfig , A_ : BlipaQFormerConfig , A_ : PretrainedConfig , **A_ : List[Any] , ): return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **A_ , ) def UpperCAmelCase__ ( self : Tuple): lowerCAmelCase_ : List[Any] = copy.deepcopy(self.__dict__) lowerCAmelCase_ : Union[str, Any] = self.vision_config.to_dict() lowerCAmelCase_ : List[Any] = self.qformer_config.to_dict() lowerCAmelCase_ : Any = self.text_config.to_dict() lowerCAmelCase_ : Any = self.__class__.model_type return output
103
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A__ : Union[str, Any] = logging.get_logger(__name__) A__ : Tuple = { '''facebook/xlm-roberta-xl''': '''https://huggingface.co/facebook/xlm-roberta-xl/resolve/main/config.json''', '''facebook/xlm-roberta-xxl''': '''https://huggingface.co/facebook/xlm-roberta-xxl/resolve/main/config.json''', # See all XLM-RoBERTa-XL models at https://huggingface.co/models?filter=xlm-roberta-xl } class __snake_case ( UpperCamelCase_ ): _a = '''xlm-roberta-xl''' def __init__( self : int , A_ : List[str]=2_5_0_8_8_0 , A_ : List[str]=2_5_6_0 , A_ : Optional[int]=3_6 , A_ : List[Any]=3_2 , A_ : Optional[int]=1_0_2_4_0 , A_ : Dict="gelu" , A_ : int=0.1 , A_ : Optional[Any]=0.1 , A_ : int=5_1_4 , A_ : Any=1 , A_ : Optional[Any]=0.02 , A_ : str=1e-05 , A_ : Dict=1 , A_ : Any=0 , A_ : Tuple=2 , A_ : str="absolute" , A_ : str=True , A_ : List[str]=None , **A_ : Dict , ): super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_) lowerCAmelCase_ : Tuple = vocab_size lowerCAmelCase_ : List[str] = hidden_size lowerCAmelCase_ : int = num_hidden_layers lowerCAmelCase_ : int = num_attention_heads lowerCAmelCase_ : Dict = hidden_act lowerCAmelCase_ : int = intermediate_size lowerCAmelCase_ : Tuple = hidden_dropout_prob lowerCAmelCase_ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase_ : Union[str, Any] = max_position_embeddings lowerCAmelCase_ : Dict = type_vocab_size lowerCAmelCase_ : str = initializer_range lowerCAmelCase_ : str = layer_norm_eps lowerCAmelCase_ : Optional[Any] = position_embedding_type lowerCAmelCase_ : Optional[Any] = use_cache lowerCAmelCase_ : List[str] = classifier_dropout class __snake_case ( UpperCamelCase_ ): @property def UpperCAmelCase__ ( self : List[str]): if self.task == "multiple-choice": lowerCAmelCase_ : Union[str, Any] = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowerCAmelCase_ : List[Any] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ])
103
1
'''simple docstring''' import os from pathlib import Path def UpperCamelCase( ): from torch.utils.cpp_extension import load UpperCAmelCase : Optional[Any] = Path(_a ).resolve().parent.parent.parent / """kernels""" / """deformable_detr""" UpperCAmelCase : Tuple = [ root / filename for filename in [ """vision.cpp""", os.path.join('cpu' , 'ms_deform_attn_cpu.cpp' ), os.path.join('cuda' , 'ms_deform_attn_cuda.cu' ), ] ] load( 'MultiScaleDeformableAttention' , _a , with_cuda=_a , extra_include_paths=[str(_a )] , extra_cflags=['-DWITH_CUDA=1'] , extra_cuda_cflags=[ '-DCUDA_HAS_FP16=1', '-D__CUDA_NO_HALF_OPERATORS__', '-D__CUDA_NO_HALF_CONVERSIONS__', '-D__CUDA_NO_HALF2_OPERATORS__', ] , ) import MultiScaleDeformableAttention as MSDA return MSDA
365
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...feature_extraction_utils import FeatureExtractionMixin from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType, logging lowercase__ = logging.get_logger(__name__) lowercase__ = { "deepmind/language-perceiver": "https://huggingface.co/deepmind/language-perceiver/resolve/main/config.json", # See all Perceiver models at https://huggingface.co/models?filter=perceiver } class A_ ( _snake_case ): '''simple docstring''' UpperCAmelCase_ : Optional[int] = """perceiver""" def __init__( self : str , lowercase_ : List[str]=256 , lowercase_ : List[str]=1_280 , lowercase_ : str=768 , lowercase_ : Tuple=1 , lowercase_ : str=26 , lowercase_ : List[Any]=8 , lowercase_ : int=8 , lowercase_ : List[str]=None , lowercase_ : Dict=None , lowercase_ : int="kv" , lowercase_ : Union[str, Any]=1 , lowercase_ : List[str]=1 , lowercase_ : Any="gelu" , lowercase_ : Optional[Any]=0.1 , lowercase_ : str=0.02 , lowercase_ : Optional[Any]=1E-12 , lowercase_ : Any=True , lowercase_ : Union[str, Any]=262 , lowercase_ : Union[str, Any]=2_048 , lowercase_ : Optional[int]=56 , lowercase_ : int=[368, 496] , lowercase_ : str=16 , lowercase_ : Optional[int]=1_920 , lowercase_ : Tuple=16 , lowercase_ : int=[1, 16, 224, 224] , **lowercase_ : Union[str, Any] , ) -> List[str]: super().__init__(**lowercase_ ) UpperCAmelCase : Union[str, Any] = num_latents UpperCAmelCase : List[Any] = d_latents UpperCAmelCase : Dict = d_model UpperCAmelCase : Dict = num_blocks UpperCAmelCase : Optional[int] = num_self_attends_per_block UpperCAmelCase : Optional[Any] = num_self_attention_heads UpperCAmelCase : Optional[Any] = num_cross_attention_heads UpperCAmelCase : Tuple = qk_channels UpperCAmelCase : Optional[int] = v_channels UpperCAmelCase : str = cross_attention_shape_for_attention UpperCAmelCase : Union[str, Any] = self_attention_widening_factor UpperCAmelCase : List[Any] = cross_attention_widening_factor UpperCAmelCase : Optional[Any] = hidden_act UpperCAmelCase : int = attention_probs_dropout_prob UpperCAmelCase : Optional[int] = initializer_range UpperCAmelCase : Dict = layer_norm_eps UpperCAmelCase : Any = use_query_residual # masked language modeling attributes UpperCAmelCase : Any = vocab_size UpperCAmelCase : List[Any] = max_position_embeddings # image classification attributes UpperCAmelCase : str = image_size # flow attributes UpperCAmelCase : Any = train_size # multimodal autoencoding attributes UpperCAmelCase : Any = num_frames UpperCAmelCase : List[Any] = audio_samples_per_frame UpperCAmelCase : Tuple = samples_per_patch UpperCAmelCase : Union[str, Any] = output_shape class A_ ( _snake_case ): '''simple docstring''' @property def UpperCAmelCase_ ( self : Union[str, Any] ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": UpperCAmelCase : Tuple = {0: 'batch', 1: 'choice', 2: 'sequence'} else: UpperCAmelCase : Tuple = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('inputs', dynamic_axis), ('attention_mask', dynamic_axis), ] ) @property def UpperCAmelCase_ ( self : str ) -> float: return 1E-4 def UpperCAmelCase_ ( self : str , lowercase_ : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , lowercase_ : int = -1 , lowercase_ : int = -1 , lowercase_ : int = -1 , lowercase_ : bool = False , lowercase_ : Optional[TensorType] = None , lowercase_ : int = 3 , lowercase_ : int = 40 , lowercase_ : int = 40 , ) -> Mapping[str, Any]: # copied from `transformers.onnx.config.OnnxConfig` and slightly altered/simplified if isinstance(lowercase_ , lowercase_ ): # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX UpperCAmelCase : Tuple = compute_effective_axis_dimension( lowercase_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX UpperCAmelCase : int = preprocessor.num_special_tokens_to_add(lowercase_ ) UpperCAmelCase : Union[str, Any] = compute_effective_axis_dimension( lowercase_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=lowercase_ ) # Generate dummy inputs according to compute batch and sequence UpperCAmelCase : List[Any] = [' '.join(['a'] ) * seq_length] * batch_size UpperCAmelCase : Union[str, Any] = dict(preprocessor(lowercase_ , return_tensors=lowercase_ ) ) UpperCAmelCase : Union[str, Any] = inputs.pop('input_ids' ) return inputs elif isinstance(lowercase_ , lowercase_ ) and preprocessor.model_input_names[0] == "pixel_values": # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX UpperCAmelCase : Tuple = compute_effective_axis_dimension(lowercase_ , fixed_dimension=OnnxConfig.default_fixed_batch ) UpperCAmelCase : Any = self._generate_dummy_images(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) UpperCAmelCase : Tuple = dict(preprocessor(images=lowercase_ , return_tensors=lowercase_ ) ) UpperCAmelCase : Dict = inputs.pop('pixel_values' ) return inputs else: raise ValueError( 'Unable to generate dummy inputs for the model. Please provide a tokenizer or a preprocessor.' )
280
0
import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, ) from transformers.models.albert.modeling_albert import ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST class A__ : def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=16 , A_=36 , A_=6 , A_=6 , A_=6 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.02 , A_=3 , A_=4 , A_=None , ): '''simple docstring''' UpperCamelCase : Tuple = parent UpperCamelCase : Dict = batch_size UpperCamelCase : Tuple = seq_length UpperCamelCase : Tuple = is_training UpperCamelCase : str = use_input_mask UpperCamelCase : Tuple = use_token_type_ids UpperCamelCase : Union[str, Any] = use_labels UpperCamelCase : Optional[Any] = vocab_size UpperCamelCase : Any = embedding_size UpperCamelCase : int = hidden_size UpperCamelCase : Union[str, Any] = num_hidden_layers UpperCamelCase : Union[str, Any] = num_hidden_groups UpperCamelCase : str = num_attention_heads UpperCamelCase : Optional[int] = intermediate_size UpperCamelCase : Any = hidden_act UpperCamelCase : str = hidden_dropout_prob UpperCamelCase : int = attention_probs_dropout_prob UpperCamelCase : Union[str, Any] = max_position_embeddings UpperCamelCase : Optional[Any] = type_vocab_size UpperCamelCase : Tuple = type_sequence_label_size UpperCamelCase : Optional[int] = initializer_range UpperCamelCase : List[str] = num_labels UpperCamelCase : Union[str, Any] = num_choices UpperCamelCase : str = scope def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase : List[Any] = None if self.use_input_mask: UpperCamelCase : Optional[int] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase : str = None if self.use_token_type_ids: UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCamelCase : List[Any] = None UpperCamelCase : Dict = None UpperCamelCase : int = None if self.use_labels: UpperCamelCase : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCamelCase : Any = ids_tensor([self.batch_size] , self.num_choices ) UpperCamelCase : List[str] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCamelCase( self ): '''simple docstring''' return AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , num_hidden_groups=self.num_hidden_groups , ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = AlbertModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() UpperCamelCase : Optional[Any] = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase ) UpperCamelCase : List[str] = model(__UpperCamelCase , token_type_ids=__UpperCamelCase ) UpperCamelCase : Tuple = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = AlbertForPreTraining(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() UpperCamelCase : List[str] = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase , sentence_order_label=__UpperCamelCase , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.sop_logits.shape , (self.batch_size, config.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Tuple = AlbertForMaskedLM(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() UpperCamelCase : Optional[Any] = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : int = AlbertForQuestionAnswering(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() UpperCamelCase : List[str] = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , start_positions=__UpperCamelCase , end_positions=__UpperCamelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[str] = self.num_labels UpperCamelCase : str = AlbertForSequenceClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() UpperCamelCase : Union[str, Any] = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.num_labels UpperCamelCase : str = AlbertForTokenClassification(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() UpperCamelCase : Optional[int] = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.num_choices UpperCamelCase : Tuple = AlbertForMultipleChoice(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() UpperCamelCase : Dict = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase : List[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase : Any = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase : Tuple = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) : Optional[Any] = config_and_inputs UpperCamelCase : Tuple = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class A__ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCAmelCase :Any = ( ( AlbertModel, AlbertForPreTraining, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertForQuestionAnswering, ) if is_torch_available() else () ) _UpperCAmelCase :Optional[Any] = ( { 'feature-extraction': AlbertModel, 'fill-mask': AlbertForMaskedLM, 'question-answering': AlbertForQuestionAnswering, 'text-classification': AlbertForSequenceClassification, 'token-classification': AlbertForTokenClassification, 'zero-shot': AlbertForSequenceClassification, } if is_torch_available() else {} ) _UpperCAmelCase :Dict = True def __UpperCamelCase( self , A_ , A_ , A_=False ): '''simple docstring''' UpperCamelCase : List[str] = super()._prepare_for_class(__UpperCamelCase , __UpperCamelCase , return_labels=__UpperCamelCase ) if return_labels: if model_class in get_values(__UpperCamelCase ): UpperCamelCase : Any = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=__UpperCamelCase ) UpperCamelCase : List[str] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__UpperCamelCase ) return inputs_dict def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = AlbertModelTester(self ) UpperCamelCase : Optional[int] = ConfigTester(self , config_class=__UpperCamelCase , hidden_size=37 ) def __UpperCamelCase( self ): '''simple docstring''' self.config_tester.run_common_tests() def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__UpperCamelCase ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCamelCase ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCamelCase ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCamelCase ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCamelCase ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCamelCase : Any = type self.model_tester.create_and_check_model(*__UpperCamelCase ) @slow def __UpperCamelCase( self ): '''simple docstring''' for model_name in ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase : Optional[Any] = AlbertModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) @require_torch class A__ ( unittest.TestCase ): @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = AlbertModel.from_pretrained("albert-base-v2" ) UpperCamelCase : Union[str, Any] = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) UpperCamelCase : Any = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): UpperCamelCase : List[str] = model(__UpperCamelCase , attention_mask=__UpperCamelCase )[0] UpperCamelCase : List[str] = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , __UpperCamelCase ) UpperCamelCase : List[str] = torch.tensor( [[[-0.65_13, 1.50_35, -0.27_66], [-0.65_15, 1.50_46, -0.27_80], [-0.65_12, 1.50_49, -0.27_84]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __UpperCamelCase , atol=1e-4 ) )
52
"""simple docstring""" def lowercase ( _SCREAMING_SNAKE_CASE : list ): '''simple docstring''' if len(_SCREAMING_SNAKE_CASE ) <= 1: return lst _UpperCAmelCase = 1 while i < len(_SCREAMING_SNAKE_CASE ): if lst[i - 1] <= lst[i]: i += 1 else: _UpperCAmelCase , _UpperCAmelCase = lst[i], lst[i - 1] i -= 1 if i == 0: _UpperCAmelCase = 1 return lst if __name__ == "__main__": __A : Dict = input("Enter numbers separated by a comma:\n").strip() __A : List[Any] = [int(item) for item in user_input.split(",")] print(gnome_sort(unsorted))
260
0
"""simple docstring""" from __future__ import annotations def _lowerCamelCase(__UpperCamelCase ) -> list[int]: return [ord(__UpperCamelCase ) - 96 for elem in plain] def _lowerCamelCase(__UpperCamelCase ) -> str: return "".join(chr(elem + 96 ) for elem in encoded ) def _lowerCamelCase() -> None: _lowerCAmelCase =encode(input("""-> """ ).strip().lower() ) print("""Encoded: """ , __UpperCamelCase ) print("""Decoded:""" , decode(__UpperCamelCase ) ) if __name__ == "__main__": main()
341
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCamelCase__ ( __magic_name__ ): '''simple docstring''' lowerCamelCase = ['''image_processor''', '''tokenizer'''] lowerCamelCase = '''CLIPImageProcessor''' lowerCamelCase = ('''XLMRobertaTokenizer''', '''XLMRobertaTokenizerFast''') def __init__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , **__UpperCAmelCase ) -> Union[str, Any]: _lowerCAmelCase =None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , __UpperCAmelCase , ) _lowerCAmelCase =kwargs.pop("""feature_extractor""" ) _lowerCAmelCase =image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(__UpperCAmelCase , __UpperCAmelCase ) def __call__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=None , **__UpperCAmelCase ) -> Optional[Any]: if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: _lowerCAmelCase =self.tokenizer(__UpperCAmelCase , return_tensors=__UpperCAmelCase , **__UpperCAmelCase ) if images is not None: _lowerCAmelCase =self.image_processor(__UpperCAmelCase , return_tensors=__UpperCAmelCase , **__UpperCAmelCase ) if text is not None and images is not None: _lowerCAmelCase =image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__UpperCAmelCase ) , tensor_type=__UpperCAmelCase ) def _lowerCAmelCase ( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> List[Any]: return self.tokenizer.batch_decode(*__UpperCAmelCase , **__UpperCAmelCase ) def _lowerCAmelCase ( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[int]: return self.tokenizer.decode(*__UpperCAmelCase , **__UpperCAmelCase ) @property def _lowerCAmelCase ( self ) -> int: _lowerCAmelCase =self.tokenizer.model_input_names _lowerCAmelCase =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
341
1
"""simple docstring""" def lowercase ( A_ )-> int: '''simple docstring''' a : Any = 0 while num > 0: digit_sum += num % 10 num //= 10 return digit_sum def lowercase ( A_ = 100 )-> int: '''simple docstring''' a : Tuple = 1 a : Optional[int] = 2 for i in range(2 , max_n + 1 ): a : Optional[Any] = pre_numerator a : Optional[int] = 2 * i // 3 if i % 3 == 0 else 1 a : Dict = cur_numerator a : List[str] = e_cont * pre_numerator + temp return sum_digits(A_ ) if __name__ == "__main__": print(f'''{solution() = }''')
40
import json import os from pathlib import Path import pytest from datasets.download.download_config import DownloadConfig from datasets.download.download_manager import DownloadManager from datasets.utils.file_utils import hash_url_to_filename lowerCAmelCase__ : Union[str, Any] = '''http://www.mocksite.com/file1.txt''' lowerCAmelCase__ : Optional[Any] = '''"text": ["foo", "foo"]''' lowerCAmelCase__ : List[str] = '''6d8ce9aa78a471c7477201efbeabd3bb01ac2e7d100a6dc024ba1608361f90a8''' class __snake_case : __lowerCamelCase = 200 __lowerCamelCase = {"""Content-Length""": """100"""} __lowerCamelCase = {} def __a ( self , **__UpperCamelCase ) -> Union[str, Any]: '''simple docstring''' return [bytes(__UpperCamelCase , 'utf-8' )] def UpperCamelCase__ ( *A__ , **A__ ) -> Optional[Any]: return MockResponse() @pytest.mark.parametrize('urls_type' , [str, list, dict] ) def UpperCamelCase__ ( A__ , A__ , A__ ) -> Any: import requests monkeypatch.setattr(A__ , 'request' , A__ ) snake_case__ : Any = URL if issubclass(A__ , A__ ): snake_case__ : Optional[Any] = url elif issubclass(A__ , A__ ): snake_case__ : Dict = [url] elif issubclass(A__ , A__ ): snake_case__ : Any = {'train': url} snake_case__ : Union[str, Any] = 'dummy' snake_case__ : List[str] = 'downloads' snake_case__ : int = tmp_path snake_case__ : Tuple = DownloadConfig( cache_dir=os.path.join(A__ , A__ ) , use_etag=A__ , ) snake_case__ : Any = DownloadManager(dataset_name=A__ , download_config=A__ ) snake_case__ : Any = dl_manager.download(A__ ) snake_case__ : Dict = urls for downloaded_paths in [downloaded_paths]: if isinstance(A__ , A__ ): snake_case__ : int = [downloaded_paths] snake_case__ : Any = [urls] elif isinstance(A__ , A__ ): assert "train" in downloaded_paths.keys() snake_case__ : Union[str, Any] = downloaded_paths.values() snake_case__ : Any = urls.values() assert downloaded_paths for downloaded_path, input_url in zip(A__ , A__ ): assert downloaded_path == dl_manager.downloaded_paths[input_url] snake_case__ : int = Path(A__ ) snake_case__ : Optional[int] = downloaded_path.parts assert parts[-1] == HASH assert parts[-2] == cache_subdir assert downloaded_path.exists() snake_case__ : Optional[Any] = downloaded_path.read_text() assert content == CONTENT snake_case__ : int = downloaded_path.with_suffix('.json' ) assert metadata_downloaded_path.exists() snake_case__ : int = json.loads(metadata_downloaded_path.read_text() ) assert metadata_content == {"url": URL, "etag": None} @pytest.mark.parametrize('paths_type' , [str, list, dict] ) def UpperCamelCase__ ( A__ , A__ , A__ ) -> Any: snake_case__ : Tuple = str(A__ ) if issubclass(A__ , A__ ): snake_case__ : Dict = filename elif issubclass(A__ , A__ ): snake_case__ : Any = [filename] elif issubclass(A__ , A__ ): snake_case__ : Dict = {'train': filename} snake_case__ : Union[str, Any] = 'dummy' snake_case__ : List[Any] = xz_file.parent snake_case__ : Dict = 'extracted' snake_case__ : List[Any] = DownloadConfig( cache_dir=A__ , use_etag=A__ , ) snake_case__ : Optional[int] = DownloadManager(dataset_name=A__ , download_config=A__ ) snake_case__ : Optional[Any] = dl_manager.extract(A__ ) snake_case__ : Union[str, Any] = paths for extracted_paths in [extracted_paths]: if isinstance(A__ , A__ ): snake_case__ : str = [extracted_paths] snake_case__ : Dict = [paths] elif isinstance(A__ , A__ ): assert "train" in extracted_paths.keys() snake_case__ : Any = extracted_paths.values() snake_case__ : Dict = paths.values() assert extracted_paths for extracted_path, input_path in zip(A__ , A__ ): assert extracted_path == dl_manager.extracted_paths[input_path] snake_case__ : Optional[int] = Path(A__ ) snake_case__ : Any = extracted_path.parts assert parts[-1] == hash_url_to_filename(A__ , etag=A__ ) assert parts[-2] == extracted_subdir assert extracted_path.exists() snake_case__ : Dict = extracted_path.read_text() snake_case__ : Union[str, Any] = text_file.read_text() assert extracted_file_content == expected_file_content def UpperCamelCase__ ( A__ , A__ ) -> Union[str, Any]: assert path.endswith('.jsonl' ) for num_items, line in enumerate(A__ , start=1 ): snake_case__ : Optional[int] = json.loads(line.decode('utf-8' ) ) assert item.keys() == {"col_1", "col_2", "col_3"} assert num_items == 4 @pytest.mark.parametrize('archive_jsonl' , ['tar_jsonl_path', 'zip_jsonl_path'] ) def UpperCamelCase__ ( A__ , A__ ) -> Optional[Any]: snake_case__ : Tuple = request.getfixturevalue(A__ ) snake_case__ : Optional[int] = DownloadManager() for num_jsonl, (path, file) in enumerate(dl_manager.iter_archive(A__ ) , start=1 ): _test_jsonl(A__ , A__ ) assert num_jsonl == 2 @pytest.mark.parametrize('archive_nested_jsonl' , ['tar_nested_jsonl_path', 'zip_nested_jsonl_path'] ) def UpperCamelCase__ ( A__ , A__ ) -> int: snake_case__ : List[Any] = request.getfixturevalue(A__ ) snake_case__ : str = DownloadManager() for num_tar, (path, file) in enumerate(dl_manager.iter_archive(A__ ) , start=1 ): for num_jsonl, (subpath, subfile) in enumerate(dl_manager.iter_archive(A__ ) , start=1 ): _test_jsonl(A__ , A__ ) assert num_tar == 1 assert num_jsonl == 2 def UpperCamelCase__ ( A__ ) -> Union[str, Any]: snake_case__ : Dict = DownloadManager() for num_file, file in enumerate(dl_manager.iter_files(A__ ) , start=1 ): assert os.path.basename(A__ ) == ("test.txt" if num_file == 1 else "train.txt") assert num_file == 2
143
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) UpperCAmelCase__ = { 'configuration_falcon': ['FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP', 'FalconConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ = [ 'FALCON_PRETRAINED_MODEL_ARCHIVE_LIST', 'FalconForCausalLM', 'FalconModel', 'FalconPreTrainedModel', 'FalconForSequenceClassification', 'FalconForTokenClassification', 'FalconForQuestionAnswering', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys UpperCAmelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
40
"""simple docstring""" from __future__ import annotations class lowerCAmelCase__ : def __init__( self : Optional[int] , _lowerCamelCase : int = 0 ): _snake_case = key def lowercase ( self : Tuple , _lowerCamelCase : str , _lowerCamelCase : int ): assert isinstance(_lowerCamelCase , _lowerCamelCase ) and isinstance(_lowerCamelCase , _lowerCamelCase ) _snake_case = key or self.__key or 1 # make sure key is an appropriate size key %= 255 return [chr(ord(_lowerCamelCase ) ^ key ) for ch in content] def lowercase ( self : List[Any] , _lowerCamelCase : str , _lowerCamelCase : int ): assert isinstance(_lowerCamelCase , _lowerCamelCase ) and isinstance(_lowerCamelCase , _lowerCamelCase ) _snake_case = key or self.__key or 1 # make sure key is an appropriate size key %= 255 return [chr(ord(_lowerCamelCase ) ^ key ) for ch in content] def lowercase ( self : Union[str, Any] , _lowerCamelCase : str , _lowerCamelCase : int = 0 ): assert isinstance(_lowerCamelCase , _lowerCamelCase ) and isinstance(_lowerCamelCase , _lowerCamelCase ) _snake_case = key or self.__key or 1 # make sure key can be any size while key > 255: key -= 255 # This will be returned _snake_case = '''''' for ch in content: ans += chr(ord(_lowerCamelCase ) ^ key ) return ans def lowercase ( self : List[str] , _lowerCamelCase : str , _lowerCamelCase : int = 0 ): assert isinstance(_lowerCamelCase , _lowerCamelCase ) and isinstance(_lowerCamelCase , _lowerCamelCase ) _snake_case = key or self.__key or 1 # make sure key can be any size while key > 255: key -= 255 # This will be returned _snake_case = '''''' for ch in content: ans += chr(ord(_lowerCamelCase ) ^ key ) return ans def lowercase ( self : List[Any] , _lowerCamelCase : str , _lowerCamelCase : int = 0 ): assert isinstance(_lowerCamelCase , _lowerCamelCase ) and isinstance(_lowerCamelCase , _lowerCamelCase ) try: with open(_lowerCamelCase ) as fin, open('''encrypt.out''' , '''w+''' ) as fout: # actual encrypt-process for line in fin: fout.write(self.encrypt_string(_lowerCamelCase , _lowerCamelCase ) ) except OSError: return False return True def lowercase ( self : Optional[Any] , _lowerCamelCase : str , _lowerCamelCase : int ): assert isinstance(_lowerCamelCase , _lowerCamelCase ) and isinstance(_lowerCamelCase , _lowerCamelCase ) try: with open(_lowerCamelCase ) as fin, open('''decrypt.out''' , '''w+''' ) as fout: # actual encrypt-process for line in fin: fout.write(self.decrypt_string(_lowerCamelCase , _lowerCamelCase ) ) except OSError: return False return True # Tests # crypt = XORCipher() # key = 67 # # test encrypt # print(crypt.encrypt("hallo welt",key)) # # test decrypt # print(crypt.decrypt(crypt.encrypt("hallo welt",key), key)) # # test encrypt_string # print(crypt.encrypt_string("hallo welt",key)) # # test decrypt_string # print(crypt.decrypt_string(crypt.encrypt_string("hallo welt",key),key)) # if (crypt.encrypt_file("test.txt",key)): # print("encrypt successful") # else: # print("encrypt unsuccessful") # if (crypt.decrypt_file("encrypt.out",key)): # print("decrypt successful") # else: # print("decrypt unsuccessful")
40
1
'''simple docstring''' import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer, DPRReaderTokenizer SCREAMING_SNAKE_CASE_: List[str] =logging.get_logger(__name__) SCREAMING_SNAKE_CASE_: Union[str, Any] ={"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} SCREAMING_SNAKE_CASE_: Optional[int] ={ """vocab_file""": { """facebook/dpr-ctx_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt""" ), """facebook/dpr-ctx_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """facebook/dpr-ctx_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json""" ), """facebook/dpr-ctx_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json""" ), }, } SCREAMING_SNAKE_CASE_: Union[str, Any] ={ """vocab_file""": { """facebook/dpr-question_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt""" ), """facebook/dpr-question_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """facebook/dpr-question_encoder-single-nq-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json""" ), """facebook/dpr-question_encoder-multiset-base""": ( """https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json""" ), }, } SCREAMING_SNAKE_CASE_: List[str] ={ """vocab_file""": { """facebook/dpr-reader-single-nq-base""": ( """https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt""" ), """facebook/dpr-reader-multiset-base""": ( """https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """facebook/dpr-reader-single-nq-base""": ( """https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json""" ), """facebook/dpr-reader-multiset-base""": ( """https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json""" ), }, } SCREAMING_SNAKE_CASE_: str ={ """facebook/dpr-ctx_encoder-single-nq-base""": 5_12, """facebook/dpr-ctx_encoder-multiset-base""": 5_12, } SCREAMING_SNAKE_CASE_: str ={ """facebook/dpr-question_encoder-single-nq-base""": 5_12, """facebook/dpr-question_encoder-multiset-base""": 5_12, } SCREAMING_SNAKE_CASE_: List[Any] ={ """facebook/dpr-reader-single-nq-base""": 5_12, """facebook/dpr-reader-multiset-base""": 5_12, } SCREAMING_SNAKE_CASE_: int ={ """facebook/dpr-ctx_encoder-single-nq-base""": {"""do_lower_case""": True}, """facebook/dpr-ctx_encoder-multiset-base""": {"""do_lower_case""": True}, } SCREAMING_SNAKE_CASE_: Tuple ={ """facebook/dpr-question_encoder-single-nq-base""": {"""do_lower_case""": True}, """facebook/dpr-question_encoder-multiset-base""": {"""do_lower_case""": True}, } SCREAMING_SNAKE_CASE_: int ={ """facebook/dpr-reader-single-nq-base""": {"""do_lower_case""": True}, """facebook/dpr-reader-multiset-base""": {"""do_lower_case""": True}, } class __A ( UpperCAmelCase__ ): a__ : List[Any] = VOCAB_FILES_NAMES a__ : List[str] = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP a__ : int = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ : Tuple = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION a__ : Any = DPRContextEncoderTokenizer class __A ( UpperCAmelCase__ ): a__ : str = VOCAB_FILES_NAMES a__ : Optional[int] = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP a__ : Tuple = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ : List[str] = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION a__ : Dict = DPRQuestionEncoderTokenizer SCREAMING_SNAKE_CASE_: Dict =collections.namedtuple( 'DPRSpanPrediction', ['span_score', 'relevance_score', 'doc_id', 'start_index', 'end_index', 'text'] ) SCREAMING_SNAKE_CASE_: Optional[int] =collections.namedtuple('DPRReaderOutput', ['start_logits', 'end_logits', 'relevance_logits']) SCREAMING_SNAKE_CASE_: List[Any] =R""" Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`. It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers), using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)` with the format: [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids> Args: questions (`str` or `List[str]`): The questions to be encoded. You can specify one question for many passages. In this case, the question will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in `titles` or `texts`. titles (`str` or `List[str]`): The passages titles to be encoded. This can be a string or a list of strings if there are several passages. texts (`str` or `List[str]`): The passages texts to be encoded. This can be a string or a list of strings if there are several passages. padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`): Activates and controls padding. Accepts the following values: - `True` or `'longest'`: Pad to the longest sequence in the batch (or no padding if only a single sequence if provided). - `'max_length'`: Pad to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. - `False` or `'do_not_pad'` (default): No padding (i.e., can output a batch with sequences of different lengths). truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`): Activates and controls truncation. Accepts the following values: - `True` or `'longest_first'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. This will truncate token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch of pairs) is provided. - `'only_first'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. This will only truncate the first sequence of a pair if a pair of sequences (or a batch of pairs) is provided. - `'only_second'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. This will only truncate the second sequence of a pair if a pair of sequences (or a batch of pairs) is provided. - `False` or `'do_not_truncate'` (default): No truncation (i.e., can output batch with sequence lengths greater than the model maximum admissible input size). max_length (`int`, *optional*): Controls the maximum length to use by one of the truncation/padding parameters. If left unset or set to `None`, this will use the predefined model maximum length if a maximum length is required by one of the truncation/padding parameters. If the model has no specific maximum input length (like XLNet) truncation/padding to a maximum length will be deactivated. return_tensors (`str` or [`~utils.TensorType`], *optional*): If set, will return tensors instead of list of python integers. Acceptable values are: - `'tf'`: Return TensorFlow `tf.constant` objects. - `'pt'`: Return PyTorch `torch.Tensor` objects. - `'np'`: Return Numpy `np.ndarray` objects. return_attention_mask (`bool`, *optional*): Whether or not to return the attention mask. If not set, will return the attention mask according to the specific tokenizer's default, defined by the `return_outputs` attribute. [What are attention masks?](../glossary#attention-mask) Return: `Dict[str, List[List[int]]]`: A dictionary with the following keys: - `input_ids`: List of token ids to be fed to a model. - `attention_mask`: List of indices specifying which tokens should be attended to by the model. """ @add_start_docstrings(UpperCAmelCase__ ) class __A : def __call__(self : str , __a : Any , __a : str = None , __a : Tuple = None , __a : Optional[int] = False , __a : Tuple = False , __a : Any = None , __a : Any = None , __a : Tuple = None , **__a : Optional[int] , ): if titles is None and texts is None: return super().__call__( __lowercase , padding=__lowercase , truncation=__lowercase , max_length=__lowercase , return_tensors=__lowercase , return_attention_mask=__lowercase , **__lowercase , ) elif titles is None or texts is None: UpperCAmelCase_ = titles if texts is None else texts return super().__call__( __lowercase , __lowercase , padding=__lowercase , truncation=__lowercase , max_length=__lowercase , return_tensors=__lowercase , return_attention_mask=__lowercase , **__lowercase , ) UpperCAmelCase_ = titles if not isinstance(__lowercase , __lowercase ) else [titles] UpperCAmelCase_ = texts if not isinstance(__lowercase , __lowercase ) else [texts] UpperCAmelCase_ = len(__lowercase ) UpperCAmelCase_ = questions if not isinstance(__lowercase , __lowercase ) else [questions] * n_passages assert len(__lowercase ) == len( __lowercase ), f"""There should be as many titles than texts but got {len(__lowercase )} titles and {len(__lowercase )} texts.""" UpperCAmelCase_ = super().__call__(__lowercase , __lowercase , padding=__lowercase , truncation=__lowercase )['''input_ids'''] UpperCAmelCase_ = super().__call__(__lowercase , add_special_tokens=__lowercase , padding=__lowercase , truncation=__lowercase )['''input_ids'''] UpperCAmelCase_ = { '''input_ids''': [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(__lowercase , __lowercase ) ] } if return_attention_mask is not False: UpperCAmelCase_ = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) UpperCAmelCase_ = attention_mask return self.pad(__lowercase , padding=__lowercase , max_length=__lowercase , return_tensors=__lowercase ) def _lowercase (self : Union[str, Any] , __a : Dict , __a : Optional[int] , __a : List[str] = 16 , __a : List[Any] = 64 , __a : int = 4 , ): UpperCAmelCase_ = reader_input['''input_ids'''] UpperCAmelCase_ = reader_output[:3] UpperCAmelCase_ = len(__lowercase ) UpperCAmelCase_ = sorted(range(__lowercase ) , reverse=__lowercase , key=relevance_logits.__getitem__ ) UpperCAmelCase_ = [] for doc_id in sorted_docs: UpperCAmelCase_ = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence UpperCAmelCase_ = sequence_ids.index(self.sep_token_id , 2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: UpperCAmelCase_ = sequence_ids.index(self.pad_token_id ) else: UpperCAmelCase_ = len(__lowercase ) UpperCAmelCase_ = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] , end_logits=end_logits[doc_id][passage_offset:sequence_len] , max_answer_length=__lowercase , top_spans=__lowercase , ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] , relevance_score=relevance_logits[doc_id] , doc_id=__lowercase , start_index=__lowercase , end_index=__lowercase , text=self.decode(sequence_ids[start_index : end_index + 1] ) , ) ) if len(__lowercase ) >= num_spans: break return nbest_spans_predictions[:num_spans] def _lowercase (self : Dict , __a : Union[str, Any] , __a : Any , __a : List[str] , __a : Dict , ): UpperCAmelCase_ = [] for start_index, start_score in enumerate(__lowercase ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) UpperCAmelCase_ = sorted(__lowercase , key=lambda __a : x[1] , reverse=__lowercase ) UpperCAmelCase_ = [] for (start_index, end_index), score in scores: assert start_index <= end_index, f"""Wrong span indices: [{start_index}:{end_index}]""" UpperCAmelCase_ = end_index - start_index + 1 assert length <= max_answer_length, f"""Span is too long: {length} > {max_answer_length}""" if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(__lowercase ) == top_spans: break return chosen_span_intervals @add_end_docstrings(UpperCAmelCase__ ) class __A ( UpperCAmelCase__ , UpperCAmelCase__ ): a__ : Optional[int] = VOCAB_FILES_NAMES a__ : Tuple = READER_PRETRAINED_VOCAB_FILES_MAP a__ : Any = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ : Optional[int] = READER_PRETRAINED_INIT_CONFIGURATION a__ : Union[str, Any] = ["""input_ids""", """attention_mask"""] a__ : Optional[Any] = DPRReaderTokenizer
1
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def lowerCAmelCase ( )-> int: lowerCAmelCase_ : int = ArgumentParser('''Transformers CLI tool''' , usage='''transformers-cli <command> [<args>]''' ) lowerCAmelCase_ : Dict = parser.add_subparsers(help='''transformers-cli command helpers''' ) # Register commands ConvertCommand.register_subcommand(lowerCAmelCase_ ) DownloadCommand.register_subcommand(lowerCAmelCase_ ) EnvironmentCommand.register_subcommand(lowerCAmelCase_ ) RunCommand.register_subcommand(lowerCAmelCase_ ) ServeCommand.register_subcommand(lowerCAmelCase_ ) UserCommands.register_subcommand(lowerCAmelCase_ ) AddNewModelCommand.register_subcommand(lowerCAmelCase_ ) AddNewModelLikeCommand.register_subcommand(lowerCAmelCase_ ) LfsCommands.register_subcommand(lowerCAmelCase_ ) PTtoTFCommand.register_subcommand(lowerCAmelCase_ ) # Let's go lowerCAmelCase_ : Union[str, Any] = parser.parse_args() if not hasattr(lowerCAmelCase_ , '''func''' ): parser.print_help() exit(1 ) # Run lowerCAmelCase_ : List[Any] = args.func(lowerCAmelCase_ ) service.run() if __name__ == "__main__": main()
262
0
'''simple docstring''' import argparse from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import BigBirdPegasusConfig, BigBirdPegasusForConditionalGeneration UpperCAmelCase : Any = [ # tf -> hf ('/', '.'), ('layer_', 'layers.'), ('kernel', 'weight'), ('beta', 'bias'), ('gamma', 'weight'), ('pegasus', 'model'), ] UpperCAmelCase : Optional[Any] = [ ('.output.dense', '.fc2'), ('intermediate.LayerNorm', 'final_layer_norm'), ('intermediate.dense', 'fc1'), ] UpperCAmelCase : Optional[int] = ( INIT_COMMON + [ ('attention.self.LayerNorm', 'self_attn_layer_norm'), ('attention.output.dense', 'self_attn.out_proj'), ('attention.self', 'self_attn'), ('attention.encdec.LayerNorm', 'encoder_attn_layer_norm'), ('attention.encdec_output.dense', 'encoder_attn.out_proj'), ('attention.encdec', 'encoder_attn'), ('key', 'k_proj'), ('value', 'v_proj'), ('query', 'q_proj'), ('decoder.LayerNorm', 'decoder.layernorm_embedding'), ] + END_COMMON ) UpperCAmelCase : List[str] = ( INIT_COMMON + [ ('embeddings.word_embeddings', 'shared.weight'), ('embeddings.position_embeddings', 'embed_positions.weight'), ('attention.self.LayerNorm', 'self_attn_layer_norm'), ('attention.output.dense', 'self_attn.output'), ('attention.self', 'self_attn.self'), ('encoder.LayerNorm', 'encoder.layernorm_embedding'), ] + END_COMMON ) UpperCAmelCase : List[Any] = [ 'encdec/key/bias', 'encdec/query/bias', 'encdec/value/bias', 'self/key/bias', 'self/query/bias', 'self/value/bias', 'encdec_output/dense/bias', 'attention/output/dense/bias', ] def a__ ( a__ , a__ ): """simple docstring""" for tf_name, hf_name in patterns: __SCREAMING_SNAKE_CASE = k.replace(a__ , a__ ) return k def a__ ( a__ , a__ ): """simple docstring""" __SCREAMING_SNAKE_CASE = BigBirdPegasusConfig(**a__ ) __SCREAMING_SNAKE_CASE = BigBirdPegasusForConditionalGeneration(a__ ) __SCREAMING_SNAKE_CASE = torch_model.state_dict() __SCREAMING_SNAKE_CASE = {} # separating decoder weights __SCREAMING_SNAKE_CASE = {k: tf_weights[k] for k in tf_weights if k.startswith("""pegasus/decoder""" )} __SCREAMING_SNAKE_CASE = {k: tf_weights[k] for k in tf_weights if not k.startswith("""pegasus/decoder""" )} for k, v in tqdm(decoder_weights.items() , """tf -> hf conversion""" ): __SCREAMING_SNAKE_CASE = [k.endswith(a__ ) for ending in KEYS_TO_IGNORE] if any(a__ ): continue __SCREAMING_SNAKE_CASE = DECODER_PATTERNS __SCREAMING_SNAKE_CASE = rename_state_dict_key(a__ , a__ ) if new_k not in state_dict: raise ValueError(F'could not find new key {new_k} in state dict. (converted from {k})' ) if any(True if i in k else False for i in ["""dense""", """query""", """key""", """value"""] ): __SCREAMING_SNAKE_CASE = v.T __SCREAMING_SNAKE_CASE = torch.from_numpy(a__ ) assert v.shape == state_dict[new_k].shape, F'{new_k}, {k}, {v.shape}, {state_dict[new_k].shape}' for k, v in tqdm(remaining_weights.items() , """tf -> hf conversion""" ): __SCREAMING_SNAKE_CASE = [k.endswith(a__ ) for ending in KEYS_TO_IGNORE] if any(a__ ): continue __SCREAMING_SNAKE_CASE = REMAINING_PATTERNS __SCREAMING_SNAKE_CASE = rename_state_dict_key(a__ , a__ ) if new_k not in state_dict and k != "pegasus/embeddings/position_embeddings": raise ValueError(F'could not find new key {new_k} in state dict. (converted from {k})' ) if any(True if i in k else False for i in ["""dense""", """query""", """key""", """value"""] ): __SCREAMING_SNAKE_CASE = v.T __SCREAMING_SNAKE_CASE = torch.from_numpy(a__ ) if k != "pegasus/embeddings/position_embeddings": assert v.shape == state_dict[new_k].shape, F'{new_k}, {k}, {v.shape}, {state_dict[new_k].shape}' __SCREAMING_SNAKE_CASE = mapping["""model.embed_positions.weight"""] __SCREAMING_SNAKE_CASE = mapping.pop("""model.embed_positions.weight""" ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = torch_model.load_state_dict(a__ , strict=a__ ) __SCREAMING_SNAKE_CASE = [ k for k in missing if k not in [ """final_logits_bias""", """model.encoder.embed_tokens.weight""", """model.decoder.embed_tokens.weight""", """lm_head.weight""", ] ] assert unexpected_missing == [], F'no matches found for the following torch keys {unexpected_missing}' assert extra == [], F'no matches found for the following tf keys {extra}' return torch_model def a__ ( a__ ): """simple docstring""" __SCREAMING_SNAKE_CASE = tf.train.list_variables(a__ ) __SCREAMING_SNAKE_CASE = {} __SCREAMING_SNAKE_CASE = ["""global_step"""] for name, shape in tqdm(a__ , desc="""converting tf checkpoint to dict""" ): __SCREAMING_SNAKE_CASE = any(pat in name for pat in ignore_name ) if skip_key: continue __SCREAMING_SNAKE_CASE = tf.train.load_variable(a__ , a__ ) __SCREAMING_SNAKE_CASE = array return tf_weights def a__ ( a__ , a__ , a__ ): """simple docstring""" __SCREAMING_SNAKE_CASE = get_tf_weights_as_numpy(a__ ) __SCREAMING_SNAKE_CASE = convert_bigbird_pegasus(a__ , a__ ) torch_model.save_pretrained(a__ ) if __name__ == "__main__": UpperCAmelCase : Any = argparse.ArgumentParser() parser.add_argument('--tf_ckpt_path', type=str, help='passed to tf.train.list_variables') parser.add_argument('--save_dir', default=None, type=str, help='Path to the output PyTorch model.') UpperCAmelCase : int = parser.parse_args() UpperCAmelCase : Dict = {} convert_bigbird_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir, config_update=config_update)
331
'''simple docstring''' import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roformer.modeling_flax_roformer import ( FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, ) class lowerCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __init__( self : Tuple , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Optional[Any]=13 , __SCREAMING_SNAKE_CASE : Any=7 , __SCREAMING_SNAKE_CASE : Optional[Any]=True , __SCREAMING_SNAKE_CASE : List[str]=True , __SCREAMING_SNAKE_CASE : Optional[Any]=True , __SCREAMING_SNAKE_CASE : List[str]=True , __SCREAMING_SNAKE_CASE : List[Any]=99 , __SCREAMING_SNAKE_CASE : Union[str, Any]=32 , __SCREAMING_SNAKE_CASE : Dict=5 , __SCREAMING_SNAKE_CASE : str=4 , __SCREAMING_SNAKE_CASE : Tuple=37 , __SCREAMING_SNAKE_CASE : List[Any]="gelu" , __SCREAMING_SNAKE_CASE : Tuple=0.1 , __SCREAMING_SNAKE_CASE : Optional[int]=0.1 , __SCREAMING_SNAKE_CASE : Optional[Any]=512 , __SCREAMING_SNAKE_CASE : Optional[Any]=16 , __SCREAMING_SNAKE_CASE : Optional[Any]=2 , __SCREAMING_SNAKE_CASE : Tuple=0.02 , __SCREAMING_SNAKE_CASE : List[Any]=4 , ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = seq_length __SCREAMING_SNAKE_CASE = is_training __SCREAMING_SNAKE_CASE = use_attention_mask __SCREAMING_SNAKE_CASE = use_token_type_ids __SCREAMING_SNAKE_CASE = use_labels __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = max_position_embeddings __SCREAMING_SNAKE_CASE = type_vocab_size __SCREAMING_SNAKE_CASE = type_sequence_label_size __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = num_choices def UpperCAmelCase__ ( self : Dict ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __SCREAMING_SNAKE_CASE = None if self.use_attention_mask: __SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) __SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __SCREAMING_SNAKE_CASE = RoFormerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def UpperCAmelCase__ ( self : List[Any] ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = config_and_inputs __SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class lowerCAmelCase__ ( a , unittest.TestCase ): """simple docstring""" lowerCAmelCase__ = True lowerCAmelCase__ = ( ( FlaxRoFormerModel, FlaxRoFormerForMaskedLM, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, ) if is_flax_available() else () ) def UpperCAmelCase__ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = FlaxRoFormerModelTester(self ) @slow def UpperCAmelCase__ ( self : int ) -> Any: """simple docstring""" for model_class_name in self.all_model_classes: __SCREAMING_SNAKE_CASE = model_class_name.from_pretrained("""junnyu/roformer_chinese_small""" , from_pt=__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = model(np.ones((1, 1) ) ) self.assertIsNotNone(__SCREAMING_SNAKE_CASE ) @require_flax class lowerCAmelCase__ ( unittest.TestCase ): """simple docstring""" @slow def UpperCAmelCase__ ( self : Any ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = FlaxRoFormerForMaskedLM.from_pretrained("""junnyu/roformer_chinese_base""" ) __SCREAMING_SNAKE_CASE = jnp.array([[0, 1, 2, 3, 4, 5]] ) __SCREAMING_SNAKE_CASE = model(__SCREAMING_SNAKE_CASE )[0] __SCREAMING_SNAKE_CASE = 50_000 __SCREAMING_SNAKE_CASE = (1, 6, vocab_size) self.assertEqual(output.shape , __SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = jnp.array( [[[-0.1205, -1.0265, 0.2922], [-1.5134, 0.1974, 0.1519], [-5.0135, -3.9003, -0.8404]]] ) self.assertTrue(jnp.allclose(output[:, :3, :3] , __SCREAMING_SNAKE_CASE , atol=1E-4 ) )
331
1
"""simple docstring""" import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_=13 , lowercase_=30 , lowercase_=2 , lowercase_=3 , lowercase_=True , lowercase_=True , lowercase_=32 , lowercase_=5 , lowercase_=4 , lowercase_=37 , lowercase_="gelu" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=10 , lowercase_=0.02 , lowercase_=3 , lowercase_=None , lowercase_=2 , ): """simple docstring""" UpperCAmelCase_ : List[str] = parent UpperCAmelCase_ : int = batch_size UpperCAmelCase_ : int = image_size UpperCAmelCase_ : List[Any] = patch_size UpperCAmelCase_ : Any = num_channels UpperCAmelCase_ : Optional[int] = is_training UpperCAmelCase_ : Union[str, Any] = use_labels UpperCAmelCase_ : Union[str, Any] = hidden_size UpperCAmelCase_ : str = num_hidden_layers UpperCAmelCase_ : List[str] = num_attention_heads UpperCAmelCase_ : str = intermediate_size UpperCAmelCase_ : str = hidden_act UpperCAmelCase_ : List[Any] = hidden_dropout_prob UpperCAmelCase_ : Union[str, Any] = attention_probs_dropout_prob UpperCAmelCase_ : str = type_sequence_label_size UpperCAmelCase_ : str = initializer_range UpperCAmelCase_ : Union[str, Any] = scope UpperCAmelCase_ : str = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase_ : int = (image_size // patch_size) ** 2 UpperCAmelCase_ : Optional[Any] = num_patches + 2 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ : Tuple = None if self.use_labels: UpperCAmelCase_ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase_ : Union[str, Any] = self.get_config() return config, pixel_values, labels def UpperCamelCase__ ( self ): """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowercase_ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = DeiTModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() UpperCAmelCase_ : List[Any] = model(lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = DeiTForMaskedImageModeling(config=lowercase_ ) model.to(lowercase_ ) model.eval() UpperCAmelCase_ : List[Any] = model(lowercase_ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase_ : List[str] = 1 UpperCAmelCase_ : Optional[Any] = DeiTForMaskedImageModeling(lowercase_ ) model.to(lowercase_ ) model.eval() UpperCAmelCase_ : Dict = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase_ : Optional[int] = model(lowercase_ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Tuple = self.type_sequence_label_size UpperCAmelCase_ : Union[str, Any] = DeiTForImageClassification(lowercase_ ) model.to(lowercase_ ) model.eval() UpperCAmelCase_ : List[str] = model(lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase_ : Union[str, Any] = 1 UpperCAmelCase_ : Optional[int] = DeiTForImageClassification(lowercase_ ) model.to(lowercase_ ) model.eval() UpperCAmelCase_ : Dict = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase_ : List[Any] = model(lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[str] = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ) : Dict = config_and_inputs UpperCAmelCase_ : Optional[int] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class A_ (lowercase__ ,lowercase__ ,unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE__ : Tuple = ( { """feature-extraction""": DeiTModel, """image-classification""": (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE__ : List[Any] = False SCREAMING_SNAKE_CASE__ : Optional[Any] = False SCREAMING_SNAKE_CASE__ : List[str] = False def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Any = DeiTModelTester(self ) UpperCAmelCase_ : Optional[int] = ConfigTester(self , config_class=lowercase_ , has_text_modality=lowercase_ , hidden_size=37 ) def UpperCamelCase__ ( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="DeiT does not use inputs_embeds" ) def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : List[Any] = model_class(lowercase_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCAmelCase_ : Any = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase_ , nn.Linear ) ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : Dict = model_class(lowercase_ ) UpperCAmelCase_ : Optional[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ : str = [*signature.parameters.keys()] UpperCAmelCase_ : Optional[int] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_=False ): """simple docstring""" UpperCAmelCase_ : Tuple = super()._prepare_for_class(lowercase_ , lowercase_ , return_labels=lowercase_ ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def UpperCamelCase__ ( self ): """simple docstring""" if not self.model_tester.is_training: return UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Union[str, Any] = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowercase_ ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue UpperCAmelCase_ : Optional[int] = model_class(lowercase_ ) model.to(lowercase_ ) model.train() UpperCAmelCase_ : List[Any] = self._prepare_for_class(lowercase_ , lowercase_ , return_labels=lowercase_ ) UpperCAmelCase_ : Dict = model(**lowercase_ ).loss loss.backward() def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : int = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return UpperCAmelCase_ : Dict = False UpperCAmelCase_ : Optional[int] = True for model_class in self.all_model_classes: if model_class in get_values(lowercase_ ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue UpperCAmelCase_ : List[str] = model_class(lowercase_ ) model.gradient_checkpointing_enable() model.to(lowercase_ ) model.train() UpperCAmelCase_ : Optional[int] = self._prepare_for_class(lowercase_ , lowercase_ , return_labels=lowercase_ ) UpperCAmelCase_ : Any = model(**lowercase_ ).loss loss.backward() def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ : Dict = [ {"title": "multi_label_classification", "num_labels": 2, "dtype": torch.float}, {"title": "single_label_classification", "num_labels": 1, "dtype": torch.long}, {"title": "regression", "num_labels": 1, "dtype": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowercase_ ), *get_values(lowercase_ ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F"""Testing {model_class} with {problem_type["title"]}""" ): UpperCAmelCase_ : str = problem_type["title"] UpperCAmelCase_ : List[Any] = problem_type["num_labels"] UpperCAmelCase_ : Union[str, Any] = model_class(lowercase_ ) model.to(lowercase_ ) model.train() UpperCAmelCase_ : int = self._prepare_for_class(lowercase_ , lowercase_ , return_labels=lowercase_ ) if problem_type["num_labels"] > 1: UpperCAmelCase_ : List[Any] = inputs["labels"].unsqueeze(1 ).repeat(1 , problem_type["num_labels"] ) UpperCAmelCase_ : Tuple = inputs["labels"].to(problem_type["dtype"] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowercase_ ) as warning_list: UpperCAmelCase_ : List[str] = model(**lowercase_ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F"""Something is going wrong in the regression problem: intercepted {w.message}""" ) loss.backward() @slow def UpperCamelCase__ ( self ): """simple docstring""" for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : Union[str, Any] = DeiTModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) def __a ( ): UpperCAmelCase_ : Any = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class A_ (unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase__ ( self ): """simple docstring""" return ( DeiTImageProcessor.from_pretrained("facebook/deit-base-distilled-patch16-224" ) if is_vision_available() else None ) @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = DeiTForImageClassificationWithTeacher.from_pretrained("facebook/deit-base-distilled-patch16-224" ).to( lowercase_ ) UpperCAmelCase_ : List[str] = self.default_image_processor UpperCAmelCase_ : List[str] = prepare_img() UpperCAmelCase_ : int = image_processor(images=lowercase_ , return_tensors="pt" ).to(lowercase_ ) # forward pass with torch.no_grad(): UpperCAmelCase_ : Dict = model(**lowercase_ ) # verify the logits UpperCAmelCase_ : List[str] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowercase_ ) UpperCAmelCase_ : str = torch.tensor([-1.02_66, 0.19_12, -1.28_61] ).to(lowercase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase_ , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[str] = DeiTModel.from_pretrained( "facebook/deit-base-distilled-patch16-224" , torch_dtype=torch.floataa , device_map="auto" ) UpperCAmelCase_ : str = self.default_image_processor UpperCAmelCase_ : Union[str, Any] = prepare_img() UpperCAmelCase_ : List[Any] = image_processor(images=lowercase_ , return_tensors="pt" ) UpperCAmelCase_ : List[str] = inputs.pixel_values.to(lowercase_ ) # forward pass to make sure inference works in fp16 with torch.no_grad(): UpperCAmelCase_ : int = model(lowercase_ )
61
"""simple docstring""" import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder __A = "base_with_context" def UpperCamelCase__ ( lowercase__ : Optional[Any] , lowercase__ : List[Any] ): snake_case : Dict = nn.Parameter(torch.FloatTensor(weights["token_embedder"]["embedding"] ) ) snake_case : Tuple = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ) , requires_grad=lowercase__ ) for lyr_num, lyr in enumerate(model.encoders ): snake_case : Tuple = weights[F'''layers_{lyr_num}'''] snake_case : List[Any] = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) snake_case : List[Any] = ly_weight["attention"] snake_case : List[Any] = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) snake_case : List[Any] = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) snake_case : Any = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) snake_case : List[str] = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) snake_case : Tuple = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) snake_case : Dict = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) snake_case : str = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) snake_case : Union[str, Any] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) snake_case : List[Any] = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def UpperCamelCase__ ( lowercase__ : Tuple , lowercase__ : List[Any] ): snake_case : Optional[Any] = nn.Parameter(torch.FloatTensor(weights["input_proj"]["kernel"].T ) ) snake_case : Union[str, Any] = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ) , requires_grad=lowercase__ ) for lyr_num, lyr in enumerate(model.encoders ): snake_case : str = weights[F'''layers_{lyr_num}'''] snake_case : Any = ly_weight["attention"] snake_case : Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) snake_case : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) snake_case : Dict = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) snake_case : int = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) snake_case : List[Any] = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) snake_case : Optional[int] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) snake_case : str = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) snake_case : Optional[Any] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) snake_case : Optional[Any] = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) snake_case : int = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def UpperCamelCase__ ( lowercase__ : str , lowercase__ : Union[str, Any] ): snake_case : int = nn.Parameter(torch.FloatTensor(weights["time_emb_dense0"]["kernel"].T ) ) snake_case : List[Any] = nn.Parameter(torch.FloatTensor(weights["time_emb_dense1"]["kernel"].T ) ) snake_case : Tuple = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ) , requires_grad=lowercase__ ) snake_case : Tuple = nn.Parameter( torch.FloatTensor(weights["continuous_inputs_projection"]["kernel"].T ) ) for lyr_num, lyr in enumerate(model.decoders ): snake_case : Union[str, Any] = weights[F'''layers_{lyr_num}'''] snake_case : List[str] = nn.Parameter( torch.FloatTensor(ly_weight["pre_self_attention_layer_norm"]["scale"] ) ) snake_case : Any = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_0"]["DenseGeneral_0"]["kernel"].T ) ) snake_case : Union[str, Any] = ly_weight["self_attention"] snake_case : List[str] = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) snake_case : Any = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) snake_case : Optional[Any] = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) snake_case : str = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) snake_case : List[str] = ly_weight["MultiHeadDotProductAttention_0"] snake_case : Tuple = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) snake_case : Dict = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) snake_case : str = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) snake_case : Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) snake_case : Union[str, Any] = nn.Parameter( torch.FloatTensor(ly_weight["pre_cross_attention_layer_norm"]["scale"] ) ) snake_case : Dict = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) snake_case : Tuple = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_1"]["DenseGeneral_0"]["kernel"].T ) ) snake_case : Tuple = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) snake_case : Any = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) snake_case : str = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) snake_case : List[str] = nn.Parameter(torch.FloatTensor(weights["decoder_norm"]["scale"] ) ) snake_case : List[Any] = nn.Parameter(torch.FloatTensor(weights["spec_out_dense"]["kernel"].T ) ) return model def UpperCamelCase__ ( lowercase__ : Any ): snake_case : Union[str, Any] = checkpoints.load_tax_checkpoint(args.checkpoint_path ) snake_case : List[Any] = jnp.tree_util.tree_map(onp.array , lowercase__ ) snake_case : Tuple = [ "from __gin__ import dynamic_registration", "from music_spectrogram_diffusion.models.diffusion import diffusion_utils", "diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0", "diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()", ] snake_case : List[str] = os.path.join(args.checkpoint_path , ".." , "config.gin" ) snake_case : List[str] = inference.parse_training_gin_file(lowercase__ , lowercase__ ) snake_case : List[Any] = inference.InferenceModel(args.checkpoint_path , lowercase__ ) snake_case : str = DDPMScheduler(beta_schedule="squaredcos_cap_v2" , variance_type="fixed_large" ) snake_case : int = SpectrogramNotesEncoder( max_length=synth_model.sequence_length["inputs"] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj="gated-gelu" , ) snake_case : Tuple = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length["targets_context"] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj="gated-gelu" , ) snake_case : str = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length["targets_context"] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) snake_case : Optional[int] = load_notes_encoder(ta_checkpoint["target"]["token_encoder"] , lowercase__ ) snake_case : Any = load_continuous_encoder(ta_checkpoint["target"]["continuous_encoder"] , lowercase__ ) snake_case : List[Any] = load_decoder(ta_checkpoint["target"]["decoder"] , lowercase__ ) snake_case : int = OnnxRuntimeModel.from_pretrained("kashif/soundstream_mel_decoder" ) snake_case : Tuple = SpectrogramDiffusionPipeline( notes_encoder=lowercase__ , continuous_encoder=lowercase__ , decoder=lowercase__ , scheduler=lowercase__ , melgan=lowercase__ , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": __A = argparse.ArgumentParser() parser.add_argument("--output_path", default=None, type=str, required=True, help="Path to the converted model.") parser.add_argument( "--save", default=True, type=bool, required=False, help="Whether to save the converted model or not." ) parser.add_argument( "--checkpoint_path", default=f'{MODEL}/checkpoint_500000', type=str, required=False, help="Path to the original jax model checkpoint.", ) __A = parser.parse_args() main(args)
148
0
from typing import List import jiwer import jiwer.transforms as tr from packaging import version import datasets from datasets.config import PY_VERSION if PY_VERSION < version.parse('3.8'): import importlib_metadata else: import importlib.metadata as importlib_metadata a_ = '' if version.parse(importlib_metadata.version('jiwer')) < version.parse('2.3.0'): class _lowercase ( tr.AbstractTransform ): def __init__( self : str , snake_case : str = " " ) -> List[str]: """simple docstring""" UpperCamelCase_ : Optional[Any] = sentence_delimiter def SCREAMING_SNAKE_CASE__ ( self : List[Any] , snake_case : str ) -> Optional[int]: """simple docstring""" return list(snake_case ) def SCREAMING_SNAKE_CASE__ ( self : List[str] , snake_case : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCamelCase_ : Tuple = [] for sent_idx, sentence in enumerate(snake_case ): chars.extend(self.process_string(snake_case ) ) if self.sentence_delimiter is not None and self.sentence_delimiter != "" and sent_idx < len(snake_case ) - 1: chars.append(self.sentence_delimiter ) return chars a_ = tr.Compose( [tr.RemoveMultipleSpaces(), tr.Strip(), SentencesToListOfCharacters(SENTENCE_DELIMITER)] ) else: a_ = tr.Compose( [ tr.RemoveMultipleSpaces(), tr.Strip(), tr.ReduceToSingleSentence(SENTENCE_DELIMITER), tr.ReduceToListOfListOfChars(), ] ) a_ = '\\n@inproceedings{inproceedings,\n author = {Morris, Andrew and Maier, Viktoria and Green, Phil},\n year = {2004},\n month = {01},\n pages = {},\n title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.}\n}\n' a_ = '\\nCharacter error rate (CER) is a common metric of the performance of an automatic speech recognition system.\n\nCER is similar to Word Error Rate (WER), but operates on character instead of word. Please refer to docs of WER for further information.\n\nCharacter error rate can be computed as:\n\nCER = (S + D + I) / N = (S + D + I) / (S + D + C)\n\nwhere\n\nS is the number of substitutions,\nD is the number of deletions,\nI is the number of insertions,\nC is the number of correct characters,\nN is the number of characters in the reference (N=S+D+C).\n\nCER\'s output is not always a number between 0 and 1, in particular when there is a high number of insertions. This value is often associated to the percentage of characters that were incorrectly predicted. The lower the value, the better the\nperformance of the ASR system with a CER of 0 being a perfect score.\n' a_ = '\nComputes CER score of transcribed segments against references.\nArgs:\n references: list of references for each speech input.\n predictions: list of transcribtions to score.\n concatenate_texts: Whether or not to concatenate sentences before evaluation, set to True for more accurate result.\nReturns:\n (float): the character error rate\n\nExamples:\n\n >>> predictions = ["this is the prediction", "there is an other sample"]\n >>> references = ["this is the reference", "there is another one"]\n >>> cer = datasets.load_metric("cer")\n >>> cer_score = cer.compute(predictions=predictions, references=references)\n >>> print(cer_score)\n 0.34146341463414637\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowercase ( datasets.Metric ): def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/jitsi/jiwer/'] , reference_urls=[ 'https://en.wikipedia.org/wiki/Word_error_rate', 'https://sites.google.com/site/textdigitisation/qualitymeasures/computingerrorrates', ] , ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , snake_case : Tuple , snake_case : Tuple , snake_case : List[str]=False ) -> Any: """simple docstring""" if concatenate_texts: return jiwer.compute_measures( snake_case , snake_case , truth_transform=snake_case , hypothesis_transform=snake_case , )["wer"] UpperCamelCase_ : Dict = 0 UpperCamelCase_ : Union[str, Any] = 0 for prediction, reference in zip(snake_case , snake_case ): UpperCamelCase_ : str = jiwer.compute_measures( snake_case , snake_case , truth_transform=snake_case , hypothesis_transform=snake_case , ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
351
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class _lowercase : def __init__( self : List[Any] , snake_case : int , snake_case : Any=9_9 , snake_case : Tuple=1_3 , snake_case : str=7 , snake_case : List[str]=9 , snake_case : Optional[Any]=True , snake_case : Any=True , snake_case : Optional[Any]=False , snake_case : List[str]=3_2 , snake_case : str=5 , snake_case : Any=4 , snake_case : List[str]=3_7 , snake_case : Optional[Any]=8 , snake_case : Optional[Any]=0.1 , snake_case : Dict=0.002 , snake_case : Any=1 , snake_case : Optional[int]=0 , snake_case : List[str]=0 , snake_case : List[str]=None , snake_case : List[str]=None , ) -> Optional[int]: """simple docstring""" UpperCamelCase_ : int = parent UpperCamelCase_ : List[Any] = batch_size UpperCamelCase_ : int = encoder_seq_length UpperCamelCase_ : int = decoder_seq_length # For common tests UpperCamelCase_ : List[Any] = self.decoder_seq_length UpperCamelCase_ : Optional[Any] = is_training UpperCamelCase_ : Tuple = use_attention_mask UpperCamelCase_ : int = use_labels UpperCamelCase_ : List[str] = vocab_size UpperCamelCase_ : Dict = hidden_size UpperCamelCase_ : Any = num_hidden_layers UpperCamelCase_ : Any = num_attention_heads UpperCamelCase_ : Dict = d_ff UpperCamelCase_ : List[Any] = relative_attention_num_buckets UpperCamelCase_ : List[Any] = dropout_rate UpperCamelCase_ : Dict = initializer_factor UpperCamelCase_ : Union[str, Any] = eos_token_id UpperCamelCase_ : Optional[int] = pad_token_id UpperCamelCase_ : List[str] = decoder_start_token_id UpperCamelCase_ : str = None UpperCamelCase_ : int = decoder_layers def SCREAMING_SNAKE_CASE__ ( self : List[Any] ) -> List[str]: """simple docstring""" return TaConfig.from_pretrained('google/umt5-base' ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , snake_case : Optional[int] , snake_case : Any , snake_case : Optional[int] , snake_case : Optional[int]=None , snake_case : List[Any]=None , snake_case : int=None , snake_case : Optional[int]=None , snake_case : Tuple=None , ) -> List[str]: """simple docstring""" if attention_mask is None: UpperCamelCase_ : Optional[Any] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: UpperCamelCase_ : Optional[int] = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: UpperCamelCase_ : Optional[int] = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=snake_case ) if decoder_head_mask is None: UpperCamelCase_ : Dict = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=snake_case ) if cross_attn_head_mask is None: UpperCamelCase_ : Optional[Any] = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=snake_case ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def SCREAMING_SNAKE_CASE__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" UpperCamelCase_ : Union[str, Any] = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) UpperCamelCase_ : Any = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input UpperCamelCase_ : Union[str, Any] = input_ids.clamp(self.pad_token_id + 1 ) UpperCamelCase_ : Any = decoder_input_ids.clamp(self.pad_token_id + 1 ) UpperCamelCase_ : Dict = self.get_config() UpperCamelCase_ : Dict = config.num_attention_heads UpperCamelCase_ : Optional[int] = self.prepare_inputs_dict(snake_case , snake_case , snake_case ) return config, input_dict def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCamelCase_, UpperCamelCase_ : Any = self.prepare_config_and_inputs() return config, inputs_dict def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> Dict: """simple docstring""" return TaConfig( vocab_size=1_6_6 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def SCREAMING_SNAKE_CASE__ ( self : int ) -> Optional[int]: """simple docstring""" return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def SCREAMING_SNAKE_CASE__ ( self : str , snake_case : Dict , snake_case : List[str] , snake_case : Tuple , snake_case : int , snake_case : List[str] , snake_case : Optional[Any] , ) -> Tuple: """simple docstring""" UpperCamelCase_ : int = UMTaModel(config=snake_case ) model.to(snake_case ) model.eval() UpperCamelCase_ : Any = model( input_ids=snake_case , decoder_input_ids=snake_case , attention_mask=snake_case , decoder_attention_mask=snake_case , ) UpperCamelCase_ : List[str] = model(input_ids=snake_case , decoder_input_ids=snake_case ) UpperCamelCase_ : Optional[Any] = result.last_hidden_state UpperCamelCase_ : Optional[Any] = result.past_key_values UpperCamelCase_ : Optional[int] = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(snake_case ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def SCREAMING_SNAKE_CASE__ ( self : Any , snake_case : Tuple , snake_case : List[Any] , snake_case : Optional[int] , snake_case : Any , snake_case : Tuple , snake_case : str , ) -> Union[str, Any]: """simple docstring""" UpperCamelCase_ : int = UMTaModel(config=snake_case ).get_decoder().to(snake_case ).eval() # first forward pass UpperCamelCase_ : str = model(snake_case , use_cache=snake_case ) UpperCamelCase_ : List[Any] = model(snake_case ) UpperCamelCase_ : Dict = model(snake_case , use_cache=snake_case ) self.parent.assertTrue(len(snake_case ) == len(snake_case ) ) self.parent.assertTrue(len(snake_case ) == len(snake_case ) + 1 ) UpperCamelCase_, UpperCamelCase_ : Optional[Any] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids UpperCamelCase_ : Any = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and UpperCamelCase_ : List[Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) UpperCamelCase_ : List[Any] = model(snake_case )['last_hidden_state'] UpperCamelCase_ : List[str] = model(snake_case , past_key_values=snake_case )['last_hidden_state'] # select random slice UpperCamelCase_ : List[str] = ids_tensor((1,) , output_from_past.shape[-1] ).item() UpperCamelCase_ : Union[str, Any] = output_from_no_past[:, -1, random_slice_idx].detach() UpperCamelCase_ : Any = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , snake_case : Tuple , snake_case : int , ) -> Union[str, Any]: """simple docstring""" UpperCamelCase_ : Optional[int] = UMTaModel(config=snake_case ).to(snake_case ).half().eval() UpperCamelCase_ : Union[str, Any] = model(**snake_case )['last_hidden_state'] self.parent.assertFalse(torch.isnan(snake_case ).any().item() ) @require_torch class _lowercase ( snake_case_ , snake_case_ , snake_case_ , unittest.TestCase ): lowercase = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) lowercase = (UMTaForConditionalGeneration,) if is_torch_available() else () lowercase = ( { 'conversational': UMTaForConditionalGeneration, 'feature-extraction': UMTaModel, 'summarization': UMTaForConditionalGeneration, 'text2text-generation': UMTaForConditionalGeneration, 'translation': UMTaForConditionalGeneration, 'question-answering': UMTaForQuestionAnswering, } if is_torch_available() else {} ) lowercase = True lowercase = False lowercase = False lowercase = True lowercase = True # The small UMT5 model needs higher percentages for CPU/MP tests lowercase = [0.8, 0.9] def SCREAMING_SNAKE_CASE__ ( self : Dict ) -> List[Any]: """simple docstring""" UpperCamelCase_ : Optional[Any] = UMTaModelTester(self ) @unittest.skip('Test has a segmentation fault on torch 1.8.0' ) def SCREAMING_SNAKE_CASE__ ( self : str ) -> List[str]: """simple docstring""" UpperCamelCase_ : List[str] = self.model_tester.prepare_config_and_inputs() UpperCamelCase_ : List[str] = UMTaModel(config_and_inputs[0] ).to(snake_case ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( snake_case , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , f"{tmpdirname}/t5_test.onnx" , export_params=snake_case , opset_version=9 , input_names=['input_ids', 'decoder_input_ids'] , ) @unittest.skipIf(torch_device == 'cpu' , 'Cant do half precision' ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" UpperCamelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] ) -> Optional[int]: """simple docstring""" UpperCamelCase_ : Tuple = ['encoder_attentions', 'decoder_attentions', 'cross_attentions'] UpperCamelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() UpperCamelCase_ : Union[str, Any] = config_and_inputs[0] UpperCamelCase_ : Tuple = UMTaForConditionalGeneration(snake_case ).eval() model.to(snake_case ) UpperCamelCase_ : str = { 'head_mask': torch.zeros(config.num_layers , config.num_heads , device=snake_case ), 'decoder_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=snake_case ), 'cross_attn_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=snake_case ), } for attn_name, (name, mask) in zip(snake_case , head_masking.items() ): UpperCamelCase_ : Optional[int] = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": UpperCamelCase_ : Union[str, Any] = torch.ones( config.num_decoder_layers , config.num_heads , device=snake_case ) UpperCamelCase_ : Any = model.generate( config_and_inputs[1]['input_ids'] , num_beams=1 , max_length=3 , output_attentions=snake_case , return_dict_in_generate=snake_case , **snake_case , ) # We check the state of decoder_attentions and cross_attentions just from the last step UpperCamelCase_ : int = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('Does not work on the tiny model as we keep hitting edge cases.' ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> Optional[Any]: """simple docstring""" pass @require_torch @require_sentencepiece @require_tokenizers class _lowercase ( unittest.TestCase ): @slow @unittest.skip( 'Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged' ) def SCREAMING_SNAKE_CASE__ ( self : Any ) -> List[Any]: """simple docstring""" UpperCamelCase_ : str = UMTaForConditionalGeneration.from_pretrained('google/umt5-small' , return_dict=snake_case ).to(snake_case ) UpperCamelCase_ : int = AutoTokenizer.from_pretrained('google/umt5-small' , use_fast=snake_case , legacy=snake_case ) UpperCamelCase_ : Dict = [ 'Bonjour monsieur <extra_id_0> bien <extra_id_1>.', 'No se como puedo <extra_id_0>.', 'This is the reason why we <extra_id_0> them.', 'The <extra_id_0> walks in <extra_id_1>, seats', 'A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.', ] UpperCamelCase_ : Dict = tokenizer(snake_case , return_tensors='pt' , padding=snake_case ).input_ids # fmt: off UpperCamelCase_ : List[str] = torch.tensor( [ [ 3_8_5_3_0, 2_1_0_7_0_3, 2_5_6_2_9_9, 1_4_1_0, 2_5_6_2_9_8, 2_7_4, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 8_2_6, 3_2_1, 6_7_1, 2_5_9_2_2, 2_5_6_2_9_9, 2_7_4, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1_4_6_0, 3_3_9, 3_1_2, 1_9_0_1_4, 1_0_6_2_0, 7_5_8, 2_5_6_2_9_9, 2_3_5_5,2_7_4, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 5_1_7, 2_5_6_2_9_9, 1_4_8_6_9, 2_8_1, 3_0_1, 2_5_6_2_9_8, 2_7_5, 1_1_9_9_8_3,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 3_2_0, 2_5_6_2_9_9, 1_4_8_6_9, 2_8_1, 2_2_3_4, 2_8_9, 2_2_7_5, 3_3_3,6_1_3_9_1, 2_8_9, 2_5_6_2_9_8, 5_4_3, 2_5_6_2_9_7, 1_6_8_7_1_4, 3_2_9, 2_5_6_2_9_6,2_7_4, 1], ] ) # fmt: on torch.testing.assert_allclose(snake_case , snake_case ) UpperCamelCase_ : int = model.generate(input_ids.to(snake_case ) ) UpperCamelCase_ : List[Any] = [ '<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>', '<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', ] UpperCamelCase_ : Dict = tokenizer.batch_decode(snake_case ) self.assertEqual(snake_case , snake_case )
50
0
"""simple docstring""" # This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny vocab first, and then a tiny model - so the outcome is truly tiny - # all files ~60KB. As compared to taking a full-size model, reducing to the minimum its layers and # emb dimensions, but keeping the full vocab + merges files, leading to ~3MB in total for all files. # The latter is done by `fsmt-make-super-tiny-model.py`. # # It will be used then as "stas/tiny-wmt19-en-ru" from pathlib import Path import json import tempfile from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES UpperCAmelCase__ = 'tiny-wmt19-en-ru' # Build # borrowed from a test UpperCAmelCase__ = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] UpperCAmelCase__ = dict(zip(vocab, range(len(vocab)))) UpperCAmelCase__ = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase__ = Path(tmpdirname) UpperCAmelCase__ = build_dir / VOCAB_FILES_NAMES['src_vocab_file'] UpperCAmelCase__ = build_dir / VOCAB_FILES_NAMES['tgt_vocab_file'] UpperCAmelCase__ = build_dir / VOCAB_FILES_NAMES['merges_file'] with open(src_vocab_file, 'w') as fp: fp.write(json.dumps(vocab_tokens)) with open(tgt_vocab_file, 'w') as fp: fp.write(json.dumps(vocab_tokens)) with open(merges_file, 'w') as fp: fp.write('\n'.join(merges)) UpperCAmelCase__ = FSMTTokenizer( langs=['en', 'ru'], src_vocab_size=len(vocab), tgt_vocab_size=len(vocab), src_vocab_file=src_vocab_file, tgt_vocab_file=tgt_vocab_file, merges_file=merges_file, ) UpperCAmelCase__ = FSMTConfig( langs=['ru', 'en'], src_vocab_size=1000, tgt_vocab_size=1000, d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) UpperCAmelCase__ = FSMTForConditionalGeneration(config) print(F"num of params {tiny_model.num_parameters()}") # Test UpperCAmelCase__ = tokenizer(['Making tiny model'], return_tensors='pt') UpperCAmelCase__ = tiny_model(**batch) print('test output:', len(outputs.logits[0])) # Save tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(F"Generated {mname_tiny}") # Upload # transformers-cli upload tiny-wmt19-en-ru
288
"""simple docstring""" from __future__ import annotations from bisect import bisect_left from functools import total_ordering from heapq import merge @total_ordering class lowerCAmelCase__ ( A_ ): def __lt__( self : Any , _lowerCamelCase : int ): return self[-1] < other[-1] def __eq__( self : int , _lowerCamelCase : Optional[Any] ): return self[-1] == other[-1] def _UpperCAmelCase ( __lowerCamelCase : list ) -> list: _snake_case = [] # sort into stacks for element in collection: _snake_case = Stack([element] ) _snake_case = bisect_left(__lowerCamelCase , __lowerCamelCase ) if i != len(__lowerCamelCase ): stacks[i].append(__lowerCamelCase ) else: stacks.append(__lowerCamelCase ) # use a heap-based merge to merge stack efficiently _snake_case = merge(*(reversed(__lowerCamelCase ) for stack in stacks) ) return collection if __name__ == "__main__": UpperCAmelCase__ = input('Enter numbers separated by a comma:\n').strip() UpperCAmelCase__ = [int(item) for item in user_input.split(',')] print(patience_sort(unsorted))
288
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase : int = logging.get_logger(__name__) __UpperCamelCase : Union[str, Any] = { 'microsoft/biogpt': 'https://huggingface.co/microsoft/biogpt/resolve/main/config.json', # See all BioGPT models at https://huggingface.co/models?filter=biogpt } class lowercase__ ( UpperCamelCase_): UpperCamelCase_ = """biogpt""" def __init__( self : str , UpperCamelCase__ : Any=4_2384 , UpperCamelCase__ : str=1024 , UpperCamelCase__ : str=24 , UpperCamelCase__ : int=16 , UpperCamelCase__ : Optional[Any]=4096 , UpperCamelCase__ : List[Any]="gelu" , UpperCamelCase__ : Optional[int]=0.1 , UpperCamelCase__ : Dict=0.1 , UpperCamelCase__ : List[Any]=1024 , UpperCamelCase__ : Optional[Any]=0.02 , UpperCamelCase__ : str=1E-12 , UpperCamelCase__ : Dict=True , UpperCamelCase__ : Dict=True , UpperCamelCase__ : List[Any]=0.0 , UpperCamelCase__ : Optional[Any]=0.0 , UpperCamelCase__ : List[Any]=1 , UpperCamelCase__ : List[str]=0 , UpperCamelCase__ : List[str]=2 , **UpperCamelCase__ : Tuple , ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = vocab_size SCREAMING_SNAKE_CASE : List[str] = max_position_embeddings SCREAMING_SNAKE_CASE : Optional[int] = hidden_size SCREAMING_SNAKE_CASE : List[str] = num_hidden_layers SCREAMING_SNAKE_CASE : Tuple = num_attention_heads SCREAMING_SNAKE_CASE : Optional[Any] = intermediate_size SCREAMING_SNAKE_CASE : Any = hidden_act SCREAMING_SNAKE_CASE : List[Any] = hidden_dropout_prob SCREAMING_SNAKE_CASE : str = attention_probs_dropout_prob SCREAMING_SNAKE_CASE : Union[str, Any] = initializer_range SCREAMING_SNAKE_CASE : Union[str, Any] = layer_norm_eps SCREAMING_SNAKE_CASE : List[Any] = scale_embedding SCREAMING_SNAKE_CASE : Tuple = use_cache SCREAMING_SNAKE_CASE : Optional[int] = layerdrop SCREAMING_SNAKE_CASE : Optional[int] = activation_dropout super().__init__(pad_token_id=UpperCamelCase__ , bos_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , **UpperCamelCase__ )
360
# # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def A ( *_lowercase ): with open(_lowercase , '''r''' ) as fh: fcntl.flock(_lowercase , fcntl.LOCK_EX ) try: print(*_lowercase ) finally: fcntl.flock(_lowercase , fcntl.LOCK_UN ) __UpperCamelCase : Union[str, Any] = int(os.environ['LOCAL_RANK']) torch.cuda.set_device(local_rank) __UpperCamelCase : Any = torch.device('cuda', local_rank) __UpperCamelCase : Union[str, Any] = socket.gethostname() __UpperCamelCase : Tuple = f"""[{hostname}-{local_rank}]""" try: # test distributed dist.init_process_group('nccl') dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank __UpperCamelCase : List[Any] = dist.get_rank() __UpperCamelCase : List[Any] = dist.get_world_size() printflock(f"""{gpu} is OK (global rank: {rank}/{world_size})""") dist.barrier() if rank == 0: printflock(f"""pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}""") except Exception: printflock(f"""{gpu} is broken""") raise
258
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase =logging.get_logger(__name__) _lowerCamelCase ={ "facebook/data2vec-text-base": "https://huggingface.co/data2vec/resolve/main/config.json", } class a_ ( lowerCamelCase_ ): """simple docstring""" __UpperCAmelCase = 'data2vec-text' def __init__( self : int ,snake_case : Dict=30522 ,snake_case : str=768 ,snake_case : Any=12 ,snake_case : Tuple=12 ,snake_case : Optional[int]=3072 ,snake_case : List[Any]="gelu" ,snake_case : Optional[int]=0.1 ,snake_case : List[Any]=0.1 ,snake_case : List[str]=512 ,snake_case : Optional[Any]=2 ,snake_case : Optional[Any]=0.02 ,snake_case : Dict=1e-12 ,snake_case : Optional[int]=1 ,snake_case : Dict=0 ,snake_case : Any=2 ,snake_case : List[str]="absolute" ,snake_case : str=True ,snake_case : int=None ,**snake_case : Union[str, Any] ,): super().__init__(pad_token_id=snake_case ,bos_token_id=snake_case ,eos_token_id=snake_case ,**snake_case ) SCREAMING_SNAKE_CASE =vocab_size SCREAMING_SNAKE_CASE =hidden_size SCREAMING_SNAKE_CASE =num_hidden_layers SCREAMING_SNAKE_CASE =num_attention_heads SCREAMING_SNAKE_CASE =hidden_act SCREAMING_SNAKE_CASE =intermediate_size SCREAMING_SNAKE_CASE =hidden_dropout_prob SCREAMING_SNAKE_CASE =attention_probs_dropout_prob SCREAMING_SNAKE_CASE =max_position_embeddings SCREAMING_SNAKE_CASE =type_vocab_size SCREAMING_SNAKE_CASE =initializer_range SCREAMING_SNAKE_CASE =layer_norm_eps SCREAMING_SNAKE_CASE =position_embedding_type SCREAMING_SNAKE_CASE =use_cache SCREAMING_SNAKE_CASE =classifier_dropout class a_ ( lowerCamelCase_ ): """simple docstring""" @property def _lowerCAmelCase ( self : str ): if self.task == "multiple-choice": SCREAMING_SNAKE_CASE ={0: 'batch', 1: 'choice', 2: 'sequence'} else: SCREAMING_SNAKE_CASE ={0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
334
import inspect import unittest from datasets import load_dataset from packaging import version from transformers import BeitConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_MAPPING, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, ) from transformers.models.beit.modeling_beit import BEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): import PIL from PIL import Image from transformers import BeitImageProcessor class a_ : """simple docstring""" def __init__( self : Optional[int] ,snake_case : Any ,snake_case : Dict=100 ,snake_case : List[Any]=13 ,snake_case : str=30 ,snake_case : List[str]=2 ,snake_case : List[Any]=3 ,snake_case : Tuple=True ,snake_case : Optional[Any]=True ,snake_case : int=32 ,snake_case : Tuple=4 ,snake_case : List[Any]=4 ,snake_case : Optional[Any]=37 ,snake_case : Optional[Any]="gelu" ,snake_case : Tuple=0.1 ,snake_case : Union[str, Any]=0.1 ,snake_case : List[Any]=10 ,snake_case : Tuple=0.02 ,snake_case : List[str]=3 ,snake_case : Any=None ,snake_case : int=[0, 1, 2, 3] ,): SCREAMING_SNAKE_CASE =parent SCREAMING_SNAKE_CASE =100 SCREAMING_SNAKE_CASE =batch_size SCREAMING_SNAKE_CASE =image_size SCREAMING_SNAKE_CASE =patch_size SCREAMING_SNAKE_CASE =num_channels SCREAMING_SNAKE_CASE =is_training SCREAMING_SNAKE_CASE =use_labels SCREAMING_SNAKE_CASE =hidden_size SCREAMING_SNAKE_CASE =num_hidden_layers SCREAMING_SNAKE_CASE =num_attention_heads SCREAMING_SNAKE_CASE =intermediate_size SCREAMING_SNAKE_CASE =hidden_act SCREAMING_SNAKE_CASE =hidden_dropout_prob SCREAMING_SNAKE_CASE =attention_probs_dropout_prob SCREAMING_SNAKE_CASE =type_sequence_label_size SCREAMING_SNAKE_CASE =initializer_range SCREAMING_SNAKE_CASE =scope SCREAMING_SNAKE_CASE =out_indices SCREAMING_SNAKE_CASE =num_labels # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) SCREAMING_SNAKE_CASE =(image_size // patch_size) ** 2 SCREAMING_SNAKE_CASE =num_patches + 1 def _lowerCAmelCase ( self : List[Any] ): SCREAMING_SNAKE_CASE =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE =None SCREAMING_SNAKE_CASE =None if self.use_labels: SCREAMING_SNAKE_CASE =ids_tensor([self.batch_size] ,self.type_sequence_label_size ) SCREAMING_SNAKE_CASE =ids_tensor([self.batch_size, self.image_size, self.image_size] ,self.num_labels ) SCREAMING_SNAKE_CASE =self.get_config() return config, pixel_values, labels, pixel_labels def _lowerCAmelCase ( self : Dict ): return BeitConfig( vocab_size=self.vocab_size ,image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=snake_case ,initializer_range=self.initializer_range ,out_indices=self.out_indices ,) def _lowerCAmelCase ( self : Union[str, Any] ,snake_case : Tuple ,snake_case : Optional[Any] ,snake_case : Union[str, Any] ,snake_case : Optional[int] ): SCREAMING_SNAKE_CASE =BeitModel(config=snake_case ) model.to(snake_case ) model.eval() SCREAMING_SNAKE_CASE =model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCAmelCase ( self : Union[str, Any] ,snake_case : Optional[int] ,snake_case : Dict ,snake_case : Any ,snake_case : List[str] ): SCREAMING_SNAKE_CASE =BeitForMaskedImageModeling(config=snake_case ) model.to(snake_case ) model.eval() SCREAMING_SNAKE_CASE =model(snake_case ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length - 1, self.vocab_size) ) def _lowerCAmelCase ( self : Optional[Any] ,snake_case : Any ,snake_case : str ,snake_case : Any ,snake_case : str ): SCREAMING_SNAKE_CASE =self.type_sequence_label_size SCREAMING_SNAKE_CASE =BeitForImageClassification(snake_case ) model.to(snake_case ) model.eval() SCREAMING_SNAKE_CASE =model(snake_case ,labels=snake_case ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) # test greyscale images SCREAMING_SNAKE_CASE =1 SCREAMING_SNAKE_CASE =BeitForImageClassification(snake_case ) model.to(snake_case ) model.eval() SCREAMING_SNAKE_CASE =floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE =model(snake_case ,labels=snake_case ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def _lowerCAmelCase ( self : List[str] ,snake_case : Tuple ,snake_case : str ,snake_case : Optional[int] ,snake_case : int ): SCREAMING_SNAKE_CASE =self.num_labels SCREAMING_SNAKE_CASE =BeitForSemanticSegmentation(snake_case ) model.to(snake_case ) model.eval() SCREAMING_SNAKE_CASE =model(snake_case ) self.parent.assertEqual( result.logits.shape ,(self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) SCREAMING_SNAKE_CASE =model(snake_case ,labels=snake_case ) self.parent.assertEqual( result.logits.shape ,(self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) def _lowerCAmelCase ( self : str ): SCREAMING_SNAKE_CASE =self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE =config_and_inputs SCREAMING_SNAKE_CASE ={'pixel_values': pixel_values} return config, inputs_dict @require_torch class a_ ( lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): """simple docstring""" __UpperCAmelCase = ( (BeitModel, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase = ( { 'feature-extraction': BeitModel, 'image-classification': BeitForImageClassification, 'image-segmentation': BeitForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase = False __UpperCAmelCase = False __UpperCAmelCase = False def _lowerCAmelCase ( self : List[Any] ): SCREAMING_SNAKE_CASE =BeitModelTester(self ) SCREAMING_SNAKE_CASE =ConfigTester(self ,config_class=snake_case ,has_text_modality=snake_case ,hidden_size=37 ) def _lowerCAmelCase ( self : List[str] ): self.config_tester.run_common_tests() @unittest.skip(reason='BEiT does not use inputs_embeds' ) def _lowerCAmelCase ( self : List[Any] ): pass @require_torch_multi_gpu @unittest.skip(reason='BEiT has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def _lowerCAmelCase ( self : Union[str, Any] ): pass def _lowerCAmelCase ( self : Tuple ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE =model_class(snake_case ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) SCREAMING_SNAKE_CASE =model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case ,nn.Linear ) ) def _lowerCAmelCase ( self : int ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE =model_class(snake_case ) SCREAMING_SNAKE_CASE =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE =[*signature.parameters.keys()] SCREAMING_SNAKE_CASE =['pixel_values'] self.assertListEqual(arg_names[:1] ,snake_case ) def _lowerCAmelCase ( self : List[str] ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def _lowerCAmelCase ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case ) def _lowerCAmelCase ( self : Dict ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case ) def _lowerCAmelCase ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*snake_case ) def _lowerCAmelCase ( self : Any ): if not self.model_tester.is_training: return SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE =True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if model_class in [*get_values(snake_case ), BeitForMaskedImageModeling]: continue SCREAMING_SNAKE_CASE =model_class(snake_case ) model.to(snake_case ) model.train() SCREAMING_SNAKE_CASE =self._prepare_for_class(snake_case ,snake_case ,return_labels=snake_case ) SCREAMING_SNAKE_CASE =model(**snake_case ).loss loss.backward() def _lowerCAmelCase ( self : int ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return SCREAMING_SNAKE_CASE =False SCREAMING_SNAKE_CASE =True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if ( model_class in [*get_values(snake_case ), BeitForMaskedImageModeling] or not model_class.supports_gradient_checkpointing ): continue SCREAMING_SNAKE_CASE =model_class(snake_case ) model.gradient_checkpointing_enable() model.to(snake_case ) model.train() SCREAMING_SNAKE_CASE =self._prepare_for_class(snake_case ,snake_case ,return_labels=snake_case ) SCREAMING_SNAKE_CASE =model(**snake_case ).loss loss.backward() def _lowerCAmelCase ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE =self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE =_config_zero_init(snake_case ) for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE =model_class(config=snake_case ) for name, param in model.named_parameters(): # we skip lambda parameters as these require special initial values # determined by config.layer_scale_init_value if "lambda" in name: continue if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() ,[0.0, 1.0] ,msg=f'Parameter {name} of model {model_class} seems not properly initialized' ,) @slow def _lowerCAmelCase ( self : List[str] ): for model_name in BEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE =BeitModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) def snake_case__ ( ): """simple docstring""" SCREAMING_SNAKE_CASE =Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class a_ ( unittest.TestCase ): """simple docstring""" @cached_property def _lowerCAmelCase ( self : Tuple ): return BeitImageProcessor.from_pretrained('microsoft/beit-base-patch16-224' ) if is_vision_available() else None @slow def _lowerCAmelCase ( self : int ): SCREAMING_SNAKE_CASE =BeitForMaskedImageModeling.from_pretrained('microsoft/beit-base-patch16-224-pt22k' ).to(snake_case ) SCREAMING_SNAKE_CASE =self.default_image_processor SCREAMING_SNAKE_CASE =prepare_img() SCREAMING_SNAKE_CASE =image_processor(images=snake_case ,return_tensors='pt' ).pixel_values.to(snake_case ) # prepare bool_masked_pos SCREAMING_SNAKE_CASE =torch.ones((1, 196) ,dtype=torch.bool ).to(snake_case ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE =model(pixel_values=snake_case ,bool_masked_pos=snake_case ) SCREAMING_SNAKE_CASE =outputs.logits # verify the logits SCREAMING_SNAKE_CASE =torch.Size((1, 196, 8192) ) self.assertEqual(logits.shape ,snake_case ) SCREAMING_SNAKE_CASE =torch.tensor( [[-3.2_437, 0.5_072, -13.9_174], [-3.2_456, 0.4_948, -13.9_401], [-3.2_033, 0.5_121, -13.8_550]] ).to(snake_case ) self.assertTrue(torch.allclose(logits[bool_masked_pos][:3, :3] ,snake_case ,atol=1e-2 ) ) @slow def _lowerCAmelCase ( self : List[str] ): SCREAMING_SNAKE_CASE =BeitForImageClassification.from_pretrained('microsoft/beit-base-patch16-224' ).to(snake_case ) SCREAMING_SNAKE_CASE =self.default_image_processor SCREAMING_SNAKE_CASE =prepare_img() SCREAMING_SNAKE_CASE =image_processor(images=snake_case ,return_tensors='pt' ).to(snake_case ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE =model(**snake_case ) SCREAMING_SNAKE_CASE =outputs.logits # verify the logits SCREAMING_SNAKE_CASE =torch.Size((1, 1000) ) self.assertEqual(logits.shape ,snake_case ) SCREAMING_SNAKE_CASE =torch.tensor([-1.2_385, -1.0_987, -1.0_108] ).to(snake_case ) self.assertTrue(torch.allclose(logits[0, :3] ,snake_case ,atol=1e-4 ) ) SCREAMING_SNAKE_CASE =281 self.assertEqual(logits.argmax(-1 ).item() ,snake_case ) @slow def _lowerCAmelCase ( self : int ): SCREAMING_SNAKE_CASE =BeitForImageClassification.from_pretrained('microsoft/beit-large-patch16-224-pt22k-ft22k' ).to( snake_case ) SCREAMING_SNAKE_CASE =self.default_image_processor SCREAMING_SNAKE_CASE =prepare_img() SCREAMING_SNAKE_CASE =image_processor(images=snake_case ,return_tensors='pt' ).to(snake_case ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE =model(**snake_case ) SCREAMING_SNAKE_CASE =outputs.logits # verify the logits SCREAMING_SNAKE_CASE =torch.Size((1, 21841) ) self.assertEqual(logits.shape ,snake_case ) SCREAMING_SNAKE_CASE =torch.tensor([1.6_881, -0.2_787, 0.5_901] ).to(snake_case ) self.assertTrue(torch.allclose(logits[0, :3] ,snake_case ,atol=1e-4 ) ) SCREAMING_SNAKE_CASE =2396 self.assertEqual(logits.argmax(-1 ).item() ,snake_case ) @slow def _lowerCAmelCase ( self : Tuple ): SCREAMING_SNAKE_CASE =BeitForSemanticSegmentation.from_pretrained('microsoft/beit-base-finetuned-ade-640-640' ) SCREAMING_SNAKE_CASE =model.to(snake_case ) SCREAMING_SNAKE_CASE =BeitImageProcessor(do_resize=snake_case ,size=640 ,do_center_crop=snake_case ) SCREAMING_SNAKE_CASE =load_dataset('hf-internal-testing/fixtures_ade20k' ,split='test' ) SCREAMING_SNAKE_CASE =Image.open(ds[0]['file'] ) SCREAMING_SNAKE_CASE =image_processor(images=snake_case ,return_tensors='pt' ).to(snake_case ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE =model(**snake_case ) SCREAMING_SNAKE_CASE =outputs.logits # verify the logits SCREAMING_SNAKE_CASE =torch.Size((1, 150, 160, 160) ) self.assertEqual(logits.shape ,snake_case ) SCREAMING_SNAKE_CASE =version.parse(PIL.__version__ ) < version.parse('9.0.0' ) if is_pillow_less_than_a: SCREAMING_SNAKE_CASE =torch.tensor( [ [[-4.9_225, -2.3_954, -3.0_522], [-2.8_822, -1.0_046, -1.7_561], [-2.9_549, -1.3_228, -2.1_347]], [[-5.8_168, -3.4_129, -4.0_778], [-3.8_651, -2.2_214, -3.0_277], [-3.8_356, -2.4_643, -3.3_535]], [[-0.0_078, 3.9_952, 4.0_754], [2.9_856, 4.6_944, 5.0_035], [3.2_413, 4.7_813, 4.9_969]], ] ,device=snake_case ,) else: SCREAMING_SNAKE_CASE =torch.tensor( [ [[-4.8_960, -2.3_688, -3.0_355], [-2.8_478, -0.9_836, -1.7_418], [-2.9_449, -1.3_332, -2.1_456]], [[-5.8_081, -3.4_124, -4.1_006], [-3.8_561, -2.2_081, -3.0_323], [-3.8_365, -2.4_601, -3.3_669]], [[-0.0_309, 3.9_868, 4.0_540], [2.9_640, 4.6_877, 4.9_976], [3.2_081, 4.7_690, 4.9_942]], ] ,device=snake_case ,) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] ,snake_case ,atol=1e-4 ) ) @slow def _lowerCAmelCase ( self : int ): SCREAMING_SNAKE_CASE =BeitForSemanticSegmentation.from_pretrained('microsoft/beit-base-finetuned-ade-640-640' ) SCREAMING_SNAKE_CASE =model.to(snake_case ) SCREAMING_SNAKE_CASE =BeitImageProcessor(do_resize=snake_case ,size=640 ,do_center_crop=snake_case ) SCREAMING_SNAKE_CASE =load_dataset('hf-internal-testing/fixtures_ade20k' ,split='test' ) SCREAMING_SNAKE_CASE =Image.open(ds[0]['file'] ) SCREAMING_SNAKE_CASE =image_processor(images=snake_case ,return_tensors='pt' ).to(snake_case ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE =model(**snake_case ) SCREAMING_SNAKE_CASE =outputs.logits.detach().cpu() SCREAMING_SNAKE_CASE =image_processor.post_process_semantic_segmentation(outputs=snake_case ,target_sizes=[(500, 300)] ) SCREAMING_SNAKE_CASE =torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape ,snake_case ) SCREAMING_SNAKE_CASE =image_processor.post_process_semantic_segmentation(outputs=snake_case ) SCREAMING_SNAKE_CASE =torch.Size((160, 160) ) self.assertEqual(segmentation[0].shape ,snake_case )
334
1
'''simple docstring''' from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class lowercase ( _lowerCamelCase ): """simple docstring""" def __init__( self ,a_ ,a_ = None ,a_ = None ,a_ = True ,a_ = None ,a_ = False ,a_ = None ,a_ = True ,a_ = "arrow" ,**a_ ,) -> str: super().__init__( split=a_ ,features=a_ ,cache_dir=a_ ,keep_in_memory=a_ ,streaming=a_ ,**a_ ,) _UpperCAmelCase : Any = load_from_cache_file _UpperCAmelCase : Optional[int] = file_format _UpperCAmelCase : int = Spark( df=a_ ,features=a_ ,cache_dir=a_ ,working_dir=a_ ,**a_ ,) def _snake_case ( self ) -> int: if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) _UpperCAmelCase : str = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=a_ ,file_format=self._file_format ,) return self.builder.as_dataset(split=self.split )
349
'''simple docstring''' import argparse import copy def snake_case_ ( lowerCAmelCase_ )-> Dict: '''simple docstring''' _UpperCAmelCase : Dict = {} with open(lowerCAmelCase_ ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: _UpperCAmelCase : Optional[int] = [] _list.append([line.split()[1], line.split()[2]] ) _UpperCAmelCase : List[str] = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: _UpperCAmelCase : List[str] = [] _list.append([line.split()[0], line.split()[2]] ) _UpperCAmelCase : Optional[int] = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ )-> List[Any]: '''simple docstring''' with open(lowerCAmelCase_ ) as f: _UpperCAmelCase : List[Any] = f.read(1 ) _UpperCAmelCase : int = start_node _UpperCAmelCase : List[Any] = [] _UpperCAmelCase : Dict = start_node _UpperCAmelCase : Any = 0 while visiting not in first_solution: _UpperCAmelCase : Optional[int] = 10000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(lowerCAmelCase_ ) and k[0] not in first_solution: _UpperCAmelCase : Optional[int] = k[1] _UpperCAmelCase : List[str] = k[0] first_solution.append(lowerCAmelCase_ ) _UpperCAmelCase : Optional[int] = distance_of_first_solution + int(lowerCAmelCase_ ) _UpperCAmelCase : Dict = best_node first_solution.append(lowerCAmelCase_ ) _UpperCAmelCase : List[str] = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 _UpperCAmelCase : int = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 10000 ) return first_solution, distance_of_first_solution def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ )-> int: '''simple docstring''' _UpperCAmelCase : int = [] for n in solution[1:-1]: _UpperCAmelCase : Tuple = solution.index(lowerCAmelCase_ ) for kn in solution[1:-1]: _UpperCAmelCase : int = solution.index(lowerCAmelCase_ ) if n == kn: continue _UpperCAmelCase : Tuple = copy.deepcopy(lowerCAmelCase_ ) _UpperCAmelCase : Union[str, Any] = kn _UpperCAmelCase : List[str] = n _UpperCAmelCase : Optional[int] = 0 for k in _tmp[:-1]: _UpperCAmelCase : List[str] = _tmp[_tmp.index(lowerCAmelCase_ ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: _UpperCAmelCase : Dict = distance + int(i[1] ) _tmp.append(lowerCAmelCase_ ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) _UpperCAmelCase : Dict = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda lowerCAmelCase_ : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )-> int: '''simple docstring''' _UpperCAmelCase : List[Any] = 1 _UpperCAmelCase : Optional[Any] = first_solution _UpperCAmelCase : List[Any] = [] _UpperCAmelCase : List[Any] = distance_of_first_solution _UpperCAmelCase : Dict = solution while count <= iters: _UpperCAmelCase : Any = find_neighborhood(lowerCAmelCase_ , lowerCAmelCase_ ) _UpperCAmelCase : Dict = 0 _UpperCAmelCase : Optional[Any] = neighborhood[index_of_best_solution] _UpperCAmelCase : Optional[Any] = len(lowerCAmelCase_ ) - 1 _UpperCAmelCase : Optional[Any] = False while not found: _UpperCAmelCase : Tuple = 0 while i < len(lowerCAmelCase_ ): if best_solution[i] != solution[i]: _UpperCAmelCase : Any = best_solution[i] _UpperCAmelCase : str = solution[i] break _UpperCAmelCase : int = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) _UpperCAmelCase : Tuple = True _UpperCAmelCase : List[Any] = best_solution[:-1] _UpperCAmelCase : str = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: _UpperCAmelCase : Tuple = cost _UpperCAmelCase : List[Any] = solution else: _UpperCAmelCase : Any = index_of_best_solution + 1 _UpperCAmelCase : Dict = neighborhood[index_of_best_solution] if len(lowerCAmelCase_ ) >= size: tabu_list.pop(0 ) _UpperCAmelCase : Optional[Any] = count + 1 return best_solution_ever, best_cost def snake_case_ ( lowerCAmelCase_=None )-> Optional[int]: '''simple docstring''' _UpperCAmelCase : Tuple = generate_neighbours(args.File ) _UpperCAmelCase ,_UpperCAmelCase : Tuple = generate_first_solution( args.File , lowerCAmelCase_ ) _UpperCAmelCase ,_UpperCAmelCase : str = tabu_search( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , args.Iterations , args.Size , ) print(F'''Best solution: {best_sol}, with total distance: {best_cost}.''' ) if __name__ == "__main__": A_ : Optional[int] = argparse.ArgumentParser(description="""Tabu Search""") parser.add_argument( """-f""", """--File""", type=str, help="""Path to the file containing the data""", required=True, ) parser.add_argument( """-i""", """--Iterations""", type=int, help="""How many iterations the algorithm should perform""", required=True, ) parser.add_argument( """-s""", """--Size""", type=int, help="""Size of the tabu list""", required=True ) # Pass the arguments to main method main(parser.parse_args())
349
1
"""simple docstring""" import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->UnCLIP class __snake_case ( _lowercase): snake_case__ : torch.FloatTensor snake_case__ : Optional[torch.FloatTensor] = None def snake_case_ ( A_ : List[Any], A_ : Union[str, Any]=0.999, A_ : str="cosine", ): '''simple docstring''' if alpha_transform_type == "cosine": def alpha_bar_fn(A_ : Any ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(A_ : int ): return math.exp(t * -12.0 ) else: raise ValueError(F'''Unsupported alpha_tranform_type: {alpha_transform_type}''' ) _lowerCamelCase : List[str] = [] for i in range(A_ ): _lowerCamelCase : List[str] = i / num_diffusion_timesteps _lowerCamelCase : Any = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(A_ ) / alpha_bar_fn(A_ ), A_ ) ) return torch.tensor(A_, dtype=torch.floataa ) class __snake_case ( _lowercase , _lowercase): @register_to_config def __init__( self : int , __lowerCAmelCase : int = 1_0_0_0 , __lowerCAmelCase : str = "fixed_small_log" , __lowerCAmelCase : bool = True , __lowerCAmelCase : Optional[float] = 1.0 , __lowerCAmelCase : str = "epsilon" , __lowerCAmelCase : str = "squaredcos_cap_v2" , ): """simple docstring""" if beta_schedule != "squaredcos_cap_v2": raise ValueError('''UnCLIPScheduler only supports `beta_schedule`: \'squaredcos_cap_v2\'''' ) _lowerCamelCase : str = betas_for_alpha_bar(__lowerCAmelCase ) _lowerCamelCase : Any = 1.0 - self.betas _lowerCamelCase : List[str] = torch.cumprod(self.alphas , dim=0 ) _lowerCamelCase : Dict = torch.tensor(1.0 ) # standard deviation of the initial noise distribution _lowerCamelCase : Any = 1.0 # setable values _lowerCamelCase : str = None _lowerCamelCase : int = torch.from_numpy(np.arange(0 , __lowerCAmelCase )[::-1].copy() ) _lowerCamelCase : int = variance_type def SCREAMING_SNAKE_CASE ( self : str , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : Optional[int] = None ): """simple docstring""" return sample def SCREAMING_SNAKE_CASE ( self : List[Any] , __lowerCAmelCase : int , __lowerCAmelCase : Union[str, torch.device] = None ): """simple docstring""" _lowerCamelCase : Dict = num_inference_steps _lowerCamelCase : Any = (self.config.num_train_timesteps - 1) / (self.num_inference_steps - 1) _lowerCamelCase : Union[str, Any] = (np.arange(0 , __lowerCAmelCase ) * step_ratio).round()[::-1].copy().astype(np.intaa ) _lowerCamelCase : str = torch.from_numpy(__lowerCAmelCase ).to(__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Any=None , __lowerCAmelCase : List[str]=None , __lowerCAmelCase : int=None ): """simple docstring""" if prev_timestep is None: _lowerCamelCase : int = t - 1 _lowerCamelCase : int = self.alphas_cumprod[t] _lowerCamelCase : List[str] = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one _lowerCamelCase : Optional[int] = 1 - alpha_prod_t _lowerCamelCase : Optional[int] = 1 - alpha_prod_t_prev if prev_timestep == t - 1: _lowerCamelCase : Optional[int] = self.betas[t] else: _lowerCamelCase : Union[str, Any] = 1 - alpha_prod_t / alpha_prod_t_prev # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample _lowerCamelCase : Dict = beta_prod_t_prev / beta_prod_t * beta if variance_type is None: _lowerCamelCase : List[Any] = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small_log": _lowerCamelCase : Tuple = torch.log(torch.clamp(__lowerCAmelCase , min=1E-20 ) ) _lowerCamelCase : Optional[Any] = torch.exp(0.5 * variance ) elif variance_type == "learned_range": # NOTE difference with DDPM scheduler _lowerCamelCase : Dict = variance.log() _lowerCamelCase : Optional[int] = beta.log() _lowerCamelCase : Dict = (predicted_variance + 1) / 2 _lowerCamelCase : str = frac * max_log + (1 - frac) * min_log return variance def SCREAMING_SNAKE_CASE ( self : Any , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : int , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : Optional[int] = None , __lowerCAmelCase : Optional[int]=None , __lowerCAmelCase : bool = True , ): """simple docstring""" _lowerCamelCase : Dict = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type == "learned_range": _lowerCamelCase , _lowerCamelCase : int = torch.split(__lowerCAmelCase , sample.shape[1] , dim=1 ) else: _lowerCamelCase : Any = None # 1. compute alphas, betas if prev_timestep is None: _lowerCamelCase : List[str] = t - 1 _lowerCamelCase : Any = self.alphas_cumprod[t] _lowerCamelCase : Dict = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one _lowerCamelCase : List[str] = 1 - alpha_prod_t _lowerCamelCase : List[str] = 1 - alpha_prod_t_prev if prev_timestep == t - 1: _lowerCamelCase : Optional[Any] = self.betas[t] _lowerCamelCase : List[Any] = self.alphas[t] else: _lowerCamelCase : Tuple = 1 - alpha_prod_t / alpha_prod_t_prev _lowerCamelCase : Any = 1 - beta # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": _lowerCamelCase : Optional[int] = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": _lowerCamelCase : Dict = model_output else: raise ValueError( f'''prediction_type given as {self.config.prediction_type} must be one of `epsilon` or `sample`''' ''' for the UnCLIPScheduler.''' ) # 3. Clip "predicted x_0" if self.config.clip_sample: _lowerCamelCase : Dict = torch.clamp( __lowerCAmelCase , -self.config.clip_sample_range , self.config.clip_sample_range ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _lowerCamelCase : Dict = (alpha_prod_t_prev ** 0.5 * beta) / beta_prod_t _lowerCamelCase : int = alpha ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _lowerCamelCase : Union[str, Any] = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise _lowerCamelCase : Optional[Any] = 0 if t > 0: _lowerCamelCase : Dict = randn_tensor( model_output.shape , dtype=model_output.dtype , generator=__lowerCAmelCase , device=model_output.device ) _lowerCamelCase : Any = self._get_variance( __lowerCAmelCase , predicted_variance=__lowerCAmelCase , prev_timestep=__lowerCAmelCase , ) if self.variance_type == "fixed_small_log": _lowerCamelCase : Optional[int] = variance elif self.variance_type == "learned_range": _lowerCamelCase : Optional[int] = (0.5 * variance).exp() else: raise ValueError( f'''variance_type given as {self.variance_type} must be one of `fixed_small_log` or `learned_range`''' ''' for the UnCLIPScheduler.''' ) _lowerCamelCase : Optional[int] = variance * variance_noise _lowerCamelCase : int = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return UnCLIPSchedulerOutput(prev_sample=__lowerCAmelCase , pred_original_sample=__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Any , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : torch.IntTensor , ): """simple docstring""" _lowerCamelCase : Dict = self.alphas_cumprod.to(device=original_samples.device , dtype=original_samples.dtype ) _lowerCamelCase : Optional[Any] = timesteps.to(original_samples.device ) _lowerCamelCase : str = alphas_cumprod[timesteps] ** 0.5 _lowerCamelCase : int = sqrt_alpha_prod.flatten() while len(sqrt_alpha_prod.shape ) < len(original_samples.shape ): _lowerCamelCase : List[str] = sqrt_alpha_prod.unsqueeze(-1 ) _lowerCamelCase : Optional[int] = (1 - alphas_cumprod[timesteps]) ** 0.5 _lowerCamelCase : str = sqrt_one_minus_alpha_prod.flatten() while len(sqrt_one_minus_alpha_prod.shape ) < len(original_samples.shape ): _lowerCamelCase : List[Any] = sqrt_one_minus_alpha_prod.unsqueeze(-1 ) _lowerCamelCase : Tuple = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples
72
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) UpperCAmelCase : Any = { '''configuration_falcon''': ['''FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''FalconConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Any = [ '''FALCON_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FalconForCausalLM''', '''FalconModel''', '''FalconPreTrainedModel''', '''FalconForSequenceClassification''', '''FalconForTokenClassification''', '''FalconForQuestionAnswering''', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys UpperCAmelCase : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
280
0
"""simple docstring""" import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_MAPPING, LevitForImageClassification, LevitForImageClassificationWithTeacher, LevitModel, ) from transformers.models.levit.modeling_levit import LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ): def _UpperCAmelCase ( self : int): """simple docstring""" lowercase_ = self.config_class(**self.inputs_dict) self.parent.assertTrue(hasattr(lowerCAmelCase_ , """hidden_sizes""")) self.parent.assertTrue(hasattr(lowerCAmelCase_ , """num_attention_heads""")) class SCREAMING_SNAKE_CASE__ : def __init__( self : List[str] , lowerCAmelCase_ : int , lowerCAmelCase_ : Dict=1_3 , lowerCAmelCase_ : Optional[int]=6_4 , lowerCAmelCase_ : Optional[Any]=3 , lowerCAmelCase_ : Optional[int]=3 , lowerCAmelCase_ : Tuple=2 , lowerCAmelCase_ : int=1 , lowerCAmelCase_ : Optional[int]=1_6 , lowerCAmelCase_ : List[Any]=[1_2_8, 2_5_6, 3_8_4] , lowerCAmelCase_ : Union[str, Any]=[4, 6, 8] , lowerCAmelCase_ : Union[str, Any]=[2, 3, 4] , lowerCAmelCase_ : Union[str, Any]=[1_6, 1_6, 1_6] , lowerCAmelCase_ : Any=0 , lowerCAmelCase_ : Optional[Any]=[2, 2, 2] , lowerCAmelCase_ : List[str]=[2, 2, 2] , lowerCAmelCase_ : Optional[int]=0.02 , lowerCAmelCase_ : Optional[int]=True , lowerCAmelCase_ : Optional[Any]=True , lowerCAmelCase_ : List[Any]=2 , ): """simple docstring""" lowercase_ = parent lowercase_ = batch_size lowercase_ = image_size lowercase_ = num_channels lowercase_ = kernel_size lowercase_ = stride lowercase_ = padding lowercase_ = hidden_sizes lowercase_ = num_attention_heads lowercase_ = depths lowercase_ = key_dim lowercase_ = drop_path_rate lowercase_ = patch_size lowercase_ = attention_ratio lowercase_ = mlp_ratio lowercase_ = initializer_range lowercase_ = [ ["""Subsample""", key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ["""Subsample""", key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] lowercase_ = is_training lowercase_ = use_labels lowercase_ = num_labels lowercase_ = initializer_range def _UpperCAmelCase ( self : Dict): """simple docstring""" lowercase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) lowercase_ = None if self.use_labels: lowercase_ = ids_tensor([self.batch_size] , self.num_labels) lowercase_ = self.get_config() return config, pixel_values, labels def _UpperCAmelCase ( self : List[Any]): """simple docstring""" return LevitConfig( image_size=self.image_size , num_channels=self.num_channels , kernel_size=self.kernel_size , stride=self.stride , padding=self.padding , patch_size=self.patch_size , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , depths=self.depths , key_dim=self.key_dim , drop_path_rate=self.drop_path_rate , mlp_ratio=self.mlp_ratio , attention_ratio=self.attention_ratio , initializer_range=self.initializer_range , down_ops=self.down_ops , ) def _UpperCAmelCase ( self : Optional[Any] , lowerCAmelCase_ : str , lowerCAmelCase_ : str , lowerCAmelCase_ : Tuple): """simple docstring""" lowercase_ = LevitModel(config=lowerCAmelCase_) model.to(lowerCAmelCase_) model.eval() lowercase_ = model(lowerCAmelCase_) lowercase_ = (self.image_size, self.image_size) lowercase_ , lowercase_ = image_size[0], image_size[1] for _ in range(4): lowercase_ = floor(((height + 2 * self.padding - self.kernel_size) / self.stride) + 1) lowercase_ = floor(((width + 2 * self.padding - self.kernel_size) / self.stride) + 1) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, ceil(height / 4) * ceil(width / 4), self.hidden_sizes[-1]) , ) def _UpperCAmelCase ( self : List[Any] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Union[str, Any]): """simple docstring""" lowercase_ = self.num_labels lowercase_ = LevitForImageClassification(lowerCAmelCase_) model.to(lowerCAmelCase_) model.eval() lowercase_ = model(lowerCAmelCase_ , labels=lowerCAmelCase_) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def _UpperCAmelCase ( self : str): """simple docstring""" lowercase_ = self.prepare_config_and_inputs() lowercase_ , lowercase_ , lowercase_ = config_and_inputs lowercase_ = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): lowercase__ = ( (LevitModel, LevitForImageClassification, LevitForImageClassificationWithTeacher) if is_torch_available() else () ) lowercase__ = ( { "feature-extraction": LevitModel, "image-classification": (LevitForImageClassification, LevitForImageClassificationWithTeacher), } if is_torch_available() else {} ) lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False def _UpperCAmelCase ( self : Optional[int]): """simple docstring""" lowercase_ = LevitModelTester(self) lowercase_ = ConfigTester(self , config_class=lowerCAmelCase_ , has_text_modality=lowerCAmelCase_ , hidden_size=3_7) def _UpperCAmelCase ( self : Tuple): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _UpperCAmelCase ( self : int): """simple docstring""" return @unittest.skip(reason="""Levit does not use inputs_embeds""") def _UpperCAmelCase ( self : List[str]): """simple docstring""" pass @unittest.skip(reason="""Levit does not support input and output embeddings""") def _UpperCAmelCase ( self : List[Any]): """simple docstring""" pass @unittest.skip(reason="""Levit does not output attentions""") def _UpperCAmelCase ( self : List[Any]): """simple docstring""" pass def _UpperCAmelCase ( self : str): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ = model_class(lowerCAmelCase_) lowercase_ = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase_ = [*signature.parameters.keys()] lowercase_ = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowerCAmelCase_) def _UpperCAmelCase ( self : str): """simple docstring""" def check_hidden_states_output(lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : Tuple): lowercase_ = model_class(lowerCAmelCase_) model.to(lowerCAmelCase_) model.eval() with torch.no_grad(): lowercase_ = model(**self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_)) lowercase_ = outputs.hidden_states lowercase_ = len(self.model_tester.depths) + 1 self.assertEqual(len(lowerCAmelCase_) , lowerCAmelCase_) lowercase_ = (self.model_tester.image_size, self.model_tester.image_size) lowercase_ , lowercase_ = image_size[0], image_size[1] for _ in range(4): lowercase_ = floor( ( (height + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1) lowercase_ = floor( ( (width + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-2:]) , [ height * width, self.model_tester.hidden_sizes[0], ] , ) lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase_ = True check_hidden_states_output(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase_ = True check_hidden_states_output(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""") def _UpperCAmelCase ( self : Tuple): """simple docstring""" pass def _UpperCAmelCase ( self : int , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Optional[int]=False): """simple docstring""" lowercase_ = super()._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ , return_labels=lowerCAmelCase_) if return_labels: if model_class.__name__ == "LevitForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def _UpperCAmelCase ( self : Dict): """simple docstring""" lowercase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase_) def _UpperCAmelCase ( self : str): """simple docstring""" lowercase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase_) def _UpperCAmelCase ( self : Union[str, Any]): """simple docstring""" if not self.model_tester.is_training: return lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ = True for model_class in self.all_model_classes: # LevitForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowerCAmelCase_) or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue lowercase_ = model_class(lowerCAmelCase_) model.to(lowerCAmelCase_) model.train() lowercase_ = self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ , return_labels=lowerCAmelCase_) lowercase_ = model(**lowerCAmelCase_).loss loss.backward() def _UpperCAmelCase ( self : List[Any]): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return lowercase_ = False lowercase_ = True for model_class in self.all_model_classes: if model_class in get_values(lowerCAmelCase_) or not model_class.supports_gradient_checkpointing: continue # LevitForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "LevitForImageClassificationWithTeacher": continue lowercase_ = model_class(lowerCAmelCase_) model.gradient_checkpointing_enable() model.to(lowerCAmelCase_) model.train() lowercase_ = self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ , return_labels=lowerCAmelCase_) lowercase_ = model(**lowerCAmelCase_).loss loss.backward() def _UpperCAmelCase ( self : Any): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() lowercase_ = [ {"""title""": """multi_label_classification""", """num_labels""": 2, """dtype""": torch.float}, {"""title""": """single_label_classification""", """num_labels""": 1, """dtype""": torch.long}, {"""title""": """regression""", """num_labels""": 1, """dtype""": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowerCAmelCase_), ] or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F'''Testing {model_class} with {problem_type["title"]}'''): lowercase_ = problem_type["""title"""] lowercase_ = problem_type["""num_labels"""] lowercase_ = model_class(lowerCAmelCase_) model.to(lowerCAmelCase_) model.train() lowercase_ = self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ , return_labels=lowerCAmelCase_) if problem_type["num_labels"] > 1: lowercase_ = inputs["""labels"""].unsqueeze(1).repeat(1 , problem_type["""num_labels"""]) lowercase_ = inputs["""labels"""].to(problem_type["""dtype"""]) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowerCAmelCase_) as warning_list: lowercase_ = model(**lowerCAmelCase_).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message): raise ValueError( F'''Something is going wrong in the regression problem: intercepted {w.message}''') loss.backward() @slow def _UpperCAmelCase ( self : List[str]): """simple docstring""" for model_name in LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase_ = LevitModel.from_pretrained(lowerCAmelCase_) self.assertIsNotNone(lowerCAmelCase_) def _SCREAMING_SNAKE_CASE () -> str: '''simple docstring''' lowercase_ = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @cached_property def _UpperCAmelCase ( self : Optional[Any]): """simple docstring""" return LevitImageProcessor.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0]) @slow def _UpperCAmelCase ( self : int): """simple docstring""" lowercase_ = LevitForImageClassificationWithTeacher.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0]).to( lowerCAmelCase_) lowercase_ = self.default_image_processor lowercase_ = prepare_img() lowercase_ = image_processor(images=lowerCAmelCase_ , return_tensors="""pt""").to(lowerCAmelCase_) # forward pass with torch.no_grad(): lowercase_ = model(**lowerCAmelCase_) # verify the logits lowercase_ = torch.Size((1, 1_0_0_0)) self.assertEqual(outputs.logits.shape , lowerCAmelCase_) lowercase_ = torch.tensor([1.0_448, -0.3_745, -1.8_317]).to(lowerCAmelCase_) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase_ , atol=1E-4))
313
"""simple docstring""" import unittest from transformers import PegasusConfig, PegasusTokenizer, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html UpperCAmelCase : Optional[Any] = "platform" import jax import jax.numpy as jnp import numpy as np from transformers import FlaxPegasusForConditionalGeneration, FlaxPegasusModel @require_flax class SCREAMING_SNAKE_CASE__ : lowercase__ = PegasusConfig lowercase__ = {} lowercase__ = "gelu" def __init__( self : Any , lowerCAmelCase_ : str , lowerCAmelCase_ : List[Any]=1_3 , lowerCAmelCase_ : Any=7 , lowerCAmelCase_ : Any=True , lowerCAmelCase_ : Tuple=False , lowerCAmelCase_ : str=9_9 , lowerCAmelCase_ : Tuple=3_2 , lowerCAmelCase_ : Dict=5 , lowerCAmelCase_ : Union[str, Any]=4 , lowerCAmelCase_ : Dict=3_7 , lowerCAmelCase_ : Any=0.1 , lowerCAmelCase_ : Any=0.1 , lowerCAmelCase_ : Optional[int]=2_0 , lowerCAmelCase_ : Tuple=2 , lowerCAmelCase_ : List[str]=1 , lowerCAmelCase_ : Optional[Any]=0 , ): """simple docstring""" lowercase_ = parent lowercase_ = batch_size lowercase_ = seq_length lowercase_ = is_training lowercase_ = use_labels lowercase_ = vocab_size lowercase_ = hidden_size lowercase_ = num_hidden_layers lowercase_ = num_attention_heads lowercase_ = intermediate_size lowercase_ = hidden_dropout_prob lowercase_ = attention_probs_dropout_prob lowercase_ = max_position_embeddings lowercase_ = eos_token_id lowercase_ = pad_token_id lowercase_ = bos_token_id def _UpperCAmelCase ( self : Optional[Any]): """simple docstring""" lowercase_ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size).clip(3 , self.vocab_size) lowercase_ = np.expand_dims(np.array([self.eos_token_id] * self.batch_size) , 1) lowercase_ = np.concatenate([input_ids, eos_tensor] , axis=1) lowercase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) lowercase_ = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) lowercase_ = prepare_pegasus_inputs_dict(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_) return config, inputs_dict def _UpperCAmelCase ( self : Optional[Any] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Optional[Any]): """simple docstring""" lowercase_ = 2_0 lowercase_ = model_class_name(lowerCAmelCase_) lowercase_ = model.encode(inputs_dict["""input_ids"""]) lowercase_ , lowercase_ = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) lowercase_ = model.init_cache(decoder_input_ids.shape[0] , lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype="""i4""") lowercase_ = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1)[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) lowercase_ = model.decode( decoder_input_ids[:, :-1] , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , past_key_values=lowerCAmelCase_ , decoder_position_ids=lowerCAmelCase_ , ) lowercase_ = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""") lowercase_ = model.decode( decoder_input_ids[:, -1:] , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , past_key_values=outputs_cache.past_key_values , decoder_position_ids=lowerCAmelCase_ , ) lowercase_ = model.decode(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]))) self.parent.assertTrue(diff < 1E-3 , msg=F'''Max diff is {diff}''') def _UpperCAmelCase ( self : Optional[int] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Dict): """simple docstring""" lowercase_ = 2_0 lowercase_ = model_class_name(lowerCAmelCase_) lowercase_ = model.encode(inputs_dict["""input_ids"""]) lowercase_ , lowercase_ = ( inputs_dict["""decoder_input_ids"""], inputs_dict["""decoder_attention_mask"""], ) lowercase_ = jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1])), ] , axis=-1 , ) lowercase_ = model.init_cache(decoder_input_ids.shape[0] , lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1)[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) lowercase_ = model.decode( decoder_input_ids[:, :-1] , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , past_key_values=lowerCAmelCase_ , decoder_position_ids=lowerCAmelCase_ , ) lowercase_ = jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype="""i4""") lowercase_ = model.decode( decoder_input_ids[:, -1:] , lowerCAmelCase_ , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=lowerCAmelCase_ , decoder_position_ids=lowerCAmelCase_ , ) lowercase_ = model.decode(lowerCAmelCase_ , lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_) lowercase_ = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]))) self.parent.assertTrue(diff < 1E-3 , msg=F'''Max diff is {diff}''') def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None , __lowerCAmelCase=None , ) -> Optional[Any]: '''simple docstring''' if attention_mask is None: lowercase_ = np.not_equal(__lowerCAmelCase , config.pad_token_id ).astype(np.inta ) if decoder_attention_mask is None: lowercase_ = np.concatenate( [ np.ones(decoder_input_ids[:, :1].shape , dtype=np.inta ), np.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ).astype(np.inta ), ] , axis=-1 , ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, } @require_flax class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , unittest.TestCase ): lowercase__ = ( ( FlaxPegasusForConditionalGeneration, FlaxPegasusModel, ) if is_flax_available() else () ) lowercase__ = (FlaxPegasusForConditionalGeneration,) if is_flax_available() else () lowercase__ = True lowercase__ = False lowercase__ = False lowercase__ = False def _UpperCAmelCase ( self : Tuple): """simple docstring""" lowercase_ = FlaxPegasusModelTester(self) lowercase_ = ConfigTester(self , config_class=lowerCAmelCase_) def _UpperCAmelCase ( self : Any): """simple docstring""" self.config_tester.run_common_tests() def _UpperCAmelCase ( self : List[str]): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_) def _UpperCAmelCase ( self : Dict): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_) def _UpperCAmelCase ( self : Dict): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): lowercase_ = self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = model_class(lowerCAmelCase_) @jax.jit def encode_jitted(lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[int]=None , **lowerCAmelCase_ : Optional[int]): return model.encode(input_ids=lowerCAmelCase_ , attention_mask=lowerCAmelCase_) with self.subTest("""JIT Enabled"""): lowercase_ = encode_jitted(**lowerCAmelCase_).to_tuple() with self.subTest("""JIT Disabled"""): with jax.disable_jit(): lowercase_ = encode_jitted(**lowerCAmelCase_).to_tuple() self.assertEqual(len(lowerCAmelCase_) , len(lowerCAmelCase_)) for jitted_output, output in zip(lowerCAmelCase_ , lowerCAmelCase_): self.assertEqual(jitted_output.shape , output.shape) def _UpperCAmelCase ( self : Tuple): """simple docstring""" lowercase_ , lowercase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): lowercase_ = model_class(lowerCAmelCase_) lowercase_ = model.encode(inputs_dict["""input_ids"""] , inputs_dict["""attention_mask"""]) lowercase_ = { """decoder_input_ids""": inputs_dict["""decoder_input_ids"""], """decoder_attention_mask""": inputs_dict["""decoder_attention_mask"""], """encoder_outputs""": encoder_outputs, } @jax.jit def decode_jitted(lowerCAmelCase_ : int , lowerCAmelCase_ : int , lowerCAmelCase_ : Dict): return model.decode( decoder_input_ids=lowerCAmelCase_ , decoder_attention_mask=lowerCAmelCase_ , encoder_outputs=lowerCAmelCase_ , ) with self.subTest("""JIT Enabled"""): lowercase_ = decode_jitted(**lowerCAmelCase_).to_tuple() with self.subTest("""JIT Disabled"""): with jax.disable_jit(): lowercase_ = decode_jitted(**lowerCAmelCase_).to_tuple() self.assertEqual(len(lowerCAmelCase_) , len(lowerCAmelCase_)) for jitted_output, output in zip(lowerCAmelCase_ , lowerCAmelCase_): self.assertEqual(jitted_output.shape , output.shape) @slow def _UpperCAmelCase ( self : Tuple): """simple docstring""" for model_class_name in self.all_model_classes: lowercase_ = model_class_name.from_pretrained("""google/pegasus-large""" , from_pt=lowerCAmelCase_) lowercase_ = np.ones((1, 1)) lowercase_ = model(lowerCAmelCase_) self.assertIsNotNone(lowerCAmelCase_) @slow def _UpperCAmelCase ( self : Any): """simple docstring""" lowercase_ = FlaxPegasusForConditionalGeneration.from_pretrained("""google/pegasus-xsum""") lowercase_ = PegasusTokenizer.from_pretrained("""google/pegasus-xsum""") lowercase_ = [ """ PG&E stated it scheduled the blackouts in response to forecasts for high winds amid dry conditions. The aim is to reduce the risk of wildfires. Nearly 800 thousand customers were scheduled to be affected by the shutoffs which were expected to last through at least midday tomorrow.""", """ The London trio are up for best UK act and best album, as well as getting two nominations in the best song category.\"We got told like this morning 'Oh I think you're nominated'\", said Dappy.\"And I was like 'Oh yeah, which one?' And now we've got nominated for four awards. I mean, wow!\"Bandmate Fazer added: \"We thought it's best of us to come down and mingle with everyone and say hello to the cameras. And now we find we've got four nominations.\"The band have two shots at the best song prize, getting the nod for their Tynchy Stryder collaboration Number One, and single Strong Again.Their album Uncle B will also go up against records by the likes of Beyonce and Kanye West.N-Dubz picked up the best newcomer Mobo in 2007, but female member Tulisa said they wouldn't be too disappointed if they didn't win this time around.\"At the end of the day we're grateful to be where we are in our careers.\"If it don't happen then it don't happen - live to fight another day and keep on making albums and hits for the fans.\"Dappy also revealed they could be performing live several times on the night.The group will be doing Number One and also a possible rendition of the War Child single, I Got Soul.The charity song is a re-working of The Killers' All These Things That I've Done and is set to feature artists like Chipmunk, Ironik and Pixie Lott.This year's Mobos will be held outside of London for the first time, in Glasgow on 30 September.N-Dubz said they were looking forward to performing for their Scottish fans and boasted about their recent shows north of the border.\"We just done Edinburgh the other day,\" said Dappy.\"We smashed up an N-Dubz show over there. We done Aberdeen about three or four months ago - we smashed up that show over there! Everywhere we go we smash it up!\" """, ] lowercase_ = [ """California's largest electricity provider has turned off power to hundreds of thousands of customers.""", """Pop group N-Dubz have revealed they were surprised to get four nominations for this year's Mobo Awards.""", ] lowercase_ = tokenizer(lowerCAmelCase_ , return_tensors="""np""" , truncation=lowerCAmelCase_ , max_length=5_1_2 , padding=lowerCAmelCase_) lowercase_ = model.generate(**lowerCAmelCase_ , num_beams=2).sequences lowercase_ = tokenizer.batch_decode(lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_) assert tgt_text == decoded
313
1
'''simple docstring''' from __future__ import annotations def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): return [ord(_SCREAMING_SNAKE_CASE ) - 96 for elem in plain] def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): return "".join(chr(elem + 96 ) for elem in encoded ) def __SCREAMING_SNAKE_CASE ( ): _snake_case = encode(input("""-> """ ).strip().lower() ) print("""Encoded: """ , _SCREAMING_SNAKE_CASE ) print("""Decoded:""" , decode(_SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": main()
341
'''simple docstring''' def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): if discount_rate < 0: raise ValueError("""Discount rate cannot be negative""" ) if not cash_flows: raise ValueError("""Cash flows list cannot be empty""" ) _snake_case = sum( cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(_SCREAMING_SNAKE_CASE ) ) return round(_SCREAMING_SNAKE_CASE , ndigits=2 ) if __name__ == "__main__": import doctest doctest.testmod()
341
1
'''simple docstring''' import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 _A : Dict =get_tests_dir('''fixtures''') _A : List[Any] =get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') _A : List[Any] =get_tests_dir('''fixtures/dummy-config.json''') class _lowercase ( unittest.TestCase ): def lowerCamelCase_ ( self: Tuple ): lowerCamelCase__ : List[Any] = 0 def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : Dict = AutoFeatureExtractor.from_pretrained("""facebook/wav2vec2-base-960h""" ) self.assertIsInstance(__lowerCAmelCase , __lowerCAmelCase ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : Union[str, Any] = AutoFeatureExtractor.from_pretrained(__lowerCAmelCase ) self.assertIsInstance(__lowerCAmelCase , __lowerCAmelCase ) def lowerCamelCase_ ( self: List[str] ): with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase__ : str = WavaVecaConfig() # remove feature_extractor_type to make sure config.json alone is enough to load feature processor locally lowerCamelCase__ : Optional[int] = AutoFeatureExtractor.from_pretrained(__lowerCAmelCase ).to_dict() config_dict.pop("""feature_extractor_type""" ) lowerCamelCase__ : Tuple = WavaVecaFeatureExtractor(**__lowerCAmelCase ) # save in new folder model_config.save_pretrained(__lowerCAmelCase ) config.save_pretrained(__lowerCAmelCase ) lowerCamelCase__ : Union[str, Any] = AutoFeatureExtractor.from_pretrained(__lowerCAmelCase ) # make sure private variable is not incorrectly saved lowerCamelCase__ : str = json.loads(config.to_json_string() ) self.assertTrue("""_processor_class""" not in dict_as_saved ) self.assertIsInstance(__lowerCAmelCase , __lowerCAmelCase ) def lowerCamelCase_ ( self: int ): lowerCamelCase__ : int = AutoFeatureExtractor.from_pretrained(__lowerCAmelCase ) self.assertIsInstance(__lowerCAmelCase , __lowerCAmelCase ) def lowerCamelCase_ ( self: Dict ): with self.assertRaisesRegex( __lowerCAmelCase , """bert-base is not a local folder and is not a valid model identifier""" ): lowerCamelCase__ : str = AutoFeatureExtractor.from_pretrained("""bert-base""" ) def lowerCamelCase_ ( self: str ): with self.assertRaisesRegex( __lowerCAmelCase , R"""aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)""" ): lowerCamelCase__ : Union[str, Any] = AutoFeatureExtractor.from_pretrained(__lowerCAmelCase , revision="""aaaaaa""" ) def lowerCamelCase_ ( self: Optional[Any] ): with self.assertRaisesRegex( __lowerCAmelCase , """hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.""" , ): lowerCamelCase__ : Tuple = AutoFeatureExtractor.from_pretrained("""hf-internal-testing/config-no-model""" ) def lowerCamelCase_ ( self: Union[str, Any] ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(__lowerCAmelCase ): lowerCamelCase__ : Dict = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" ) # If remote code is disabled, we can't load this config. with self.assertRaises(__lowerCAmelCase ): lowerCamelCase__ : str = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=__lowerCAmelCase ) lowerCamelCase__ : List[Any] = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=__lowerCAmelCase ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) # Test feature extractor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(__lowerCAmelCase ) lowerCamelCase__ : Dict = AutoFeatureExtractor.from_pretrained(__lowerCAmelCase , trust_remote_code=__lowerCAmelCase ) self.assertEqual(reloaded_feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) def lowerCamelCase_ ( self: Union[str, Any] ): try: AutoConfig.register("""custom""" , __lowerCAmelCase ) AutoFeatureExtractor.register(__lowerCAmelCase , __lowerCAmelCase ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__lowerCAmelCase ): AutoFeatureExtractor.register(__lowerCAmelCase , __lowerCAmelCase ) # Now that the config is registered, it can be used as any other config with the auto-API lowerCamelCase__ : Optional[Any] = CustomFeatureExtractor.from_pretrained(__lowerCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(__lowerCAmelCase ) lowerCamelCase__ : Dict = AutoFeatureExtractor.from_pretrained(__lowerCAmelCase ) self.assertIsInstance(__lowerCAmelCase , __lowerCAmelCase ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] def lowerCamelCase_ ( self: str ): class _lowercase ( lowerCAmelCase_ ): a = True try: AutoConfig.register("""custom""" , __lowerCAmelCase ) AutoFeatureExtractor.register(__lowerCAmelCase , __lowerCAmelCase ) # If remote code is not set, the default is to use local lowerCamelCase__ : Tuple = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) self.assertTrue(feature_extractor.is_local ) # If remote code is disabled, we load the local one. lowerCamelCase__ : Dict = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=__lowerCAmelCase ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) self.assertTrue(feature_extractor.is_local ) # If remote is enabled, we load from the Hub lowerCamelCase__ : int = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=__lowerCAmelCase ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) self.assertTrue(not hasattr(__lowerCAmelCase , """is_local""" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig]
358
'''simple docstring''' from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging _A : List[str] =logging.get_logger(__name__) class _lowercase ( _lowercase ): a = ["""input_features""", """attention_mask"""] def __init__( self: Optional[int] , UpperCamelCase__: str=80 , UpperCamelCase__: Union[str, Any]=16_000 , UpperCamelCase__: str=0.0 , UpperCamelCase__: Tuple=10 , UpperCamelCase__: Dict=25 , UpperCamelCase__: List[Any]="hamming_window" , UpperCamelCase__: Tuple=32_768.0 , UpperCamelCase__: List[Any]=0.97 , UpperCamelCase__: Optional[int]=1.0 , UpperCamelCase__: str=True , UpperCamelCase__: Optional[Any]=True , UpperCamelCase__: Optional[int]=False , **UpperCamelCase__: Optional[int] , ): super().__init__(feature_size=UpperCamelCase__ , sampling_rate=UpperCamelCase__ , padding_value=UpperCamelCase__ , **UpperCamelCase__ ) lowerCamelCase__ : Dict = feature_size lowerCamelCase__ : Tuple = sampling_rate lowerCamelCase__ : Union[str, Any] = padding_value lowerCamelCase__ : Union[str, Any] = hop_length lowerCamelCase__ : Any = win_length lowerCamelCase__ : Tuple = frame_signal_scale lowerCamelCase__ : Dict = preemphasis_coeff lowerCamelCase__ : Optional[Any] = mel_floor lowerCamelCase__ : Any = normalize_means lowerCamelCase__ : Optional[int] = normalize_vars lowerCamelCase__ : Dict = win_function lowerCamelCase__ : Dict = return_attention_mask lowerCamelCase__ : Optional[Any] = win_length * sampling_rate // 1_000 lowerCamelCase__ : Tuple = hop_length * sampling_rate // 1_000 lowerCamelCase__ : List[Any] = optimal_fft_length(self.sample_size ) lowerCamelCase__ : List[str] = (self.n_fft // 2) + 1 def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: np.array ): if self.win_function == "hamming_window": lowerCamelCase__ : Dict = window_function(window_length=self.sample_size , name=self.win_function , periodic=UpperCamelCase__ ) else: lowerCamelCase__ : Union[str, Any] = window_function(window_length=self.sample_size , name=self.win_function ) lowerCamelCase__ : List[Any] = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.feature_size , min_frequency=0.0 , max_frequency=self.sampling_rate / 2.0 , sampling_rate=self.sampling_rate , ) lowerCamelCase__ : Tuple = spectrogram( one_waveform * self.frame_signal_scale , window=UpperCamelCase__ , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , center=UpperCamelCase__ , preemphasis=self.preemphasis_coeff , mel_filters=UpperCamelCase__ , mel_floor=self.mel_floor , log_mel="""log""" , ) return msfc_features.T def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase__: List[str] , UpperCamelCase__: List[Any] , UpperCamelCase__: Union[str, Any] ): # make sure we normalize float32 arrays if self.normalize_means: lowerCamelCase__ : List[str] = x[:input_length].mean(axis=0 ) lowerCamelCase__ : List[str] = np.subtract(UpperCamelCase__ , UpperCamelCase__ ) if self.normalize_vars: lowerCamelCase__ : Union[str, Any] = x[:input_length].std(axis=0 ) lowerCamelCase__ : str = np.divide(UpperCamelCase__ , UpperCamelCase__ ) if input_length < x.shape[0]: lowerCamelCase__ : List[Any] = padding_value # make sure array is in float32 lowerCamelCase__ : str = x.astype(np.floataa ) return x def lowerCamelCase_ ( self: str , UpperCamelCase__: List[np.ndarray] , UpperCamelCase__: Optional[np.ndarray] = None ): lowerCamelCase__ : Optional[int] = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(UpperCamelCase__ , UpperCamelCase__ , self.padding_value ) for x, n in zip(UpperCamelCase__ , UpperCamelCase__ )] def __call__( self: List[str] , UpperCamelCase__: Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , UpperCamelCase__: Union[bool, str, PaddingStrategy] = False , UpperCamelCase__: Optional[int] = None , UpperCamelCase__: bool = False , UpperCamelCase__: Optional[int] = None , UpperCamelCase__: Optional[bool] = None , UpperCamelCase__: Optional[Union[str, TensorType]] = None , UpperCamelCase__: Optional[int] = None , **UpperCamelCase__: Tuple , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' F''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' F''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( """It is strongly recommended to pass the ``sampling_rate`` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) lowerCamelCase__ : List[str] = isinstance(UpperCamelCase__ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'''Only mono-channel audio is supported for input to {self}''' ) lowerCamelCase__ : List[str] = is_batched_numpy or ( isinstance(UpperCamelCase__ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCamelCase__ : str = [np.asarray(UpperCamelCase__ , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(UpperCamelCase__ , np.ndarray ): lowerCamelCase__ : List[Any] = np.asarray(UpperCamelCase__ , dtype=np.floataa ) elif isinstance(UpperCamelCase__ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCamelCase__ : Optional[Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCamelCase__ : Optional[Any] = [raw_speech] # extract fbank features lowerCamelCase__ : str = [self._extract_mfsc_features(UpperCamelCase__ ) for one_waveform in raw_speech] # convert into correct format for padding lowerCamelCase__ : str = BatchFeature({"""input_features""": features} ) lowerCamelCase__ : str = self.pad( UpperCamelCase__ , padding=UpperCamelCase__ , max_length=UpperCamelCase__ , truncation=UpperCamelCase__ , pad_to_multiple_of=UpperCamelCase__ , return_attention_mask=UpperCamelCase__ , **UpperCamelCase__ , ) # make sure list is in array format lowerCamelCase__ : str = padded_inputs.get("""input_features""" ) if isinstance(input_features[0] , UpperCamelCase__ ): lowerCamelCase__ : Dict = [np.asarray(UpperCamelCase__ , dtype=np.floataa ) for feature in input_features] lowerCamelCase__ : List[str] = padded_inputs.get("""attention_mask""" ) if attention_mask is not None: lowerCamelCase__ : Any = [np.asarray(UpperCamelCase__ , dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: lowerCamelCase__ : str = ( np.array(UpperCamelCase__ , dtype=np.intaa ) if self._get_padding_strategies(UpperCamelCase__ , max_length=UpperCamelCase__ ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) lowerCamelCase__ : Union[str, Any] = self.normalize( padded_inputs["""input_features"""] , attention_mask=UpperCamelCase__ ) if return_tensors is not None: lowerCamelCase__ : Dict = padded_inputs.convert_to_tensors(UpperCamelCase__ ) return padded_inputs
129
0
"""simple docstring""" import requests from bsa import BeautifulSoup def lowercase ( A_ , A_ )-> str: '''simple docstring''' a : List[Any] = BeautifulSoup(requests.get(A_ , params=A_ ).content , "html.parser" ) a : Tuple = soup.find("div" , attrs={"class": "gs_ri"} ) a : Any = div.find("div" , attrs={"class": "gs_fl"} ).find_all("a" ) return anchors[2].get_text() if __name__ == "__main__": __lowercase = { """title""": ( """Precisely geometry controlled microsupercapacitors for ultrahigh areal """ """capacitance, volumetric capacitance, and energy density""" ), """journal""": """Chem. Mater.""", """volume""": 30, """pages""": """3979-3990""", """year""": 2018, """hl""": """en""", } print(get_citation("""https://scholar.google.com/scholar_lookup""", params=params))
40
"""simple docstring""" import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _A ( _a ,unittest.TestCase ): """simple docstring""" UpperCAmelCase : str = LayoutLMTokenizer UpperCAmelCase : int = LayoutLMTokenizerFast UpperCAmelCase : Union[str, Any] = True UpperCAmelCase : Optional[Any] = True def __snake_case ( self : Optional[int]): super().setUp() a : Tuple = [ "[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] a : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens])) def __snake_case ( self : Optional[int] , **__UpperCAmelCase : Tuple): return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **__UpperCAmelCase) def __snake_case ( self : Union[str, Any] , __UpperCAmelCase : str): a : Tuple = "UNwant\u00E9d,running" a : Dict = "unwanted, running" return input_text, output_text def __snake_case ( self : Any): a : List[Any] = self.tokenizer_class(self.vocab_file) a : str = tokenizer.tokenize("UNwant\u00E9d,running") self.assertListEqual(__UpperCAmelCase , ["un", "##want", "##ed", ",", "runn", "##ing"]) self.assertListEqual(tokenizer.convert_tokens_to_ids(__UpperCAmelCase) , [7, 4, 5, 10, 8, 9]) def __snake_case ( self : Dict): pass
40
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_tf_available, is_torch_available, ) A__ : Optional[int] = { 'configuration_speech_to_text': ['SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Speech2TextConfig'], 'processing_speech_to_text': ['Speech2TextProcessor'], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : Dict = ['Speech2TextTokenizer'] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : Optional[Any] = ['Speech2TextFeatureExtractor'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : Dict = [ 'TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFSpeech2TextForConditionalGeneration', 'TFSpeech2TextModel', 'TFSpeech2TextPreTrainedModel', ] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : Optional[int] = [ 'SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'Speech2TextForConditionalGeneration', 'Speech2TextModel', 'Speech2TextPreTrainedModel', ] if TYPE_CHECKING: from .configuration_speech_to_text import SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, SpeechaTextConfig from .processing_speech_to_text import SpeechaTextProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speech_to_text import SpeechaTextTokenizer try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_speech_to_text import SpeechaTextFeatureExtractor try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_speech_to_text import ( TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, TFSpeechaTextForConditionalGeneration, TFSpeechaTextModel, TFSpeechaTextPreTrainedModel, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_to_text import ( SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechaTextForConditionalGeneration, SpeechaTextModel, SpeechaTextPreTrainedModel, ) else: import sys A__ : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
209
"""simple docstring""" import torch def _snake_case ( ) -> Union[str, Any]: if torch.cuda.is_available(): lowerCamelCase_ : int =torch.cuda.device_count() else: lowerCamelCase_ : List[str] =0 print(F"""Successfully ran on {num_gpus} GPUs""" ) if __name__ == "__main__": main()
209
1
from __future__ import annotations from scipy.special import comb # type: ignore class snake_case__ : """simple docstring""" def __init__( self : Any , __lowerCamelCase : list[tuple[float, float]] ) -> Tuple: a = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. a = len(__lowerCamelCase ) - 1 def __UpperCAmelCase ( self : Union[str, Any] , __lowerCamelCase : float ) -> list[float]: assert 0 <= t <= 1, "Time t must be between 0 and 1." a = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , __lowerCamelCase ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(__lowerCamelCase ) , 5 ) == 1 return output_values def __UpperCAmelCase ( self : Optional[Any] , __lowerCamelCase : float ) -> tuple[float, float]: assert 0 <= t <= 1, "Time t must be between 0 and 1." a = self.basis_function(__lowerCamelCase ) a = 0.0 a = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def __UpperCAmelCase ( self : Optional[Any] , __lowerCamelCase : float = 0.01 ) -> List[str]: from matplotlib import pyplot as plt # type: ignore a = [] # x coordinates of points to plot a = [] # y coordinates of points to plot a = 0.0 while t <= 1: a = self.bezier_curve_function(__lowerCamelCase ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size a = [i[0] for i in self.list_of_points] a = [i[1] for i in self.list_of_points] plt.plot( __lowerCamelCase , __lowerCamelCase , color="blue" , label="Curve of Degree " + str(self.degree ) , ) plt.scatter(__lowerCamelCase , __lowerCamelCase , color="red" , label="Control Points" ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
107
"""simple docstring""" def __lowerCAmelCase ( lowercase : list[int] ) -> float: """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError("List is empty" ) snake_case : List[str] = sum(lowercase ) / len(lowercase ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(lowercase ) if __name__ == "__main__": import doctest doctest.testmod()
203
0
from collections.abc import Sequence def _snake_case ( lowerCamelCase__ : Sequence[float] , lowerCamelCase__ : float ) -> float: return sum(c * (x**i) for i, c in enumerate(lowerCamelCase__ ) ) def _snake_case ( lowerCamelCase__ : Sequence[float] , lowerCamelCase__ : float ) -> float: lowerCamelCase_ : List[str] =0.0 for coeff in reversed(lowerCamelCase__ ): lowerCamelCase_ : Any =result * x + coeff return result if __name__ == "__main__": A__ : str = (0.0, 0.0, 5.0, 9.3, 7.0) A__ : List[str] = 10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
360
"""simple docstring""" import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class lowercase__ : def __init__( self : List[Any] , snake_case__ : Optional[Any] , snake_case__ : Tuple=13 , snake_case__ : str=7 , snake_case__ : Union[str, Any]=6 , snake_case__ : str=17 , snake_case__ : Any=23 , snake_case__ : int=11 , snake_case__ : Tuple=True , ): lowerCamelCase_ : str =parent lowerCamelCase_ : Union[str, Any] =batch_size lowerCamelCase_ : List[Any] =seq_length lowerCamelCase_ : Union[str, Any] =act_dim lowerCamelCase_ : Optional[Any] =state_dim lowerCamelCase_ : Optional[Any] =hidden_size lowerCamelCase_ : Tuple =max_length lowerCamelCase_ : List[Any] =is_training def UpperCAmelCase__ ( self : Dict ): lowerCamelCase_ : Optional[Any] =floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) lowerCamelCase_ : Optional[Any] =floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) lowerCamelCase_ : List[Any] =floats_tensor((self.batch_size, self.seq_length, 1) ) lowerCamelCase_ : Optional[Any] =floats_tensor((self.batch_size, self.seq_length, 1) ) lowerCamelCase_ : List[Any] =ids_tensor((self.batch_size, self.seq_length) , vocab_size=1000 ) lowerCamelCase_ : Optional[int] =random_attention_mask((self.batch_size, self.seq_length) ) lowerCamelCase_ : List[str] =self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def UpperCAmelCase__ ( self : Any ): return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def UpperCAmelCase__ ( self : Optional[Any] , snake_case__ : Tuple , snake_case__ : Dict , snake_case__ : int , snake_case__ : Dict , snake_case__ : Optional[int] , snake_case__ : List[str] , snake_case__ : List[str] , ): lowerCamelCase_ : Tuple =DecisionTransformerModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() lowerCamelCase_ : str =model(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def UpperCAmelCase__ ( self : List[str] ): lowerCamelCase_ : List[str] =self.prepare_config_and_inputs() ( ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ) : Optional[int] =config_and_inputs lowerCamelCase_ : Optional[int] ={ "states": states, "actions": actions, "rewards": rewards, "returns_to_go": returns_to_go, "timesteps": timesteps, "attention_mask": attention_mask, } return config, inputs_dict @require_torch class lowercase__ ( snake_case__, snake_case__, snake_case__, unittest.TestCase ): _UpperCAmelCase :Optional[Any] = (DecisionTransformerModel,) if is_torch_available() else () _UpperCAmelCase :int = () _UpperCAmelCase :int = {"feature-extraction": DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids _UpperCAmelCase :Union[str, Any] = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features _UpperCAmelCase :Optional[Any] = False _UpperCAmelCase :Tuple = False _UpperCAmelCase :Tuple = False _UpperCAmelCase :List[Any] = False _UpperCAmelCase :Dict = False _UpperCAmelCase :Any = False _UpperCAmelCase :List[Any] = False _UpperCAmelCase :int = False _UpperCAmelCase :str = False def UpperCAmelCase__ ( self : str ): lowerCamelCase_ : Dict =DecisionTransformerModelTester(self ) lowerCamelCase_ : str =ConfigTester(self , config_class=snake_case__ , hidden_size=37 ) def UpperCAmelCase__ ( self : Union[str, Any] ): self.config_tester.run_common_tests() def UpperCAmelCase__ ( self : int ): lowerCamelCase_ : Tuple =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) @slow def UpperCAmelCase__ ( self : List[str] ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase_ : str =DecisionTransformerModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) def UpperCAmelCase__ ( self : str ): lowerCamelCase_ , lowerCamelCase_ : Dict =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ : List[Any] =model_class(snake_case__ ) lowerCamelCase_ : int =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase_ : List[Any] =[*signature.parameters.keys()] lowerCamelCase_ : List[str] =[ "states", "actions", "rewards", "returns_to_go", "timesteps", "attention_mask", ] self.assertListEqual(arg_names[: len(snake_case__ )] , snake_case__ ) @require_torch class lowercase__ ( unittest.TestCase ): @slow def UpperCAmelCase__ ( self : Any ): lowerCamelCase_ : Optional[int] =2 # number of steps of autoregressive prediction we will perform lowerCamelCase_ : int =10 # defined by the RL environment, may be normalized lowerCamelCase_ : List[Any] =DecisionTransformerModel.from_pretrained("edbeeching/decision-transformer-gym-hopper-expert" ) lowerCamelCase_ : Union[str, Any] =model.to(snake_case__ ) lowerCamelCase_ : Any =model.config torch.manual_seed(0 ) lowerCamelCase_ : Optional[Any] =torch.randn(1 , 1 , config.state_dim ).to(device=snake_case__ , dtype=torch.floataa ) # env.reset() lowerCamelCase_ : Optional[Any] =torch.tensor( [[0.242_793, -0.28_693_074, 0.8_742_613], [0.67_815_274, -0.08_101_085, -0.12_952_147]] , device=snake_case__ ) lowerCamelCase_ : int =torch.tensor(snake_case__ , device=snake_case__ , dtype=torch.floataa ).reshape(1 , 1 , 1 ) lowerCamelCase_ : str =state lowerCamelCase_ : Optional[int] =torch.zeros(1 , 0 , config.act_dim , device=snake_case__ , dtype=torch.floataa ) lowerCamelCase_ : int =torch.zeros(1 , 0 , device=snake_case__ , dtype=torch.floataa ) lowerCamelCase_ : Tuple =torch.tensor(0 , device=snake_case__ , dtype=torch.long ).reshape(1 , 1 ) for step in range(snake_case__ ): lowerCamelCase_ : str =torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=snake_case__ )] , dim=1 ) lowerCamelCase_ : Union[str, Any] =torch.cat([rewards, torch.zeros(1 , 1 , device=snake_case__ )] , dim=1 ) lowerCamelCase_ : Optional[int] =torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ : Dict =model( states=snake_case__ , actions=snake_case__ , rewards=snake_case__ , returns_to_go=snake_case__ , timesteps=snake_case__ , attention_mask=snake_case__ , return_dict=snake_case__ , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1E-4 ) ) lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ : Optional[Any] =( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=snake_case__ , dtype=torch.floataa ), 1.0, False, {}, ) lowerCamelCase_ : str =action_pred[0, -1] lowerCamelCase_ : Optional[int] =torch.cat([states, state] , dim=1 ) lowerCamelCase_ : Optional[Any] =returns_to_go[0, -1] - reward lowerCamelCase_ : str =torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) lowerCamelCase_ : int =torch.cat( [timesteps, torch.ones((1, 1) , device=snake_case__ , dtype=torch.long ) * (step + 1)] , dim=1 )
209
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a_ :List[str] = logging.get_logger(__name__) a_ :Union[str, Any] = { """facebook/xlm-roberta-xl""": """https://huggingface.co/facebook/xlm-roberta-xl/resolve/main/config.json""", """facebook/xlm-roberta-xxl""": """https://huggingface.co/facebook/xlm-roberta-xxl/resolve/main/config.json""", # See all XLM-RoBERTa-XL models at https://huggingface.co/models?filter=xlm-roberta-xl } class snake_case__ ( __UpperCamelCase ): """simple docstring""" _SCREAMING_SNAKE_CASE = """xlm-roberta-xl""" def __init__( self : List[str], _snake_case : Dict=2_5_0_8_8_0, _snake_case : int=2_5_6_0, _snake_case : Tuple=3_6, _snake_case : Dict=3_2, _snake_case : Dict=1_0_2_4_0, _snake_case : List[Any]="gelu", _snake_case : Optional[Any]=0.1, _snake_case : Optional[int]=0.1, _snake_case : int=5_1_4, _snake_case : Any=1, _snake_case : Dict=0.0_2, _snake_case : Union[str, Any]=1e-05, _snake_case : List[Any]=1, _snake_case : Any=0, _snake_case : List[str]=2, _snake_case : List[Any]="absolute", _snake_case : Tuple=True, _snake_case : Optional[Any]=None, **_snake_case : Any, ) ->List[str]: super().__init__(pad_token_id=_snake_case, bos_token_id=_snake_case, eos_token_id=_snake_case, **_snake_case ) snake_case__ : Dict = vocab_size snake_case__ : str = hidden_size snake_case__ : str = num_hidden_layers snake_case__ : List[Any] = num_attention_heads snake_case__ : List[str] = hidden_act snake_case__ : Dict = intermediate_size snake_case__ : Any = hidden_dropout_prob snake_case__ : List[str] = attention_probs_dropout_prob snake_case__ : Tuple = max_position_embeddings snake_case__ : Any = type_vocab_size snake_case__ : Optional[Any] = initializer_range snake_case__ : List[str] = layer_norm_eps snake_case__ : Any = position_embedding_type snake_case__ : int = use_cache snake_case__ : int = classifier_dropout class snake_case__ ( __UpperCamelCase ): """simple docstring""" @property def lowercase_ ( self : Optional[Any] ) ->Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": snake_case__ : Optional[Any] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: snake_case__ : Dict = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
277
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : Optional[int] = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', 'decoder.output_projection.weight', '_float_tensor', 'encoder.embed_positions._float_tensor', 'decoder.embed_positions._float_tensor', ] for k in ignore_keys: state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> List[Any]: lowerCamelCase__ , lowerCamelCase__ : List[str] = emb.weight.shape lowerCamelCase__ : Tuple = nn.Linear(_UpperCAmelCase , _UpperCAmelCase , bias=_UpperCAmelCase ) lowerCamelCase__ : Dict = emb.weight.data return lin_layer def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> str: lowerCamelCase__ : Tuple = torch.load(_UpperCAmelCase , map_location='cpu' ) lowerCamelCase__ : List[str] = mam_aaa['args'] or mam_aaa['cfg']['model'] lowerCamelCase__ : Optional[int] = mam_aaa['model'] remove_ignore_keys_(_UpperCAmelCase ) lowerCamelCase__ : str = state_dict['encoder.embed_tokens.weight'].shape[0] lowerCamelCase__ : Union[str, Any] = MaMaaaConfig( vocab_size=_UpperCAmelCase , max_position_embeddings=1024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='relu' , ) lowerCamelCase__ : Optional[Any] = state_dict['decoder.embed_tokens.weight'] lowerCamelCase__ : Union[str, Any] = MaMaaaForConditionalGeneration(_UpperCAmelCase ) model.model.load_state_dict(_UpperCAmelCase , strict=_UpperCAmelCase ) lowerCamelCase__ : List[str] = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": _UpperCAmelCase : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument("""fairseq_path""", type=str, help="""path to a model.pt on local filesystem.""") parser.add_argument("""pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") _UpperCAmelCase : str = parser.parse_args() _UpperCAmelCase : Optional[Any] = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
50
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A_ : Optional[Any] = logging.get_logger(__name__) A_ : Any = { "google/pegasus-large": "https://huggingface.co/google/pegasus-large/resolve/main/config.json", # See all PEGASUS models at https://huggingface.co/models?filter=pegasus } class a_ ( snake_case_ ): '''simple docstring''' lowerCamelCase__ : Dict = 'pegasus' lowerCamelCase__ : Optional[Any] = ['past_key_values'] lowerCamelCase__ : List[str] = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__(self, lowerCamelCase_=5_0_2_6_5, lowerCamelCase_=1_0_2_4, lowerCamelCase_=1_2, lowerCamelCase_=4_0_9_6, lowerCamelCase_=1_6, lowerCamelCase_=1_2, lowerCamelCase_=4_0_9_6, lowerCamelCase_=1_6, lowerCamelCase_=0.0, lowerCamelCase_=0.0, lowerCamelCase_=True, lowerCamelCase_=True, lowerCamelCase_="gelu", lowerCamelCase_=1_0_2_4, lowerCamelCase_=0.1, lowerCamelCase_=0.0, lowerCamelCase_=0.0, lowerCamelCase_=0.02, lowerCamelCase_=0, lowerCamelCase_=False, lowerCamelCase_=0, lowerCamelCase_=1, lowerCamelCase_=1, **lowerCamelCase_, ): '''simple docstring''' lowerCamelCase__ : Any = vocab_size lowerCamelCase__ : str = max_position_embeddings lowerCamelCase__ : Optional[int] = d_model lowerCamelCase__ : Union[str, Any] = encoder_ffn_dim lowerCamelCase__ : List[Any] = encoder_layers lowerCamelCase__ : Union[str, Any] = encoder_attention_heads lowerCamelCase__ : Union[str, Any] = decoder_ffn_dim lowerCamelCase__ : str = decoder_layers lowerCamelCase__ : List[str] = decoder_attention_heads lowerCamelCase__ : Dict = dropout lowerCamelCase__ : Any = attention_dropout lowerCamelCase__ : Tuple = activation_dropout lowerCamelCase__ : Any = activation_function lowerCamelCase__ : int = init_std lowerCamelCase__ : List[str] = encoder_layerdrop lowerCamelCase__ : Union[str, Any] = decoder_layerdrop lowerCamelCase__ : Tuple = use_cache lowerCamelCase__ : Any = encoder_layers lowerCamelCase__ : Tuple = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=lowerCamelCase_, eos_token_id=lowerCamelCase_, is_encoder_decoder=lowerCamelCase_, decoder_start_token_id=lowerCamelCase_, forced_eos_token_id=lowerCamelCase_, **lowerCamelCase_, ) @property def a__ (self ): '''simple docstring''' return self.encoder_attention_heads @property def a__ (self ): '''simple docstring''' return self.d_model
316
"""simple docstring""" from __future__ import annotations def lowerCamelCase_ ( _lowerCamelCase , _lowerCamelCase ): lowerCamelCase__ : list[list[int]] = [] lowerCamelCase__ : list[int] = [] lowerCamelCase__ : List[str] = 0 lowerCamelCase__ : List[Any] = sum(_lowerCamelCase ) create_state_space_tree(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) return result def lowerCamelCase_ ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ): if sum(_lowerCamelCase ) > max_sum or (remaining_nums_sum + sum(_lowerCamelCase )) < max_sum: return if sum(_lowerCamelCase ) == max_sum: result.append(_lowerCamelCase ) return for index in range(_lowerCamelCase , len(_lowerCamelCase ) ): create_state_space_tree( _lowerCamelCase , _lowerCamelCase , index + 1 , [*path, nums[index]] , _lowerCamelCase , remaining_nums_sum - nums[index] , ) A_ : Optional[Any] = [3, 34, 4, 12, 5, 2] A_ : List[str] = 9 A_ : List[Any] = generate_sum_of_subsets_soln(nums, max_sum) print(*result)
316
1
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import AutoencoderKL, DDIMScheduler, DiTPipeline, DPMSolverMultistepScheduler, TransformeraDModel from diffusers.utils import is_xformers_available, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS, CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class UpperCamelCase_ ( __magic_name__ , unittest.TestCase ): lowercase = DiTPipeline lowercase = CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS lowercase = PipelineTesterMixin.required_optional_params - { 'latents', 'num_images_per_prompt', 'callback', 'callback_steps', } lowercase = CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS lowercase = False def _lowercase( self ) -> Optional[int]: torch.manual_seed(0 ) UpperCAmelCase : Dict = TransformeraDModel( sample_size=16 , num_layers=2 , patch_size=4 , attention_head_dim=8 , num_attention_heads=2 , in_channels=4 , out_channels=8 , attention_bias=A , activation_fn="""gelu-approximate""" , num_embeds_ada_norm=1000 , norm_type="""ada_norm_zero""" , norm_elementwise_affine=A , ) UpperCAmelCase : Optional[int] = AutoencoderKL() UpperCAmelCase : int = DDIMScheduler() UpperCAmelCase : str = {"""transformer""": transformer.eval(), """vae""": vae.eval(), """scheduler""": scheduler} return components def _lowercase( self , A , A=0 ) -> str: if str(A ).startswith("""mps""" ): UpperCAmelCase : Optional[Any] = torch.manual_seed(A ) else: UpperCAmelCase : Tuple = torch.Generator(device=A ).manual_seed(A ) UpperCAmelCase : Union[str, Any] = { """class_labels""": [1], """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs def _lowercase( self ) -> Dict: UpperCAmelCase : Any = """cpu""" UpperCAmelCase : List[Any] = self.get_dummy_components() UpperCAmelCase : List[str] = self.pipeline_class(**A ) pipe.to(A ) pipe.set_progress_bar_config(disable=A ) UpperCAmelCase : int = self.get_dummy_inputs(A ) UpperCAmelCase : Any = pipe(**A ).images UpperCAmelCase : Any = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 16, 16, 3) ) UpperCAmelCase : int = np.array([0.2_9_4_6, 0.6_6_0_1, 0.4_3_2_9, 0.3_2_9_6, 0.4_1_4_4, 0.5_3_1_9, 0.7_2_7_3, 0.5_0_1_3, 0.4_4_5_7] ) UpperCAmelCase : int = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(A , 1e-3 ) def _lowercase( self ) -> List[str]: self._test_inference_batch_single_identical(relax_max_difference=A , expected_max_diff=1e-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def _lowercase( self ) -> Union[str, Any]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) @require_torch_gpu @slow class UpperCamelCase_ ( unittest.TestCase ): def _lowercase( self ) -> Union[str, Any]: super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase( self ) -> List[Any]: UpperCAmelCase : Any = torch.manual_seed(0 ) UpperCAmelCase : List[str] = DiTPipeline.from_pretrained("""facebook/DiT-XL-2-256""" ) pipe.to("""cuda""" ) UpperCAmelCase : Tuple = ["""vase""", """umbrella""", """white shark""", """white wolf"""] UpperCAmelCase : str = pipe.get_label_ids(A ) UpperCAmelCase : Optional[int] = pipe(A , generator=A , num_inference_steps=40 , output_type="""np""" ).images for word, image in zip(A , A ): UpperCAmelCase : Dict = load_numpy( f'''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/dit/{word}.npy''' ) assert np.abs((expected_image - image).max() ) < 1e-2 def _lowercase( self ) -> Tuple: UpperCAmelCase : Optional[Any] = DiTPipeline.from_pretrained("""facebook/DiT-XL-2-512""" ) UpperCAmelCase : Optional[Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.to("""cuda""" ) UpperCAmelCase : Dict = ["""vase""", """umbrella"""] UpperCAmelCase : Dict = pipe.get_label_ids(A ) UpperCAmelCase : List[Any] = torch.manual_seed(0 ) UpperCAmelCase : List[str] = pipe(A , generator=A , num_inference_steps=25 , output_type="""np""" ).images for word, image in zip(A , A ): UpperCAmelCase : List[Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" f'''/dit/{word}_512.npy''' ) assert np.abs((expected_image - image).max() ) < 1e-1
265
'''simple docstring''' from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCamelCase_ : def __init__( self , A , A=12 , A=7 , A=True , A=True , A=True , A=99 , A=32 , A=32 , A=2 , A=4 , A=37 , A=0.1 , A=0.1 , A=512 , A=0.0_2 , A=0 , A=None , ) -> Any: UpperCAmelCase : Optional[Any] = parent UpperCAmelCase : str = batch_size UpperCAmelCase : Union[str, Any] = seq_length UpperCAmelCase : Optional[Any] = is_training UpperCAmelCase : int = use_input_mask UpperCAmelCase : List[Any] = use_labels UpperCAmelCase : Dict = vocab_size UpperCAmelCase : str = hidden_size UpperCAmelCase : List[Any] = projection_dim UpperCAmelCase : Tuple = num_hidden_layers UpperCAmelCase : Dict = num_attention_heads UpperCAmelCase : Optional[Any] = intermediate_size UpperCAmelCase : Any = dropout UpperCAmelCase : List[Any] = attention_dropout UpperCAmelCase : Optional[Any] = max_position_embeddings UpperCAmelCase : Tuple = initializer_range UpperCAmelCase : Optional[Any] = scope UpperCAmelCase : Union[str, Any] = bos_token_id def _lowercase( self ) -> Tuple: UpperCAmelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase : Union[str, Any] = None if self.use_input_mask: UpperCAmelCase : str = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: UpperCAmelCase : Tuple = input_mask.numpy() UpperCAmelCase , UpperCAmelCase : int = input_mask.shape UpperCAmelCase : Optional[int] = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(A ): UpperCAmelCase : Tuple = 1 UpperCAmelCase : Optional[Any] = 0 UpperCAmelCase : int = self.get_config() return config, input_ids, tf.convert_to_tensor(A ) def _lowercase( self ) -> int: return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def _lowercase( self , A , A , A ) -> Union[str, Any]: UpperCAmelCase : int = TFBlipTextModel(config=A ) UpperCAmelCase : Union[str, Any] = model(A , attention_mask=A , training=A ) UpperCAmelCase : int = model(A , training=A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _lowercase( self ) -> Optional[int]: UpperCAmelCase : Dict = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Any = config_and_inputs UpperCAmelCase : Optional[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class UpperCamelCase_ ( __magic_name__ , unittest.TestCase ): lowercase = (TFBlipTextModel,) if is_tf_available() else () lowercase = False lowercase = False lowercase = False def _lowercase( self ) -> int: UpperCAmelCase : Union[str, Any] = BlipTextModelTester(self ) UpperCAmelCase : List[str] = ConfigTester(self , config_class=A , hidden_size=37 ) def _lowercase( self ) -> Tuple: self.config_tester.run_common_tests() def _lowercase( self ) -> List[Any]: UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def _lowercase( self ) -> List[str]: pass def _lowercase( self ) -> Optional[int]: pass @unittest.skip(reason="""Blip does not use inputs_embeds""" ) def _lowercase( self ) -> Union[str, Any]: pass @unittest.skip(reason="""BlipTextModel has no base class and is not available in MODEL_MAPPING""" ) def _lowercase( self ) -> Optional[int]: pass @unittest.skip(reason="""BlipTextModel has no base class and is not available in MODEL_MAPPING""" ) def _lowercase( self ) -> Dict: pass @slow def _lowercase( self ) -> Dict: for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase : Any = TFBlipTextModel.from_pretrained(A ) self.assertIsNotNone(A ) def _lowercase( self , A=True ) -> str: super().test_pt_tf_model_equivalence(allow_missing_keys=A )
265
1
"""simple docstring""" from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo a : List[Any] = '''\ @misc{wu2016googles, title={Google\'s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation}, author={Yonghui Wu and Mike Schuster and Zhifeng Chen and Quoc V. Le and Mohammad Norouzi and Wolfgang Macherey and Maxim Krikun and Yuan Cao and Qin Gao and Klaus Macherey and Jeff Klingner and Apurva Shah and Melvin Johnson and Xiaobing Liu and Łukasz Kaiser and Stephan Gouws and Yoshikiyo Kato and Taku Kudo and Hideto Kazawa and Keith Stevens and George Kurian and Nishant Patil and Wei Wang and Cliff Young and Jason Smith and Jason Riesa and Alex Rudnick and Oriol Vinyals and Greg Corrado and Macduff Hughes and Jeffrey Dean}, year={2016}, eprint={1609.08144}, archivePrefix={arXiv}, primaryClass={cs.CL} } ''' a : Union[str, Any] = '''\ The BLEU score has some undesirable properties when used for single sentences, as it was designed to be a corpus measure. We therefore use a slightly different score for our RL experiments which we call the \'GLEU score\'. For the GLEU score, we record all sub-sequences of 1, 2, 3 or 4 tokens in output and target sequence (n-grams). We then compute a recall, which is the ratio of the number of matching n-grams to the number of total n-grams in the target (ground truth) sequence, and a precision, which is the ratio of the number of matching n-grams to the number of total n-grams in the generated output sequence. Then GLEU score is simply the minimum of recall and precision. This GLEU score\'s range is always between 0 (no matches) and 1 (all match) and it is symmetrical when switching output and target. According to our experiments, GLEU score correlates quite well with the BLEU metric on a corpus level but does not have its drawbacks for our per sentence reward objective. ''' a : Tuple = '''\ Computes corpus-level Google BLEU (GLEU) score of translated segments against one or more references. Instead of averaging the sentence level GLEU scores (i.e. macro-average precision), Wu et al. (2016) sum up the matching tokens and the max of hypothesis and reference tokens for each sentence, then compute using the aggregate values. Args: predictions (list of str): list of translations to score. Each translation should be tokenized into a list of tokens. references (list of list of str): list of lists of references for each translation. Each reference should be tokenized into a list of tokens. min_len (int): The minimum order of n-gram this function should extract. Defaults to 1. max_len (int): The maximum order of n-gram this function should extract. Defaults to 4. Returns: \'google_bleu\': google_bleu score Examples: Example 1: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references) >>> print(round(results["google_bleu"], 2)) 0.44 Example 2: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\', ... \'heed\', \'the\', \'cat\', \'commands\'] >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\', ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\', ... \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references) >>> print(round(results["google_bleu"], 2)) 0.61 Example 3: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\', ... \'heed\', \'the\', \'cat\', \'commands\'] >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\', ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\', ... \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references, min_len=2) >>> print(round(results["google_bleu"], 2)) 0.53 Example 4: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\', ... \'heed\', \'the\', \'cat\', \'commands\'] >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\', ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\', ... \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses,references=list_of_references, min_len=2, max_len=6) >>> print(round(results["google_bleu"], 2)) 0.4 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def __a ( self ) -> MetricInfo: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Sequence(datasets.Value("string" , id="token" ) , id="sequence" ), "references": datasets.Sequence( datasets.Sequence(datasets.Value("string" , id="token" ) , id="sequence" ) , id="references" ), } ) , ) def __a ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 1 , lowerCAmelCase__ = 4 , ) -> Dict[str, float]: return { "google_bleu": gleu_score.corpus_gleu( list_of_references=lowerCAmelCase__ , hypotheses=lowerCAmelCase__ , min_len=lowerCAmelCase__ , max_len=lowerCAmelCase__ ) }
79
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging a : Dict = logging.get_logger(__name__) a : List[Any] = { '''s-JoL/Open-Llama-V1''': '''https://huggingface.co/s-JoL/Open-Llama-V1/blob/main/config.json''', } class __UpperCamelCase ( a__ ): lowerCamelCase : Tuple ="""open-llama""" def __init__( self , lowerCAmelCase__=10_0000 , lowerCAmelCase__=4096 , lowerCAmelCase__=1_1008 , lowerCAmelCase__=32 , lowerCAmelCase__=32 , lowerCAmelCase__="silu" , lowerCAmelCase__=2048 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-6 , lowerCAmelCase__=True , lowerCAmelCase__=0 , lowerCAmelCase__=1 , lowerCAmelCase__=2 , lowerCAmelCase__=False , lowerCAmelCase__=True , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=None , **lowerCAmelCase__ , ) -> Tuple: a : Any = vocab_size a : List[str] = max_position_embeddings a : int = hidden_size a : str = intermediate_size a : List[str] = num_hidden_layers a : int = num_attention_heads a : Dict = hidden_act a : Union[str, Any] = initializer_range a : Tuple = rms_norm_eps a : Union[str, Any] = use_cache a : Union[str, Any] = kwargs.pop( "use_memorry_efficient_attention" , lowerCAmelCase__ ) a : int = hidden_dropout_prob a : Tuple = attention_dropout_prob a : Optional[Any] = use_stable_embedding a : str = shared_input_output_embedding a : str = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , tie_word_embeddings=lowerCAmelCase__ , **lowerCAmelCase__ , ) def __a ( self ) -> Union[str, Any]: if self.rope_scaling is None: return if not isinstance(self.rope_scaling , lowerCAmelCase__ ) or len(self.rope_scaling ) != 2: raise ValueError( "`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, " f"""got {self.rope_scaling}""" ) a : Any = self.rope_scaling.get("type" , lowerCAmelCase__ ) a : List[str] = self.rope_scaling.get("factor" , lowerCAmelCase__ ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"""`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}""" ) if rope_scaling_factor is None or not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) or rope_scaling_factor <= 1.0: raise ValueError(f"""`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}""" )
79
1
import argparse import logging import pickle from collections import Counter logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) _a = logging.getLogger(__name__) if __name__ == "__main__": _a = argparse.ArgumentParser( description='''Token Counts for smoothing the masking probabilities in MLM (cf XLM/word2vec)''' ) parser.add_argument( '''--data_file''', type=str, default='''data/dump.bert-base-uncased.pickle''', help='''The binarized dataset.''' ) parser.add_argument( '''--token_counts_dump''', type=str, default='''data/token_counts.bert-base-uncased.pickle''', help='''The dump file.''' ) parser.add_argument('''--vocab_size''', default=30522, type=int) _a = parser.parse_args() logger.info(F'''Loading data from {args.data_file}''') with open(args.data_file, '''rb''') as fp: _a = pickle.load(fp) logger.info('''Counting occurrences for MLM.''') _a = Counter() for tk_ids in data: counter.update(tk_ids) _a = [0] * args.vocab_size for k, v in counter.items(): _a = v logger.info(F'''Dump to {args.token_counts_dump}''') with open(args.token_counts_dump, '''wb''') as handle: pickle.dump(counts, handle, protocol=pickle.HIGHEST_PROTOCOL)
39
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class a ( unittest.TestCase ): """simple docstring""" def UpperCamelCase ( self: Any ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @property def UpperCamelCase ( self: Dict ): """simple docstring""" A__ = 1 A__ = 3 A__ = (32, 32) A__ = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(UpperCamelCase ) return image @property def UpperCamelCase ( self: int ): """simple docstring""" torch.manual_seed(0 ) A__ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) return model @property def UpperCamelCase ( self: Tuple ): """simple docstring""" torch.manual_seed(0 ) A__ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) return model @property def UpperCamelCase ( self: List[Any] ): """simple docstring""" torch.manual_seed(0 ) A__ = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=50_06 , ) return RobertaSeriesModelWithTransformation(UpperCamelCase ) @property def UpperCamelCase ( self: str ): """simple docstring""" def extract(*UpperCamelCase: List[str] , **UpperCamelCase: Any ): class a : """simple docstring""" def __init__( self: Any ): """simple docstring""" A__ = torch.ones([0] ) def UpperCamelCase ( self: Dict , UpperCamelCase: Optional[Any] ): """simple docstring""" self.pixel_values.to(UpperCamelCase ) return self return Out() return extract def UpperCamelCase ( self: str ): """simple docstring""" A__ = """cpu""" # ensure determinism for the device-dependent torch.Generator A__ = self.dummy_cond_unet A__ = PNDMScheduler(skip_prk_steps=UpperCamelCase ) A__ = self.dummy_vae A__ = self.dummy_text_encoder A__ = XLMRobertaTokenizer.from_pretrained("""hf-internal-testing/tiny-xlm-roberta""" ) A__ = 77 A__ = self.dummy_image.to(UpperCamelCase ) A__ = init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk A__ = AltDiffusionImgaImgPipeline( unet=UpperCamelCase , scheduler=UpperCamelCase , vae=UpperCamelCase , text_encoder=UpperCamelCase , tokenizer=UpperCamelCase , safety_checker=UpperCamelCase , feature_extractor=self.dummy_extractor , ) A__ = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=UpperCamelCase ) A__ = alt_pipe.to(UpperCamelCase ) alt_pipe.set_progress_bar_config(disable=UpperCamelCase ) A__ = """A painting of a squirrel eating a burger""" A__ = torch.Generator(device=UpperCamelCase ).manual_seed(0 ) A__ = alt_pipe( [prompt] , generator=UpperCamelCase , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" , image=UpperCamelCase , ) A__ = output.images A__ = torch.Generator(device=UpperCamelCase ).manual_seed(0 ) A__ = alt_pipe( [prompt] , generator=UpperCamelCase , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" , image=UpperCamelCase , return_dict=UpperCamelCase , )[0] A__ = image[0, -3:, -3:, -1] A__ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) A__ = np.array([0.4_427, 0.3_731, 0.4_249, 0.4_941, 0.4_546, 0.4_148, 0.4_193, 0.4_666, 0.4_499] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5e-3 @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""" ) def UpperCamelCase ( self: int ): """simple docstring""" A__ = self.dummy_cond_unet A__ = PNDMScheduler(skip_prk_steps=UpperCamelCase ) A__ = self.dummy_vae A__ = self.dummy_text_encoder A__ = XLMRobertaTokenizer.from_pretrained("""hf-internal-testing/tiny-xlm-roberta""" ) A__ = 77 A__ = self.dummy_image.to(UpperCamelCase ) # put models in fp16 A__ = unet.half() A__ = vae.half() A__ = bert.half() # make sure here that pndm scheduler skips prk A__ = AltDiffusionImgaImgPipeline( unet=UpperCamelCase , scheduler=UpperCamelCase , vae=UpperCamelCase , text_encoder=UpperCamelCase , tokenizer=UpperCamelCase , safety_checker=UpperCamelCase , feature_extractor=self.dummy_extractor , ) A__ = VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=UpperCamelCase ) A__ = alt_pipe.to(UpperCamelCase ) alt_pipe.set_progress_bar_config(disable=UpperCamelCase ) A__ = """A painting of a squirrel eating a burger""" A__ = torch.manual_seed(0 ) A__ = alt_pipe( [prompt] , generator=UpperCamelCase , num_inference_steps=2 , output_type="""np""" , image=UpperCamelCase , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""" ) def UpperCamelCase ( self: str ): """simple docstring""" A__ = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) # resize to resolution that is divisible by 8 but not 16 or 32 A__ = init_image.resize((7_60, 5_04) ) A__ = """BAAI/AltDiffusion""" A__ = AltDiffusionImgaImgPipeline.from_pretrained( UpperCamelCase , safety_checker=UpperCamelCase , ) pipe.to(UpperCamelCase ) pipe.set_progress_bar_config(disable=UpperCamelCase ) pipe.enable_attention_slicing() A__ = """A fantasy landscape, trending on artstation""" A__ = torch.manual_seed(0 ) A__ = pipe( prompt=UpperCamelCase , image=UpperCamelCase , strength=0.75 , guidance_scale=7.5 , generator=UpperCamelCase , output_type="""np""" , ) A__ = output.images[0] A__ = image[2_55:2_58, 3_83:3_86, -1] assert image.shape == (5_04, 7_60, 3) A__ = np.array([0.9_358, 0.9_397, 0.9_599, 0.9_901, 1.0_000, 1.0_000, 0.9_882, 1.0_000, 1.0_000] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class a ( unittest.TestCase ): """simple docstring""" def UpperCamelCase ( self: Any ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self: List[str] ): """simple docstring""" A__ = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) A__ = init_image.resize((7_68, 5_12) ) A__ = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy""" ) A__ = """BAAI/AltDiffusion""" A__ = AltDiffusionImgaImgPipeline.from_pretrained( UpperCamelCase , safety_checker=UpperCamelCase , ) pipe.to(UpperCamelCase ) pipe.set_progress_bar_config(disable=UpperCamelCase ) pipe.enable_attention_slicing() A__ = """A fantasy landscape, trending on artstation""" A__ = torch.manual_seed(0 ) A__ = pipe( prompt=UpperCamelCase , image=UpperCamelCase , strength=0.75 , guidance_scale=7.5 , generator=UpperCamelCase , output_type="""np""" , ) A__ = output.images[0] assert image.shape == (5_12, 7_68, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1e-2
335
0
"""simple docstring""" def UpperCamelCase_ ( lowerCAmelCase__ : str , lowerCAmelCase__ : bool = False ) -> str: """simple docstring""" if not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): lowerCAmelCase_ : Optional[int] = f"Expected string as input, found {type(lowerCAmelCase__ )}" raise ValueError(lowerCAmelCase__ ) if not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): lowerCAmelCase_ : Optional[Any] = f"Expected boolean as use_pascal parameter, found {type(lowerCAmelCase__ )}" raise ValueError(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = input_str.split('_' ) lowerCAmelCase_ : List[str] = 0 if use_pascal else 1 lowerCAmelCase_ : Union[str, Any] = words[start_index:] lowerCAmelCase_ : Union[str, Any] = [word[0].upper() + word[1:] for word in words_to_capitalize] lowerCAmelCase_ : Union[str, Any] = '' if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
289
"""simple docstring""" import re def UpperCamelCase_ ( lowerCAmelCase__ : str ) -> bool: """simple docstring""" lowerCAmelCase_ : str = re.compile( R'^(?:0|94|\+94|0{2}94)' R'7(0|1|2|4|5|6|7|8)' R'(-| |)' R'\d{7}$' ) return bool(re.search(lowerCAmelCase__ , lowerCAmelCase__ ) ) if __name__ == "__main__": lowercase__ : Optional[int] = """0094702343221""" print(is_sri_lankan_phone_number(phone))
289
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( lowercase_ = 10_00 ) -> int: A__ = 2**power A__ = 0 while n: A__, A__ = r + n % 10, n // 10 return r if __name__ == "__main__": print(solution(int(str(input()).strip())))
247
"""simple docstring""" from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = { "Salesforce/codegen-350M-nl": "https://huggingface.co/Salesforce/codegen-350M-nl/resolve/main/config.json", "Salesforce/codegen-350M-multi": "https://huggingface.co/Salesforce/codegen-350M-multi/resolve/main/config.json", "Salesforce/codegen-350M-mono": "https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/config.json", "Salesforce/codegen-2B-nl": "https://huggingface.co/Salesforce/codegen-2B-nl/resolve/main/config.json", "Salesforce/codegen-2B-multi": "https://huggingface.co/Salesforce/codegen-2B-multi/resolve/main/config.json", "Salesforce/codegen-2B-mono": "https://huggingface.co/Salesforce/codegen-2B-mono/resolve/main/config.json", "Salesforce/codegen-6B-nl": "https://huggingface.co/Salesforce/codegen-6B-nl/resolve/main/config.json", "Salesforce/codegen-6B-multi": "https://huggingface.co/Salesforce/codegen-6B-multi/resolve/main/config.json", "Salesforce/codegen-6B-mono": "https://huggingface.co/Salesforce/codegen-6B-mono/resolve/main/config.json", "Salesforce/codegen-16B-nl": "https://huggingface.co/Salesforce/codegen-16B-nl/resolve/main/config.json", "Salesforce/codegen-16B-multi": "https://huggingface.co/Salesforce/codegen-16B-multi/resolve/main/config.json", "Salesforce/codegen-16B-mono": "https://huggingface.co/Salesforce/codegen-16B-mono/resolve/main/config.json", } class UpperCAmelCase_ ( A_ ): lowercase__ = '''codegen''' lowercase__ = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : Any , snake_case_ : int=50_400 , snake_case_ : Any=2_048 , snake_case_ : List[str]=2_048 , snake_case_ : List[Any]=4_096 , snake_case_ : Union[str, Any]=28 , snake_case_ : List[str]=16 , snake_case_ : Dict=64 , snake_case_ : Dict=None , snake_case_ : List[Any]="gelu_new" , snake_case_ : List[str]=0.0 , snake_case_ : int=0.0 , snake_case_ : List[str]=0.0 , snake_case_ : Optional[Any]=1e-5 , snake_case_ : Optional[int]=0.02 , snake_case_ : Dict=True , snake_case_ : Union[str, Any]=50_256 , snake_case_ : Optional[Any]=50_256 , snake_case_ : Optional[Any]=False , **snake_case_ : Any , ) -> List[Any]: '''simple docstring''' A__ = vocab_size A__ = n_ctx A__ = n_positions A__ = n_embd A__ = n_layer A__ = n_head A__ = n_inner A__ = rotary_dim A__ = activation_function A__ = resid_pdrop A__ = embd_pdrop A__ = attn_pdrop A__ = layer_norm_epsilon A__ = initializer_range A__ = use_cache A__ = bos_token_id A__ = eos_token_id super().__init__( bos_token_id=snake_case_ , eos_token_id=snake_case_ , tie_word_embeddings=snake_case_ , **snake_case_ ) class UpperCAmelCase_ ( A_ ): def __init__( self : Any , snake_case_ : PretrainedConfig , snake_case_ : str = "default" , snake_case_ : List[PatchingSpec] = None , snake_case_ : bool = False , ) -> int: '''simple docstring''' super().__init__(snake_case_ , task=snake_case_ , patching_specs=snake_case_ , use_past=snake_case_ ) if not getattr(self._config , "pad_token_id" , snake_case_ ): # TODO: how to do that better? A__ = 0 @property def __magic_name__ ( self : Any ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' A__ = OrderedDict({"input_ids": {0: "batch", 1: "sequence"}} ) if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction="inputs" ) A__ = {0: "batch", 1: "past_sequence + sequence"} else: A__ = {0: "batch", 1: "sequence"} return common_inputs @property def __magic_name__ ( self : List[Any] ) -> int: '''simple docstring''' return self._config.n_layer @property def __magic_name__ ( self : Optional[int] ) -> int: '''simple docstring''' return self._config.n_head def __magic_name__ ( self : str , snake_case_ : PreTrainedTokenizer , snake_case_ : int = -1 , snake_case_ : int = -1 , snake_case_ : bool = False , snake_case_ : Optional[TensorType] = None , ) -> Mapping[str, Any]: '''simple docstring''' A__ = super(snake_case_ , self ).generate_dummy_inputs( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) # We need to order the input in the way they appears in the forward() A__ = OrderedDict({"input_ids": common_inputs["input_ids"]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch A__, A__ = common_inputs["input_ids"].shape # Not using the same length for past_key_values A__ = seqlen + 2 A__ = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) A__ = [ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(self.num_layers ) ] A__ = common_inputs["attention_mask"] if self.use_past: A__ = ordered_inputs["attention_mask"].dtype A__ = torch.cat( [ordered_inputs["attention_mask"], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) return ordered_inputs @property def __magic_name__ ( self : Optional[int] ) -> int: '''simple docstring''' return 13
247
1
from __future__ import annotations from collections.abc import Generator def _UpperCAmelCase ( ): __UpperCamelCase ={} __UpperCamelCase =2 while True: __UpperCamelCase =factor_map.pop(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if factor: __UpperCamelCase =factor + prime while x in factor_map: x += factor __UpperCamelCase =factor else: __UpperCamelCase =prime yield prime prime += 1 def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : float = 1E10 ): __UpperCamelCase =sieve() __UpperCamelCase =1 while True: __UpperCamelCase =next(SCREAMING_SNAKE_CASE__ ) if (2 * prime * n) > limit: return n # Ignore the next prime as the reminder will be 2. next(SCREAMING_SNAKE_CASE__ ) n += 2 if __name__ == "__main__": print(solution())
117
from __future__ import annotations from typing import Any def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : list[Any] ): create_state_space_tree(SCREAMING_SNAKE_CASE__ , [] , 0 ) def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : list[Any] , SCREAMING_SNAKE_CASE__ : list[Any] , SCREAMING_SNAKE_CASE__ : int ): if index == len(SCREAMING_SNAKE_CASE__ ): print(SCREAMING_SNAKE_CASE__ ) return create_state_space_tree(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _A = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(['A', 'B', 'C']) generate_all_subsequences(seq)
117
1
'''simple docstring''' def _A ( lowercase__ , lowercase__ ): while second != 0: lowercase__ = first & second first ^= second lowercase__ = c << 1 return first if __name__ == "__main__": import doctest doctest.testmod() __A = int(input("Enter the first number: ").strip()) __A = int(input("Enter the second number: ").strip()) print(F'''{add(first, second) = }''')
164
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer __A = logging.get_logger(__name__) __A = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} __A = { "vocab_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json" ), }, } __A = { "yjernite/retribert-base-uncased": 512, } __A = { "yjernite/retribert-base-uncased": {"do_lower_case": True}, } class A ( __UpperCAmelCase ): lowerCamelCase : int = VOCAB_FILES_NAMES lowerCamelCase : str = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Optional[Any] = PRETRAINED_INIT_CONFIGURATION lowerCamelCase : str = RetriBertTokenizer lowerCamelCase : Optional[int] = ["""input_ids""", """attention_mask"""] def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__="[UNK]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[PAD]" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__=True , lowerCamelCase__=None , **lowerCamelCase__ , ) -> Union[str, Any]: '''simple docstring''' super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) lowercase__ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , lowerCamelCase__ ) != do_lower_case or normalizer_state.get("""strip_accents""" , lowerCamelCase__ ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , lowerCamelCase__ ) != tokenize_chinese_chars ): lowercase__ = getattr(lowerCamelCase__ , normalizer_state.pop("""type""" ) ) lowercase__ = do_lower_case lowercase__ = strip_accents lowercase__ = tokenize_chinese_chars lowercase__ = normalizer_class(**lowerCamelCase__ ) lowercase__ = do_lower_case def A__ ( self , lowerCamelCase__ , lowerCamelCase__=None ) -> Dict: '''simple docstring''' lowercase__ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ) -> List[int]: '''simple docstring''' lowercase__ = [self.sep_token_id] lowercase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ) -> Tuple[str]: '''simple docstring''' lowercase__ = self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
164
1
'''simple docstring''' import gc import random import tempfile import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline from diffusers.utils import floats_tensor, nightly, torch_device from diffusers.utils.testing_utils import require_torch_gpu class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _lowercase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowercase ( self ): """simple docstring""" _lowerCAmelCase = 1 _lowerCAmelCase = 3 _lowerCAmelCase = (32, 32) _lowerCAmelCase = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(_a ) return image @property def _lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) _lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) return model @property def _lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) _lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) return model @property def _lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) _lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) return CLIPTextModel(_a ) @property def _lowercase ( self ): """simple docstring""" def extract(*_lowercase , **_lowercase ): class UpperCAmelCase_ : '''simple docstring''' def __init__( self ): """simple docstring""" _lowerCAmelCase = torch.ones([0] ) def _lowercase ( self , _lowercase ): """simple docstring""" self.pixel_values.to(_a ) return self return Out() return extract def _lowercase ( self ): """simple docstring""" _lowerCAmelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCAmelCase = self.dummy_cond_unet _lowerCAmelCase = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=_a , set_alpha_to_one=_a , ) _lowerCAmelCase = self.dummy_vae _lowerCAmelCase = self.dummy_text_encoder _lowerCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # make sure here that pndm scheduler skips prk _lowerCAmelCase = StableDiffusionPipeline( unet=_a , scheduler=_a , vae=_a , text_encoder=_a , tokenizer=_a , safety_checker=_a , feature_extractor=self.dummy_extractor , ) _lowerCAmelCase = sd_pipe.to(_a ) sd_pipe.set_progress_bar_config(disable=_a ) _lowerCAmelCase = '''A painting of a squirrel eating a burger''' _lowerCAmelCase = torch.Generator(device=_a ).manual_seed(0 ) _lowerCAmelCase = sd_pipe([prompt] , generator=_a , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" ) _lowerCAmelCase = output.images _lowerCAmelCase = torch.Generator(device=_a ).manual_seed(0 ) _lowerCAmelCase = sd_pipe( [prompt] , generator=_a , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" , return_dict=_a , )[0] _lowerCAmelCase = image[0, -3:, -3:, -1] _lowerCAmelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _lowerCAmelCase = np.array([0.5756, 0.6118, 0.5005, 0.5041, 0.5471, 0.4726, 0.4976, 0.4865, 0.4864] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self ): """simple docstring""" _lowerCAmelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCAmelCase = self.dummy_cond_unet _lowerCAmelCase = PNDMScheduler(skip_prk_steps=_a ) _lowerCAmelCase = self.dummy_vae _lowerCAmelCase = self.dummy_text_encoder _lowerCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # make sure here that pndm scheduler skips prk _lowerCAmelCase = StableDiffusionPipeline( unet=_a , scheduler=_a , vae=_a , text_encoder=_a , tokenizer=_a , safety_checker=_a , feature_extractor=self.dummy_extractor , ) _lowerCAmelCase = sd_pipe.to(_a ) sd_pipe.set_progress_bar_config(disable=_a ) _lowerCAmelCase = '''A painting of a squirrel eating a burger''' _lowerCAmelCase = torch.Generator(device=_a ).manual_seed(0 ) _lowerCAmelCase = sd_pipe([prompt] , generator=_a , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" ) _lowerCAmelCase = output.images _lowerCAmelCase = torch.Generator(device=_a ).manual_seed(0 ) _lowerCAmelCase = sd_pipe( [prompt] , generator=_a , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" , return_dict=_a , )[0] _lowerCAmelCase = image[0, -3:, -3:, -1] _lowerCAmelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _lowerCAmelCase = np.array([0.5125, 0.5716, 0.4828, 0.5060, 0.5650, 0.4768, 0.5185, 0.4895, 0.4993] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self ): """simple docstring""" _lowerCAmelCase = StableDiffusionPipeline.from_pretrained( """hf-internal-testing/tiny-stable-diffusion-lms-pipe""" , safety_checker=_a ) assert isinstance(_a , _a ) assert isinstance(pipe.scheduler , _a ) assert pipe.safety_checker is None _lowerCAmelCase = pipe("""example prompt""" , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_a ) _lowerCAmelCase = StableDiffusionPipeline.from_pretrained(_a ) # sanity check that the pipeline still works assert pipe.safety_checker is None _lowerCAmelCase = pipe("""example prompt""" , num_inference_steps=2 ).images[0] assert image is not None @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""" ) def _lowercase ( self ): """simple docstring""" _lowerCAmelCase = self.dummy_cond_unet _lowerCAmelCase = PNDMScheduler(skip_prk_steps=_a ) _lowerCAmelCase = self.dummy_vae _lowerCAmelCase = self.dummy_text_encoder _lowerCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # put models in fp16 _lowerCAmelCase = unet.half() _lowerCAmelCase = vae.half() _lowerCAmelCase = bert.half() # make sure here that pndm scheduler skips prk _lowerCAmelCase = StableDiffusionPipeline( unet=_a , scheduler=_a , vae=_a , text_encoder=_a , tokenizer=_a , safety_checker=_a , feature_extractor=self.dummy_extractor , ) _lowerCAmelCase = sd_pipe.to(_a ) sd_pipe.set_progress_bar_config(disable=_a ) _lowerCAmelCase = '''A painting of a squirrel eating a burger''' _lowerCAmelCase = sd_pipe([prompt] , num_inference_steps=2 , output_type="""np""" ).images assert image.shape == (1, 64, 64, 3) @nightly @require_torch_gpu class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _lowercase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase ( self ): """simple docstring""" _lowerCAmelCase = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" , safety_checker=_a ) _lowerCAmelCase = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) _lowerCAmelCase = sd_pipe.to(_a ) sd_pipe.set_progress_bar_config(disable=_a ) _lowerCAmelCase = ( '''portrait of girl with smokey eyes makeup in abandoned hotel, grange clothes, redshift, wide high angle''' ''' coloured polaroid photograph with flash, kodak film, hyper real, stunning moody cinematography, with''' ''' anamorphic lenses, by maripol, fallen angels by wong kar - wai, style of suspiria and neon demon and''' ''' children from bahnhof zoo, detailed ''' ) _lowerCAmelCase = 4_003_660_346 _lowerCAmelCase = 7 # without safety guidance (sld_guidance_scale = 0) _lowerCAmelCase = torch.manual_seed(_a ) _lowerCAmelCase = sd_pipe( [prompt] , generator=_a , guidance_scale=_a , num_inference_steps=50 , output_type="""np""" , width=512 , height=512 , sld_guidance_scale=0 , ) _lowerCAmelCase = output.images _lowerCAmelCase = image[0, -3:, -3:, -1] _lowerCAmelCase = [0.2278, 0.2231, 0.2249, 0.2333, 0.2303, 0.1885, 0.2273, 0.2144, 0.2176] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 # without safety guidance (strong configuration) _lowerCAmelCase = torch.manual_seed(_a ) _lowerCAmelCase = sd_pipe( [prompt] , generator=_a , guidance_scale=_a , num_inference_steps=50 , output_type="""np""" , width=512 , height=512 , sld_guidance_scale=2_000 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) _lowerCAmelCase = output.images _lowerCAmelCase = image[0, -3:, -3:, -1] _lowerCAmelCase = [0.2383, 0.2276, 0.236, 0.2192, 0.2186, 0.2053, 0.1971, 0.1901, 0.1719] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self ): """simple docstring""" _lowerCAmelCase = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" , safety_checker=_a ) _lowerCAmelCase = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) _lowerCAmelCase = sd_pipe.to(_a ) sd_pipe.set_progress_bar_config(disable=_a ) _lowerCAmelCase = '''padme amidala taking a bath artwork, safe for work, no nudity''' _lowerCAmelCase = 2_734_971_755 _lowerCAmelCase = 7 _lowerCAmelCase = torch.manual_seed(_a ) _lowerCAmelCase = sd_pipe( [prompt] , generator=_a , guidance_scale=_a , num_inference_steps=50 , output_type="""np""" , width=512 , height=512 , sld_guidance_scale=0 , ) _lowerCAmelCase = output.images _lowerCAmelCase = image[0, -3:, -3:, -1] _lowerCAmelCase = [0.3502, 0.3622, 0.3396, 0.3642, 0.3478, 0.3318, 0.35, 0.3348, 0.3297] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 _lowerCAmelCase = torch.manual_seed(_a ) _lowerCAmelCase = sd_pipe( [prompt] , generator=_a , guidance_scale=_a , num_inference_steps=50 , output_type="""np""" , width=512 , height=512 , sld_guidance_scale=2_000 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) _lowerCAmelCase = output.images _lowerCAmelCase = image[0, -3:, -3:, -1] _lowerCAmelCase = [0.5531, 0.5206, 0.4895, 0.5156, 0.5182, 0.4751, 0.4802, 0.4803, 0.4443] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self ): """simple docstring""" _lowerCAmelCase = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" ) _lowerCAmelCase = sd_pipe.to(_a ) sd_pipe.set_progress_bar_config(disable=_a ) _lowerCAmelCase = ( '''the four horsewomen of the apocalypse, painting by tom of finland, gaston bussiere, craig mullins, j. c.''' ''' leyendecker''' ) _lowerCAmelCase = 1_044_355_234 _lowerCAmelCase = 12 _lowerCAmelCase = torch.manual_seed(_a ) _lowerCAmelCase = sd_pipe( [prompt] , generator=_a , guidance_scale=_a , num_inference_steps=50 , output_type="""np""" , width=512 , height=512 , sld_guidance_scale=0 , ) _lowerCAmelCase = output.images _lowerCAmelCase = image[0, -3:, -3:, -1] _lowerCAmelCase = np.array([0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0] ) assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-7 _lowerCAmelCase = torch.manual_seed(_a ) _lowerCAmelCase = sd_pipe( [prompt] , generator=_a , guidance_scale=_a , num_inference_steps=50 , output_type="""np""" , width=512 , height=512 , sld_guidance_scale=2_000 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) _lowerCAmelCase = output.images _lowerCAmelCase = image[0, -3:, -3:, -1] _lowerCAmelCase = np.array([0.5818, 0.6285, 0.6835, 0.6019, 0.625, 0.6754, 0.6096, 0.6334, 0.6561] ) assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
365
'''simple docstring''' # Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position _lowercase = """2.13.1""" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("""3.7"""): raise ImportWarning( """To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition.""" ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( """To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n""" """If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.""" ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip _lowercase = concatenate_datasets _lowercase = DownloadConfig _lowercase = DownloadManager _lowercase = DownloadMode _lowercase = DownloadConfig _lowercase = DownloadMode _lowercase = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
229
0
from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup SCREAMING_SNAKE_CASE__ : Dict = "https://www.indeed.co.in/jobs?q=mobile+app+development&l=" def __magic_name__ ( __lowerCAmelCase : str = "mumbai" ) -> Generator[tuple[str, str], None, None]: __lowerCamelCase = BeautifulSoup(requests.get(url + location ).content , '''html.parser''' ) # This attribute finds out all the specifics listed in a job for job in soup.find_all('''div''' , attrs={'''data-tn-component''': '''organicJob'''} ): __lowerCamelCase = job.find('''a''' , attrs={'''data-tn-element''': '''jobTitle'''} ).text.strip() __lowerCamelCase = job.find('''span''' , {'''class''': '''company'''} ).text.strip() yield job_title, company_name if __name__ == "__main__": for i, job in enumerate(fetch_jobs("Bangalore"), 1): print(F'Job {i:>2} is {job[0]} at {job[1]}')
270
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE__ : Optional[Any] = {"configuration_wavlm": ["WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "WavLMConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ : int = [ "WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST", "WavLMForAudioFrameClassification", "WavLMForCTC", "WavLMForSequenceClassification", "WavLMForXVector", "WavLMModel", "WavLMPreTrainedModel", ] if TYPE_CHECKING: from .configuration_wavlm import WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP, WavLMConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavlm import ( WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST, WavLMForAudioFrameClassification, WavLMForCTC, WavLMForSequenceClassification, WavLMForXVector, WavLMModel, WavLMPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE__ : int = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
270
1
'''simple docstring''' import argparse import collections import json from pathlib import Path import requests import torch import yaml from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTImageProcessor, MobileViTVaConfig, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, ) from transformers.utils import logging logging.set_verbosity_info() lowercase__ = logging.get_logger(__name__) def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ ) -> Optional[int]: '''simple docstring''' print('''Loading config file...''' ) def flatten_yaml_as_dict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__="" , SCREAMING_SNAKE_CASE__="." ): snake_case : List[str] = [] for k, v in d.items(): snake_case : Union[str, Any] = parent_key + sep + k if parent_key else k if isinstance(SCREAMING_SNAKE_CASE__ , collections.abc.MutableMapping ): items.extend(flatten_yaml_as_dict(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , sep=SCREAMING_SNAKE_CASE__ ).items() ) else: items.append((new_key, v) ) return dict(SCREAMING_SNAKE_CASE__ ) snake_case : Tuple = argparse.Namespace() with open(SCREAMING_SNAKE_CASE__ , '''r''' ) as yaml_file: try: snake_case : Optional[Any] = yaml.load(SCREAMING_SNAKE_CASE__ , Loader=yaml.FullLoader ) snake_case : Optional[Any] = flatten_yaml_as_dict(SCREAMING_SNAKE_CASE__ ) for k, v in flat_cfg.items(): setattr(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) except yaml.YAMLError as exc: logger.error('''Error while loading config file: {}. Error message: {}'''.format(SCREAMING_SNAKE_CASE__ , str(SCREAMING_SNAKE_CASE__ ) ) ) return config def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Any: '''simple docstring''' snake_case : Optional[Any] = MobileViTVaConfig() snake_case : Tuple = False # dataset if task_name.startswith('''imagenet1k_''' ): snake_case : Optional[Any] = 1000 if int(task_name.strip().split('''_''' )[-1] ) == 384: snake_case : List[Any] = 384 else: snake_case : Tuple = 256 snake_case : List[str] = '''imagenet-1k-id2label.json''' elif task_name.startswith('''imagenet21k_to_1k_''' ): snake_case : Any = 2_1000 if int(task_name.strip().split('''_''' )[-1] ) == 384: snake_case : List[Any] = 384 else: snake_case : Dict = 256 snake_case : int = '''imagenet-22k-id2label.json''' elif task_name.startswith('''ade20k_''' ): snake_case : Optional[int] = 151 snake_case : List[Any] = 512 snake_case : int = '''ade20k-id2label.json''' snake_case : Tuple = True elif task_name.startswith('''voc_''' ): snake_case : Tuple = 21 snake_case : Union[str, Any] = 512 snake_case : List[Any] = '''pascal-voc-id2label.json''' snake_case : Dict = True # orig_config snake_case : Optional[int] = load_orig_config_file(SCREAMING_SNAKE_CASE__ ) assert getattr(SCREAMING_SNAKE_CASE__ , '''model.classification.name''' , -1 ) == "mobilevit_v2", "Invalid model" snake_case : Dict = getattr(SCREAMING_SNAKE_CASE__ , '''model.classification.mitv2.width_multiplier''' , 1.0 ) assert ( getattr(SCREAMING_SNAKE_CASE__ , '''model.classification.mitv2.attn_norm_layer''' , -1 ) == "layer_norm_2d" ), "Norm layers other than layer_norm_2d is not supported" snake_case : str = getattr(SCREAMING_SNAKE_CASE__ , '''model.classification.activation.name''' , '''swish''' ) # config.image_size == getattr(orig_config, 'sampler.bs.crop_size_width', 256) if is_segmentation_model: snake_case : List[str] = getattr(SCREAMING_SNAKE_CASE__ , '''model.segmentation.output_stride''' , 16 ) if "_deeplabv3" in task_name: snake_case : Dict = getattr(SCREAMING_SNAKE_CASE__ , '''model.segmentation.deeplabv3.aspp_rates''' , [12, 24, 36] ) snake_case : Dict = getattr(SCREAMING_SNAKE_CASE__ , '''model.segmentation.deeplabv3.aspp_out_channels''' , 512 ) snake_case : List[str] = getattr(SCREAMING_SNAKE_CASE__ , '''model.segmentation.deeplabv3.aspp_dropout''' , 0.1 ) # id2label snake_case : Tuple = '''huggingface/label-files''' snake_case : Optional[Any] = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , repo_type='''dataset''' ) , '''r''' ) ) snake_case : List[str] = {int(SCREAMING_SNAKE_CASE__ ): v for k, v in idalabel.items()} snake_case : Optional[int] = idalabel snake_case : Optional[Any] = {v: k for k, v in idalabel.items()} return config def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: '''simple docstring''' snake_case : int = dct.pop(SCREAMING_SNAKE_CASE__ ) snake_case : Dict = val def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=False ) -> Any: '''simple docstring''' if base_model: snake_case : Optional[Any] = '''''' else: snake_case : Dict = '''mobilevitv2.''' snake_case : str = [] for k in state_dict.keys(): if k[:8] == "encoder.": snake_case : Optional[int] = k[8:] else: snake_case : int = k if ".block." in k: snake_case : int = k_new.replace('''.block.''' , '''.''' ) if ".conv." in k: snake_case : int = k_new.replace('''.conv.''' , '''.convolution.''' ) if ".norm." in k: snake_case : int = k_new.replace('''.norm.''' , '''.normalization.''' ) if "conv_1." in k: snake_case : Union[str, Any] = k_new.replace('''conv_1.''' , F'{model_prefix}conv_stem.' ) for i in [1, 2]: if F'layer_{i}.' in k: snake_case : List[str] = k_new.replace(F'layer_{i}.' , F'{model_prefix}encoder.layer.{i-1}.layer.' ) if ".exp_1x1." in k: snake_case : Optional[int] = k_new.replace('''.exp_1x1.''' , '''.expand_1x1.''' ) if ".red_1x1." in k: snake_case : Optional[int] = k_new.replace('''.red_1x1.''' , '''.reduce_1x1.''' ) for i in [3, 4, 5]: if F'layer_{i}.0.' in k: snake_case : int = k_new.replace(F'layer_{i}.0.' , F'{model_prefix}encoder.layer.{i-1}.downsampling_layer.' ) if F'layer_{i}.1.local_rep.0.' in k: snake_case : Optional[Any] = k_new.replace(F'layer_{i}.1.local_rep.0.' , F'{model_prefix}encoder.layer.{i-1}.conv_kxk.' ) if F'layer_{i}.1.local_rep.1.' in k: snake_case : List[str] = k_new.replace(F'layer_{i}.1.local_rep.1.' , F'{model_prefix}encoder.layer.{i-1}.conv_1x1.' ) for i in [3, 4, 5]: if i == 3: snake_case : Union[str, Any] = [0, 1] elif i == 4: snake_case : List[str] = [0, 1, 2, 3] elif i == 5: snake_case : Optional[int] = [0, 1, 2] for j in j_in: if F'layer_{i}.1.global_rep.{j}.' in k: snake_case : Any = k_new.replace( F'layer_{i}.1.global_rep.{j}.' , F'{model_prefix}encoder.layer.{i-1}.transformer.layer.{j}.' ) if F'layer_{i}.1.global_rep.{j+1}.' in k: snake_case : str = k_new.replace( F'layer_{i}.1.global_rep.{j+1}.' , F'{model_prefix}encoder.layer.{i-1}.layernorm.' ) if F'layer_{i}.1.conv_proj.' in k: snake_case : Optional[int] = k_new.replace(F'layer_{i}.1.conv_proj.' , F'{model_prefix}encoder.layer.{i-1}.conv_projection.' ) if "pre_norm_attn.0." in k: snake_case : Dict = k_new.replace('''pre_norm_attn.0.''' , '''layernorm_before.''' ) if "pre_norm_attn.1." in k: snake_case : Tuple = k_new.replace('''pre_norm_attn.1.''' , '''attention.''' ) if "pre_norm_ffn.0." in k: snake_case : Optional[Any] = k_new.replace('''pre_norm_ffn.0.''' , '''layernorm_after.''' ) if "pre_norm_ffn.1." in k: snake_case : List[Any] = k_new.replace('''pre_norm_ffn.1.''' , '''ffn.conv1.''' ) if "pre_norm_ffn.3." in k: snake_case : Optional[Any] = k_new.replace('''pre_norm_ffn.3.''' , '''ffn.conv2.''' ) if "classifier.1." in k: snake_case : str = k_new.replace('''classifier.1.''' , '''classifier.''' ) if "seg_head." in k: snake_case : Tuple = k_new.replace('''seg_head.''' , '''segmentation_head.''' ) if ".aspp_layer." in k: snake_case : Optional[int] = k_new.replace('''.aspp_layer.''' , '''.''' ) if ".aspp_pool." in k: snake_case : List[str] = k_new.replace('''.aspp_pool.''' , '''.''' ) rename_keys.append((k, k_new) ) return rename_keys def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ ) -> int: '''simple docstring''' snake_case : Optional[int] = [] for k in state_dict.keys(): if k.startswith('''seg_head.aux_head.''' ): keys_to_ignore.append(SCREAMING_SNAKE_CASE__ ) for k in keys_to_ignore: state_dict.pop(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( ) -> Union[str, Any]: '''simple docstring''' snake_case : Optional[Any] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' # url = "https://cdn.britannica.com/86/141086-050-9D7C75EE/Gulfstream-G450-business-jet-passengers.jpg" snake_case : List[Any] = Image.open(requests.get(SCREAMING_SNAKE_CASE__ , stream=SCREAMING_SNAKE_CASE__ ).raw ) return im @torch.no_grad() def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> List[str]: '''simple docstring''' snake_case : List[str] = get_mobilevitva_config(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # load original state_dict snake_case : Optional[Any] = torch.load(SCREAMING_SNAKE_CASE__ , map_location='''cpu''' ) # load huggingface model if task_name.startswith('''ade20k_''' ) or task_name.startswith('''voc_''' ): snake_case : Optional[int] = MobileViTVaForSemanticSegmentation(SCREAMING_SNAKE_CASE__ ).eval() snake_case : Union[str, Any] = False else: snake_case : Any = MobileViTVaForImageClassification(SCREAMING_SNAKE_CASE__ ).eval() snake_case : List[Any] = False # remove and rename some keys of load the original model snake_case : Any = checkpoint remove_unused_keys(SCREAMING_SNAKE_CASE__ ) snake_case : List[Any] = create_rename_keys(SCREAMING_SNAKE_CASE__ , base_model=SCREAMING_SNAKE_CASE__ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # load modified state_dict model.load_state_dict(SCREAMING_SNAKE_CASE__ ) # Check outputs on an image, prepared by MobileViTImageProcessor snake_case : Optional[Any] = MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 32 ) snake_case : Optional[int] = image_processor(images=prepare_img() , return_tensors='''pt''' ) snake_case : Optional[Any] = model(**SCREAMING_SNAKE_CASE__ ) # verify classification model if task_name.startswith('''imagenet''' ): snake_case : str = outputs.logits snake_case : Any = logits.argmax(-1 ).item() print('''Predicted class:''' , model.config.idalabel[predicted_class_idx] ) if task_name.startswith('''imagenet1k_256''' ) and config.width_multiplier == 1.0: # expected_logits for base variant snake_case : List[str] = torch.tensor([-1.6_3_3_6E0_0, -7.3_2_0_4E-0_2, -5.1_8_8_3E-0_1] ) assert torch.allclose(logits[0, :3] , SCREAMING_SNAKE_CASE__ , atol=1E-4 ) Path(SCREAMING_SNAKE_CASE__ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE__ ) print(F'Saving model {task_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) print(F'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--task", default="imagenet1k_256", type=str, help=( "Name of the task for which the MobileViTV2 model you'd like to convert is trained on . " "\n Classification (ImageNet-1k)\n - MobileViTV2 (256x256) : imagenet1k_256\n - MobileViTV2 (Trained on 256x256 and Finetuned on 384x384) : imagenet1k_384\n - MobileViTV2 (Trained on ImageNet-21k and Finetuned on ImageNet-1k 256x256) :\n imagenet21k_to_1k_256\n - MobileViTV2 (Trained on ImageNet-21k, Finetuned on ImageNet-1k 256x256, and Finetuned on\n ImageNet-1k 384x384) : imagenet21k_to_1k_384\n Segmentation\n - ADE20K Dataset : ade20k_deeplabv3\n - Pascal VOC 2012 Dataset: voc_deeplabv3\n " ), choices=[ "imagenet1k_256", "imagenet1k_384", "imagenet21k_to_1k_256", "imagenet21k_to_1k_384", "ade20k_deeplabv3", "voc_deeplabv3", ], ) parser.add_argument( "--orig_checkpoint_path", required=True, type=str, help="Path to the original state dict (.pt file)." ) parser.add_argument("--orig_config_path", required=True, type=str, help="Path to the original config file.") parser.add_argument( "--pytorch_dump_folder_path", required=True, type=str, help="Path to the output PyTorch model directory." ) lowercase__ = parser.parse_args() convert_mobilevitva_checkpoint( args.task, args.orig_checkpoint_path, args.orig_config_path, args.pytorch_dump_folder_path )
359
'''simple docstring''' import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(__SCREAMING_SNAKE_CASE ) , """Tatoeba directory does not exist.""" ) class snake_case__ ( unittest.TestCase ): """simple docstring""" @cached_property def lowerCAmelCase ( self : List[str] ) -> str: """simple docstring""" snake_case : Tuple = tempfile.mkdtemp() return TatoebaConverter(save_dir=UpperCamelCase__ ) @slow def lowerCAmelCase ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" self.resolver.convert_models(['''heb-eng'''] ) @slow def lowerCAmelCase ( self : Tuple ) -> str: """simple docstring""" snake_case ,snake_case : Dict = self.resolver.write_model_card('''opus-mt-he-en''' , dry_run=UpperCamelCase__ ) assert mmeta["long_pair"] == "heb-eng"
83
0
import inspect import re from hashlib import shaaaa from typing import Dict, List from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql import sql # noqa F401 from .text import text def SCREAMING_SNAKE_CASE_ ( __A : List[str] ) -> str: """simple docstring""" a_ : Tuple = [] for line in lines: a_ : Any = re.sub(R'#.*' , '' , __A ) # remove comments if line: filtered_lines.append(__A ) a_ : Tuple = '\n'.join(__A ) # Make a hash from all this code a_ : Tuple = full_str.encode('utf-8' ) return shaaaa(__A ).hexdigest() # get importable module names and hash for caching UpperCAmelCase_ : List[Any] = { 'csv': (csv.__name__, _hash_python_lines(inspect.getsource(csv).splitlines())), 'json': (json.__name__, _hash_python_lines(inspect.getsource(json).splitlines())), 'pandas': (pandas.__name__, _hash_python_lines(inspect.getsource(pandas).splitlines())), 'parquet': (parquet.__name__, _hash_python_lines(inspect.getsource(parquet).splitlines())), 'arrow': (arrow.__name__, _hash_python_lines(inspect.getsource(arrow).splitlines())), 'text': (text.__name__, _hash_python_lines(inspect.getsource(text).splitlines())), 'imagefolder': (imagefolder.__name__, _hash_python_lines(inspect.getsource(imagefolder).splitlines())), 'audiofolder': (audiofolder.__name__, _hash_python_lines(inspect.getsource(audiofolder).splitlines())), } # Used to infer the module to use based on the data files extensions UpperCAmelCase_ : Dict = { '.csv': ('csv', {}), '.tsv': ('csv', {'sep': '\t'}), '.json': ('json', {}), '.jsonl': ('json', {}), '.parquet': ('parquet', {}), '.arrow': ('arrow', {}), '.txt': ('text', {}), } _EXTENSION_TO_MODULE.update({ext: ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext: ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) UpperCAmelCase_ : Optional[int] = {'imagefolder', 'audiofolder'} # Used to filter data files based on extensions given a module name UpperCAmelCase_ : Dict[str, List[str]] = {} for _ext, (_module, _) in _EXTENSION_TO_MODULE.items(): _MODULE_TO_EXTENSIONS.setdefault(_module, []).append(_ext) _MODULE_TO_EXTENSIONS["imagefolder"].append('.zip') _MODULE_TO_EXTENSIONS["audiofolder"].append('.zip')
32
"""simple docstring""" def lowercase ( a__ : float , a__ : float ) -> float: if density <= 0: raise ValueError('''Impossible fluid density''' ) if bulk_modulus <= 0: raise ValueError('''Impossible bulk modulus''' ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
256
0
import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = [ ('bert.bert', 'visual_bert'), ('bert.cls', 'cls'), ('bert.classifier', 'cls'), ('token_type_embeddings_visual', 'visual_token_type_embeddings'), ('position_embeddings_visual', 'visual_position_embeddings'), ('projection', 'visual_projection'), ] __UpperCAmelCase = [ 'nlvr2_coco_pre_trained.th', 'nlvr2_fine_tuned.th', 'nlvr2_pre_trained.th', 'vcr_coco_pre_train.th', 'vcr_fine_tune.th', 'vcr_pre_train.th', 'vqa_coco_pre_trained.th', 'vqa_fine_tuned.th', 'vqa_pre_trained.th', ] def _snake_case ( lowercase__ : Any ) -> List[str]: '''simple docstring''' lowerCAmelCase_ :List[str] = torch.load(lowercase__ , map_location="""cpu""" ) return sd def _snake_case ( lowercase__ : int , lowercase__ : Optional[Any] , lowercase__ : List[str]=rename_keys_prefix ) -> Any: '''simple docstring''' lowerCAmelCase_ :Optional[int] = OrderedDict() lowerCAmelCase_ :str = torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue lowerCAmelCase_ :List[str] = key for name_pair in rename_keys_prefix: lowerCAmelCase_ :Optional[Any] = new_key.replace(name_pair[0] , name_pair[1] ) lowerCAmelCase_ :int = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately lowerCAmelCase_ :str = new_d["""cls.predictions.bias"""] return new_d @torch.no_grad() def _snake_case ( lowercase__ : Optional[Any] , lowercase__ : Dict ) -> int: '''simple docstring''' assert ( checkpoint_path.split("""/""" )[-1] in ACCEPTABLE_CHECKPOINTS ), f"""The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.""" # Get Config if "pre" in checkpoint_path: lowerCAmelCase_ :Optional[Any] = """pretraining""" if "vcr" in checkpoint_path: lowerCAmelCase_ :List[Any] = {"""visual_embedding_dim""": 5_1_2} elif "vqa_advanced" in checkpoint_path: lowerCAmelCase_ :Optional[Any] = {"""visual_embedding_dim""": 2_0_4_8} elif "vqa" in checkpoint_path: lowerCAmelCase_ :int = {"""visual_embedding_dim""": 2_0_4_8} elif "nlvr" in checkpoint_path: lowerCAmelCase_ :List[str] = {"""visual_embedding_dim""": 1_0_2_4} else: raise NotImplementedError(f"""No implementation found for `{checkpoint_path}`.""" ) else: if "vcr" in checkpoint_path: lowerCAmelCase_ :Tuple = {"""visual_embedding_dim""": 5_1_2} lowerCAmelCase_ :List[Any] = """multichoice""" elif "vqa_advanced" in checkpoint_path: lowerCAmelCase_ :Optional[int] = {"""visual_embedding_dim""": 2_0_4_8} lowerCAmelCase_ :Tuple = """vqa_advanced""" elif "vqa" in checkpoint_path: lowerCAmelCase_ :List[Any] = {"""visual_embedding_dim""": 2_0_4_8, """num_labels""": 3_1_2_9} lowerCAmelCase_ :Dict = """vqa""" elif "nlvr" in checkpoint_path: lowerCAmelCase_ :Optional[Any] = { """visual_embedding_dim""": 1_0_2_4, """num_labels""": 2, } lowerCAmelCase_ :Any = """nlvr""" lowerCAmelCase_ :List[str] = VisualBertConfig(**lowercase__ ) # Load State Dict lowerCAmelCase_ :Optional[int] = load_state_dict(lowercase__ ) lowerCAmelCase_ :int = get_new_dict(lowercase__ , lowercase__ ) if model_type == "pretraining": lowerCAmelCase_ :List[Any] = VisualBertForPreTraining(lowercase__ ) elif model_type == "vqa": lowerCAmelCase_ :Any = VisualBertForQuestionAnswering(lowercase__ ) elif model_type == "nlvr": lowerCAmelCase_ :Any = VisualBertForVisualReasoning(lowercase__ ) elif model_type == "multichoice": lowerCAmelCase_ :Optional[int] = VisualBertForMultipleChoice(lowercase__ ) model.load_state_dict(lowercase__ ) # Save Checkpoints Path(lowercase__ ).mkdir(exist_ok=lowercase__ ) model.save_pretrained(lowercase__ ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument('orig_checkpoint_path', type=str, help='A path to .th on local filesystem.') parser.add_argument('pytorch_dump_folder_path', type=str, help='Path to the output PyTorch model.') __UpperCAmelCase = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
351
"""simple docstring""" import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class _SCREAMING_SNAKE_CASE : def __init__( self , __A ) -> Union[str, Any]: if isinstance(__A , __A ): # Don't modify user's data should they want to reuse it (e.g. in tests), because once we # modified it, it will not be accepted here again, since `auto` values would have been overridden lowerCAmelCase_ :Tuple = deepcopy(__A ) elif os.path.exists(__A ): with io.open(__A , """r""" , encoding="""utf-8""" ) as f: lowerCAmelCase_ :str = json.load(__A ) else: try: lowerCAmelCase_ :Dict = baseaa.urlsafe_baadecode(__A ).decode("""utf-8""" ) lowerCAmelCase_ :int = json.loads(__A ) except (UnicodeDecodeError, AttributeError, ValueError): raise ValueError( f"""Expected a string path to an existing deepspeed config, or a dictionary, or a base64 encoded string. Received: {config_file_or_dict}""" ) lowerCAmelCase_ :Optional[Any] = config self.set_stage_and_offload() def __lowerCAmelCase ( self ) -> Tuple: # zero stage - this is done as early as possible, before model is created, to allow # ``is_deepspeed_zero3_enabled`` query and getting to the early deepspeed config object # during ``zero.Init()`` which needs to know the dtype, and some other hparams. lowerCAmelCase_ :Tuple = self.get_value("""zero_optimization.stage""" , -1 ) # offload lowerCAmelCase_ :Dict = False if self.is_zeroa() or self.is_zeroa(): lowerCAmelCase_ :Optional[int] = set(["""cpu""", """nvme"""] ) lowerCAmelCase_ :Union[str, Any] = set( [ self.get_value("""zero_optimization.offload_optimizer.device""" ), self.get_value("""zero_optimization.offload_param.device""" ), ] ) if len(offload_devices & offload_devices_valid ) > 0: lowerCAmelCase_ :Optional[int] = True def __lowerCAmelCase ( self , __A ) -> Optional[Any]: lowerCAmelCase_ :str = self.config # find the config node of interest if it exists lowerCAmelCase_ :Tuple = ds_key_long.split(""".""" ) lowerCAmelCase_ :List[str] = nodes.pop() for node in nodes: lowerCAmelCase_ :Tuple = config.get(__A ) if config is None: return None, ds_key return config, ds_key def __lowerCAmelCase ( self , __A , __A=None ) -> Optional[Any]: lowerCAmelCase_ , lowerCAmelCase_ :Optional[Any] = self.find_config_node(__A ) if config is None: return default return config.get(__A , __A ) def __lowerCAmelCase ( self , __A , __A=False ) -> Optional[Any]: lowerCAmelCase_ :Tuple = self.config # find the config node of interest if it exists lowerCAmelCase_ :Union[str, Any] = ds_key_long.split(""".""" ) for node in nodes: lowerCAmelCase_ :int = config lowerCAmelCase_ :Any = config.get(__A ) if config is None: if must_exist: raise ValueError(f"""Can't find {ds_key_long} entry in the config: {self.config}""" ) else: return # if found remove it if parent_config is not None: parent_config.pop(__A ) def __lowerCAmelCase ( self , __A ) -> Union[str, Any]: lowerCAmelCase_ :Optional[int] = self.get_value(__A ) return False if value is None else bool(__A ) def __lowerCAmelCase ( self , __A ) -> Optional[int]: lowerCAmelCase_ :List[str] = self.get_value(__A ) return False if value is None else not bool(__A ) def __lowerCAmelCase ( self ) -> str: return self._stage == 2 def __lowerCAmelCase ( self ) -> Union[str, Any]: return self._stage == 3 def __lowerCAmelCase ( self ) -> Union[str, Any]: return self._offload class _SCREAMING_SNAKE_CASE : def __init__( self , __A ) -> Optional[int]: lowerCAmelCase_ :Dict = engine def __lowerCAmelCase ( self , __A , **__A ) -> str: # runs backpropagation and handles mixed precision self.engine.backward(__A , **__A ) # Deepspeed's `engine.step` performs the following operations: # - gradient accumulation check # - gradient clipping # - optimizer step # - zero grad # - checking overflow # - lr_scheduler step (only if engine.lr_scheduler is not None) self.engine.step() # and this plugin overrides the above calls with no-ops when Accelerate runs under # Deepspeed, but allows normal functionality for non-Deepspeed cases thus enabling a simple # training loop that works transparently under many training regimes. class _SCREAMING_SNAKE_CASE ( A__ ): def __init__( self , __A ) -> List[str]: super().__init__(__A , device_placement=__A , scaler=__A ) lowerCAmelCase_ :List[str] = hasattr(self.optimizer , """overflow""" ) def __lowerCAmelCase ( self , __A=None ) -> Optional[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed def __lowerCAmelCase ( self ) -> List[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed @property def __lowerCAmelCase ( self ) -> int: if self.__has_overflow__: return self.optimizer.overflow return False class _SCREAMING_SNAKE_CASE ( A__ ): def __init__( self , __A , __A ) -> Optional[int]: super().__init__(__A , __A ) def __lowerCAmelCase ( self ) -> Any: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed class _SCREAMING_SNAKE_CASE : def __init__( self , __A , __A=0.0_0_1 , __A=0 , **__A ) -> List[Any]: lowerCAmelCase_ :str = params lowerCAmelCase_ :Any = lr lowerCAmelCase_ :List[Any] = weight_decay lowerCAmelCase_ :Any = kwargs class _SCREAMING_SNAKE_CASE : def __init__( self , __A , __A=None , __A=0 , **__A ) -> List[str]: lowerCAmelCase_ :Optional[int] = optimizer lowerCAmelCase_ :int = total_num_steps lowerCAmelCase_ :List[Any] = warmup_num_steps lowerCAmelCase_ :int = kwargs
1
0
"""simple docstring""" import fire from utils import calculate_rouge, save_json def lowercase ( lowerCAmelCase__ : Tuple , lowerCAmelCase__ : str , lowerCAmelCase__ : List[str]=None , **lowerCAmelCase__ : Dict ) -> int: __a = [x.strip() for x in open(lowerCAmelCase__ ).readlines()] __a = [x.strip() for x in open(lowerCAmelCase__ ).readlines()][: len(lowerCAmelCase__ )] __a = calculate_rouge(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) if save_path is not None: save_json(lowerCAmelCase__ , lowerCAmelCase__ , indent=lowerCAmelCase__ ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
45
"""simple docstring""" from math import isqrt, loga def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> list[int]: '''simple docstring''' lowercase_ = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , __lowerCAmelCase , __lowerCAmelCase ): lowercase_ = False return [i for i in range(2 , __lowerCAmelCase ) if is_prime[i]] def _SCREAMING_SNAKE_CASE (__lowerCAmelCase = 80_08_00 , __lowerCAmelCase = 80_08_00 ) -> int: '''simple docstring''' lowercase_ = degree * loga(__lowerCAmelCase ) lowercase_ = int(__lowerCAmelCase ) lowercase_ = calculate_prime_numbers(__lowerCAmelCase ) lowercase_ = 0 lowercase_ = 0 lowercase_ = len(__lowerCAmelCase ) - 1 while left < right: while ( prime_numbers[right] * loga(prime_numbers[left] ) + prime_numbers[left] * loga(prime_numbers[right] ) > upper_bound ): right -= 1 hybrid_integers_count += right - left left += 1 return hybrid_integers_count if __name__ == "__main__": print(F"{solution() = }")
136
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCamelCase : Tuple = logging.get_logger(__name__) __lowerCamelCase : Dict = { "facebook/data2vec-text-base": "https://huggingface.co/data2vec/resolve/main/config.json", } class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ): lowercase__ = "data2vec-text" def __init__( self : List[Any] , lowerCAmelCase_ : str=3_0_5_2_2 , lowerCAmelCase_ : Any=7_6_8 , lowerCAmelCase_ : int=1_2 , lowerCAmelCase_ : List[str]=1_2 , lowerCAmelCase_ : Union[str, Any]=3_0_7_2 , lowerCAmelCase_ : Any="gelu" , lowerCAmelCase_ : int=0.1 , lowerCAmelCase_ : str=0.1 , lowerCAmelCase_ : Optional[Any]=5_1_2 , lowerCAmelCase_ : Optional[Any]=2 , lowerCAmelCase_ : Optional[int]=0.02 , lowerCAmelCase_ : Union[str, Any]=1E-12 , lowerCAmelCase_ : Union[str, Any]=1 , lowerCAmelCase_ : Union[str, Any]=0 , lowerCAmelCase_ : Optional[Any]=2 , lowerCAmelCase_ : Optional[int]="absolute" , lowerCAmelCase_ : List[str]=True , lowerCAmelCase_ : List[str]=None , **lowerCAmelCase_ : Tuple , ): """simple docstring""" super().__init__(pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_) lowercase_ = vocab_size lowercase_ = hidden_size lowercase_ = num_hidden_layers lowercase_ = num_attention_heads lowercase_ = hidden_act lowercase_ = intermediate_size lowercase_ = hidden_dropout_prob lowercase_ = attention_probs_dropout_prob lowercase_ = max_position_embeddings lowercase_ = type_vocab_size lowercase_ = initializer_range lowercase_ = layer_norm_eps lowercase_ = position_embedding_type lowercase_ = use_cache lowercase_ = classifier_dropout class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ): @property def _UpperCAmelCase ( self : Optional[Any]): """simple docstring""" if self.task == "multiple-choice": lowercase_ = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowercase_ = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ])
371
"""simple docstring""" import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , unittest.TestCase ): lowercase__ = BarthezTokenizer lowercase__ = BarthezTokenizerFast lowercase__ = True lowercase__ = True def _UpperCAmelCase ( self : List[Any]): """simple docstring""" super().setUp() lowercase_ = BarthezTokenizerFast.from_pretrained("""moussaKam/mbarthez""") tokenizer.save_pretrained(self.tmpdirname) tokenizer.save_pretrained(self.tmpdirname , legacy_format=lowerCAmelCase_) lowercase_ = tokenizer def _UpperCAmelCase ( self : Any): """simple docstring""" lowercase_ = """<pad>""" lowercase_ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCAmelCase_) , lowerCAmelCase_) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCAmelCase_) , lowerCAmelCase_) def _UpperCAmelCase ( self : Dict): """simple docstring""" lowercase_ = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , """<s>""") self.assertEqual(vocab_keys[1] , """<pad>""") self.assertEqual(vocab_keys[-1] , """<mask>""") self.assertEqual(len(lowerCAmelCase_) , 1_0_1_1_2_2) def _UpperCAmelCase ( self : Optional[int]): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_0_1_1_2_2) @require_torch def _UpperCAmelCase ( self : List[str]): """simple docstring""" lowercase_ = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] lowercase_ = [0, 5_7, 3_0_1_8, 7_0_3_0_7, 9_1, 2] lowercase_ = self.tokenizer( lowerCAmelCase_ , max_length=len(lowerCAmelCase_) , padding=lowerCAmelCase_ , truncation=lowerCAmelCase_ , return_tensors="""pt""") self.assertIsInstance(lowerCAmelCase_ , lowerCAmelCase_) self.assertEqual((2, 6) , batch.input_ids.shape) self.assertEqual((2, 6) , batch.attention_mask.shape) lowercase_ = batch.input_ids.tolist()[0] self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_) def _UpperCAmelCase ( self : List[Any]): """simple docstring""" if not self.test_rust_tokenizer: return lowercase_ = self.get_tokenizer() lowercase_ = self.get_rust_tokenizer() lowercase_ = """I was born in 92000, and this is falsé.""" lowercase_ = tokenizer.tokenize(lowerCAmelCase_) lowercase_ = rust_tokenizer.tokenize(lowerCAmelCase_) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_) lowercase_ = rust_tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_) lowercase_ = self.get_rust_tokenizer() lowercase_ = tokenizer.encode(lowerCAmelCase_) lowercase_ = rust_tokenizer.encode(lowerCAmelCase_) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_) @slow def _UpperCAmelCase ( self : int): """simple docstring""" lowercase_ = {"""input_ids""": [[0, 4_9_0, 1_4_3_2_8, 4_5_0_7, 3_5_4, 4_7, 4_3_6_6_9, 9_5, 2_5, 7_8_1_1_7, 2_0_2_1_5, 1_9_7_7_9, 1_9_0, 2_2, 4_0_0, 4, 3_5_3_4_3, 8_0_3_1_0, 6_0_3, 8_6, 2_4_9_3_7, 1_0_5, 3_3_4_3_8, 9_4_7_6_2, 1_9_6, 3_9_6_4_2, 7, 1_5, 1_5_9_3_3, 1_7_3, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_0_5_3_4, 8_7, 2_5, 6_6, 3_3_5_8, 1_9_6, 5_5_2_8_9, 8, 8_2_9_6_1, 8_1, 2_2_0_4, 7_5_2_0_3, 7, 1_5, 7_6_3, 1_2_9_5_6, 2_1_6, 1_7_8, 1_4_3_2_8, 9_5_9_5, 1_3_7_7, 6_9_6_9_3, 7, 4_4_8, 7_1_0_2_1, 1_9_6, 1_8_1_0_6, 1_4_3_7, 1_3_9_7_4, 1_0_8, 9_0_8_3, 4, 4_9_3_1_5, 7, 3_9, 8_6, 1_3_2_6, 2_7_9_3, 4_6_3_3_3, 4, 4_4_8, 1_9_6, 7_4_5_8_8, 7, 4_9_3_1_5, 7, 3_9, 2_1, 8_2_2, 3_8_4_7_0, 7_4, 2_1, 6_6_7_2_3, 6_2_4_8_0, 8, 2_2_0_5_0, 5, 2]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. lowercase_ = [ """Le transformeur est un modèle d'apprentissage profond introduit en 2017, """ """utilisé principalement dans le domaine du traitement automatique des langues (TAL).""", """À l'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus """ """pour gérer des données séquentielles, telles que le langage naturel, pour des tâches """ """telles que la traduction et la synthèse de texte.""", ] self.tokenizer_integration_test_util( expected_encoding=lowerCAmelCase_ , model_name="""moussaKam/mbarthez""" , revision="""c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6""" , sequences=lowerCAmelCase_ , )
313
0