code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _A : Dict ={ '''configuration_clip''': [ '''CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''CLIPConfig''', '''CLIPOnnxConfig''', '''CLIPTextConfig''', '''CLIPVisionConfig''', ], '''processing_clip''': ['''CLIPProcessor'''], '''tokenization_clip''': ['''CLIPTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Dict =['''CLIPTokenizerFast'''] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[int] =['''CLIPFeatureExtractor'''] _A : List[Any] =['''CLIPImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : List[str] =[ '''CLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''CLIPModel''', '''CLIPPreTrainedModel''', '''CLIPTextModel''', '''CLIPTextModelWithProjection''', '''CLIPVisionModel''', '''CLIPVisionModelWithProjection''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Union[str, Any] =[ '''TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFCLIPModel''', '''TFCLIPPreTrainedModel''', '''TFCLIPTextModel''', '''TFCLIPVisionModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Dict =[ '''FlaxCLIPModel''', '''FlaxCLIPPreTrainedModel''', '''FlaxCLIPTextModel''', '''FlaxCLIPTextPreTrainedModel''', '''FlaxCLIPVisionModel''', '''FlaxCLIPVisionPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_clip import ( CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPConfig, CLIPOnnxConfig, CLIPTextConfig, CLIPVisionConfig, ) from .processing_clip import CLIPProcessor from .tokenization_clip import CLIPTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_clip_fast import CLIPTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clip import CLIPFeatureExtractor from .image_processing_clip import CLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clip import ( CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPModel, CLIPPreTrainedModel, CLIPTextModel, CLIPTextModelWithProjection, CLIPVisionModel, CLIPVisionModelWithProjection, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_clip import ( TF_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFCLIPModel, TFCLIPPreTrainedModel, TFCLIPTextModel, TFCLIPVisionModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_clip import ( FlaxCLIPModel, FlaxCLIPPreTrainedModel, FlaxCLIPTextModel, FlaxCLIPTextPreTrainedModel, FlaxCLIPVisionModel, FlaxCLIPVisionPreTrainedModel, ) else: import sys _A : Optional[Any] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
41
'''simple docstring''' def UpperCAmelCase_ (__a : list , __a : list , __a : int ): """simple docstring""" _a : Optional[Any] = len(__a ) _a : int = [[0] * n for i in range(__a )] for i in range(__a ): _a : Tuple = y_points[i] for i in range(2 , __a ): for j in range(__a , __a ): _a : Tuple = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
271
0
'''simple docstring''' import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __UpperCAmelCase ( _lowerCamelCase , unittest.TestCase ): __lowercase = GPTSanJapaneseTokenizer __lowercase = False __lowercase = {"""do_clean_text""": False, """add_prefix_space""": False} def lowerCamelCase ( self ): """simple docstring""" super().setUp() # fmt: off _snake_case = ['こん', 'こんに', 'にちは', 'ばんは', '世界,㔺界', '、', '。', '<BR>', '<SP>', '<TAB>', '<URL>', '<EMAIL>', '<TEL>', '<DATE>', '<PRICE>', '<BLOCK>', '<KIGOU>', '<U2000U2BFF>', '<|emoji1|>', '<unk>', '<|bagoftoken|>', '<|endoftext|>'] # fmt: on _snake_case = {'emoji': {'\ud83d\ude00': '<|emoji1|>'}, 'emoji_inv': {'<|emoji1|>': '\ud83d\ude00'}} # 😀 _snake_case = {'unk_token': '<unk>'} _snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) _snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['emoji_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) with open(self.emoji_file , 'w' ) as emoji_writer: emoji_writer.write(json.dumps(lowerCAmelCase_ ) ) def lowerCamelCase ( self , **lowerCAmelCase_ ): """simple docstring""" kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = 'こんにちは、世界。 \nこんばんは、㔺界。😀' _snake_case = 'こんにちは、世界。 \nこんばんは、世界。😀' return input_text, output_text def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case , _snake_case = self.get_input_output_texts(lowerCAmelCase_ ) _snake_case = tokenizer.encode(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ) _snake_case = tokenizer.decode(lowerCAmelCase_ , clean_up_tokenization_spaces=lowerCAmelCase_ ) return text, ids def lowerCamelCase ( self ): """simple docstring""" pass # TODO add if relevant def lowerCamelCase ( self ): """simple docstring""" pass # TODO add if relevant def lowerCamelCase ( self ): """simple docstring""" pass # TODO add if relevant def lowerCamelCase ( self ): """simple docstring""" _snake_case = self.get_tokenizer() # Testing tokenization _snake_case = 'こんにちは、世界。 こんばんは、㔺界。' _snake_case = ['こん', 'にちは', '、', '世界', '。', '<SP>', 'こん', 'ばんは', '、', '㔺界', '。'] _snake_case = tokenizer.tokenize(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) # Testing conversion to ids without special tokens _snake_case = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] _snake_case = tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) # Testing conversion to ids with special tokens _snake_case = tokens + [tokenizer.unk_token] _snake_case = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] _snake_case = tokenizer.convert_tokens_to_ids(lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCamelCase ( self ): """simple docstring""" _snake_case = self.get_tokenizer() # Testing tokenization _snake_case = 'こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。' _snake_case = 'こんにちは、、、、世界。こんばんは、、、、世界。' _snake_case = tokenizer.encode(lowerCAmelCase_ ) _snake_case = tokenizer.decode(lowerCAmelCase_ ) self.assertEqual(lowerCAmelCase_ , lowerCAmelCase_ ) @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization _snake_case = 'こんにちは、世界。' _snake_case = 'こんばんは、㔺界。😀' _snake_case = 'こんにちは、世界。こんばんは、世界。😀' _snake_case = tokenizer.encode(prefix_text + input_text ) _snake_case = tokenizer.encode('' , prefix_text=prefix_text + input_text ) _snake_case = tokenizer.encode(lowerCAmelCase_ , prefix_text=lowerCAmelCase_ ) _snake_case = tokenizer.decode(lowerCAmelCase_ ) _snake_case = tokenizer.decode(lowerCAmelCase_ ) _snake_case = tokenizer.decode(lowerCAmelCase_ ) self.assertEqual(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertEqual(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertEqual(lowerCAmelCase_ , lowerCAmelCase_ ) @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization _snake_case = 'こんにちは、世界。' _snake_case = 'こんばんは、㔺界。😀' _snake_case = len(tokenizer.encode(lowerCAmelCase_ ) ) - 2 _snake_case = len(tokenizer.encode(lowerCAmelCase_ ) ) - 2 _snake_case = [1] + [0] * (len_prefix + len_text + 1) _snake_case = [1] * (len_prefix + len_text + 1) + [0] _snake_case = [1] + [1] * (len_prefix) + [0] * (len_text + 1) _snake_case = tokenizer(prefix_text + input_text ).token_type_ids _snake_case = tokenizer('' , prefix_text=prefix_text + input_text ).token_type_ids _snake_case = tokenizer(lowerCAmelCase_ , prefix_text=lowerCAmelCase_ ).token_type_ids self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertListEqual(lowerCAmelCase_ , lowerCAmelCase_ ) @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) _snake_case = tokenizer.encode('あンいワ' ) _snake_case = tokenizer.encode('' , prefix_text='あンいワ' ) _snake_case = tokenizer.encode('いワ' , prefix_text='あン' ) self.assertEqual(tokenizer.decode(lowerCAmelCase_ ) , tokenizer.decode(lowerCAmelCase_ ) ) self.assertEqual(tokenizer.decode(lowerCAmelCase_ ) , tokenizer.decode(lowerCAmelCase_ ) ) self.assertNotEqual(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertNotEqual(lowerCAmelCase_ , lowerCAmelCase_ ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) _snake_case = [['武田信玄', 'は、'], ['織田信長', 'の配下の、']] _snake_case = tokenizer(lowerCAmelCase_ , padding=lowerCAmelCase_ ) _snake_case = tokenizer.batch_encode_plus(lowerCAmelCase_ , padding=lowerCAmelCase_ ) # fmt: off _snake_case = [[3_59_93, 86_40, 2_59_48, 3_59_98, 3_06_47, 3_56_75, 3_59_99, 3_59_99], [3_59_93, 1_03_82, 98_68, 3_59_98, 3_06_46, 94_59, 3_06_46, 3_56_75]] _snake_case = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] _snake_case = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , lowerCAmelCase_ ) self.assertListEqual(x_token.token_type_ids , lowerCAmelCase_ ) self.assertListEqual(x_token.attention_mask , lowerCAmelCase_ ) self.assertListEqual(x_token_a.input_ids , lowerCAmelCase_ ) self.assertListEqual(x_token_a.token_type_ids , lowerCAmelCase_ ) self.assertListEqual(x_token_a.attention_mask , lowerCAmelCase_ ) def lowerCamelCase ( self ): """simple docstring""" pass def lowerCamelCase ( self ): """simple docstring""" pass
42
'''simple docstring''' import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = inspect.getfile(accelerate.test_utils ) __UpperCAmelCase : List[str] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_cli.py'''] ) __UpperCAmelCase : Dict = ['''accelerate''', '''launch'''] __UpperCAmelCase : Dict = Path.home() / '''.cache/huggingface/accelerate''' __UpperCAmelCase : Dict = '''default_config.yaml''' __UpperCAmelCase : Optional[Any] = config_folder / config_file __UpperCAmelCase : Dict = config_folder / '''_default_config.yaml''' __UpperCAmelCase : Any = Path('''tests/test_configs''' ) @classmethod def __lowercase ( cls : int ): '''simple docstring''' if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def __lowercase ( cls : List[Any] ): '''simple docstring''' if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Dict = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' for config in sorted(self.test_config_path.glob('**/*.yaml' ) ): with self.subTest(config_file=_a ): execute_subprocess_async( self.base_cmd + ['--config_file', str(_a ), self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[int] ): '''simple docstring''' execute_subprocess_async(['accelerate', 'test'] ,env=os.environ.copy() ) class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = '''test-tpu''' __UpperCAmelCase : Any = '''us-central1-a''' __UpperCAmelCase : List[Any] = '''ls''' __UpperCAmelCase : Any = ['''accelerate''', '''tpu-config'''] __UpperCAmelCase : Dict = '''cd /usr/share''' __UpperCAmelCase : Any = '''tests/test_samples/test_command_file.sh''' __UpperCAmelCase : List[Any] = '''Running gcloud compute tpus tpu-vm ssh''' def __lowercase ( self : Dict ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Any = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--debug'] ,return_stdout=_a ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : int ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : str ): '''simple docstring''' _a : List[str] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--command', 'echo "Hello World"', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo \"Hello World\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Any = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command_file', self.command_file, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Union[str, Any] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command_file', self.command_file, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Any ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--accelerate_version', '12.0.0', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,)
271
0
# Algorithm for the pigeonhole sorting def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = min(SCREAMING_SNAKE_CASE ) # min() finds the minimum value __UpperCamelCase :List[Any] = max(SCREAMING_SNAKE_CASE ) # max() finds the maximum value __UpperCamelCase :List[Any] = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size __UpperCamelCase :List[Any] = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. __UpperCamelCase :Optional[Any] = 0 for count in range(SCREAMING_SNAKE_CASE ): while holes[count] > 0: holes[count] -= 1 __UpperCamelCase :int = count + min_val i += 1 def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Optional[Any] = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(SCREAMING_SNAKE_CASE ) print('''Sorted order is:''' , ''' '''.join(SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": main()
43
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __lowerCAmelCase = TypeVar("""T""") class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Tuple ,_a : T ): '''simple docstring''' _a : List[str] = data _a : Node[T] | None = None def __str__( self : Dict ): '''simple docstring''' return F"""{self.data}""" class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Optional[int] ): '''simple docstring''' _a : Node[T] | None = None def __iter__( self : str ): '''simple docstring''' _a : Tuple = self.top while node: yield node.data _a : int = node.next def __str__( self : str ): '''simple docstring''' return "->".join([str(_a ) for item in self] ) def __len__( self : Optional[Any] ): '''simple docstring''' return len(tuple(iter(self ) ) ) def __lowercase ( self : str ): '''simple docstring''' return self.top is None def __lowercase ( self : List[Any] ,_a : T ): '''simple docstring''' _a : int = Node(_a ) if not self.is_empty(): _a : Optional[Any] = self.top _a : List[str] = node def __lowercase ( self : Tuple ): '''simple docstring''' if self.is_empty(): raise IndexError('pop from empty stack' ) assert isinstance(self.top ,_a ) _a : List[Any] = self.top _a : int = self.top.next return pop_node.data def __lowercase ( self : List[str] ): '''simple docstring''' if self.is_empty(): raise IndexError('peek from empty stack' ) assert self.top is not None return self.top.data def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = None if __name__ == "__main__": from doctest import testmod testmod()
271
0
"""simple docstring""" import argparse import os import re import tensorflow as tf import torch from transformers import BertConfig, BertModel from transformers.utils import logging logging.set_verbosity_info() _a : Optional[int] = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( _lowerCamelCase : Optional[Any] ,_lowerCamelCase : str ,_lowerCamelCase : List[str] ) -> int: _lowerCAmelCase : List[str] = os.path.abspath(_lowerCamelCase ) logger.info(f"Converting TensorFlow checkpoint from {tf_path}" ) # Load weights from TF model _lowerCAmelCase : int = tf.train.list_variables(_lowerCamelCase ) _lowerCAmelCase : List[Any] = [] _lowerCAmelCase : Tuple = [] _lowerCAmelCase : Optional[int] = [] for full_name, shape in init_vars: # logger.info(f"Loading TF weight {name} with shape {shape}") _lowerCAmelCase : Dict = full_name.split("""/""" ) if full_name == "_CHECKPOINTABLE_OBJECT_GRAPH" or name[0] in ["global_step", "save_counter"]: logger.info(f"Skipping non-model layer {full_name}" ) continue if "optimizer" in full_name: logger.info(f"Skipping optimization layer {full_name}" ) continue if name[0] == "model": # ignore initial 'model' _lowerCAmelCase : Dict = name[1:] # figure out how many levels deep the name is _lowerCAmelCase : Optional[Any] = 0 for _name in name: if _name.startswith("""layer_with_weights""" ): depth += 1 else: break layer_depth.append(_lowerCamelCase ) # read data _lowerCAmelCase : List[Any] = tf.train.load_variable(_lowerCamelCase ,_lowerCamelCase ) names.append("""/""".join(_lowerCamelCase ) ) arrays.append(_lowerCamelCase ) logger.info(f"Read a total of {len(_lowerCamelCase ):,} layers" ) # Sanity check if len(set(_lowerCamelCase ) ) != 1: raise ValueError(f"Found layer names with different depths (layer depth {list(set(_lowerCamelCase ) )})" ) _lowerCAmelCase : Union[str, Any] = list(set(_lowerCamelCase ) )[0] if layer_depth != 1: raise ValueError( """The model contains more than just the embedding/encoder layers. This script does not handle MLM/NSP""" """ heads.""" ) # convert layers logger.info("""Converting weights...""" ) for full_name, array in zip(_lowerCamelCase ,_lowerCamelCase ): _lowerCAmelCase : Optional[Any] = full_name.split("""/""" ) _lowerCAmelCase : str = model _lowerCAmelCase : Optional[int] = [] for i, m_name in enumerate(_lowerCamelCase ): if m_name == ".ATTRIBUTES": # variable names end with .ATTRIBUTES/VARIABLE_VALUE break if m_name.startswith("""layer_with_weights""" ): _lowerCAmelCase : Dict = int(m_name.split("""-""" )[-1] ) if layer_num <= 2: # embedding layers # layer_num 0: word_embeddings # layer_num 1: position_embeddings # layer_num 2: token_type_embeddings continue elif layer_num == 3: # embedding LayerNorm trace.extend(["""embeddings""", """LayerNorm"""] ) _lowerCAmelCase : int = getattr(_lowerCamelCase ,"""embeddings""" ) _lowerCAmelCase : List[Any] = getattr(_lowerCamelCase ,"""LayerNorm""" ) elif layer_num > 3 and layer_num < config.num_hidden_layers + 4: # encoder layers trace.extend(["""encoder""", """layer""", str(layer_num - 4 )] ) _lowerCAmelCase : int = getattr(_lowerCamelCase ,"""encoder""" ) _lowerCAmelCase : Optional[int] = getattr(_lowerCamelCase ,"""layer""" ) _lowerCAmelCase : List[Any] = pointer[layer_num - 4] elif layer_num == config.num_hidden_layers + 4: # pooler layer trace.extend(["""pooler""", """dense"""] ) _lowerCAmelCase : int = getattr(_lowerCamelCase ,"""pooler""" ) _lowerCAmelCase : Optional[Any] = getattr(_lowerCamelCase ,"""dense""" ) elif m_name == "embeddings": trace.append("""embeddings""" ) _lowerCAmelCase : Tuple = getattr(_lowerCamelCase ,"""embeddings""" ) if layer_num == 0: trace.append("""word_embeddings""" ) _lowerCAmelCase : Optional[int] = getattr(_lowerCamelCase ,"""word_embeddings""" ) elif layer_num == 1: trace.append("""position_embeddings""" ) _lowerCAmelCase : List[Any] = getattr(_lowerCamelCase ,"""position_embeddings""" ) elif layer_num == 2: trace.append("""token_type_embeddings""" ) _lowerCAmelCase : str = getattr(_lowerCamelCase ,"""token_type_embeddings""" ) else: raise ValueError(f"Unknown embedding layer with name {full_name}" ) trace.append("""weight""" ) _lowerCAmelCase : Dict = getattr(_lowerCamelCase ,"""weight""" ) elif m_name == "_attention_layer": # self-attention layer trace.extend(["""attention""", """self"""] ) _lowerCAmelCase : Tuple = getattr(_lowerCamelCase ,"""attention""" ) _lowerCAmelCase : str = getattr(_lowerCamelCase ,"""self""" ) elif m_name == "_attention_layer_norm": # output attention norm trace.extend(["""attention""", """output""", """LayerNorm"""] ) _lowerCAmelCase : List[str] = getattr(_lowerCamelCase ,"""attention""" ) _lowerCAmelCase : Tuple = getattr(_lowerCamelCase ,"""output""" ) _lowerCAmelCase : Dict = getattr(_lowerCamelCase ,"""LayerNorm""" ) elif m_name == "_attention_output_dense": # output attention dense trace.extend(["""attention""", """output""", """dense"""] ) _lowerCAmelCase : int = getattr(_lowerCamelCase ,"""attention""" ) _lowerCAmelCase : int = getattr(_lowerCamelCase ,"""output""" ) _lowerCAmelCase : List[Any] = getattr(_lowerCamelCase ,"""dense""" ) elif m_name == "_output_dense": # output dense trace.extend(["""output""", """dense"""] ) _lowerCAmelCase : List[str] = getattr(_lowerCamelCase ,"""output""" ) _lowerCAmelCase : Any = getattr(_lowerCamelCase ,"""dense""" ) elif m_name == "_output_layer_norm": # output dense trace.extend(["""output""", """LayerNorm"""] ) _lowerCAmelCase : Tuple = getattr(_lowerCamelCase ,"""output""" ) _lowerCAmelCase : List[str] = getattr(_lowerCamelCase ,"""LayerNorm""" ) elif m_name == "_key_dense": # attention key trace.append("""key""" ) _lowerCAmelCase : str = getattr(_lowerCamelCase ,"""key""" ) elif m_name == "_query_dense": # attention query trace.append("""query""" ) _lowerCAmelCase : Optional[int] = getattr(_lowerCamelCase ,"""query""" ) elif m_name == "_value_dense": # attention value trace.append("""value""" ) _lowerCAmelCase : Optional[Any] = getattr(_lowerCamelCase ,"""value""" ) elif m_name == "_intermediate_dense": # attention intermediate dense trace.extend(["""intermediate""", """dense"""] ) _lowerCAmelCase : str = getattr(_lowerCamelCase ,"""intermediate""" ) _lowerCAmelCase : Optional[Any] = getattr(_lowerCamelCase ,"""dense""" ) elif m_name == "_output_layer_norm": # output layer norm trace.append("""output""" ) _lowerCAmelCase : Any = getattr(_lowerCamelCase ,"""output""" ) # weights & biases elif m_name in ["bias", "beta"]: trace.append("""bias""" ) _lowerCAmelCase : List[str] = getattr(_lowerCamelCase ,"""bias""" ) elif m_name in ["kernel", "gamma"]: trace.append("""weight""" ) _lowerCAmelCase : str = getattr(_lowerCamelCase ,"""weight""" ) else: logger.warning(f"Ignored {m_name}" ) # for certain layers reshape is necessary _lowerCAmelCase : Any = """.""".join(_lowerCamelCase ) if re.match(r"""(\S+)\.attention\.self\.(key|value|query)\.(bias|weight)""" ,_lowerCamelCase ) or re.match( r"""(\S+)\.attention\.output\.dense\.weight""" ,_lowerCamelCase ): _lowerCAmelCase : str = array.reshape(pointer.data.shape ) if "kernel" in full_name: _lowerCAmelCase : Optional[int] = array.transpose() if pointer.shape == array.shape: _lowerCAmelCase : Optional[int] = torch.from_numpy(_lowerCamelCase ) else: raise ValueError( f"Shape mismatch in layer {full_name}: Model expects shape {pointer.shape} but layer contains shape:" f" {array.shape}" ) logger.info(f"Successfully set variable {full_name} to PyTorch layer {trace}" ) return model def SCREAMING_SNAKE_CASE ( _lowerCamelCase : Union[str, Any] ,_lowerCamelCase : str ,_lowerCamelCase : Any ) -> Tuple: # Instantiate model logger.info(f"Loading model based on config from {config_path}..." ) _lowerCAmelCase : List[Any] = BertConfig.from_json_file(_lowerCamelCase ) _lowerCAmelCase : Tuple = BertModel(_lowerCamelCase ) # Load weights from checkpoint logger.info(f"Loading weights from checkpoint {tf_checkpoint_path}..." ) load_tfa_weights_in_bert(_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ) # Save pytorch-model logger.info(f"Saving PyTorch model to {pytorch_dump_path}..." ) torch.save(model.state_dict() ,_lowerCamelCase ) if __name__ == "__main__": _a : Union[str, Any] = argparse.ArgumentParser() parser.add_argument( '--tf_checkpoint_path', type=str, required=True, help='Path to the TensorFlow 2.x checkpoint path.' ) parser.add_argument( '--bert_config_file', type=str, required=True, help='The config json file corresponding to the BERT model. This specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', type=str, required=True, help='Path to the output PyTorch model (must include filename).', ) _a : List[str] = parser.parse_args() convert_tfa_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
44
'''simple docstring''' import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : List[str] ): '''simple docstring''' torch.manual_seed(0 ) _a : int = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : str = self.dummy_uncond_unet _a : int = PNDMScheduler() _a : str = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Any = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ,return_dict=_a )[0] _a : List[Any] = image[0, -3:, -3:, -1] _a : Any = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : List[Any] = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[str] = 'google/ddpm-cifar10-32' _a : str = UNetaDModel.from_pretrained(_a ) _a : Union[str, Any] = PNDMScheduler() _a : Tuple = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : str = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,output_type='numpy' ).images _a : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : Tuple = np.array([0.1564, 0.1_4645, 0.1406, 0.1_4715, 0.1_2425, 0.1_4045, 0.1_3115, 0.1_2175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
271
0
"""simple docstring""" import pytest from datasets.splits import SplitDict, SplitInfo from datasets.utils.py_utils import asdict @pytest.mark.parametrize( '''split_dict''' , [ SplitDict(), SplitDict({'''train''': SplitInfo(name='''train''' , num_bytes=1337 , num_examples=42 , dataset_name='''my_dataset''' )} ), SplitDict({'''train''': SplitInfo(name='''train''' , num_bytes=1337 , num_examples=42 )} ), SplitDict({'''train''': SplitInfo()} ), ] , ) def lowercase ( lowerCAmelCase__ : SplitDict ) -> List[Any]: __a = split_dict._to_yaml_list() assert len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ) __a = SplitDict._from_yaml_list(lowerCAmelCase__ ) for split_name, split_info in split_dict.items(): # dataset_name field is deprecated, and is therefore not part of the YAML dump __a = None # the split name of split_dict takes over the name of the split info object __a = split_name assert split_dict == reloaded @pytest.mark.parametrize( '''split_info''' , [SplitInfo(), SplitInfo(dataset_name=lowerCAmelCase__ ), SplitInfo(dataset_name='''my_dataset''' )] ) def lowercase ( lowerCAmelCase__ : int ) -> List[str]: # For backward compatibility, we need asdict(split_dict) to return split info dictrionaries with the "dataset_name" # field even if it's deprecated. This way old versionso of `datasets` can still reload dataset_infos.json files __a = asdict(SplitDict({'''train''': split_info} ) ) assert "dataset_name" in split_dict_asdict["train"] assert split_dict_asdict["train"]["dataset_name"] == split_info.dataset_name
45
'''simple docstring''' import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __lowerCAmelCase = logging.getLogger() @unittest.skip('''Temporarily disable the doc tests.''' ) @require_torch @require_tf @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : str ,_a : Path ,_a : Union[str, None] = None ,_a : Union[List[str], None] = None ,_a : Union[str, List[str], None] = None ,_a : bool = True ,): '''simple docstring''' _a : Optional[int] = [file for file in os.listdir(_a ) if os.path.isfile(os.path.join(_a ,_a ) )] if identifier is not None: _a : List[str] = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(_a ,_a ): for n_ in n_identifier: _a : Tuple = [file for file in files if n_ not in file] else: _a : Optional[Any] = [file for file in files if n_identifier not in file] _a : List[str] = ignore_files or [] ignore_files.append('__init__.py' ) _a : Tuple = [file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' ,_a ) if only_modules: _a : Any = file.split('.' )[0] try: _a : List[str] = getattr(_a ,_a ) _a : int = doctest.DocTestSuite(_a ) _a : Any = unittest.TextTestRunner().run(_a ) self.assertIs(len(result.failures ) ,0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: _a : Union[str, Any] = doctest.testfile(str('..' / directory / file ) ,optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed ,0 ) def __lowercase ( self : Any ): '''simple docstring''' _a : int = Path('src/transformers' ) _a : List[Any] = 'modeling' _a : Optional[Any] = [ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(_a ,identifier=_a ,ignore_files=_a ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Optional[Any] = Path('src/transformers' ) _a : Optional[Any] = 'tokenization' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Dict = Path('src/transformers' ) _a : str = 'configuration' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : Any ): '''simple docstring''' _a : Tuple = Path('src/transformers' ) _a : List[Any] = ['configuration', 'modeling', 'tokenization'] self.analyze_directory(_a ,n_identifier=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = Path('docs/source' ) _a : List[str] = ['favicon.ico'] self.analyze_directory(_a ,ignore_files=_a ,only_modules=_a )
271
0
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, TensorType SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ = { "openai/imagegpt-small": "", "openai/imagegpt-medium": "", "openai/imagegpt-large": "", } class lowercase ( _UpperCAmelCase ): _SCREAMING_SNAKE_CASE = 'imagegpt' _SCREAMING_SNAKE_CASE = ['past_key_values'] _SCREAMING_SNAKE_CASE = { 'hidden_size': 'n_embd', 'max_position_embeddings': 'n_positions', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self , lowercase=512 + 1 , lowercase=32 * 32 , lowercase=512 , lowercase=24 , lowercase=8 , lowercase=None , lowercase="quick_gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=0.1 , lowercase=1e-5 , lowercase=0.02 , lowercase=True , lowercase=True , lowercase=False , lowercase=False , lowercase=False , **lowercase , ) -> Any: lowerCAmelCase = vocab_size lowerCAmelCase = n_positions lowerCAmelCase = n_embd lowerCAmelCase = n_layer lowerCAmelCase = n_head lowerCAmelCase = n_inner lowerCAmelCase = activation_function lowerCAmelCase = resid_pdrop lowerCAmelCase = embd_pdrop lowerCAmelCase = attn_pdrop lowerCAmelCase = layer_norm_epsilon lowerCAmelCase = initializer_range lowerCAmelCase = scale_attn_weights lowerCAmelCase = use_cache lowerCAmelCase = scale_attn_by_inverse_layer_idx lowerCAmelCase = reorder_and_upcast_attn lowerCAmelCase = tie_word_embeddings super().__init__(tie_word_embeddings=lowercase , **lowercase ) class lowercase ( _UpperCAmelCase ): @property def _snake_case ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """sequence"""}), ] ) def _snake_case ( self , lowercase , lowercase = 1 , lowercase = -1 , lowercase = False , lowercase = None , lowercase = 3 , lowercase = 32 , lowercase = 32 , ) -> Mapping[str, Any]: lowerCAmelCase = self._generate_dummy_images(lowercase , lowercase , lowercase , lowercase ) lowerCAmelCase = dict(preprocessor(images=lowercase , return_tensors=lowercase ) ) return inputs
46
'''simple docstring''' import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def UpperCAmelCase_ (__a : Optional[Any] , __a : str , __a : Optional[Any]=None ): """simple docstring""" assert torch_layer.weight.shape == weight.shape, f"""{torch_layer} layer.weight does not match""" _a : str = nn.Parameter(__a ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f"""{torch_layer} layer.bias does not match""" _a : Any = nn.Parameter(__a ) def UpperCAmelCase_ (__a : int , __a : Optional[Any] , __a : int ): """simple docstring""" _a : Tuple = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : Dict = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Optional[Any] , __a : Optional[int] , __a : List[str] ): """simple docstring""" _a : Dict = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : str = np.asarray(weights[2] ) _a : int = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Any , __a : Any , __a : Optional[Any] ): """simple docstring""" _a : List[str] = weights[0][0][0] _a : List[Any] = np.asarray(layer_norm_a[0] ) _a : List[str] = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # lsh weights + output _a : List[str] = weights[0][1] if len(__a ) < 4: set_layer_weights_in_torch_lsh(__a , torch_block.attention , __a ) else: set_layer_weights_in_torch_local(__a , torch_block.attention , __a ) # intermediate weighs _a : Optional[Any] = weights[2][0][1][2] # Chunked Feed Forward if len(__a ) == 4: _a : Union[str, Any] = intermediate_weights[2] # layernorm 2 _a : Any = np.asarray(intermediate_weights[0][0] ) _a : List[Any] = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # intermediate dense _a : Any = np.asarray(intermediate_weights[1][0] ) _a : Any = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) # intermediate out _a : Optional[int] = np.asarray(intermediate_weights[4][0] ) _a : int = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Dict , __a : Dict , __a : List[Any] ): """simple docstring""" _a : Optional[int] = torch_model.reformer # word embeds _a : Tuple = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(__a ) , ) if isinstance(weights[3] , __a ): _a : Any = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _a : List[Any] = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f"""{position_embeddings[emb_idx]} emb does not match""" _a : Any = nn.Parameter(torch.tensor(__a ) ) _a : List[str] = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( __a ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _a : Tuple = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(__a , __a , __a ) # output layer norm _a : Optional[Any] = np.asarray(weights[7][0] ) _a : int = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # output embeddings _a : List[str] = np.asarray(weights[9][0] ) _a : int = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Tuple , __a : Optional[Any] , __a : Dict ): """simple docstring""" _a : List[Any] = ReformerConfig.from_json_file(__a ) print(f"""Building PyTorch model from configuration: {config}""" ) _a : int = ReformerModelWithLMHead(__a ) with open(__a , 'rb' ) as f: _a : Optional[Any] = pickle.load(__a )['weights'] set_model_weights_in_torch(__a , __a , config.hidden_size ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , __a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--trax_model_pkl_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained Reformer model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __lowerCAmelCase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
271
0
'''simple docstring''' import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class A__ ( unittest.TestCase ): def A ( self : Dict ) -> int: '''simple docstring''' _SCREAMING_SNAKE_CASE =torch.nn.Linear(10 , 10 ) _SCREAMING_SNAKE_CASE =torch.optim.SGD(model.parameters() , 0.1 ) _SCREAMING_SNAKE_CASE =Accelerator() _SCREAMING_SNAKE_CASE =accelerator.prepare(_a ) try: pickle.loads(pickle.dumps(_a ) ) except Exception as e: self.fail(f"Accelerated optimizer pickling failed with {e}" ) AcceleratorState._reset_state()
47
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Union[str, Any] = VQModel( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=3 ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1000 ,) return CLIPTextModel(_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : Dict = self.dummy_uncond_unet _a : List[Any] = DDIMScheduler() _a : List[Any] = self.dummy_vq_model _a : str = LDMPipeline(unet=_a ,vqvae=_a ,scheduler=_a ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : List[str] = torch.manual_seed(0 ) _a : List[str] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Union[str, Any] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ,return_dict=_a )[0] _a : Tuple = image[0, -3:, -3:, -1] _a : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _a : int = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) _a : Any = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : List[str] = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Dict = ldm(generator=_a ,num_inference_steps=5 ,output_type='numpy' ).images _a : str = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _a : Optional[int] = np.array([0.4399, 0.4_4975, 0.4_6825, 0.474, 0.4359, 0.4581, 0.4_5095, 0.4341, 0.4447] ) _a : int = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
271
0
import numpy # List of input, output pairs SCREAMING_SNAKE_CASE__ : List[str] = ( ((5, 2, 3), 15), ((6, 5, 9), 25), ((11, 12, 13), 41), ((1, 1, 1), 8), ((11, 12, 13), 41), ) SCREAMING_SNAKE_CASE__ : List[Any] = (((515, 22, 13), 555), ((61, 35, 49), 150)) SCREAMING_SNAKE_CASE__ : Optional[int] = [2, 4, 1, 5] SCREAMING_SNAKE_CASE__ : List[Any] = len(train_data) SCREAMING_SNAKE_CASE__ : Any = 0.009 def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE="train" ) -> str: return calculate_hypothesis_value(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) - output( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) def A ( _SCREAMING_SNAKE_CASE ) -> List[str]: lowerCamelCase : Optional[Any] = 0 for i in range(len(_SCREAMING_SNAKE_CASE ) - 1 ): hyp_val += data_input_tuple[i] * parameter_vector[i + 1] hyp_val += parameter_vector[0] return hyp_val def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Optional[int]: if data_set == "train": return train_data[example_no][1] elif data_set == "test": return test_data[example_no][1] return None def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> Union[str, Any]: if data_set == "train": return _hypothesis_value(train_data[example_no][0] ) elif data_set == "test": return _hypothesis_value(test_data[example_no][0] ) return None def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE=m ) -> str: lowerCamelCase : List[Any] = 0 for i in range(_SCREAMING_SNAKE_CASE ): if index == -1: summation_value += _error(_SCREAMING_SNAKE_CASE ) else: summation_value += _error(_SCREAMING_SNAKE_CASE ) * train_data[i][0][index] return summation_value def A ( _SCREAMING_SNAKE_CASE ) -> Tuple: lowerCamelCase : str = summation_of_cost_derivative(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) / m return cost_derivative_value def A ( ) -> str: global parameter_vector # Tune these values to set a tolerance value for predicted output lowerCamelCase : str = 0.000002 lowerCamelCase : List[str] = 0 lowerCamelCase : List[Any] = 0 while True: j += 1 lowerCamelCase : Dict = [0, 0, 0, 0] for i in range(0 ,len(_SCREAMING_SNAKE_CASE ) ): lowerCamelCase : List[Any] = get_cost_derivative(i - 1 ) lowerCamelCase : Dict = ( parameter_vector[i] - LEARNING_RATE * cost_derivative ) if numpy.allclose( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,atol=_SCREAMING_SNAKE_CASE ,rtol=_SCREAMING_SNAKE_CASE ,): break lowerCamelCase : Dict = temp_parameter_vector print(("Number of iterations:", j) ) def A ( ) -> Dict: for i in range(len(_SCREAMING_SNAKE_CASE ) ): print(("Actual output value:", output(_SCREAMING_SNAKE_CASE ,"test" )) ) print(("Hypothesis output:", calculate_hypothesis_value(_SCREAMING_SNAKE_CASE ,"test" )) ) if __name__ == "__main__": run_gradient_descent() print('\nTesting gradient descent for a linear hypothesis function.\n') test_gradient_descent()
48
'''simple docstring''' import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : int ,*_a : Optional[int] ,**_a : str ): '''simple docstring''' warnings.warn( 'The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use BeitImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case :Optional[int] = logging.get_logger(__name__) __snake_case :Union[str, Any] = { '''huggingface/autoformer-tourism-monthly''': '''https://huggingface.co/huggingface/autoformer-tourism-monthly/resolve/main/config.json''', } class _A ( __UpperCAmelCase ): UpperCamelCase__ : Optional[int] = '''autoformer''' UpperCamelCase__ : List[str] = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', '''num_hidden_layers''': '''encoder_layers''', } def __init__( self : Optional[int] , __SCREAMING_SNAKE_CASE : Optional[int] = None , __SCREAMING_SNAKE_CASE : Optional[int] = None , __SCREAMING_SNAKE_CASE : str = "student_t" , __SCREAMING_SNAKE_CASE : str = "nll" , __SCREAMING_SNAKE_CASE : int = 1 , __SCREAMING_SNAKE_CASE : List[int] = [1, 2, 3, 4, 5, 6, 7] , __SCREAMING_SNAKE_CASE : bool = True , __SCREAMING_SNAKE_CASE : int = 0 , __SCREAMING_SNAKE_CASE : int = 0 , __SCREAMING_SNAKE_CASE : int = 0 , __SCREAMING_SNAKE_CASE : int = 0 , __SCREAMING_SNAKE_CASE : Optional[List[int]] = None , __SCREAMING_SNAKE_CASE : Optional[List[int]] = None , __SCREAMING_SNAKE_CASE : int = 64 , __SCREAMING_SNAKE_CASE : int = 2 , __SCREAMING_SNAKE_CASE : int = 2 , __SCREAMING_SNAKE_CASE : int = 2 , __SCREAMING_SNAKE_CASE : int = 2 , __SCREAMING_SNAKE_CASE : int = 32 , __SCREAMING_SNAKE_CASE : int = 32 , __SCREAMING_SNAKE_CASE : str = "gelu" , __SCREAMING_SNAKE_CASE : float = 0.1 , __SCREAMING_SNAKE_CASE : float = 0.1 , __SCREAMING_SNAKE_CASE : float = 0.1 , __SCREAMING_SNAKE_CASE : float = 0.1 , __SCREAMING_SNAKE_CASE : float = 0.1 , __SCREAMING_SNAKE_CASE : int = 100 , __SCREAMING_SNAKE_CASE : float = 0.02 , __SCREAMING_SNAKE_CASE : bool = True , __SCREAMING_SNAKE_CASE : Optional[Any]=True , __SCREAMING_SNAKE_CASE : int = 10 , __SCREAMING_SNAKE_CASE : int = 25 , __SCREAMING_SNAKE_CASE : int = 3 , **__SCREAMING_SNAKE_CASE : Tuple , ): '''simple docstring''' __a = prediction_length __a = context_length if context_length is not None else prediction_length __a = distribution_output __a = loss __a = input_size __a = num_time_features __a = lags_sequence __a = scaling __a = num_dynamic_real_features __a = num_static_real_features __a = num_static_categorical_features if cardinality is not None and num_static_categorical_features > 0: if len(__SCREAMING_SNAKE_CASE) != num_static_categorical_features: raise ValueError( '''The cardinality should be a list of the same length as `num_static_categorical_features`''') __a = cardinality else: __a = [0] if embedding_dimension is not None and num_static_categorical_features > 0: if len(__SCREAMING_SNAKE_CASE) != num_static_categorical_features: raise ValueError( '''The embedding dimension should be a list of the same length as `num_static_categorical_features`''') __a = embedding_dimension else: __a = [min(50 , (cat + 1) // 2) for cat in self.cardinality] __a = num_parallel_samples # Transformer architecture configuration __a = input_size * len(self.lags_sequence) + self._number_of_features __a = d_model __a = encoder_attention_heads __a = decoder_attention_heads __a = encoder_ffn_dim __a = decoder_ffn_dim __a = encoder_layers __a = decoder_layers __a = dropout __a = attention_dropout __a = activation_dropout __a = encoder_layerdrop __a = decoder_layerdrop __a = activation_function __a = init_std __a = use_cache # Autoformer __a = label_length __a = moving_average __a = autocorrelation_factor super().__init__(is_encoder_decoder=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE) @property def _lowerCamelCase ( self : Dict): '''simple docstring''' return ( sum(self.embedding_dimension) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
49
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { """linear""": get_linear_schedule_with_warmup, """cosine""": get_cosine_schedule_with_warmup, """cosine_w_restarts""": get_cosine_with_hard_restarts_schedule_with_warmup, """polynomial""": get_polynomial_decay_schedule_with_warmup, """constant""": get_constant_schedule, """constant_w_warmup""": get_constant_schedule_with_warmup, } class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Optional[int] ,_a : Optional[Any]=None ,_a : Dict=None ,*_a : int ,**_a : str ): '''simple docstring''' super().__init__(*_a ,**_a ) if config is None: assert isinstance(self.model ,_a ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F""" {self.model.__class__}""" ) _a : List[Any] = self.model.config else: _a : Optional[int] = config _a : List[str] = data_args _a : List[Any] = self.config.tgt_vocab_size if isinstance(self.config ,_a ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ' padding..' ) if self.args.label_smoothing == 0: _a : List[str] = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss _a : Tuple = label_smoothed_nll_loss def __lowercase ( self : List[str] ,_a : int ): '''simple docstring''' if self.optimizer is None: _a : Union[str, Any] = ['bias', 'LayerNorm.weight'] _a : Tuple = [ { 'params': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], 'weight_decay': self.args.weight_decay, }, { 'params': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], 'weight_decay': 0.0, }, ] _a : Optional[int] = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: _a : Any = Adafactor _a : Dict = {'scale_parameter': False, 'relative_step': False} else: _a : Union[str, Any] = AdamW _a : str = { 'betas': (self.args.adam_betaa, self.args.adam_betaa), 'eps': self.args.adam_epsilon, } _a : Union[str, Any] = self.args.learning_rate if self.sharded_ddp: _a : str = OSS( params=_a ,optim=_a ,**_a ,) else: _a : Tuple = optimizer_cls(_a ,**_a ) if self.lr_scheduler is None: _a : List[Any] = self._get_lr_scheduler(_a ) else: # ignoring --lr_scheduler logger.warning('scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.' ) def __lowercase ( self : List[Any] ,_a : List[Any] ): '''simple docstring''' _a : str = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": _a : int = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": _a : List[str] = schedule_func(self.optimizer ,num_warmup_steps=self.args.warmup_steps ) else: _a : Optional[int] = schedule_func( self.optimizer ,num_warmup_steps=self.args.warmup_steps ,num_training_steps=_a ) return scheduler def __lowercase ( self : Tuple ): '''simple docstring''' if isinstance(self.train_dataset ,torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size ,distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) ,) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def __lowercase ( self : Dict ,_a : Dict ,_a : Any ,_a : Dict ): '''simple docstring''' if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Union[str, Any] = self.loss_fn(logits.view(-1 ,logits.shape[-1] ) ,labels.view(-1 ) ) else: # compute usual loss via models _a, _a : Union[str, Any] = model(**_a ,labels=_a ,use_cache=_a )[:2] else: # compute label smoothed loss _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Any = torch.nn.functional.log_softmax(_a ,dim=-1 ) _a, _a : List[str] = self.loss_fn(_a ,_a ,self.args.label_smoothing ,ignore_index=self.config.pad_token_id ) return loss, logits def __lowercase ( self : Optional[int] ,_a : Union[str, Any] ,_a : List[Any] ): '''simple docstring''' _a : Optional[int] = inputs.pop('labels' ) _a, _a : int = self._compute_loss(_a ,_a ,_a ) return loss def __lowercase ( self : Optional[Any] ,_a : nn.Module ,_a : Dict[str, Union[torch.Tensor, Any]] ,_a : bool ,_a : Optional[List[str]] = None ,): '''simple docstring''' _a : int = self._prepare_inputs(_a ) _a : Any = { 'max_length': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, 'num_beams': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: _a : int = self.model.generate( inputs['input_ids'] ,attention_mask=inputs['attention_mask'] ,**_a ,) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: _a : int = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) _a : Union[str, Any] = inputs.pop('labels' ) with torch.no_grad(): # compute loss on predict data _a, _a : Optional[int] = self._compute_loss(_a ,_a ,_a ) _a : Optional[Any] = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) _a : Optional[Any] = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: _a : Dict = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) return (loss, logits, labels) def __lowercase ( self : str ,_a : Tuple ,_a : Tuple ): '''simple docstring''' _a : List[Any] = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( 'Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be' F""" padded to `max_length`={max_length}""" ) _a : int = pad_token_id * torch.ones( (tensor.shape[0], max_length) ,dtype=tensor.dtype ,device=tensor.device ) _a : Union[str, Any] = tensor return padded_tensor
271
0
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: if n == 1 or not isinstance(_UpperCAmelCase , _UpperCAmelCase ): return 0 elif n == 2: return 1 else: lowerCamelCase__ : List[str] = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> int: lowerCamelCase__ : List[str] = 0 lowerCamelCase__ : Dict = 2 while digits < n: index += 1 lowerCamelCase__ : Union[str, Any] = len(str(fibonacci(_UpperCAmelCase ) ) ) return index def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 1000 ) -> int: return fibonacci_digits_index(_UpperCAmelCase ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
50
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser __lowerCAmelCase = re.compile(r"""\s+""") def UpperCAmelCase_ (__a : Any ): """simple docstring""" return {"hash": hashlib.mda(re.sub(__a , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : List[str] = [len(__a ) for line in example['content'].splitlines()] return {"line_mean": np.mean(__a ), "line_max": max(__a )} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Union[str, Any] = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase_ (__a : Optional[int] , __a : Any ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def UpperCAmelCase_ (__a : int , __a : Union[str, Any]=5 ): """simple docstring""" _a : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] _a : List[str] = example['content'].splitlines() for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase_ (__a : List[str] , __a : Dict=5 , __a : Tuple=0.05 ): """simple docstring""" _a : Optional[int] = ['unit tests', 'test file', 'configuration file'] _a : int = example['content'].splitlines() _a : int = 0 _a : Dict = 0 # first test for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test _a : int = example['content'].count('\n' ) _a : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" _a : List[str] = ['def ', 'class ', 'for ', 'while '] _a : str = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase_ (__a : int , __a : Any=4 ): """simple docstring""" _a : List[str] = example['content'].splitlines() _a : Dict = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Optional[Any] = tokenizer(example['content'] , truncation=__a )['input_ids'] _a : Optional[int] = len(example['content'] ) / len(__a ) return {"ratio": ratio} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Dict = {} results.update(get_hash(__a ) ) results.update(line_stats(__a ) ) results.update(alpha_stats(__a ) ) results.update(char_token_ratio(__a ) ) results.update(is_autogenerated(__a ) ) results.update(is_config_or_test(__a ) ) results.update(has_no_keywords(__a ) ) results.update(has_few_assignments(__a ) ) return results def UpperCAmelCase_ (__a : Any , __a : Any , __a : str ): """simple docstring""" if not check_uniques(__a , __a ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase_ (__a : Union[str, Any] ): """simple docstring""" with open(__a , 'rb' ) as f_in: with gzip.open(str(__a ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(__a , __a ) os.unlink(__a ) # Settings __lowerCAmelCase = HfArgumentParser(PreprocessingArguments) __lowerCAmelCase = parser.parse_args() if args.num_workers is None: __lowerCAmelCase = multiprocessing.cpu_count() __lowerCAmelCase = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset __lowerCAmelCase = time.time() __lowerCAmelCase = load_dataset(args.dataset_name, split="""train""") print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing __lowerCAmelCase = time.time() __lowerCAmelCase = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes __lowerCAmelCase = set(ds.unique("""hash""")) __lowerCAmelCase = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics __lowerCAmelCase = time.time() __lowerCAmelCase = ds.filter(filter, fn_kwargs={"""uniques""": uniques, """args""": args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: __lowerCAmelCase = time.time() __lowerCAmelCase , __lowerCAmelCase = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file __lowerCAmelCase = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / """duplicate_clusters.json""", """w""") as f: json.dump(duplicate_clusters, f) __lowerCAmelCase = output_dir / """data""" data_dir.mkdir(exist_ok=True) __lowerCAmelCase = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): __lowerCAmelCase = str(data_dir / f'''file-{file_number+1:012}.json''') __lowerCAmelCase = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
271
0
from argparse import ArgumentParser, Namespace from typing import Any, List, Optional from ..pipelines import Pipeline, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand try: from fastapi import Body, FastAPI, HTTPException from fastapi.routing import APIRoute from pydantic import BaseModel from starlette.responses import JSONResponse from uvicorn import run snake_case_ : Union[str, Any] = True except (ImportError, AttributeError): snake_case_ : str = object def A (*__A : Tuple , **__A : List[str] ) -> List[str]: """simple docstring""" pass snake_case_ : List[str] = False snake_case_ : Tuple = logging.get_logger("transformers-cli/serving") def A (__A : Namespace ) -> Optional[int]: """simple docstring""" UpperCAmelCase_ = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) return ServeCommand(__A , args.host , args.port , args.workers ) class __snake_case ( a ): UpperCAmelCase__ : dict class __snake_case ( a ): UpperCAmelCase__ : List[str] UpperCAmelCase__ : Optional[List[int]] class __snake_case ( a ): UpperCAmelCase__ : str class __snake_case ( a ): UpperCAmelCase__ : Any class __snake_case ( a ): @staticmethod def lowerCamelCase ( _snake_case : ArgumentParser): """simple docstring""" UpperCAmelCase_ = parser.add_parser( '''serve''' , help='''CLI tool to run inference requests through REST and GraphQL endpoints.''') serve_parser.add_argument( '''--task''' , type=_snake_case , choices=get_supported_tasks() , help='''The task to run the pipeline on''' , ) serve_parser.add_argument('''--host''' , type=_snake_case , default='''localhost''' , help='''Interface the server will listen on.''') serve_parser.add_argument('''--port''' , type=_snake_case , default=8888 , help='''Port the serving will listen to.''') serve_parser.add_argument('''--workers''' , type=_snake_case , default=1 , help='''Number of http workers''') serve_parser.add_argument('''--model''' , type=_snake_case , help='''Model\'s name or path to stored model.''') serve_parser.add_argument('''--config''' , type=_snake_case , help='''Model\'s config name or path to stored model.''') serve_parser.add_argument('''--tokenizer''' , type=_snake_case , help='''Tokenizer name to use.''') serve_parser.add_argument( '''--device''' , type=_snake_case , default=-1 , help='''Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)''' , ) serve_parser.set_defaults(func=_snake_case) def __init__( self : Tuple , _snake_case : Pipeline , _snake_case : str , _snake_case : int , _snake_case : int): """simple docstring""" UpperCAmelCase_ = pipeline UpperCAmelCase_ = host UpperCAmelCase_ = port UpperCAmelCase_ = workers if not _serve_dependencies_installed: raise RuntimeError( '''Using serve command requires FastAPI and uvicorn. ''' '''Please install transformers with [serving]: pip install "transformers[serving]".''' '''Or install FastAPI and uvicorn separately.''') else: logger.info(F"""Serving model over {host}:{port}""") UpperCAmelCase_ = FastAPI( routes=[ APIRoute( '''/''' , self.model_info , response_model=_snake_case , response_class=_snake_case , methods=['''GET'''] , ), APIRoute( '''/tokenize''' , self.tokenize , response_model=_snake_case , response_class=_snake_case , methods=['''POST'''] , ), APIRoute( '''/detokenize''' , self.detokenize , response_model=_snake_case , response_class=_snake_case , methods=['''POST'''] , ), APIRoute( '''/forward''' , self.forward , response_model=_snake_case , response_class=_snake_case , methods=['''POST'''] , ), ] , timeout=600 , ) def lowerCamelCase ( self : Tuple): """simple docstring""" run(self._app , host=self.host , port=self.port , workers=self.workers) def lowerCamelCase ( self : List[str]): """simple docstring""" return ServeModelInfoResult(infos=vars(self._pipeline.model.config)) def lowerCamelCase ( self : Tuple , _snake_case : str = Body(_snake_case , embed=_snake_case) , _snake_case : bool = Body(_snake_case , embed=_snake_case)): """simple docstring""" try: UpperCAmelCase_ = self._pipeline.tokenizer.tokenize(_snake_case) if return_ids: UpperCAmelCase_ = self._pipeline.tokenizer.convert_tokens_to_ids(_snake_case) return ServeTokenizeResult(tokens=_snake_case , tokens_ids=_snake_case) else: return ServeTokenizeResult(tokens=_snake_case) except Exception as e: raise HTTPException(status_code=500 , detail={'''model''': '''''', '''error''': str(_snake_case)}) def lowerCamelCase ( self : Tuple , _snake_case : List[int] = Body(_snake_case , embed=_snake_case) , _snake_case : bool = Body(_snake_case , embed=_snake_case) , _snake_case : bool = Body(_snake_case , embed=_snake_case) , ): """simple docstring""" try: UpperCAmelCase_ = self._pipeline.tokenizer.decode(_snake_case , _snake_case , _snake_case) return ServeDeTokenizeResult(model='''''' , text=_snake_case) except Exception as e: raise HTTPException(status_code=500 , detail={'''model''': '''''', '''error''': str(_snake_case)}) async def lowerCamelCase ( self : List[Any] , _snake_case : Optional[int]=Body(_snake_case , embed=_snake_case)): """simple docstring""" if len(_snake_case) == 0: return ServeForwardResult(output=[] , attention=[]) try: # Forward through the model UpperCAmelCase_ = self._pipeline(_snake_case) return ServeForwardResult(output=_snake_case) except Exception as e: raise HTTPException(500 , {'''error''': str(_snake_case)})
51
'''simple docstring''' import argparse from typing import List import evaluate import numpy as np import torch from datasets import DatasetDict, load_dataset # New Code # # We'll be using StratifiedKFold for this example from sklearn.model_selection import StratifiedKFold from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to perform Cross Validation, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __lowerCAmelCase = 1_6 __lowerCAmelCase = 3_2 def UpperCAmelCase_ (__a : Accelerator , __a : DatasetDict , __a : List[int] , __a : List[int] , __a : int = 1_6 ): """simple docstring""" _a : Union[str, Any] = AutoTokenizer.from_pretrained('bert-base-cased' ) _a : str = DatasetDict( { 'train': dataset['train'].select(__a ), 'validation': dataset['train'].select(__a ), 'test': dataset['validation'], } ) def tokenize_function(__a : List[Any] ): # max_length=None => use the model max length (it's actually the default) _a : Optional[int] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=__a , max_length=__a ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _a : List[str] = datasets.map( __a , batched=__a , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _a : List[Any] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(__a : int ): # On TPU it's best to pad everything to the same length or training will be very slow. _a : Dict = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _a : Tuple = 1_6 elif accelerator.mixed_precision != "no": _a : List[Any] = 8 else: _a : List[Any] = None return tokenizer.pad( __a , padding='longest' , max_length=__a , pad_to_multiple_of=__a , return_tensors='pt' , ) # Instantiate dataloaders. _a : Any = DataLoader( tokenized_datasets['train'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[int] = DataLoader( tokenized_datasets['validation'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[Any] = DataLoader( tokenized_datasets['test'] , shuffle=__a , collate_fn=__a , batch_size=__a ) return train_dataloader, eval_dataloader, test_dataloader def UpperCAmelCase_ (__a : Any , __a : Union[str, Any] ): """simple docstring""" _a : Dict = [] # Download the dataset _a : Tuple = load_dataset('glue' , 'mrpc' ) # Create our splits _a : Union[str, Any] = StratifiedKFold(n_splits=int(args.num_folds ) ) # Initialize accelerator _a : Any = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _a : Optional[Any] = config['lr'] _a : Optional[int] = int(config['num_epochs'] ) _a : Dict = int(config['seed'] ) _a : Dict = int(config['batch_size'] ) _a : Optional[int] = evaluate.load('glue' , 'mrpc' ) # If the batch size is too big we use gradient accumulation _a : List[Any] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: _a : Any = batch_size // MAX_GPU_BATCH_SIZE _a : List[str] = MAX_GPU_BATCH_SIZE set_seed(__a ) # New Code # # Create our folds: _a : int = kfold.split(np.zeros(datasets['train'].num_rows ) , datasets['train']['label'] ) _a : Any = [] # Iterate over them for i, (train_idxs, valid_idxs) in enumerate(__a ): _a, _a, _a : Optional[Any] = get_fold_dataloaders( __a , __a , __a , __a , ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _a : Dict = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=__a ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _a : List[Any] = model.to(accelerator.device ) # Instantiate optimizer _a : List[str] = AdamW(params=model.parameters() , lr=__a ) # Instantiate scheduler _a : List[Any] = get_linear_schedule_with_warmup( optimizer=__a , num_warmup_steps=1_0_0 , num_training_steps=(len(__a ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _a, _a, _a, _a, _a : Union[str, Any] = accelerator.prepare( __a , __a , __a , __a , __a ) # Now we train the model for epoch in range(__a ): model.train() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _a : Dict = model(**__a ) _a : int = outputs.loss _a : Any = loss / gradient_accumulation_steps accelerator.backward(__a ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Union[str, Any] = model(**__a ) _a : Tuple = outputs.logits.argmax(dim=-1 ) _a, _a : Any = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=__a , references=__a , ) _a : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , __a ) # New Code # # We also run predictions on the test set at the very end _a : Any = [] for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Tuple = model(**__a ) _a : Dict = outputs.logits _a, _a : Optional[int] = accelerator.gather_for_metrics((predictions, batch['labels']) ) fold_predictions.append(predictions.cpu() ) if i == 0: # We need all of the test predictions test_references.append(references.cpu() ) # Use accelerator.print to print only on the main process. test_predictions.append(torch.cat(__a , dim=0 ) ) # We now need to release all our memory and get rid of the current model, optimizer, etc accelerator.free_memory() # New Code # # Finally we check the accuracy of our folded results: _a : Dict = torch.cat(__a , dim=0 ) _a : Any = torch.stack(__a , dim=0 ).sum(dim=0 ).div(int(args.num_folds ) ).argmax(dim=-1 ) _a : str = metric.compute(predictions=__a , references=__a ) accelerator.print('Average test metrics from all folds:' , __a ) def UpperCAmelCase_ (): """simple docstring""" _a : Any = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=__a , default=__a , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) # New Code # parser.add_argument('--num_folds' , type=__a , default=3 , help='The number of splits to perform across the dataset' ) _a : Any = parser.parse_args() _a : int = {'lr': 2e-5, 'num_epochs': 3, 'seed': 4_2, 'batch_size': 1_6} training_function(__a , __a ) if __name__ == "__main__": main()
271
0
from typing import List, Optional, Tuple, Union import torch from ...utils import logging, randn_tensor from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline __lowerCamelCase : str = logging.get_logger(__name__) # pylint: disable=invalid-name class A__ ( __snake_case ): def __init__( self , A_ , A_ ): '''simple docstring''' super().__init__() self.register_modules(unet=A_ , scheduler=A_ ) @torch.no_grad() def __call__( self , A_ = 1 , A_ = 100 , A_ = None , A_ = None , A_ = True , ): '''simple docstring''' if audio_length_in_s is None: UpperCamelCase : str = self.unet.config.sample_size / self.unet.config.sample_rate UpperCamelCase : Optional[Any] = audio_length_in_s * self.unet.config.sample_rate UpperCamelCase : Any = 2 ** len(self.unet.up_blocks ) if sample_size < 3 * down_scale_factor: raise ValueError( F"""{audio_length_in_s} is too small. Make sure it's bigger or equal to""" F""" {3 * down_scale_factor / self.unet.config.sample_rate}.""" ) UpperCamelCase : Union[str, Any] = int(A_ ) if sample_size % down_scale_factor != 0: UpperCamelCase : List[str] = ( (audio_length_in_s * self.unet.config.sample_rate) // down_scale_factor + 1 ) * down_scale_factor logger.info( F"""{audio_length_in_s} is increased to {sample_size / self.unet.config.sample_rate} so that it can be handled""" F""" by the model. It will be cut to {original_sample_size / self.unet.config.sample_rate} after the denoising""" " process." ) UpperCamelCase : Any = int(A_ ) UpperCamelCase : Union[str, Any] = next(iter(self.unet.parameters() ) ).dtype UpperCamelCase : Optional[int] = (batch_size, self.unet.config.in_channels, sample_size) if isinstance(A_ , A_ ) and len(A_ ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(A_ )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) UpperCamelCase : Optional[Any] = randn_tensor(A_ , generator=A_ , device=self.device , dtype=A_ ) # set step values self.scheduler.set_timesteps(A_ , device=audio.device ) UpperCamelCase : Optional[int] = self.scheduler.timesteps.to(A_ ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output UpperCamelCase : Dict = self.unet(A_ , A_ ).sample # 2. compute previous image: x_t -> t_t-1 UpperCamelCase : int = self.scheduler.step(A_ , A_ , A_ ).prev_sample UpperCamelCase : Optional[Any] = audio.clamp(-1 , 1 ).float().cpu().numpy() UpperCamelCase : Dict = audio[:, :, :original_sample_size] if not return_dict: return (audio,) return AudioPipelineOutput(audios=A_ )
52
'''simple docstring''' from __future__ import annotations __lowerCAmelCase = [-1_0, -5, 0, 5, 5.1, 1_1, 1_3, 2_1, 3, 4, -2_1, -1_0, -5, -1, 0] __lowerCAmelCase = [-5, 0, 5, 5.1, 1_1, 1_3, 2_1, -1, 4, -1, -1_0, -5, -1, 0, -1] def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Optional[int] = [] _a : int = len(__a ) for i in range(__a ): _a : float = -1 for j in range(i + 1 , __a ): if arr[i] < arr[j]: _a : Any = arr[j] break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Tuple = [] for i, outer in enumerate(__a ): _a : float = -1 for inner in arr[i + 1 :]: if outer < inner: _a : Dict = inner break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : int = len(__a ) _a : list[float] = [] _a : list[float] = [-1] * arr_size for index in reversed(range(__a ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: _a : Dict = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) __lowerCAmelCase = ( """from __main__ import arr, next_greatest_element_slow, """ """next_greatest_element_fast, next_greatest_element""" ) print( """next_greatest_element_slow():""", timeit("""next_greatest_element_slow(arr)""", setup=setup), ) print( """next_greatest_element_fast():""", timeit("""next_greatest_element_fast(arr)""", setup=setup), ) print( """ next_greatest_element():""", timeit("""next_greatest_element(arr)""", setup=setup), )
271
0
'''simple docstring''' import pytest from datasets.splits import SplitDict, SplitInfo from datasets.utils.py_utils import asdict @pytest.mark.parametrize( 'split_dict' , [ SplitDict(), SplitDict({'train': SplitInfo(name='train' , num_bytes=1337 , num_examples=42 , dataset_name='my_dataset' )} ), SplitDict({'train': SplitInfo(name='train' , num_bytes=1337 , num_examples=42 )} ), SplitDict({'train': SplitInfo()} ), ] , ) def lowercase__ ( __lowercase : SplitDict ) -> int: """simple docstring""" __UpperCamelCase = split_dict._to_yaml_list() assert len(__lowercase ) == len(__lowercase ) __UpperCamelCase = SplitDict._from_yaml_list(__lowercase ) for split_name, split_info in split_dict.items(): # dataset_name field is deprecated, and is therefore not part of the YAML dump __UpperCamelCase = None # the split name of split_dict takes over the name of the split info object __UpperCamelCase = split_name assert split_dict == reloaded @pytest.mark.parametrize( 'split_info' , [SplitInfo(), SplitInfo(dataset_name=__lowercase ), SplitInfo(dataset_name='my_dataset' )] ) def lowercase__ ( __lowercase : Dict ) -> Any: """simple docstring""" __UpperCamelCase = asdict(SplitDict({'train': split_info} ) ) assert "dataset_name" in split_dict_asdict["train"] assert split_dict_asdict["train"]["dataset_name"] == split_info.dataset_name
53
'''simple docstring''' import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home __lowerCAmelCase = HUGGINGFACE_HUB_CACHE __lowerCAmelCase = """config.json""" __lowerCAmelCase = """diffusion_pytorch_model.bin""" __lowerCAmelCase = """diffusion_flax_model.msgpack""" __lowerCAmelCase = """model.onnx""" __lowerCAmelCase = """diffusion_pytorch_model.safetensors""" __lowerCAmelCase = """weights.pb""" __lowerCAmelCase = """https://huggingface.co""" __lowerCAmelCase = default_cache_path __lowerCAmelCase = """diffusers_modules""" __lowerCAmelCase = os.getenv("""HF_MODULES_CACHE""", os.path.join(hf_cache_home, """modules""")) __lowerCAmelCase = ["""fp16""", """non-ema"""] __lowerCAmelCase = """.self_attn"""
271
0
"""simple docstring""" import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" def __init__( self : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[Any]=1_3 , UpperCAmelCase__ : List[str]=7 , UpperCAmelCase__ : Union[str, Any]=True , UpperCAmelCase__ : Optional[int]=True , UpperCAmelCase__ : List[str]=False , UpperCAmelCase__ : Any=True , UpperCAmelCase__ : List[Any]=9_9 , UpperCAmelCase__ : int=3_2 , UpperCAmelCase__ : List[Any]=5 , UpperCAmelCase__ : Optional[int]=4 , UpperCAmelCase__ : Union[str, Any]=6_4 , UpperCAmelCase__ : Dict="gelu" , UpperCAmelCase__ : Tuple=0.1 , UpperCAmelCase__ : Dict=0.1 , UpperCAmelCase__ : List[str]=5_1_2 , UpperCAmelCase__ : Dict=1_6 , UpperCAmelCase__ : Optional[int]=2 , UpperCAmelCase__ : str=0.02 , UpperCAmelCase__ : List[str]=3 , UpperCAmelCase__ : Optional[Any]=4 , UpperCAmelCase__ : str=None , UpperCAmelCase__ : int=2 , UpperCAmelCase__ : str=2 , UpperCAmelCase__ : str=2 , UpperCAmelCase__ : Any=2 , UpperCAmelCase__ : List[str]=4 , UpperCAmelCase__ : Dict=1 , ) -> Optional[int]: __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = seq_length __SCREAMING_SNAKE_CASE = is_training __SCREAMING_SNAKE_CASE = use_input_mask __SCREAMING_SNAKE_CASE = use_token_type_ids __SCREAMING_SNAKE_CASE = use_labels __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = max_position_embeddings __SCREAMING_SNAKE_CASE = type_vocab_size __SCREAMING_SNAKE_CASE = type_sequence_label_size __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = num_labels __SCREAMING_SNAKE_CASE = num_choices __SCREAMING_SNAKE_CASE = scope __SCREAMING_SNAKE_CASE = q_groups __SCREAMING_SNAKE_CASE = k_groups __SCREAMING_SNAKE_CASE = v_groups __SCREAMING_SNAKE_CASE = post_attention_groups __SCREAMING_SNAKE_CASE = intermediate_groups __SCREAMING_SNAKE_CASE = output_groups def UpperCAmelCase_ ( self : str ) -> Any: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __SCREAMING_SNAKE_CASE = None if self.use_input_mask: __SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None if self.use_labels: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_choices ) __SCREAMING_SNAKE_CASE = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase_ ( self : Tuple ) -> Tuple: return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def UpperCAmelCase_ ( self : int , UpperCAmelCase__ : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : str , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : List[str] ) -> int: __SCREAMING_SNAKE_CASE = SqueezeBertModel(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase_ ( self : int , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : int , UpperCAmelCase__ : str , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int] ) -> Dict: __SCREAMING_SNAKE_CASE = SqueezeBertForMaskedLM(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase_ ( self : Optional[Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[Any] ) -> Dict: __SCREAMING_SNAKE_CASE = SqueezeBertForQuestionAnswering(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , start_positions=UpperCAmelCase__ , end_positions=UpperCAmelCase__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase_ ( self : Tuple , UpperCAmelCase__ : int , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Dict ) -> Any: __SCREAMING_SNAKE_CASE = self.num_labels __SCREAMING_SNAKE_CASE = SqueezeBertForSequenceClassification(UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self : Dict , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Union[str, Any] ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = self.num_labels __SCREAMING_SNAKE_CASE = SqueezeBertForTokenClassification(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , labels=UpperCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase_ ( self : Tuple , UpperCAmelCase__ : Any , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : List[str] ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = self.num_choices __SCREAMING_SNAKE_CASE = SqueezeBertForMultipleChoice(config=UpperCAmelCase__ ) model.to(UpperCAmelCase__ ) model.eval() __SCREAMING_SNAKE_CASE = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __SCREAMING_SNAKE_CASE = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __SCREAMING_SNAKE_CASE = model( UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , labels=UpperCAmelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCAmelCase_ ( self : int ) -> Dict: __SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ((__SCREAMING_SNAKE_CASE) , (__SCREAMING_SNAKE_CASE) , (__SCREAMING_SNAKE_CASE) , (__SCREAMING_SNAKE_CASE) , (__SCREAMING_SNAKE_CASE) , (__SCREAMING_SNAKE_CASE)) = config_and_inputs __SCREAMING_SNAKE_CASE = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class UpperCamelCase_ ( UpperCamelCase , UpperCamelCase , unittest.TestCase): """simple docstring""" snake_case__ : Optional[Any] = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) snake_case__ : Tuple = ( { "feature-extraction": SqueezeBertModel, "fill-mask": SqueezeBertForMaskedLM, "question-answering": SqueezeBertForQuestionAnswering, "text-classification": SqueezeBertForSequenceClassification, "token-classification": SqueezeBertForTokenClassification, "zero-shot": SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) snake_case__ : Tuple = False snake_case__ : List[str] = True snake_case__ : Dict = False def UpperCAmelCase_ ( self : int ) -> Optional[int]: __SCREAMING_SNAKE_CASE = SqueezeBertModelTester(self ) __SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=UpperCAmelCase__ , dim=3_7 ) def UpperCAmelCase_ ( self : List[Any] ) -> str: self.config_tester.run_common_tests() def UpperCAmelCase_ ( self : Any ) -> Any: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> int: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Any ) -> int: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Dict ) -> Dict: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Tuple: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*UpperCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] ) -> List[str]: __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*UpperCAmelCase__ ) @slow def UpperCAmelCase_ ( self : Optional[Any] ) -> str: for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __SCREAMING_SNAKE_CASE = SqueezeBertModel.from_pretrained(UpperCAmelCase__ ) self.assertIsNotNone(UpperCAmelCase__ ) @require_sentencepiece @require_tokenizers @require_torch class UpperCamelCase_ ( unittest.TestCase): """simple docstring""" @slow def UpperCAmelCase_ ( self : Optional[Any] ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = SqueezeBertForSequenceClassification.from_pretrained("squeezebert/squeezebert-mnli" ) __SCREAMING_SNAKE_CASE = torch.tensor([[1, 2_9_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 1_3, 1_5_8_8, 2]] ) __SCREAMING_SNAKE_CASE = model(UpperCAmelCase__ )[0] __SCREAMING_SNAKE_CASE = torch.Size((1, 3) ) self.assertEqual(output.shape , UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = torch.tensor([[0.6_401, -0.0_349, -0.6_041]] ) self.assertTrue(torch.allclose(UpperCAmelCase__ , UpperCAmelCase__ , atol=1E-4 ) )
54
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class UpperCAmelCase__ : """simple docstring""" def __init__( self : int ,_a : Any ,_a : Optional[int]=2 ,_a : Optional[Any]=True ,_a : Dict=False ,_a : Dict=10 ,_a : Any=3 ,_a : str=32 * 8 ,_a : Optional[int]=32 * 8 ,_a : int=4 ,_a : str=64 ,): '''simple docstring''' _a : Dict = parent _a : Union[str, Any] = batch_size _a : Tuple = is_training _a : List[str] = use_auxiliary_loss _a : Optional[Any] = num_queries _a : str = num_channels _a : List[str] = min_size _a : int = max_size _a : Optional[int] = num_labels _a : List[str] = hidden_dim _a : int = hidden_dim def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Tuple = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( _a ) _a : Optional[Any] = torch.ones([self.batch_size, self.min_size, self.max_size] ,device=_a ) _a : Union[str, Any] = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] ,device=_a ) > 0.5 ).float() _a : Tuple = (torch.rand((self.batch_size, self.num_labels) ,device=_a ) > 0.5).long() _a : Dict = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : int = MaskaFormerConfig( hidden_size=self.hidden_dim ,) _a : str = self.num_queries _a : Union[str, Any] = self.num_labels _a : Tuple = [1, 1, 1, 1] _a : Dict = self.num_channels _a : str = 64 _a : Tuple = 128 _a : Optional[Any] = self.hidden_dim _a : Union[str, Any] = self.hidden_dim _a : List[Any] = self.hidden_dim return config def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a, _a, _a, _a, _a : Optional[Any] = self.prepare_config_and_inputs() _a : str = {'pixel_values': pixel_values, 'pixel_mask': pixel_mask} return config, inputs_dict def __lowercase ( self : List[str] ,_a : Optional[Any] ,_a : str ): '''simple docstring''' _a : str = output.encoder_hidden_states _a : Any = output.pixel_decoder_hidden_states _a : Optional[Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,config.decoder_layers ) def __lowercase ( self : List[str] ,_a : str ,_a : List[Any] ,_a : Any ,_a : Union[str, Any]=False ): '''simple docstring''' with torch.no_grad(): _a : str = MaskaFormerModel(config=_a ) model.to(_a ) model.eval() _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[Any] = model(_a ,output_hidden_states=_a ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape ,(self.batch_size, self.num_queries, self.hidden_dim) ,) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(_a ,_a ) def __lowercase ( self : Tuple ,_a : List[Any] ,_a : Union[str, Any] ,_a : Tuple ,_a : List[str] ,_a : Any ): '''simple docstring''' _a : int = MaskaFormerForUniversalSegmentation(config=_a ) model.to(_a ) model.eval() def comm_check_on_output(_a : Any ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape ,(self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) ,) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape ,(self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[int] = model(_a ) comm_check_on_output(_a ) _a : List[str] = model( pixel_values=_a ,pixel_mask=_a ,mask_labels=_a ,class_labels=_a ) comm_check_on_output(_a ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape ,torch.Size([1] ) ) @require_torch class UpperCAmelCase__ ( lowercase__ , lowercase__ , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[int] = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () __UpperCAmelCase : Dict = {'''feature-extraction''': MaskaFormerModel} if is_torch_available() else {} __UpperCAmelCase : Dict = False __UpperCAmelCase : Tuple = False __UpperCAmelCase : Dict = False __UpperCAmelCase : List[Any] = False def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Union[str, Any] = MaskaFormerModelTester(self ) _a : Dict = ConfigTester(self ,config_class=_a ,has_text_modality=_a ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : Optional[int] ): '''simple docstring''' _a, _a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*_a ) @unittest.skip(reason='Mask2Former does not use inputs_embeds' ) def __lowercase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not have a get_input_embeddings method' ) def __lowercase ( self : str ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former is not a generative model' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not use token embeddings' ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason='Mask2Former has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def __lowercase ( self : Dict ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass def __lowercase ( self : int ): '''simple docstring''' _a, _a : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Union[str, Any] = model_class(_a ) _a : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a : Optional[Any] = [*signature.parameters.keys()] _a : List[Any] = ['pixel_values'] self.assertListEqual(arg_names[:1] ,_a ) @slow def __lowercase ( self : List[str] ): '''simple docstring''' for model_name in ["facebook/mask2former-swin-small-coco-instance"]: _a : Dict = MaskaFormerModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def __lowercase ( self : List[Any] ): '''simple docstring''' _a : int = (self.model_tester.min_size,) * 2 _a : Any = { 'pixel_values': torch.randn((2, 3, *size) ,device=_a ), 'mask_labels': torch.randn((2, 10, *size) ,device=_a ), 'class_labels': torch.zeros(2 ,10 ,device=_a ).long(), } _a : List[Any] = self.model_tester.get_config() _a : int = MaskaFormerForUniversalSegmentation(_a ).to(_a ) _a : str = model(**_a ) self.assertTrue(outputs.loss is not None ) def __lowercase ( self : List[str] ): '''simple docstring''' _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : int ): '''simple docstring''' _a, _a : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Any = model_class(_a ).to(_a ) _a : Optional[int] = model(**_a ,output_attentions=_a ) self.assertTrue(outputs.attentions is not None ) def __lowercase ( self : Tuple ): '''simple docstring''' if not self.model_tester.is_training: return _a : List[str] = self.all_model_classes[1] _a, _a, _a, _a, _a : List[str] = self.model_tester.prepare_config_and_inputs() _a : Any = model_class(_a ) model.to(_a ) model.train() _a : Union[str, Any] = model(_a ,mask_labels=_a ,class_labels=_a ).loss loss.backward() def __lowercase ( self : int ): '''simple docstring''' _a : int = self.all_model_classes[1] _a, _a, _a, _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs() _a : str = True _a : str = True _a : List[str] = model_class(_a ).to(_a ) model.train() _a : Optional[int] = model(_a ,mask_labels=_a ,class_labels=_a ) _a : Tuple = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _a : str = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() _a : Dict = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _a : List[str] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_a ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __lowerCAmelCase = 1e-4 def UpperCAmelCase_ (): """simple docstring""" _a : int = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_vision @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return "facebook/mask2former-swin-small-coco-instance" @cached_property def __lowercase ( self : Any ): '''simple docstring''' return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def __lowercase ( self : Any ): '''simple docstring''' _a : List[str] = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(_a ) _a : int = self.default_image_processor _a : Tuple = prepare_img() _a : Any = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Union[str, Any] = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[Any] = model(**_a ) _a : List[Any] = torch.tensor( [[-0.2790, -1.0717, -1.1668], [-0.5128, -0.3128, -0.4987], [-0.5832, 0.1971, -0.0197]] ).to(_a ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : str = torch.tensor( [[0.8973, 1.1847, 1.1776], [1.1934, 1.5040, 1.5128], [1.1153, 1.4486, 1.4951]] ).to(_a ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : Any = torch.tensor( [[2.1152, 1.7000, -0.8603], [1.5808, 1.8004, -0.9353], [1.6043, 1.7495, -0.5999]] ).to(_a ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Optional[Any] = self.default_image_processor _a : List[Any] = prepare_img() _a : str = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Any = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[int] = model(**_a ) # masks_queries_logits _a : Dict = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape ,(1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) _a : Dict = [ [-8.7839, -9.0056, -8.8121], [-7.4104, -7.0313, -6.5401], [-6.6105, -6.3427, -6.4675], ] _a : Optional[Any] = torch.tensor(_a ).to(_a ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] ,_a ,atol=_a ) ) # class_queries_logits _a : str = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape ,(1, model.config.num_queries, model.config.num_labels + 1) ) _a : str = torch.tensor( [ [1.8324, -8.0835, -4.1922], [0.8450, -9.0050, -3.6053], [0.3045, -7.7293, -3.0275], ] ).to(_a ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : Any = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Tuple = self.default_image_processor _a : Tuple = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] ,segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] ,return_tensors='pt' ,) _a : str = inputs['pixel_values'].to(_a ) _a : str = [el.to(_a ) for el in inputs['mask_labels']] _a : Dict = [el.to(_a ) for el in inputs['class_labels']] with torch.no_grad(): _a : List[str] = model(**_a ) self.assertTrue(outputs.loss is not None )
271
0
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a_ : int = logging.get_logger(__name__) a_ : List[str] = {"""vocab_file""": """sentencepiece.model"""} a_ : Union[str, Any] = { """vocab_file""": { """google/rembert""": """https://huggingface.co/google/rembert/resolve/main/sentencepiece.model""", }, } a_ : str = { """google/rembert""": 256, } class snake_case ( lowercase ): """simple docstring""" _lowerCamelCase = VOCAB_FILES_NAMES _lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP _lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , UpperCamelCase , UpperCamelCase=False , UpperCamelCase=True , UpperCamelCase=True , UpperCamelCase="[CLS]" , UpperCamelCase="[SEP]" , UpperCamelCase="[UNK]" , UpperCamelCase="[SEP]" , UpperCamelCase="[PAD]" , UpperCamelCase="[CLS]" , UpperCamelCase="[MASK]" , **UpperCamelCase , ): """simple docstring""" super().__init__( do_lower_case=UpperCamelCase , remove_space=UpperCamelCase , keep_accents=UpperCamelCase , bos_token=UpperCamelCase , eos_token=UpperCamelCase , unk_token=UpperCamelCase , sep_token=UpperCamelCase , pad_token=UpperCamelCase , cls_token=UpperCamelCase , mask_token=UpperCamelCase , **UpperCamelCase , ) lowerCamelCase_ = do_lower_case lowerCamelCase_ = remove_space lowerCamelCase_ = keep_accents lowerCamelCase_ = vocab_file lowerCamelCase_ = spm.SentencePieceProcessor() self.sp_model.Load(UpperCamelCase ) @property def snake_case ( self ): """simple docstring""" return len(self.sp_model ) def snake_case ( self ): """simple docstring""" lowerCamelCase_ = {self.convert_ids_to_tokens(UpperCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): """simple docstring""" lowerCamelCase_ = self.__dict__.copy() lowerCamelCase_ = None return state def __setstate__( self , UpperCamelCase ): """simple docstring""" lowerCamelCase_ = d lowerCamelCase_ = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def snake_case ( self , UpperCamelCase , UpperCamelCase=False ): """simple docstring""" lowerCamelCase_ = self.sp_model.EncodeAsPieces(UpperCamelCase ) return pieces def snake_case ( self , UpperCamelCase ): """simple docstring""" return self.sp_model.PieceToId(UpperCamelCase ) def snake_case ( self , UpperCamelCase ): """simple docstring""" return self.sp_model.IdToPiece(UpperCamelCase ) def snake_case ( self , UpperCamelCase ): """simple docstring""" lowerCamelCase_ = self.sp_model.decode_pieces(UpperCamelCase ) return out_string def snake_case ( self , UpperCamelCase , UpperCamelCase = None ): """simple docstring""" lowerCamelCase_ = [self.sep_token_id] lowerCamelCase_ = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def snake_case ( self , UpperCamelCase , UpperCamelCase = None , UpperCamelCase = False ): """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(UpperCamelCase )) + [1] + ([0] * len(UpperCamelCase )) + [1] return [1] + ([0] * len(UpperCamelCase )) + [1] def snake_case ( self , UpperCamelCase , UpperCamelCase = None ): """simple docstring""" lowerCamelCase_ = [self.sep_token_id] lowerCamelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def snake_case ( self , UpperCamelCase , UpperCamelCase = None ): """simple docstring""" if not os.path.isdir(UpperCamelCase ): logger.error("Vocabulary path ({}) should be a directory".format(UpperCamelCase ) ) return lowerCamelCase_ = os.path.join( UpperCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase ): copyfile(self.vocab_file , UpperCamelCase ) return (out_vocab_file,)
55
'''simple docstring''' import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def UpperCAmelCase_ (__a : List[Any] ): """simple docstring""" if ( (cp >= 0x4E_00 and cp <= 0x9F_FF) or (cp >= 0x34_00 and cp <= 0x4D_BF) # or (cp >= 0x2_00_00 and cp <= 0x2_A6_DF) # or (cp >= 0x2_A7_00 and cp <= 0x2_B7_3F) # or (cp >= 0x2_B7_40 and cp <= 0x2_B8_1F) # or (cp >= 0x2_B8_20 and cp <= 0x2_CE_AF) # or (cp >= 0xF9_00 and cp <= 0xFA_FF) or (cp >= 0x2_F8_00 and cp <= 0x2_FA_1F) # ): # return True return False def UpperCAmelCase_ (__a : str ): """simple docstring""" for char in word: _a : Union[str, Any] = ord(__a ) if not _is_chinese_char(__a ): return 0 return 1 def UpperCAmelCase_ (__a : List[str] ): """simple docstring""" _a : Dict = set() for token in tokens: _a : str = len(__a ) > 1 and is_chinese(__a ) if chinese_word: word_set.add(__a ) _a : Optional[Any] = list(__a ) return word_list def UpperCAmelCase_ (__a : List[str] , __a : set() ): """simple docstring""" if not chinese_word_set: return bert_tokens _a : Optional[Any] = max([len(__a ) for w in chinese_word_set] ) _a : Optional[int] = bert_tokens _a, _a : Any = 0, len(__a ) while start < end: _a : Tuple = True if is_chinese(bert_word[start] ): _a : Union[str, Any] = min(end - start , __a ) for i in range(__a , 1 , -1 ): _a : Optional[Any] = ''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): _a : Any = '##' + bert_word[j] _a : Union[str, Any] = start + i _a : int = False break if single_word: start += 1 return bert_word def UpperCAmelCase_ (__a : List[str] , __a : LTP , __a : BertTokenizer ): """simple docstring""" _a : int = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : Union[str, Any] = ltp_tokenizer.seg(lines[i : i + 1_0_0] )[0] _a : Optional[Any] = [get_chinese_word(__a ) for r in res] ltp_res.extend(__a ) assert len(__a ) == len(__a ) _a : str = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : List[str] = bert_tokenizer(lines[i : i + 1_0_0] , add_special_tokens=__a , truncation=__a , max_length=5_1_2 ) bert_res.extend(res['input_ids'] ) assert len(__a ) == len(__a ) _a : List[str] = [] for input_ids, chinese_word in zip(__a , __a ): _a : int = [] for id in input_ids: _a : Optional[int] = bert_tokenizer._convert_id_to_token(__a ) input_tokens.append(__a ) _a : List[str] = add_sub_symbol(__a , __a ) _a : Tuple = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(__a ): if token[:2] == "##": _a : str = token[2:] # save chinese tokens' pos if len(__a ) == 1 and _is_chinese_char(ord(__a ) ): ref_id.append(__a ) ref_ids.append(__a ) assert len(__a ) == len(__a ) return ref_ids def UpperCAmelCase_ (__a : Optional[Any] ): """simple docstring""" with open(args.file_name , 'r' , encoding='utf-8' ) as f: _a : Dict = f.readlines() _a : int = [line.strip() for line in data if len(__a ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' _a : int = LTP(args.ltp ) # faster in GPU device _a : Tuple = BertTokenizer.from_pretrained(args.bert ) _a : int = prepare_ref(__a , __a , __a ) with open(args.save_path , 'w' , encoding='utf-8' ) as f: _a : Optional[Any] = [json.dumps(__a ) + '\n' for ref in ref_ids] f.writelines(__a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser(description="""prepare_chinese_ref""") parser.add_argument( """--file_name""", type=str, default="""./resources/chinese-demo.txt""", help="""file need process, same as training data in lm""", ) parser.add_argument( """--ltp""", type=str, default="""./resources/ltp""", help="""resources for LTP tokenizer, usually a path""" ) parser.add_argument("""--bert""", type=str, default="""./resources/robert""", help="""resources for Bert tokenizer""") parser.add_argument("""--save_path""", type=str, default="""./resources/ref.txt""", help="""path to save res""") __lowerCAmelCase = parser.parse_args() main(args)
271
0
'''simple docstring''' import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def __magic_name__ ( __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' snake_case_ = VideoMAEConfig() set_architecture_configs(__UpperCAmelCase, __UpperCAmelCase ) if "finetuned" not in model_name: snake_case_ = False if "finetuned" in model_name: snake_case_ = '''huggingface/label-files''' if "kinetics" in model_name: snake_case_ = 400 snake_case_ = '''kinetics400-id2label.json''' elif "ssv2" in model_name: snake_case_ = 174 snake_case_ = '''something-something-v2-id2label.json''' else: raise ValueError('''Model name should either contain \'kinetics\' or \'ssv2\' in case it\'s fine-tuned.''' ) snake_case_ = json.load(open(hf_hub_download(__UpperCAmelCase, __UpperCAmelCase, repo_type='''dataset''' ), '''r''' ) ) snake_case_ = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} snake_case_ = idalabel snake_case_ = {v: k for k, v in idalabel.items()} return config def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> List[str]: '''simple docstring''' if "small" in model_name: snake_case_ = 384 snake_case_ = 1536 snake_case_ = 12 snake_case_ = 16 snake_case_ = 12 snake_case_ = 3 snake_case_ = 192 snake_case_ = 768 elif "large" in model_name: snake_case_ = 1024 snake_case_ = 4096 snake_case_ = 24 snake_case_ = 16 snake_case_ = 12 snake_case_ = 8 snake_case_ = 512 snake_case_ = 2048 elif "huge" in model_name: snake_case_ = 1280 snake_case_ = 5120 snake_case_ = 32 snake_case_ = 16 snake_case_ = 12 snake_case_ = 8 snake_case_ = 640 snake_case_ = 2560 elif "base" not in model_name: raise ValueError('''Model name should include either "small", "base", "large", or "huge"''' ) def __magic_name__ ( __UpperCAmelCase ) -> List[Any]: '''simple docstring''' if "encoder." in name: snake_case_ = name.replace('''encoder.''', '''''' ) if "cls_token" in name: snake_case_ = name.replace('''cls_token''', '''videomae.embeddings.cls_token''' ) if "decoder_pos_embed" in name: snake_case_ = name.replace('''decoder_pos_embed''', '''decoder.decoder_pos_embed''' ) if "pos_embed" in name and "decoder" not in name: snake_case_ = name.replace('''pos_embed''', '''videomae.embeddings.position_embeddings''' ) if "patch_embed.proj" in name: snake_case_ = name.replace('''patch_embed.proj''', '''videomae.embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: snake_case_ = name.replace('''patch_embed.norm''', '''videomae.embeddings.norm''' ) if "decoder.blocks" in name: snake_case_ = name.replace('''decoder.blocks''', '''decoder.decoder_layers''' ) if "blocks" in name: snake_case_ = name.replace('''blocks''', '''videomae.encoder.layer''' ) if "attn.proj" in name: snake_case_ = name.replace('''attn.proj''', '''attention.output.dense''' ) if "attn" in name and "bias" not in name: snake_case_ = name.replace('''attn''', '''attention.self''' ) if "attn" in name: snake_case_ = name.replace('''attn''', '''attention.attention''' ) if "norm1" in name: snake_case_ = name.replace('''norm1''', '''layernorm_before''' ) if "norm2" in name: snake_case_ = name.replace('''norm2''', '''layernorm_after''' ) if "mlp.fc1" in name: snake_case_ = name.replace('''mlp.fc1''', '''intermediate.dense''' ) if "mlp.fc2" in name: snake_case_ = name.replace('''mlp.fc2''', '''output.dense''' ) if "decoder_embed" in name: snake_case_ = name.replace('''decoder_embed''', '''decoder.decoder_embed''' ) if "decoder_norm" in name: snake_case_ = name.replace('''decoder_norm''', '''decoder.decoder_norm''' ) if "decoder_pred" in name: snake_case_ = name.replace('''decoder_pred''', '''decoder.decoder_pred''' ) if "norm.weight" in name and "decoder" not in name and "fc" not in name: snake_case_ = name.replace('''norm.weight''', '''videomae.layernorm.weight''' ) if "norm.bias" in name and "decoder" not in name and "fc" not in name: snake_case_ = name.replace('''norm.bias''', '''videomae.layernorm.bias''' ) if "head" in name and "decoder" not in name: snake_case_ = name.replace('''head''', '''classifier''' ) return name def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' for key in orig_state_dict.copy().keys(): snake_case_ = orig_state_dict.pop(__UpperCAmelCase ) if key.startswith('''encoder.''' ): snake_case_ = key.replace('''encoder.''', '''''' ) if "qkv" in key: snake_case_ = key.split('''.''' ) if key.startswith('''decoder.blocks''' ): snake_case_ = config.decoder_hidden_size snake_case_ = int(key_split[2] ) snake_case_ = '''decoder.decoder_layers.''' if "weight" in key: snake_case_ = val[:dim, :] snake_case_ = val[dim : dim * 2, :] snake_case_ = val[-dim:, :] else: snake_case_ = config.hidden_size snake_case_ = int(key_split[1] ) snake_case_ = '''videomae.encoder.layer.''' if "weight" in key: snake_case_ = val[:dim, :] snake_case_ = val[dim : dim * 2, :] snake_case_ = val[-dim:, :] else: snake_case_ = val return orig_state_dict def __magic_name__ ( ) -> List[Any]: '''simple docstring''' snake_case_ = hf_hub_download( repo_id='''hf-internal-testing/spaghetti-video''', filename='''eating_spaghetti.npy''', repo_type='''dataset''' ) snake_case_ = np.load(__UpperCAmelCase ) return list(__UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ) -> Any: '''simple docstring''' snake_case_ = get_videomae_config(__UpperCAmelCase ) if "finetuned" in model_name: snake_case_ = VideoMAEForVideoClassification(__UpperCAmelCase ) else: snake_case_ = VideoMAEForPreTraining(__UpperCAmelCase ) # download original checkpoint, hosted on Google Drive snake_case_ = '''pytorch_model.bin''' gdown.cached_download(__UpperCAmelCase, __UpperCAmelCase, quiet=__UpperCAmelCase ) snake_case_ = torch.load(__UpperCAmelCase, map_location='''cpu''' ) if "model" in files: snake_case_ = files['''model'''] else: snake_case_ = files['''module'''] snake_case_ = convert_state_dict(__UpperCAmelCase, __UpperCAmelCase ) model.load_state_dict(__UpperCAmelCase ) model.eval() # verify model on basic input snake_case_ = VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5], image_std=[0.5, 0.5, 0.5] ) snake_case_ = prepare_video() snake_case_ = image_processor(__UpperCAmelCase, return_tensors='''pt''' ) if "finetuned" not in model_name: snake_case_ = hf_hub_download(repo_id='''hf-internal-testing/bool-masked-pos''', filename='''bool_masked_pos.pt''' ) snake_case_ = torch.load(__UpperCAmelCase ) snake_case_ = model(**__UpperCAmelCase ) snake_case_ = outputs.logits snake_case_ = [ '''videomae-small-finetuned-kinetics''', '''videomae-small-finetuned-ssv2''', # Kinetics-400 checkpoints (short = pretrained only for 800 epochs instead of 1600) '''videomae-base-short''', '''videomae-base-short-finetuned-kinetics''', '''videomae-base''', '''videomae-base-finetuned-kinetics''', '''videomae-large''', '''videomae-large-finetuned-kinetics''', '''videomae-huge-finetuned-kinetics''', # Something-Something-v2 checkpoints (short = pretrained only for 800 epochs instead of 2400) '''videomae-base-short-ssv2''', '''videomae-base-short-finetuned-ssv2''', '''videomae-base-ssv2''', '''videomae-base-finetuned-ssv2''', ] # NOTE: logits were tested with image_mean and image_std equal to [0.5, 0.5, 0.5] and [0.5, 0.5, 0.5] if model_name == "videomae-small-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([-0.9_2_9_1, -0.4_0_6_1, -0.9_3_0_7] ) elif model_name == "videomae-small-finetuned-ssv2": snake_case_ = torch.Size([1, 174] ) snake_case_ = torch.tensor([0.2_6_7_1, -0.4_6_8_9, -0.8_2_3_5] ) elif model_name == "videomae-base": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.7_7_3_9, 0.7_9_6_8, 0.7_0_8_9], [0.6_7_0_1, 0.7_4_8_7, 0.6_2_0_9], [0.4_2_8_7, 0.5_1_5_8, 0.4_7_7_3]] ) elif model_name == "videomae-base-short": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.7_9_9_4, 0.9_6_1_2, 0.8_5_0_8], [0.7_4_0_1, 0.8_9_5_8, 0.8_3_0_2], [0.5_8_6_2, 0.7_4_6_8, 0.7_3_2_5]] ) # we verified the loss both for normalized and unnormalized targets for this one snake_case_ = torch.tensor([0.5_1_4_2] ) if config.norm_pix_loss else torch.tensor([0.6_4_6_9] ) elif model_name == "videomae-large": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.7_1_4_9, 0.7_9_9_7, 0.6_9_6_6], [0.6_7_6_8, 0.7_8_6_9, 0.6_9_4_8], [0.5_1_3_9, 0.6_2_2_1, 0.5_6_0_5]] ) elif model_name == "videomae-large-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([0.0_7_7_1, 0.0_0_1_1, -0.3_6_2_5] ) elif model_name == "videomae-huge-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([0.2_4_3_3, 0.1_6_3_2, -0.4_8_9_4] ) elif model_name == "videomae-base-short-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([0.6_5_8_8, 0.0_9_9_0, -0.2_4_9_3] ) elif model_name == "videomae-base-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([0.3_6_6_9, -0.0_6_8_8, -0.2_4_2_1] ) elif model_name == "videomae-base-short-ssv2": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.4_7_1_2, 0.5_2_9_6, 0.5_7_8_6], [0.2_2_7_8, 0.2_7_2_9, 0.4_0_2_6], [0.0_3_5_2, 0.0_7_3_0, 0.2_5_0_6]] ) elif model_name == "videomae-base-short-finetuned-ssv2": snake_case_ = torch.Size([1, 174] ) snake_case_ = torch.tensor([-0.0_5_3_7, -0.1_5_3_9, -0.3_2_6_6] ) elif model_name == "videomae-base-ssv2": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.8_1_3_1, 0.8_7_2_7, 0.8_5_4_6], [0.7_3_6_6, 0.9_3_7_7, 0.8_8_7_0], [0.5_9_3_5, 0.8_8_7_4, 0.8_5_6_4]] ) elif model_name == "videomae-base-finetuned-ssv2": snake_case_ = torch.Size([1, 174] ) snake_case_ = torch.tensor([0.1_9_6_1, -0.8_3_3_7, -0.6_3_8_9] ) else: raise ValueError(F"Model name not supported. Should be one of {model_names}" ) # verify logits assert logits.shape == expected_shape if "finetuned" in model_name: assert torch.allclose(logits[0, :3], __UpperCAmelCase, atol=1e-4 ) else: print('''Logits:''', logits[0, :3, :3] ) assert torch.allclose(logits[0, :3, :3], __UpperCAmelCase, atol=1e-4 ) print('''Logits ok!''' ) # verify loss, if applicable if model_name == "videomae-base-short": snake_case_ = outputs.loss assert torch.allclose(__UpperCAmelCase, __UpperCAmelCase, atol=1e-4 ) print('''Loss ok!''' ) if pytorch_dump_folder_path is not None: print(F"Saving model and image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__UpperCAmelCase ) model.save_pretrained(__UpperCAmelCase ) if push_to_hub: print('''Pushing to the hub...''' ) model.push_to_hub(__UpperCAmelCase, organization='''nielsr''' ) if __name__ == "__main__": a : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://drive.google.com/u/1/uc?id=1tEhLyskjb755TJ65ptsrafUG2llSwQE1&amp;export=download&amp;confirm=t&amp;uuid=aa3276eb-fb7e-482a-adec-dc7171df14c4', type=str, help=( 'URL of the original PyTorch checkpoint (on Google Drive) you\'d like to convert. Should be a direct' ' download link.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default='/Users/nielsrogge/Documents/VideoMAE/Test', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--model_name', default='videomae-base', type=str, help='Name of the model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) a : Union[str, Any] = parser.parse_args() convert_videomae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
56
'''simple docstring''' import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Tuple ,*_a : List[str] ,**_a : Any ): '''simple docstring''' warnings.warn( 'The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use VideoMAEImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
"""simple docstring""" import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def snake_case ( self ): __lowerCAmelCase = "laion/clap-htsat-unfused" __lowerCAmelCase = tempfile.mkdtemp() def snake_case ( self , **__a ): return RobertaTokenizer.from_pretrained(self.checkpoint , **__a ) def snake_case ( self , **__a ): return ClapFeatureExtractor.from_pretrained(self.checkpoint , **__a ) def snake_case ( self ): shutil.rmtree(self.tmpdirname ) def snake_case ( self ): __lowerCAmelCase = self.get_tokenizer() __lowerCAmelCase = self.get_feature_extractor() __lowerCAmelCase = ClapProcessor(tokenizer=__a , feature_extractor=__a ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , __a ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , __a ) def snake_case ( self ): __lowerCAmelCase = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) __lowerCAmelCase = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) __lowerCAmelCase = self.get_feature_extractor(do_normalize=__a , padding_value=1.0 ) __lowerCAmelCase = ClapProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=__a , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __a ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , __a ) def snake_case ( self ): __lowerCAmelCase = self.get_feature_extractor() __lowerCAmelCase = self.get_tokenizer() __lowerCAmelCase = ClapProcessor(tokenizer=__a , feature_extractor=__a ) __lowerCAmelCase = floats_list((3, 10_00) ) __lowerCAmelCase = feature_extractor(__a , return_tensors="np" ) __lowerCAmelCase = processor(audios=__a , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def snake_case ( self ): __lowerCAmelCase = self.get_feature_extractor() __lowerCAmelCase = self.get_tokenizer() __lowerCAmelCase = ClapProcessor(tokenizer=__a , feature_extractor=__a ) __lowerCAmelCase = "This is a test string" __lowerCAmelCase = processor(text=__a ) __lowerCAmelCase = tokenizer(__a ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def snake_case ( self ): __lowerCAmelCase = self.get_feature_extractor() __lowerCAmelCase = self.get_tokenizer() __lowerCAmelCase = ClapProcessor(tokenizer=__a , feature_extractor=__a ) __lowerCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __lowerCAmelCase = processor.batch_decode(__a ) __lowerCAmelCase = tokenizer.batch_decode(__a ) self.assertListEqual(__a , __a ) def snake_case ( self ): __lowerCAmelCase = self.get_feature_extractor() __lowerCAmelCase = self.get_tokenizer() __lowerCAmelCase = ClapProcessor(tokenizer=__a , feature_extractor=__a ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg="`processor` and `feature_extractor` model input names do not match" , )
57
'''simple docstring''' from __future__ import annotations from random import choice def UpperCAmelCase_ (__a : str ): """simple docstring""" return choice(__a ) def UpperCAmelCase_ (__a : list[int] , __a : int ): """simple docstring""" _a : Dict = random_pivot(__a ) # partition based on pivot # linear time _a : Optional[int] = [e for e in lst if e < pivot] _a : List[str] = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(__a ) == k - 1: return pivot # pivot is in elements bigger than k elif len(__a ) < k - 1: return kth_number(__a , k - len(__a ) - 1 ) # pivot is in elements smaller than k else: return kth_number(__a , __a ) if __name__ == "__main__": import doctest doctest.testmod()
271
0
'''simple docstring''' import re def lowerCamelCase ( __lowerCamelCase : str ) ->bool: _SCREAMING_SNAKE_CASE = re.compile(R"""^(\+91[\-\s]?)?[0]?(91)?[789]\d{9}$""" ) if match := re.search(__lowerCamelCase , __lowerCamelCase ): return match.string == phone return False if __name__ == "__main__": print(indian_phone_validator("""+918827897895"""))
58
'''simple docstring''' class UpperCAmelCase__ : """simple docstring""" def __init__( self : Dict ): '''simple docstring''' _a : Dict = {} def __lowercase ( self : Union[str, Any] ): '''simple docstring''' print(self.vertex ) for i in self.vertex: print(_a ,' -> ' ,' -> '.join([str(_a ) for j in self.vertex[i]] ) ) def __lowercase ( self : Dict ,_a : int ,_a : int ): '''simple docstring''' if from_vertex in self.vertex: self.vertex[from_vertex].append(_a ) else: # else make a new vertex _a : int = [to_vertex] def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Tuple = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(_a ,_a ) def __lowercase ( self : Union[str, Any] ,_a : int ,_a : list ): '''simple docstring''' _a : List[Any] = True print(_a ,end=' ' ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(_a ,_a ) if __name__ == "__main__": __lowerCAmelCase = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print("""DFS:""") g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
271
0
from ...configuration_utils import PretrainedConfig class UpperCAmelCase ( A_ ): A__ : Optional[Any] = "bert-generation" def __init__(self : Dict , snake_case__ : int=5_03_58 , snake_case__ : int=10_24 , snake_case__ : int=24 , snake_case__ : Any=16 , snake_case__ : List[str]=40_96 , snake_case__ : List[Any]="gelu" , snake_case__ : Dict=0.1 , snake_case__ : List[Any]=0.1 , snake_case__ : Dict=5_12 , snake_case__ : Dict=0.02 , snake_case__ : Optional[Any]=1e-12 , snake_case__ : List[Any]=0 , snake_case__ : Dict=2 , snake_case__ : Optional[int]=1 , snake_case__ : str="absolute" , snake_case__ : Any=True , **snake_case__ : str , ) -> Optional[Any]: '''simple docstring''' super().__init__(pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ , **snake_case__ ) snake_case : Dict = vocab_size snake_case : List[str] = hidden_size snake_case : Any = num_hidden_layers snake_case : Any = num_attention_heads snake_case : Tuple = hidden_act snake_case : Optional[Any] = intermediate_size snake_case : int = hidden_dropout_prob snake_case : Optional[int] = attention_probs_dropout_prob snake_case : List[str] = max_position_embeddings snake_case : Dict = initializer_range snake_case : Optional[Any] = layer_norm_eps snake_case : Optional[Any] = position_embedding_type snake_case : Dict = use_cache
59
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = """▁""" __lowerCAmelCase = {"""vocab_file""": """sentencepiece.bpe.model""", """monolingual_vocab_file""": """dict.txt"""} __lowerCAmelCase = { """vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model""", }, """monolingual_vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt""", }, } __lowerCAmelCase = {"""vinai/bartpho-syllable""": 1_0_2_4} class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Optional[Any] = VOCAB_FILES_NAMES __UpperCAmelCase : Dict = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ['''input_ids''', '''attention_mask'''] def __init__( self : str ,_a : str ,_a : Any ,_a : Any="<s>" ,_a : Dict="</s>" ,_a : int="</s>" ,_a : Union[str, Any]="<s>" ,_a : List[Any]="<unk>" ,_a : Optional[Any]="<pad>" ,_a : List[str]="<mask>" ,_a : Optional[Dict[str, Any]] = None ,**_a : int ,): '''simple docstring''' _a : Any = AddedToken(_a ,lstrip=_a ,rstrip=_a ) if isinstance(_a ,_a ) else mask_token _a : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_a ,eos_token=_a ,unk_token=_a ,sep_token=_a ,cls_token=_a ,pad_token=_a ,mask_token=_a ,sp_model_kwargs=self.sp_model_kwargs ,**_a ,) _a : Optional[int] = vocab_file _a : Union[str, Any] = monolingual_vocab_file _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_a ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility _a : Union[str, Any] = {} _a : int = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(_a ) not in self.fairseq_tokens_to_ids: _a : int = cnt cnt += 1 with open(_a ,'r' ,encoding='utf-8' ) as f: for line in f.readlines(): _a : str = line.strip().split()[0] _a : Tuple = len(self.fairseq_tokens_to_ids ) if str(_a ) not in self.fairseq_tokens_to_ids: _a : List[str] = len(self.fairseq_tokens_to_ids ) _a : Tuple = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Union[str, Any] ): '''simple docstring''' _a : int = self.__dict__.copy() _a : str = None _a : Optional[Any] = self.sp_model.serialized_model_proto() return state def __setstate__( self : Tuple ,_a : Tuple ): '''simple docstring''' _a : Tuple = d # for backward compatibility if not hasattr(self ,'sp_model_kwargs' ): _a : List[str] = {} _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def __lowercase ( self : Dict ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _a : Dict = [self.cls_token_id] _a : int = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __lowercase ( self : int ,_a : List[int] ,_a : Optional[List[int]] = None ,_a : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a ,token_ids_a=_a ,already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def __lowercase ( self : Tuple ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' _a : List[str] = [self.sep_token_id] _a : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __lowercase ( self : Dict ): '''simple docstring''' return len(self.fairseq_ids_to_tokens ) def __lowercase ( self : Dict ): '''simple docstring''' _a : List[str] = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __lowercase ( self : Tuple ,_a : str ): '''simple docstring''' return self.sp_model.encode(_a ,out_type=_a ) def __lowercase ( self : Union[str, Any] ,_a : Union[str, Any] ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def __lowercase ( self : Any ,_a : int ): '''simple docstring''' return self.fairseq_ids_to_tokens[index] def __lowercase ( self : Tuple ,_a : Union[str, Any] ): '''simple docstring''' _a : str = ''.join(_a ).replace(_a ,' ' ).strip() return out_string def __lowercase ( self : Union[str, Any] ,_a : str ,_a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(_a ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['monolingual_vocab_file'] ,) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,_a ) elif not os.path.isfile(self.vocab_file ): with open(_a ,'wb' ) as fi: _a : List[Any] = self.sp_model.serialized_model_proto() fi.write(_a ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( _a ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file ,_a ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(_a ,'w' ,encoding='utf-8' ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(F"""{str(_a )} \n""" ) return out_vocab_file, out_monolingual_vocab_file
271
0
"""simple docstring""" from __future__ import annotations import math def _snake_case ( _snake_case : int , _snake_case : int , _snake_case : bool , _snake_case : list[int] , _snake_case : float ): if depth < 0: raise ValueError('''Depth cannot be less than 0''' ) if not scores: raise ValueError('''Scores cannot be empty''' ) if depth == height: return scores[node_index] return ( max( minimax(depth + 1 , node_index * 2 , _snake_case , _snake_case , _snake_case ) , minimax(depth + 1 , node_index * 2 + 1 , _snake_case , _snake_case , _snake_case ) , ) if is_max else min( minimax(depth + 1 , node_index * 2 , _snake_case , _snake_case , _snake_case ) , minimax(depth + 1 , node_index * 2 + 1 , _snake_case , _snake_case , _snake_case ) , ) ) def _snake_case ( ): lowerCAmelCase : Optional[int] = [90, 23, 6, 33, 21, 65, 123, 34423] lowerCAmelCase : Union[str, Any] = math.log(len(_snake_case ) , 2 ) print(f'''Optimal value : {minimax(0 , 0 , _snake_case , _snake_case , _snake_case )}''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
60
'''simple docstring''' import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = None __UpperCAmelCase : List[Any] = None @property def __lowercase ( self : Dict ): '''simple docstring''' return self.feat_extract_tester.prepare_feat_extract_dict() def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_a ,'feature_size' ) ) self.assertTrue(hasattr(_a ,'sampling_rate' ) ) self.assertTrue(hasattr(_a ,'padding_value' ) ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_tester.prepare_inputs_for_common() _a : str = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_a ) == len(_a ) for x, y in zip(_a ,processed_features[input_name] ) ) ) _a : Any = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Union[str, Any] = BatchFeature({input_name: speech_inputs} ,tensor_type='np' ) _a : Union[str, Any] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[int] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def __lowercase ( self : Any ): '''simple docstring''' _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : str = BatchFeature({input_name: speech_inputs} ,tensor_type='pt' ) _a : str = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : str = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : int = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = feat_extract.model_input_names[0] _a : int = BatchFeature({input_name: speech_inputs} ,tensor_type='tf' ) _a : Optional[int] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[Any] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def __lowercase ( self : Dict ,_a : Any=False ): '''simple docstring''' def _inputs_have_equal_length(_a : Tuple ): _a : Tuple = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : Optional[Any] ,_a : Union[str, Any] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : int = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Tuple = BatchFeature({input_name: speech_inputs} ) _a : str = self.feat_extract_tester.seq_length_diff _a : Dict = self.feat_extract_tester.max_seq_length + pad_diff _a : Dict = self.feat_extract_tester.min_seq_length _a : Optional[Any] = self.feat_extract_tester.batch_size _a : Tuple = self.feat_extract_tester.feature_size # test padding for List[int] + numpy _a : int = feat_extract.pad(_a ,padding=_a ) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad(_a ,padding='longest' ) _a : Any = input_a[input_name] _a : Optional[Any] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[-1] ) ) _a : List[str] = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) _a : str = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' )[input_name] _a : int = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,return_tensors='np' ) _a : Optional[int] = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy _a : Tuple = feat_extract.pad(_a ,pad_to_multiple_of=10 ) _a : List[str] = input_a[input_name] _a : str = feat_extract.pad(_a ,padding='longest' ,pad_to_multiple_of=10 ) _a : Tuple = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ) _a : Any = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ,return_tensors='np' ,) _a : Dict = input_a[input_name] self.assertTrue(all(len(_a ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) _a : List[str] = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_a ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] ,(batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct _a : Any = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1E-3 ) def __lowercase ( self : List[Any] ,_a : Optional[int]=False ): '''simple docstring''' def _inputs_have_equal_length(_a : List[str] ): _a : Union[str, Any] = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : List[str] ,_a : List[str] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : str = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Any = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) # truncate to smallest _a : Union[str, Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,truncation=_a ) _a : str = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ) _a : Tuple = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to smallest with np _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ,truncation=_a ,) _a : Any = input_a[input_name] _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ) _a : int = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to middle _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ,return_tensors='np' ,) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ) _a : Tuple = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,return_tensors='np' ) _a : Dict = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' ,truncation=_a )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy _a : Optional[Any] = 12 _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,truncation=_a ,) _a : Tuple = input_a[input_name] _a : str = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,) _a : List[Any] = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of _a : List[Any] = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: _a : Union[str, Any] = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Dict ): '''simple docstring''' self._check_truncation(numpify=_a ) def __lowercase ( self : str ): '''simple docstring''' self._check_truncation(numpify=_a ) @require_torch def __lowercase ( self : Dict ): '''simple docstring''' _a : Any = self.feature_extraction_class(**self.feat_extract_dict ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Optional[int] = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='pt' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : List[str] = self.feature_extraction_class(**self.feat_extract_dict ) _a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : Dict = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : Any = feat_extract.pad(_a ,padding='longest' ,return_tensors='tf' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : str = self.feat_extract_dict _a : List[Any] = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Tuple = [len(_a ) for x in speech_inputs] _a : int = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : str = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual(list(processed.attention_mask.shape ) ,list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() ,_a ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_dict _a : Tuple = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : Dict = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = [len(_a ) for x in speech_inputs] _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Any = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = min(_a ) _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,truncation=_a ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual( list(processed_pad.attention_mask.shape ) ,[processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() ,[max_length for x in speech_inputs] )
271
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging _a = logging.get_logger(__name__) if is_vision_available(): import PIL class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = ["""pixel_values"""] def __init__( self , lowercase_ = True , lowercase_ = None , lowercase_ = PILImageResampling.BICUBIC , lowercase_ = True , lowercase_ = None , lowercase_ = True , lowercase_ = 1 / 255 , lowercase_ = True , lowercase_ = None , lowercase_ = None , lowercase_ = True , **lowercase_ , ): """simple docstring""" super().__init__(**lowercase_ ) UpperCAmelCase_ : Dict = size if size is not None else {"shortest_edge": 224} UpperCAmelCase_ : str = get_size_dict(lowercase_ , default_to_square=lowercase_ ) UpperCAmelCase_ : Dict = crop_size if crop_size is not None else {"height": 224, "width": 224} UpperCAmelCase_ : List[Any] = get_size_dict(lowercase_ , default_to_square=lowercase_ , param_name="crop_size" ) UpperCAmelCase_ : Optional[int] = do_resize UpperCAmelCase_ : int = size UpperCAmelCase_ : Tuple = resample UpperCAmelCase_ : Optional[int] = do_center_crop UpperCAmelCase_ : Union[str, Any] = crop_size UpperCAmelCase_ : Optional[int] = do_rescale UpperCAmelCase_ : Any = rescale_factor UpperCAmelCase_ : Dict = do_normalize UpperCAmelCase_ : Union[str, Any] = image_mean if image_mean is not None else OPENAI_CLIP_MEAN UpperCAmelCase_ : Optional[int] = image_std if image_std is not None else OPENAI_CLIP_STD UpperCAmelCase_ : List[Any] = do_convert_rgb def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ = PILImageResampling.BICUBIC , lowercase_ = None , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Tuple = get_size_dict(lowercase_ , default_to_square=lowercase_ ) if "shortest_edge" not in size: raise ValueError(F"""The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}""" ) UpperCAmelCase_ : Optional[Any] = get_resize_output_image_size(lowercase_ , size=size["shortest_edge"] , default_to_square=lowercase_ ) return resize(lowercase_ , size=lowercase_ , resample=lowercase_ , data_format=lowercase_ , **lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ = None , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Any = get_size_dict(lowercase_ ) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` parameter must contain the keys (height, width). Got {size.keys()}""" ) return center_crop(lowercase_ , size=(size["height"], size["width"]) , data_format=lowercase_ , **lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ = None , **lowercase_ , ): """simple docstring""" return rescale(lowercase_ , scale=lowercase_ , data_format=lowercase_ , **lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ = None , **lowercase_ , ): """simple docstring""" return normalize(lowercase_ , mean=lowercase_ , std=lowercase_ , data_format=lowercase_ , **lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = ChannelDimension.FIRST , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Any = do_resize if do_resize is not None else self.do_resize UpperCAmelCase_ : Optional[Any] = size if size is not None else self.size UpperCAmelCase_ : Any = get_size_dict(lowercase_ , param_name="size" , default_to_square=lowercase_ ) UpperCAmelCase_ : str = resample if resample is not None else self.resample UpperCAmelCase_ : List[Any] = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase_ : Optional[int] = crop_size if crop_size is not None else self.crop_size UpperCAmelCase_ : Union[str, Any] = get_size_dict(lowercase_ , param_name="crop_size" , default_to_square=lowercase_ ) UpperCAmelCase_ : Tuple = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase_ : Union[str, Any] = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase_ : Any = do_normalize if do_normalize is not None else self.do_normalize UpperCAmelCase_ : Union[str, Any] = image_mean if image_mean is not None else self.image_mean UpperCAmelCase_ : Any = image_std if image_std is not None else self.image_std UpperCAmelCase_ : List[str] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb UpperCAmelCase_ : Optional[Any] = make_list_of_images(lowercase_ ) if not valid_images(lowercase_ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: UpperCAmelCase_ : Optional[int] = [convert_to_rgb(lowercase_ ) for image in images] # All transformations expect numpy arrays. UpperCAmelCase_ : Tuple = [to_numpy_array(lowercase_ ) for image in images] if do_resize: UpperCAmelCase_ : str = [self.resize(image=lowercase_ , size=lowercase_ , resample=lowercase_ ) for image in images] if do_center_crop: UpperCAmelCase_ : List[str] = [self.center_crop(image=lowercase_ , size=lowercase_ ) for image in images] if do_rescale: UpperCAmelCase_ : Optional[int] = [self.rescale(image=lowercase_ , scale=lowercase_ ) for image in images] if do_normalize: UpperCAmelCase_ : List[str] = [self.normalize(image=lowercase_ , mean=lowercase_ , std=lowercase_ ) for image in images] UpperCAmelCase_ : Dict = [to_channel_dimension_format(lowercase_ , lowercase_ ) for image in images] UpperCAmelCase_ : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowercase_ , tensor_type=lowercase_ )
61
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import KarrasVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : UNetaDModel __UpperCAmelCase : KarrasVeScheduler def __init__( self : Union[str, Any] ,_a : UNetaDModel ,_a : KarrasVeScheduler ): '''simple docstring''' super().__init__() self.register_modules(unet=_a ,scheduler=_a ) @torch.no_grad() def __call__( self : List[Any] ,_a : int = 1 ,_a : int = 50 ,_a : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,_a : Optional[str] = "pil" ,_a : bool = True ,**_a : List[Any] ,): '''simple docstring''' _a : Any = self.unet.config.sample_size _a : Optional[int] = (batch_size, 3, img_size, img_size) _a : Dict = self.unet # sample x_0 ~ N(0, sigma_0^2 * I) _a : Dict = randn_tensor(_a ,generator=_a ,device=self.device ) * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(_a ) for t in self.progress_bar(self.scheduler.timesteps ): # here sigma_t == t_i from the paper _a : Optional[int] = self.scheduler.schedule[t] _a : List[str] = self.scheduler.schedule[t - 1] if t > 0 else 0 # 1. Select temporarily increased noise level sigma_hat # 2. Add new noise to move from sample_i to sample_hat _a, _a : List[Any] = self.scheduler.add_noise_to_input(_a ,_a ,generator=_a ) # 3. Predict the noise residual given the noise magnitude `sigma_hat` # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_hat / 2) * model((sample_hat + 1) / 2 ,sigma_hat / 2 ).sample # 4. Evaluate dx/dt at sigma_hat # 5. Take Euler step from sigma to sigma_prev _a : Tuple = self.scheduler.step(_a ,_a ,_a ,_a ) if sigma_prev != 0: # 6. Apply 2nd order correction # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_prev / 2) * model((step_output.prev_sample + 1) / 2 ,sigma_prev / 2 ).sample _a : Optional[Any] = self.scheduler.step_correct( _a ,_a ,_a ,_a ,step_output.prev_sample ,step_output['derivative'] ,) _a : Dict = step_output.prev_sample _a : Tuple = (sample / 2 + 0.5).clamp(0 ,1 ) _a : Optional[Any] = sample.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": _a : List[str] = self.numpy_to_pil(_a ) if not return_dict: return (image,) return ImagePipelineOutput(images=_a )
271
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _A = { 'configuration_convbert': ['CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ConvBertConfig', 'ConvBertOnnxConfig'], 'tokenization_convbert': ['ConvBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = ['ConvBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ 'CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'ConvBertForMaskedLM', 'ConvBertForMultipleChoice', 'ConvBertForQuestionAnswering', 'ConvBertForSequenceClassification', 'ConvBertForTokenClassification', 'ConvBertLayer', 'ConvBertModel', 'ConvBertPreTrainedModel', 'load_tf_weights_in_convbert', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ 'TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFConvBertForMaskedLM', 'TFConvBertForMultipleChoice', 'TFConvBertForQuestionAnswering', 'TFConvBertForSequenceClassification', 'TFConvBertForTokenClassification', 'TFConvBertLayer', 'TFConvBertModel', 'TFConvBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_convbert import CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvBertConfig, ConvBertOnnxConfig from .tokenization_convbert import ConvBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_convbert_fast import ConvBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convbert import ( CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvBertForMaskedLM, ConvBertForMultipleChoice, ConvBertForQuestionAnswering, ConvBertForSequenceClassification, ConvBertForTokenClassification, ConvBertLayer, ConvBertModel, ConvBertPreTrainedModel, load_tf_weights_in_convbert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convbert import ( TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertLayer, TFConvBertModel, TFConvBertPreTrainedModel, ) else: import sys _A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
62
'''simple docstring''' import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging __lowerCAmelCase = ( """https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py""" ) __lowerCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def UpperCAmelCase_ (): """simple docstring""" _a : Optional[int] = 'https://pypi.org/pypi/diffusers/json' _a : int = json.loads(request.urlopen(__a ).read() )['releases'].keys() return sorted(__a , key=lambda __a : version.Version(__a ) ) def UpperCAmelCase_ (): """simple docstring""" if HF_MODULES_CACHE in sys.path: return sys.path.append(__a ) os.makedirs(__a , exist_ok=__a ) _a : str = Path(__a ) / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : Union[str, os.PathLike] ): """simple docstring""" init_hf_modules() _a : Dict = Path(__a ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(__a , exist_ok=__a ) _a : Optional[int] = dynamic_module_path / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : str ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : int = f.read() # Imports of the form `import .xxx` _a : Tuple = re.findall('^\s*import\s+\.(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall('^\s*from\s+\.(\S+)\s+import' , __a , flags=re.MULTILINE ) # Unique-ify return list(set(__a ) ) def UpperCAmelCase_ (__a : Any ): """simple docstring""" _a : Optional[int] = False _a : Optional[int] = [module_file] _a : List[str] = [] # Let's recurse through all relative imports while not no_change: _a : str = [] for f in files_to_check: new_imports.extend(get_relative_imports(__a ) ) _a : Union[str, Any] = Path(__a ).parent _a : str = [str(module_path / m ) for m in new_imports] _a : Tuple = [f for f in new_import_files if f not in all_relative_imports] _a : Dict = [f"""{f}.py""" for f in new_import_files] _a : List[str] = len(__a ) == 0 all_relative_imports.extend(__a ) return all_relative_imports def UpperCAmelCase_ (__a : Tuple ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : Dict = f.read() # Imports of the form `import xxx` _a : Optional[int] = re.findall('^\s*import\s+(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall('^\s*from\s+(\S+)\s+import' , __a , flags=re.MULTILINE ) # Only keep the top-level module _a : List[str] = [imp.split('.' )[0] for imp in imports if not imp.startswith('.' )] # Unique-ify and test we got them all _a : Optional[int] = list(set(__a ) ) _a : List[str] = [] for imp in imports: try: importlib.import_module(__a ) except ImportError: missing_packages.append(__a ) if len(__a ) > 0: raise ImportError( 'This modeling file requires the following packages that were not found in your environment: ' f"""{', '.join(__a )}. Run `pip install {' '.join(__a )}`""" ) return get_relative_imports(__a ) def UpperCAmelCase_ (__a : Any , __a : str ): """simple docstring""" _a : Any = module_path.replace(os.path.sep , '.' ) _a : Union[str, Any] = importlib.import_module(__a ) if class_name is None: return find_pipeline_class(__a ) return getattr(__a , __a ) def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" from ..pipelines import DiffusionPipeline _a : List[str] = dict(inspect.getmembers(__a , inspect.isclass ) ) _a : str = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls , __a ) and cls.__module__.split('.' )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( f"""Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:""" f""" {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in""" f""" {loaded_module}.""" ) _a : Any = cls return pipeline_class def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , ): """simple docstring""" _a : str = str(__a ) _a : Optional[Any] = os.path.join(__a , __a ) if os.path.isfile(__a ): _a : Tuple = module_file_or_url _a : Optional[Any] = 'local' elif pretrained_model_name_or_path.count('/' ) == 0: _a : int = get_diffusers_versions() # cut ".dev0" _a : Any = 'v' + '.'.join(__version__.split('.' )[:3] ) # retrieve github version that matches if revision is None: _a : Any = latest_version if latest_version[1:] in available_versions else 'main' logger.info(f"""Defaulting to latest_version: {revision}.""" ) elif revision in available_versions: _a : Any = f"""v{revision}""" elif revision == "main": _a : Optional[int] = revision else: raise ValueError( f"""`custom_revision`: {revision} does not exist. Please make sure to choose one of""" f""" {', '.join(available_versions + ['main'] )}.""" ) # community pipeline on GitHub _a : Tuple = COMMUNITY_PIPELINES_URL.format(revision=__a , pipeline=__a ) try: _a : Any = cached_download( __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = 'git' _a : Any = pretrained_model_name_or_path + '.py' except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise else: try: # Load from URL or cache if already cached _a : Optional[Any] = hf_hub_download( __a , __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = os.path.join('local' , '--'.join(pretrained_model_name_or_path.split('/' ) ) ) except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise # Check we have all the requirements in our environment _a : Optional[int] = check_imports(__a ) # Now we move the module inside our cached dynamic modules. _a : Optional[Any] = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(__a ) _a : Any = Path(__a ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(__a , submodule_path / module_file ) for module_needed in modules_needed: _a : Dict = f"""{module_needed}.py""" shutil.copy(os.path.join(__a , __a ) , submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(__a , __a ): _a : Optional[Any] = use_auth_token elif use_auth_token is True: _a : List[Any] = HfFolder.get_token() else: _a : Dict = None _a : int = model_info(__a , revision=__a , token=__a ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. _a : Optional[int] = submodule_path / commit_hash _a : str = full_submodule + os.path.sep + commit_hash create_dynamic_module(__a ) if not (submodule_path / module_file).exists(): shutil.copy(__a , submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( __a , f"""{module_needed}.py""" , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return os.path.join(__a , __a ) def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[str] = None , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , **__a : str , ): """simple docstring""" _a : Dict = get_cached_module_file( __a , __a , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return get_class_in_module(__a , final_module.replace('.py' , '' ) )
271
0
'''simple docstring''' import argparse import copy def _lowerCamelCase ( lowercase : Dict ) -> Optional[Any]: _a = {} with open(lowercase ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: _a = [] _list.append([line.split()[1], line.split()[2]] ) _a = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: _a = [] _list.append([line.split()[0], line.split()[2]] ) _a = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def _lowerCamelCase ( lowercase : str , lowercase : List[Any] ) -> List[str]: with open(lowercase ) as f: _a = f.read(1 ) _a = start_node _a = [] _a = start_node _a = 0 while visiting not in first_solution: _a = 1_0000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(lowercase ) and k[0] not in first_solution: _a = k[1] _a = k[0] first_solution.append(lowercase ) _a = distance_of_first_solution + int(lowercase ) _a = best_node first_solution.append(lowercase ) _a = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 _a = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 1_0000 ) return first_solution, distance_of_first_solution def _lowerCamelCase ( lowercase : Optional[int] , lowercase : Optional[int] ) -> Any: _a = [] for n in solution[1:-1]: _a = solution.index(lowercase ) for kn in solution[1:-1]: _a = solution.index(lowercase ) if n == kn: continue _a = copy.deepcopy(lowercase ) _a = kn _a = n _a = 0 for k in _tmp[:-1]: _a = _tmp[_tmp.index(lowercase ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: _a = distance + int(i[1] ) _tmp.append(lowercase ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) _a = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda lowercase : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def _lowerCamelCase ( lowercase : Union[str, Any] , lowercase : Optional[Any] , lowercase : str , lowercase : List[Any] , lowercase : int ) -> Dict: _a = 1 _a = first_solution _a = [] _a = distance_of_first_solution _a = solution while count <= iters: _a = find_neighborhood(lowercase , lowercase ) _a = 0 _a = neighborhood[index_of_best_solution] _a = len(lowercase ) - 1 _a = False while not found: _a = 0 while i < len(lowercase ): if best_solution[i] != solution[i]: _a = best_solution[i] _a = solution[i] break _a = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) _a = True _a = best_solution[:-1] _a = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: _a = cost _a = solution else: _a = index_of_best_solution + 1 _a = neighborhood[index_of_best_solution] if len(lowercase ) >= size: tabu_list.pop(0 ) _a = count + 1 return best_solution_ever, best_cost def _lowerCamelCase ( lowercase : Tuple=None ) -> Tuple: _a = generate_neighbours(args.File ) _a , _a = generate_first_solution( args.File , lowercase ) _a , _a = tabu_search( lowercase , lowercase , lowercase , args.Iterations , args.Size , ) print(F'Best solution: {best_sol}, with total distance: {best_cost}.' ) if __name__ == "__main__": lowerCAmelCase_ : int = argparse.ArgumentParser(description='Tabu Search') parser.add_argument( '-f', '--File', type=str, help='Path to the file containing the data', required=True, ) parser.add_argument( '-i', '--Iterations', type=int, help='How many iterations the algorithm should perform', required=True, ) parser.add_argument( '-s', '--Size', type=int, help='Size of the tabu list', required=True ) # Pass the arguments to main method main(parser.parse_args())
63
'''simple docstring''' def UpperCAmelCase_ (__a : list , __a : list , __a : int ): """simple docstring""" _a : Optional[Any] = len(__a ) _a : int = [[0] * n for i in range(__a )] for i in range(__a ): _a : Tuple = y_points[i] for i in range(2 , __a ): for j in range(__a , __a ): _a : Tuple = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
271
0
"""simple docstring""" from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class lowercase( nn.Module ): '''simple docstring''' def __init__( self: Any, a_: int = 16, a_: int = 88, a_: Optional[int] = None, a_: int = 1, a_: float = 0.0, a_: int = 32, a_: Optional[int] = None, a_: bool = False, a_: Optional[int] = None, a_: Optional[int] = None, a_: str = "geglu", a_: Optional[int] = None, ): '''simple docstring''' super().__init__() _snake_case : Optional[int] = nn.ModuleList( [ TransformeraDModel( num_attention_heads=a_, attention_head_dim=a_, in_channels=a_, num_layers=a_, dropout=a_, norm_num_groups=a_, cross_attention_dim=a_, attention_bias=a_, sample_size=a_, num_vector_embeds=a_, activation_fn=a_, num_embeds_ada_norm=a_, ) for _ in range(2 ) ] ) # Variables that can be set by a pipeline: # The ratio of transformer1 to transformer2's output states to be combined during inference _snake_case : Optional[Any] = 0.5 # The shape of `encoder_hidden_states` is expected to be # `(batch_size, condition_lengths[0]+condition_lengths[1], num_features)` _snake_case : str = [77, 257] # Which transformer to use to encode which condition. # E.g. `(1, 0)` means that we'll use `transformers[1](conditions[0])` and `transformers[0](conditions[1])` _snake_case : Any = [1, 0] def UpperCamelCase_ ( self: Any, a_: Optional[int], a_: Optional[Any], a_: int=None, a_: str=None, a_: List[str]=None, a_: bool = True, ): '''simple docstring''' _snake_case : List[str] = hidden_states _snake_case : Tuple = [] _snake_case : Optional[int] = 0 # attention_mask is not used yet for i in range(2 ): # for each of the two transformers, pass the corresponding condition tokens _snake_case : str = encoder_hidden_states[:, tokens_start : tokens_start + self.condition_lengths[i]] _snake_case : Any = self.transformer_index_for_condition[i] _snake_case : int = self.transformers[transformer_index]( a_, encoder_hidden_states=a_, timestep=a_, cross_attention_kwargs=a_, return_dict=a_, )[0] encoded_states.append(encoded_state - input_states ) tokens_start += self.condition_lengths[i] _snake_case : List[Any] = encoded_states[0] * self.mix_ratio + encoded_states[1] * (1 - self.mix_ratio) _snake_case : List[Any] = output_states + input_states if not return_dict: return (output_states,) return TransformeraDModelOutput(sample=a_ )
64
'''simple docstring''' import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = inspect.getfile(accelerate.test_utils ) __UpperCAmelCase : List[str] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_cli.py'''] ) __UpperCAmelCase : Dict = ['''accelerate''', '''launch'''] __UpperCAmelCase : Dict = Path.home() / '''.cache/huggingface/accelerate''' __UpperCAmelCase : Dict = '''default_config.yaml''' __UpperCAmelCase : Optional[Any] = config_folder / config_file __UpperCAmelCase : Dict = config_folder / '''_default_config.yaml''' __UpperCAmelCase : Any = Path('''tests/test_configs''' ) @classmethod def __lowercase ( cls : int ): '''simple docstring''' if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def __lowercase ( cls : List[Any] ): '''simple docstring''' if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Dict = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' for config in sorted(self.test_config_path.glob('**/*.yaml' ) ): with self.subTest(config_file=_a ): execute_subprocess_async( self.base_cmd + ['--config_file', str(_a ), self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[int] ): '''simple docstring''' execute_subprocess_async(['accelerate', 'test'] ,env=os.environ.copy() ) class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = '''test-tpu''' __UpperCAmelCase : Any = '''us-central1-a''' __UpperCAmelCase : List[Any] = '''ls''' __UpperCAmelCase : Any = ['''accelerate''', '''tpu-config'''] __UpperCAmelCase : Dict = '''cd /usr/share''' __UpperCAmelCase : Any = '''tests/test_samples/test_command_file.sh''' __UpperCAmelCase : List[Any] = '''Running gcloud compute tpus tpu-vm ssh''' def __lowercase ( self : Dict ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Any = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--debug'] ,return_stdout=_a ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : int ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : str ): '''simple docstring''' _a : List[str] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--command', 'echo "Hello World"', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo \"Hello World\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Any = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command_file', self.command_file, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Union[str, Any] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command_file', self.command_file, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Any ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--accelerate_version', '12.0.0', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,)
271
0
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def lowerCAmelCase_ ( __A ) -> Union[str, Any]: '''simple docstring''' monkeypatch.setattr("datasets.utils.deprecation_utils._emitted_deprecation_warnings", set() ) @pytest.fixture def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' class A : def __init__(self : Dict , __UpperCAmelCase : List[Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = metric_id class A : __UpperCAmelCase : Union[str, Any] = [MetricMock(UpperCAmelCase_ ) for metric_id in ['accuracy', 'mse', 'precision', 'codeparrot/apps_metric']] def lowercase_ (self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" return self._metrics monkeypatch.setattr("datasets.inspect.huggingface_hub", HfhMock() ) @pytest.mark.parametrize( "func, args", [(load_metric, ("metrics/mse",)), (list_metrics, ()), (inspect_metric, ("metrics/mse", "tmp_path"))] ) def lowerCAmelCase_ ( __A, __A, __A, __A, __A ) -> Optional[int]: '''simple docstring''' if "tmp_path" in args: UpperCAmelCase__ = tuple(arg if arg != "tmp_path" else tmp_path for arg in args ) with pytest.warns(__A, match="https://huggingface.co/docs/evaluate" ): func(*__A )
65
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __lowerCAmelCase = TypeVar("""T""") class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Tuple ,_a : T ): '''simple docstring''' _a : List[str] = data _a : Node[T] | None = None def __str__( self : Dict ): '''simple docstring''' return F"""{self.data}""" class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Optional[int] ): '''simple docstring''' _a : Node[T] | None = None def __iter__( self : str ): '''simple docstring''' _a : Tuple = self.top while node: yield node.data _a : int = node.next def __str__( self : str ): '''simple docstring''' return "->".join([str(_a ) for item in self] ) def __len__( self : Optional[Any] ): '''simple docstring''' return len(tuple(iter(self ) ) ) def __lowercase ( self : str ): '''simple docstring''' return self.top is None def __lowercase ( self : List[Any] ,_a : T ): '''simple docstring''' _a : int = Node(_a ) if not self.is_empty(): _a : Optional[Any] = self.top _a : List[str] = node def __lowercase ( self : Tuple ): '''simple docstring''' if self.is_empty(): raise IndexError('pop from empty stack' ) assert isinstance(self.top ,_a ) _a : List[Any] = self.top _a : int = self.top.next return pop_node.data def __lowercase ( self : List[str] ): '''simple docstring''' if self.is_empty(): raise IndexError('peek from empty stack' ) assert self.top is not None return self.top.data def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = None if __name__ == "__main__": from doctest import testmod testmod()
271
0
"""simple docstring""" import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging __a = logging.get_logger(__name__) __a = {"vocab_file": "vocab.json", "merges_file": "merges.txt"} # See all LED models at https://huggingface.co/models?filter=LED __a = { "vocab_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json", }, "merges_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt", }, "tokenizer_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json", }, } __a = { "allenai/led-base-16384": 1_63_84, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def A_ ( ): '''simple docstring''' snake_case_ :Tuple = ( list(range(ord("""!""" ), ord("""~""" ) + 1 ) ) + list(range(ord("""¡""" ), ord("""¬""" ) + 1 ) ) + list(range(ord("""®""" ), ord("""ÿ""" ) + 1 ) ) ) snake_case_ :List[Any] = bs[:] snake_case_ :int = 0 for b in range(2**8 ): if b not in bs: bs.append(_lowercase ) cs.append(2**8 + n ) n += 1 snake_case_ :List[Any] = [chr(_lowercase ) for n in cs] return dict(zip(_lowercase, _lowercase ) ) def A_ ( _lowercase ): '''simple docstring''' snake_case_ :Any = set() snake_case_ :int = word[0] for char in word[1:]: pairs.add((prev_char, char) ) snake_case_ :Optional[int] = char return pairs class lowerCamelCase ( _lowerCAmelCase ): '''simple docstring''' _A : Dict = VOCAB_FILES_NAMES _A : Tuple = PRETRAINED_VOCAB_FILES_MAP _A : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _A : List[str] = ["""input_ids""", """attention_mask"""] def __init__( self: List[str] , snake_case: Tuple , snake_case: Optional[int] , snake_case: Optional[Any]="replace" , snake_case: Optional[Any]="<s>" , snake_case: Any="</s>" , snake_case: int="</s>" , snake_case: Optional[int]="<s>" , snake_case: int="<unk>" , snake_case: Any="<pad>" , snake_case: str="<mask>" , snake_case: Optional[int]=False , **snake_case: str , ) -> List[Any]: snake_case_ :int = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else bos_token snake_case_ :Union[str, Any] = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else eos_token snake_case_ :Union[str, Any] = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else sep_token snake_case_ :int = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else cls_token snake_case_ :List[str] = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else unk_token snake_case_ :List[str] = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else pad_token # Mask token behave like a normal word, i.e. include the space before it snake_case_ :str = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else mask_token super().__init__( errors=snake_case , bos_token=snake_case , eos_token=snake_case , unk_token=snake_case , sep_token=snake_case , cls_token=snake_case , pad_token=snake_case , mask_token=snake_case , add_prefix_space=snake_case , **snake_case , ) with open(snake_case , encoding="""utf-8""" ) as vocab_handle: snake_case_ :int = json.load(snake_case ) snake_case_ :str = {v: k for k, v in self.encoder.items()} snake_case_ :Dict = errors # how to handle errors in decoding snake_case_ :Any = bytes_to_unicode() snake_case_ :List[str] = {v: k for k, v in self.byte_encoder.items()} with open(snake_case , encoding="""utf-8""" ) as merges_handle: snake_case_ :List[str] = merges_handle.read().split("""\n""" )[1:-1] snake_case_ :str = [tuple(merge.split() ) for merge in bpe_merges] snake_case_ :Optional[int] = dict(zip(snake_case , range(len(snake_case ) ) ) ) snake_case_ :int = {} snake_case_ :Union[str, Any] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions snake_case_ :Union[str, Any] = re.compile(r"""'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+""" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def lowerCAmelCase_ ( self: str ) -> List[str]: return len(self.encoder ) def lowerCAmelCase_ ( self: Optional[int] ) -> int: return dict(self.encoder , **self.added_tokens_encoder ) def lowerCAmelCase_ ( self: Tuple , snake_case: Optional[Any] ) -> Union[str, Any]: if token in self.cache: return self.cache[token] snake_case_ :Optional[Any] = tuple(snake_case ) snake_case_ :Any = get_pairs(snake_case ) if not pairs: return token while True: snake_case_ :Union[str, Any] = min(snake_case , key=lambda snake_case : self.bpe_ranks.get(snake_case , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break snake_case_, snake_case_ :int = bigram snake_case_ :int = [] snake_case_ :str = 0 while i < len(snake_case ): try: snake_case_ :Any = word.index(snake_case , snake_case ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) snake_case_ :Any = j if word[i] == first and i < len(snake_case ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 snake_case_ :Dict = tuple(snake_case ) snake_case_ :Optional[Any] = new_word if len(snake_case ) == 1: break else: snake_case_ :Optional[int] = get_pairs(snake_case ) snake_case_ :Optional[int] = """ """.join(snake_case ) snake_case_ :Optional[int] = word return word def lowerCAmelCase_ ( self: Optional[int] , snake_case: List[str] ) -> str: snake_case_ :List[str] = [] for token in re.findall(self.pat , snake_case ): snake_case_ :str = """""".join( self.byte_encoder[b] for b in token.encode("""utf-8""" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(snake_case ).split(""" """ ) ) return bpe_tokens def lowerCAmelCase_ ( self: List[Any] , snake_case: Tuple ) -> Optional[int]: return self.encoder.get(snake_case , self.encoder.get(self.unk_token ) ) def lowerCAmelCase_ ( self: int , snake_case: List[str] ) -> Tuple: return self.decoder.get(snake_case ) def lowerCAmelCase_ ( self: Dict , snake_case: Dict ) -> str: snake_case_ :Optional[Any] = """""".join(snake_case ) snake_case_ :Dict = bytearray([self.byte_decoder[c] for c in text] ).decode("""utf-8""" , errors=self.errors ) return text def lowerCAmelCase_ ( self: str , snake_case: str , snake_case: Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(snake_case ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return snake_case_ :Dict = os.path.join( snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) snake_case_ :Dict = os.path.join( snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(snake_case , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=snake_case , ensure_ascii=snake_case ) + """\n""" ) snake_case_ :Optional[Any] = 0 with open(snake_case , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda snake_case : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" """ Please check that the tokenizer is not corrupted!""" ) snake_case_ :Optional[int] = token_index writer.write(""" """.join(snake_case ) + """\n""" ) index += 1 return vocab_file, merge_file def lowerCAmelCase_ ( self: Optional[int] , snake_case: List[int] , snake_case: Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] snake_case_ :List[str] = [self.cls_token_id] snake_case_ :List[str] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowerCAmelCase_ ( self: List[Any] , snake_case: List[int] , snake_case: Optional[List[int]] = None , snake_case: bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case , token_ids_a=snake_case , already_has_special_tokens=snake_case ) if token_ids_a is None: return [1] + ([0] * len(snake_case )) + [1] return [1] + ([0] * len(snake_case )) + [1, 1] + ([0] * len(snake_case )) + [1] def lowerCAmelCase_ ( self: Optional[Any] , snake_case: List[int] , snake_case: Optional[List[int]] = None ) -> List[int]: snake_case_ :Optional[Any] = [self.sep_token_id] snake_case_ :Dict = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowerCAmelCase_ ( self: List[Any] , snake_case: Tuple , snake_case: Optional[int]=False , **snake_case: List[Any] ) -> Dict: snake_case_ :List[Any] = kwargs.pop("""add_prefix_space""" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(snake_case ) > 0 and not text[0].isspace()): snake_case_ :Any = """ """ + text return (text, kwargs) def lowerCAmelCase_ ( self: List[str] , snake_case: Union[Dict[str, EncodedInput], BatchEncoding] , snake_case: Optional[int] = None , snake_case: PaddingStrategy = PaddingStrategy.DO_NOT_PAD , snake_case: Optional[int] = None , snake_case: Optional[bool] = None , ) -> dict: snake_case_ :Any = super()._pad( encoded_inputs=snake_case , max_length=snake_case , padding_strategy=snake_case , pad_to_multiple_of=snake_case , return_attention_mask=snake_case , ) # Load from model defaults if return_attention_mask is None: snake_case_ :List[str] = """attention_mask""" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: snake_case_ :str = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. snake_case_ :Tuple = len(encoded_inputs["""global_attention_mask"""] ) != len(snake_case ) if needs_to_be_padded: snake_case_ :Tuple = len(snake_case ) - len(encoded_inputs["""global_attention_mask"""] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` snake_case_ :Union[str, Any] = ( encoded_inputs["""global_attention_mask"""] + [-1] * difference ) elif self.padding_side == "left": snake_case_ :Optional[int] = [-1] * difference + encoded_inputs[ """global_attention_mask""" ] else: raise ValueError("""Invalid padding strategy:""" + str(self.padding_side ) ) return encoded_inputs
66
'''simple docstring''' import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : List[str] ): '''simple docstring''' torch.manual_seed(0 ) _a : int = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : str = self.dummy_uncond_unet _a : int = PNDMScheduler() _a : str = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Any = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ,return_dict=_a )[0] _a : List[Any] = image[0, -3:, -3:, -1] _a : Any = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : List[Any] = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[str] = 'google/ddpm-cifar10-32' _a : str = UNetaDModel.from_pretrained(_a ) _a : Union[str, Any] = PNDMScheduler() _a : Tuple = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : str = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,output_type='numpy' ).images _a : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : Tuple = np.array([0.1564, 0.1_4645, 0.1406, 0.1_4715, 0.1_2425, 0.1_4045, 0.1_3115, 0.1_2175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
271
0
'''simple docstring''' import argparse import os import gluonnlp as nlp import mxnet as mx import numpy as np import torch from gluonnlp.base import get_home_dir from gluonnlp.model.bert import BERTEncoder from gluonnlp.model.utils import _load_vocab from gluonnlp.vocab import Vocab from packaging import version from torch import nn from transformers import BertConfig, BertForMaskedLM, BertModel, RobertaTokenizer from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging if version.parse(nlp.__version__) != version.parse("0.8.3"): raise Exception("requires gluonnlp == 0.8.3") if version.parse(mx.__version__) != version.parse("1.5.0"): raise Exception("requires mxnet == 1.5.0") logging.set_verbosity_info() __UpperCAmelCase =logging.get_logger(__name__) __UpperCAmelCase ="The Nymphenburg Palace is a beautiful palace in Munich!" def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ) -> Optional[Any]: __lowerCamelCase = { '''attention_cell''': '''multi_head''', '''num_layers''': 4, '''units''': 10_24, '''hidden_size''': 7_68, '''max_length''': 5_12, '''num_heads''': 8, '''scaled''': True, '''dropout''': 0.1, '''use_residual''': True, '''embed_size''': 10_24, '''embed_dropout''': 0.1, '''word_embed''': None, '''layer_norm_eps''': 1E-5, '''token_type_vocab_size''': 2, } __lowerCamelCase = bort_4_8_768_1024_hparams # Let's construct the original Bort model here # Taken from official BERT implementation, see: # https://github.com/alexa/bort/blob/master/bort/bort.py __lowerCamelCase = BERTEncoder( attention_cell=predefined_args['''attention_cell'''] , num_layers=predefined_args['''num_layers'''] , units=predefined_args['''units'''] , hidden_size=predefined_args['''hidden_size'''] , max_length=predefined_args['''max_length'''] , num_heads=predefined_args['''num_heads'''] , scaled=predefined_args['''scaled'''] , dropout=predefined_args['''dropout'''] , output_attention=UpperCamelCase__ , output_all_encodings=UpperCamelCase__ , use_residual=predefined_args['''use_residual'''] , activation=predefined_args.get('''activation''' , '''gelu''' ) , layer_norm_eps=predefined_args.get('''layer_norm_eps''' , UpperCamelCase__ ) , ) # Vocab information needs to be fetched first # It's the same as RoBERTa, so RobertaTokenizer can be used later __lowerCamelCase = '''openwebtext_ccnews_stories_books_cased''' # Specify download folder to Gluonnlp's vocab __lowerCamelCase = os.path.join(get_home_dir() , '''models''' ) __lowerCamelCase = _load_vocab(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , cls=UpperCamelCase__ ) __lowerCamelCase = nlp.model.BERTModel( UpperCamelCase__ , len(UpperCamelCase__ ) , units=predefined_args['''units'''] , embed_size=predefined_args['''embed_size'''] , embed_dropout=predefined_args['''embed_dropout'''] , word_embed=predefined_args['''word_embed'''] , use_pooler=UpperCamelCase__ , use_token_type_embed=UpperCamelCase__ , token_type_vocab_size=predefined_args['''token_type_vocab_size'''] , use_classifier=UpperCamelCase__ , use_decoder=UpperCamelCase__ , ) original_bort.load_parameters(UpperCamelCase__ , cast_dtype=UpperCamelCase__ , ignore_extra=UpperCamelCase__ ) __lowerCamelCase = original_bort._collect_params_with_prefix() # Build our config 🤗 __lowerCamelCase = { '''architectures''': ['''BertForMaskedLM'''], '''attention_probs_dropout_prob''': predefined_args['''dropout'''], '''hidden_act''': '''gelu''', '''hidden_dropout_prob''': predefined_args['''dropout'''], '''hidden_size''': predefined_args['''embed_size'''], '''initializer_range''': 0.0_2, '''intermediate_size''': predefined_args['''hidden_size'''], '''layer_norm_eps''': predefined_args['''layer_norm_eps'''], '''max_position_embeddings''': predefined_args['''max_length'''], '''model_type''': '''bort''', '''num_attention_heads''': predefined_args['''num_heads'''], '''num_hidden_layers''': predefined_args['''num_layers'''], '''pad_token_id''': 1, # 2 = BERT, 1 = RoBERTa '''type_vocab_size''': 1, # 2 = BERT, 1 = RoBERTa '''vocab_size''': len(UpperCamelCase__ ), } __lowerCamelCase = BertConfig.from_dict(UpperCamelCase__ ) __lowerCamelCase = BertForMaskedLM(UpperCamelCase__ ) hf_bort_model.eval() # Parameter mapping table (Gluonnlp to Transformers) # * denotes layer index # # | Gluon Parameter | Transformers Parameter # | -------------------------------------------------------------- | ---------------------- # | `encoder.layer_norm.beta` | `bert.embeddings.LayerNorm.bias` # | `encoder.layer_norm.gamma` | `bert.embeddings.LayerNorm.weight` # | `encoder.position_weight` | `bert.embeddings.position_embeddings.weight` # | `word_embed.0.weight` | `bert.embeddings.word_embeddings.weight` # | `encoder.transformer_cells.*.attention_cell.proj_key.bias` | `bert.encoder.layer.*.attention.self.key.bias` # | `encoder.transformer_cells.*.attention_cell.proj_key.weight` | `bert.encoder.layer.*.attention.self.key.weight` # | `encoder.transformer_cells.*.attention_cell.proj_query.bias` | `bert.encoder.layer.*.attention.self.query.bias` # | `encoder.transformer_cells.*.attention_cell.proj_query.weight` | `bert.encoder.layer.*.attention.self.query.weight` # | `encoder.transformer_cells.*.attention_cell.proj_value.bias` | `bert.encoder.layer.*.attention.self.value.bias` # | `encoder.transformer_cells.*.attention_cell.proj_value.weight` | `bert.encoder.layer.*.attention.self.value.weight` # | `encoder.transformer_cells.*.ffn.ffn_2.bias` | `bert.encoder.layer.*.attention.output.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_2.weight` | `bert.encoder.layer.*.attention.output.dense.weight` # | `encoder.transformer_cells.*.layer_norm.beta` | `bert.encoder.layer.*.attention.output.LayerNorm.bias` # | `encoder.transformer_cells.*.layer_norm.gamma` | `bert.encoder.layer.*.attention.output.LayerNorm.weight` # | `encoder.transformer_cells.*.ffn.ffn_1.bias` | `bert.encoder.layer.*.intermediate.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_1.weight` | `bert.encoder.layer.*.intermediate.dense.weight` # | `encoder.transformer_cells.*.ffn.layer_norm.beta` | `bert.encoder.layer.*.output.LayerNorm.bias` # | `encoder.transformer_cells.*.ffn.layer_norm.gamma` | `bert.encoder.layer.*.output.LayerNorm.weight` # | `encoder.transformer_cells.*.proj.bias` | `bert.encoder.layer.*.output.dense.bias` # | `encoder.transformer_cells.*.proj.weight` | `bert.encoder.layer.*.output.dense.weight` # Helper function to convert MXNET Arrays to PyTorch def to_torch(UpperCamelCase__ ) -> nn.Parameter: return nn.Parameter(torch.FloatTensor(mx_array.data().asnumpy() ) ) # Check param shapes and map new HF param back def check_and_map_params(UpperCamelCase__ , UpperCamelCase__ ): __lowerCamelCase = hf_param.shape __lowerCamelCase = to_torch(params[gluon_param] ) __lowerCamelCase = gluon_param.shape assert ( shape_hf == shape_gluon ), f"""The gluon parameter {gluon_param} has shape {shape_gluon}, but expects shape {shape_hf} for Transformers""" return gluon_param __lowerCamelCase = check_and_map_params( hf_bort_model.bert.embeddings.word_embeddings.weight , '''word_embed.0.weight''' ) __lowerCamelCase = check_and_map_params( hf_bort_model.bert.embeddings.position_embeddings.weight , '''encoder.position_weight''' ) __lowerCamelCase = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.bias , '''encoder.layer_norm.beta''' ) __lowerCamelCase = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.weight , '''encoder.layer_norm.gamma''' ) # Inspired by RoBERTa conversion script, we just zero them out (Bort does not use them) __lowerCamelCase = torch.zeros_like( hf_bort_model.bert.embeddings.token_type_embeddings.weight.data ) for i in range(hf_bort_config.num_hidden_layers ): __lowerCamelCase = hf_bort_model.bert.encoder.layer[i] # self attention __lowerCamelCase = layer.attention.self __lowerCamelCase = check_and_map_params( self_attn.key.bias.data , f"""encoder.transformer_cells.{i}.attention_cell.proj_key.bias""" ) __lowerCamelCase = check_and_map_params( self_attn.key.weight.data , f"""encoder.transformer_cells.{i}.attention_cell.proj_key.weight""" ) __lowerCamelCase = check_and_map_params( self_attn.query.bias.data , f"""encoder.transformer_cells.{i}.attention_cell.proj_query.bias""" ) __lowerCamelCase = check_and_map_params( self_attn.query.weight.data , f"""encoder.transformer_cells.{i}.attention_cell.proj_query.weight""" ) __lowerCamelCase = check_and_map_params( self_attn.value.bias.data , f"""encoder.transformer_cells.{i}.attention_cell.proj_value.bias""" ) __lowerCamelCase = check_and_map_params( self_attn.value.weight.data , f"""encoder.transformer_cells.{i}.attention_cell.proj_value.weight""" ) # self attention output __lowerCamelCase = layer.attention.output __lowerCamelCase = check_and_map_params( self_output.dense.bias , f"""encoder.transformer_cells.{i}.proj.bias""" ) __lowerCamelCase = check_and_map_params( self_output.dense.weight , f"""encoder.transformer_cells.{i}.proj.weight""" ) __lowerCamelCase = check_and_map_params( self_output.LayerNorm.bias , f"""encoder.transformer_cells.{i}.layer_norm.beta""" ) __lowerCamelCase = check_and_map_params( self_output.LayerNorm.weight , f"""encoder.transformer_cells.{i}.layer_norm.gamma""" ) # intermediate __lowerCamelCase = layer.intermediate __lowerCamelCase = check_and_map_params( intermediate.dense.bias , f"""encoder.transformer_cells.{i}.ffn.ffn_1.bias""" ) __lowerCamelCase = check_and_map_params( intermediate.dense.weight , f"""encoder.transformer_cells.{i}.ffn.ffn_1.weight""" ) # output __lowerCamelCase = layer.output __lowerCamelCase = check_and_map_params( bert_output.dense.bias , f"""encoder.transformer_cells.{i}.ffn.ffn_2.bias""" ) __lowerCamelCase = check_and_map_params( bert_output.dense.weight , f"""encoder.transformer_cells.{i}.ffn.ffn_2.weight""" ) __lowerCamelCase = check_and_map_params( bert_output.LayerNorm.bias , f"""encoder.transformer_cells.{i}.ffn.layer_norm.beta""" ) __lowerCamelCase = check_and_map_params( bert_output.LayerNorm.weight , f"""encoder.transformer_cells.{i}.ffn.layer_norm.gamma""" ) # Save space and energy 🎄 hf_bort_model.half() # Compare output of both models __lowerCamelCase = RobertaTokenizer.from_pretrained('''roberta-base''' ) __lowerCamelCase = tokenizer.encode_plus(UpperCamelCase__ )['''input_ids'''] # Get gluon output __lowerCamelCase = mx.nd.array([input_ids] ) __lowerCamelCase = original_bort(inputs=UpperCamelCase__ , token_types=[] ) # Get Transformer output (save and reload model again) hf_bort_model.save_pretrained(UpperCamelCase__ ) __lowerCamelCase = BertModel.from_pretrained(UpperCamelCase__ ) hf_bort_model.eval() __lowerCamelCase = tokenizer.encode_plus(UpperCamelCase__ , return_tensors='''pt''' ) __lowerCamelCase = hf_bort_model(**UpperCamelCase__ )[0] __lowerCamelCase = output_gluon[0].asnumpy() __lowerCamelCase = output_hf[0].detach().numpy() __lowerCamelCase = np.max(np.abs(hf_layer - gluon_layer ) ).item() __lowerCamelCase = np.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1E-3 ) if success: print('''✔️ Both model do output the same tensors''' ) else: print('''❌ Both model do **NOT** output the same tensors''' ) print('''Absolute difference is:''' , UpperCamelCase__ ) if __name__ == "__main__": __UpperCAmelCase =argparse.ArgumentParser() # Required parameters parser.add_argument( "--bort_checkpoint_path", default=None, type=str, required=True, help="Path the official Bort params file." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) __UpperCAmelCase =parser.parse_args() convert_bort_checkpoint_to_pytorch(args.bort_checkpoint_path, args.pytorch_dump_folder_path)
67
'''simple docstring''' import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __lowerCAmelCase = logging.getLogger() @unittest.skip('''Temporarily disable the doc tests.''' ) @require_torch @require_tf @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : str ,_a : Path ,_a : Union[str, None] = None ,_a : Union[List[str], None] = None ,_a : Union[str, List[str], None] = None ,_a : bool = True ,): '''simple docstring''' _a : Optional[int] = [file for file in os.listdir(_a ) if os.path.isfile(os.path.join(_a ,_a ) )] if identifier is not None: _a : List[str] = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(_a ,_a ): for n_ in n_identifier: _a : Tuple = [file for file in files if n_ not in file] else: _a : Optional[Any] = [file for file in files if n_identifier not in file] _a : List[str] = ignore_files or [] ignore_files.append('__init__.py' ) _a : Tuple = [file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' ,_a ) if only_modules: _a : Any = file.split('.' )[0] try: _a : List[str] = getattr(_a ,_a ) _a : int = doctest.DocTestSuite(_a ) _a : Any = unittest.TextTestRunner().run(_a ) self.assertIs(len(result.failures ) ,0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: _a : Union[str, Any] = doctest.testfile(str('..' / directory / file ) ,optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed ,0 ) def __lowercase ( self : Any ): '''simple docstring''' _a : int = Path('src/transformers' ) _a : List[Any] = 'modeling' _a : Optional[Any] = [ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(_a ,identifier=_a ,ignore_files=_a ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Optional[Any] = Path('src/transformers' ) _a : Optional[Any] = 'tokenization' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Dict = Path('src/transformers' ) _a : str = 'configuration' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : Any ): '''simple docstring''' _a : Tuple = Path('src/transformers' ) _a : List[Any] = ['configuration', 'modeling', 'tokenization'] self.analyze_directory(_a ,n_identifier=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = Path('docs/source' ) _a : List[str] = ['favicon.ico'] self.analyze_directory(_a ,ignore_files=_a ,only_modules=_a )
271
0
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import SegformerImageProcessor, SwinConfig, UperNetConfig, UperNetForSemanticSegmentation def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: List[str] ) -> Any: '''simple docstring''' A__ = 3_8_4 A__ = 7 if "tiny" in model_name: A__ = 9_6 A__ = (2, 2, 6, 2) A__ = (3, 6, 1_2, 2_4) elif "small" in model_name: A__ = 9_6 A__ = (2, 2, 1_8, 2) A__ = (3, 6, 1_2, 2_4) elif "base" in model_name: A__ = 1_2_8 A__ = (2, 2, 1_8, 2) A__ = (4, 8, 1_6, 3_2) A__ = 1_2 A__ = 5_1_2 elif "large" in model_name: A__ = 1_9_2 A__ = (2, 2, 1_8, 2) A__ = (6, 1_2, 2_4, 4_8) A__ = 1_2 A__ = 7_6_8 # set label information A__ = 1_5_0 A__ = "huggingface/label-files" A__ = "ade20k-id2label.json" A__ = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , repo_type="dataset" ) , "r" ) ) A__ = {int(SCREAMING_SNAKE_CASE_ ): v for k, v in idalabel.items()} A__ = {v: k for k, v in idalabel.items()} A__ = SwinConfig( embed_dim=SCREAMING_SNAKE_CASE_ , depths=SCREAMING_SNAKE_CASE_ , num_heads=SCREAMING_SNAKE_CASE_ , window_size=SCREAMING_SNAKE_CASE_ , out_features=["stage1", "stage2", "stage3", "stage4"] , ) A__ = UperNetConfig( backbone_config=SCREAMING_SNAKE_CASE_ , auxiliary_in_channels=SCREAMING_SNAKE_CASE_ , num_labels=SCREAMING_SNAKE_CASE_ , idalabel=SCREAMING_SNAKE_CASE_ , labelaid=SCREAMING_SNAKE_CASE_ , ) return config def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: List[str] ) -> List[str]: '''simple docstring''' A__ = [] # fmt: off # stem rename_keys.append(("backbone.patch_embed.projection.weight", "backbone.embeddings.patch_embeddings.projection.weight") ) rename_keys.append(("backbone.patch_embed.projection.bias", "backbone.embeddings.patch_embeddings.projection.bias") ) rename_keys.append(("backbone.patch_embed.norm.weight", "backbone.embeddings.norm.weight") ) rename_keys.append(("backbone.patch_embed.norm.bias", "backbone.embeddings.norm.bias") ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F'backbone.stages.{i}.blocks.{j}.norm1.weight', F'backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.weight') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.norm1.bias', F'backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.bias') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_bias_table', F'backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_index', F'backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.weight', F'backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.bias', F'backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.norm2.weight', F'backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.weight') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.norm2.bias', F'backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.bias') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.weight', F'backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.bias', F'backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.ffn.layers.1.weight', F'backbone.encoder.layers.{i}.blocks.{j}.output.dense.weight') ) rename_keys.append((F'backbone.stages.{i}.blocks.{j}.ffn.layers.1.bias', F'backbone.encoder.layers.{i}.blocks.{j}.output.dense.bias') ) if i < 3: rename_keys.append((F'backbone.stages.{i}.downsample.reduction.weight', F'backbone.encoder.layers.{i}.downsample.reduction.weight') ) rename_keys.append((F'backbone.stages.{i}.downsample.norm.weight', F'backbone.encoder.layers.{i}.downsample.norm.weight') ) rename_keys.append((F'backbone.stages.{i}.downsample.norm.bias', F'backbone.encoder.layers.{i}.downsample.norm.bias') ) rename_keys.append((F'backbone.norm{i}.weight', F'backbone.hidden_states_norms.stage{i+1}.weight') ) rename_keys.append((F'backbone.norm{i}.bias', F'backbone.hidden_states_norms.stage{i+1}.bias') ) # decode head rename_keys.extend( [ ("decode_head.conv_seg.weight", "decode_head.classifier.weight"), ("decode_head.conv_seg.bias", "decode_head.classifier.bias"), ("auxiliary_head.conv_seg.weight", "auxiliary_head.classifier.weight"), ("auxiliary_head.conv_seg.bias", "auxiliary_head.classifier.bias"), ] ) # fmt: on return rename_keys def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Optional[int] , SCREAMING_SNAKE_CASE_: Tuple , SCREAMING_SNAKE_CASE_: int ) -> List[Any]: '''simple docstring''' A__ = dct.pop(SCREAMING_SNAKE_CASE_ ) A__ = val def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Any , SCREAMING_SNAKE_CASE_: Dict ) -> Union[str, Any]: '''simple docstring''' A__ = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): A__ = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) A__ = state_dict.pop(F'backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.weight' ) A__ = state_dict.pop(F'backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict A__ = in_proj_weight[:dim, :] A__ = in_proj_bias[: dim] A__ = in_proj_weight[ dim : dim * 2, : ] A__ = in_proj_bias[ dim : dim * 2 ] A__ = in_proj_weight[ -dim :, : ] A__ = in_proj_bias[-dim :] # fmt: on def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Dict ) -> List[str]: '''simple docstring''' A__ , A__ = x.shape A__ = x.reshape(SCREAMING_SNAKE_CASE_ , 4 , in_channel // 4 ) A__ = x[:, [0, 2, 1, 3], :].transpose(1 , 2 ).reshape(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return x def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Optional[Any] ) -> Any: '''simple docstring''' A__ , A__ = x.shape A__ = x.reshape(SCREAMING_SNAKE_CASE_ , in_channel // 4 , 4 ) A__ = x[:, :, [0, 2, 1, 3]].transpose(1 , 2 ).reshape(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return x def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Tuple ) -> int: '''simple docstring''' A__ = x.shape[0] A__ = x.reshape(4 , in_channel // 4 ) A__ = x[[0, 2, 1, 3], :].transpose(0 , 1 ).reshape(SCREAMING_SNAKE_CASE_ ) return x def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: List[str] ) -> Any: '''simple docstring''' A__ = x.shape[0] A__ = x.reshape(in_channel // 4 , 4 ) A__ = x[:, [0, 2, 1, 3]].transpose(0 , 1 ).reshape(SCREAMING_SNAKE_CASE_ ) return x def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: Any , SCREAMING_SNAKE_CASE_: List[Any] , SCREAMING_SNAKE_CASE_: Tuple ) -> str: '''simple docstring''' A__ = { "upernet-swin-tiny": "https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210531_112542-e380ad3e.pth", "upernet-swin-small": "https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210526_192015-ee2fff1c.pth", "upernet-swin-base": "https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K_20210531_125459-429057bf.pth", "upernet-swin-large": "https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k_20220318_091743-9ba68901.pth", } A__ = model_name_to_url[model_name] A__ = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE_ , map_location="cpu" , file_name=SCREAMING_SNAKE_CASE_ )[ "state_dict" ] for name, param in state_dict.items(): print(SCREAMING_SNAKE_CASE_ , param.shape ) A__ = get_upernet_config(SCREAMING_SNAKE_CASE_ ) A__ = UperNetForSemanticSegmentation(SCREAMING_SNAKE_CASE_ ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): A__ = state_dict.pop(SCREAMING_SNAKE_CASE_ ) if "bn" in key: A__ = key.replace("bn" , "batch_norm" ) A__ = val # rename keys A__ = create_rename_keys(SCREAMING_SNAKE_CASE_ ) for src, dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) read_in_q_k_v(SCREAMING_SNAKE_CASE_ , config.backbone_config ) # fix downsample parameters for key, value in state_dict.items(): if "downsample" in key: if "reduction" in key: A__ = reverse_correct_unfold_reduction_order(SCREAMING_SNAKE_CASE_ ) if "norm" in key: A__ = reverse_correct_unfold_norm_order(SCREAMING_SNAKE_CASE_ ) model.load_state_dict(SCREAMING_SNAKE_CASE_ ) # verify on image A__ = "https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg" A__ = Image.open(requests.get(SCREAMING_SNAKE_CASE_ , stream=SCREAMING_SNAKE_CASE_ ).raw ).convert("RGB" ) A__ = SegformerImageProcessor() A__ = processor(SCREAMING_SNAKE_CASE_ , return_tensors="pt" ).pixel_values with torch.no_grad(): A__ = model(SCREAMING_SNAKE_CASE_ ) A__ = outputs.logits print(logits.shape ) print("First values of logits:" , logits[0, 0, :3, :3] ) # assert values if model_name == "upernet-swin-tiny": A__ = torch.tensor( [[-7.5958, -7.5958, -7.4302], [-7.5958, -7.5958, -7.4302], [-7.4797, -7.4797, -7.3068]] ) elif model_name == "upernet-swin-small": A__ = torch.tensor( [[-7.1921, -7.1921, -6.9532], [-7.1921, -7.1921, -6.9532], [-7.0908, -7.0908, -6.8534]] ) elif model_name == "upernet-swin-base": A__ = torch.tensor( [[-6.5851, -6.5851, -6.4330], [-6.5851, -6.5851, -6.4330], [-6.4763, -6.4763, -6.3254]] ) elif model_name == "upernet-swin-large": A__ = torch.tensor( [[-7.5297, -7.5297, -7.3802], [-7.5297, -7.5297, -7.3802], [-7.4044, -7.4044, -7.2586]] ) print("Logits:" , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , SCREAMING_SNAKE_CASE_ , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: print(F'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) print(F'Saving processor to {pytorch_dump_folder_path}' ) processor.save_pretrained(SCREAMING_SNAKE_CASE_ ) if push_to_hub: print(F'Pushing model and processor for {model_name} to hub' ) model.push_to_hub(F'openmmlab/{model_name}' ) processor.push_to_hub(F'openmmlab/{model_name}' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""upernet-swin-tiny""", type=str, choices=[f"""upernet-swin-{size}""" for size in ["""tiny""", """small""", """base""", """large"""]], help="""Name of the Swin + UperNet model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) lowerCAmelCase__ = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
68
'''simple docstring''' import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def UpperCAmelCase_ (__a : Optional[Any] , __a : str , __a : Optional[Any]=None ): """simple docstring""" assert torch_layer.weight.shape == weight.shape, f"""{torch_layer} layer.weight does not match""" _a : str = nn.Parameter(__a ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f"""{torch_layer} layer.bias does not match""" _a : Any = nn.Parameter(__a ) def UpperCAmelCase_ (__a : int , __a : Optional[Any] , __a : int ): """simple docstring""" _a : Tuple = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : Dict = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Optional[Any] , __a : Optional[int] , __a : List[str] ): """simple docstring""" _a : Dict = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : str = np.asarray(weights[2] ) _a : int = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Any , __a : Any , __a : Optional[Any] ): """simple docstring""" _a : List[str] = weights[0][0][0] _a : List[Any] = np.asarray(layer_norm_a[0] ) _a : List[str] = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # lsh weights + output _a : List[str] = weights[0][1] if len(__a ) < 4: set_layer_weights_in_torch_lsh(__a , torch_block.attention , __a ) else: set_layer_weights_in_torch_local(__a , torch_block.attention , __a ) # intermediate weighs _a : Optional[Any] = weights[2][0][1][2] # Chunked Feed Forward if len(__a ) == 4: _a : Union[str, Any] = intermediate_weights[2] # layernorm 2 _a : Any = np.asarray(intermediate_weights[0][0] ) _a : List[Any] = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # intermediate dense _a : Any = np.asarray(intermediate_weights[1][0] ) _a : Any = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) # intermediate out _a : Optional[int] = np.asarray(intermediate_weights[4][0] ) _a : int = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Dict , __a : Dict , __a : List[Any] ): """simple docstring""" _a : Optional[int] = torch_model.reformer # word embeds _a : Tuple = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(__a ) , ) if isinstance(weights[3] , __a ): _a : Any = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _a : List[Any] = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f"""{position_embeddings[emb_idx]} emb does not match""" _a : Any = nn.Parameter(torch.tensor(__a ) ) _a : List[str] = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( __a ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _a : Tuple = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(__a , __a , __a ) # output layer norm _a : Optional[Any] = np.asarray(weights[7][0] ) _a : int = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # output embeddings _a : List[str] = np.asarray(weights[9][0] ) _a : int = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Tuple , __a : Optional[Any] , __a : Dict ): """simple docstring""" _a : List[Any] = ReformerConfig.from_json_file(__a ) print(f"""Building PyTorch model from configuration: {config}""" ) _a : int = ReformerModelWithLMHead(__a ) with open(__a , 'rb' ) as f: _a : Optional[Any] = pickle.load(__a )['weights'] set_model_weights_in_torch(__a , __a , config.hidden_size ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , __a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--trax_model_pkl_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained Reformer model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __lowerCAmelCase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
271
0
"""simple docstring""" import itertools import os from collections import Counter, defaultdict from concurrent.futures import ThreadPoolExecutor, as_completed import numpy as np import datasets from .execute import check_correctness __UpperCamelCase = '''\ @misc{chen2021evaluating, title={Evaluating Large Language Models Trained on Code}, author={Mark Chen and Jerry Tworek and Heewoo Jun and Qiming Yuan \ and Henrique Ponde de Oliveira Pinto and Jared Kaplan and Harri Edwards \ and Yuri Burda and Nicholas Joseph and Greg Brockman and Alex Ray \ and Raul Puri and Gretchen Krueger and Michael Petrov and Heidy Khlaaf \ and Girish Sastry and Pamela Mishkin and Brooke Chan and Scott Gray \ and Nick Ryder and Mikhail Pavlov and Alethea Power and Lukasz Kaiser \ and Mohammad Bavarian and Clemens Winter and Philippe Tillet \ and Felipe Petroski Such and Dave Cummings and Matthias Plappert \ and Fotios Chantzis and Elizabeth Barnes and Ariel Herbert-Voss \ and William Hebgen Guss and Alex Nichol and Alex Paino and Nikolas Tezak \ and Jie Tang and Igor Babuschkin and Suchir Balaji and Shantanu Jain \ and William Saunders and Christopher Hesse and Andrew N. Carr \ and Jan Leike and Josh Achiam and Vedant Misra and Evan Morikawa \ and Alec Radford and Matthew Knight and Miles Brundage and Mira Murati \ and Katie Mayer and Peter Welinder and Bob McGrew and Dario Amodei \ and Sam McCandlish and Ilya Sutskever and Wojciech Zaremba}, year={2021}, eprint={2107.03374}, archivePrefix={arXiv}, primaryClass={cs.LG} } ''' __UpperCamelCase = '''\ This metric implements the evaluation harness for the HumanEval problem solving dataset described in the paper "Evaluating Large Language Models Trained on Code" (https://arxiv.org/abs/2107.03374). ''' __UpperCamelCase = ''' Calculates how good are predictions given some references, using certain scores Args: predictions: list of candidates to evaluate. Each candidates should be a list of strings with several code candidates to solve the problem. references: a list with a test for each prediction. Each test should evaluate the correctness of a code candidate. k: number of code candidates to consider in the evaluation (Default: [1, 10, 100]) num_workers: number of workers used to evaluate the canidate programs (Default: 4). timeout: Returns: pass_at_k: dict with pass rates for each k results: dict with granular results of each unittest Examples: >>> code_eval = datasets.load_metric("code_eval") >>> test_cases = ["assert add(2,3)==5"] >>> candidates = [["def add(a,b): return a*b", "def add(a, b): return a+b"]] >>> pass_at_k, results = code_eval.compute(references=test_cases, predictions=candidates, k=[1, 2]) >>> print(pass_at_k) {\'pass@1\': 0.5, \'pass@2\': 1.0} ''' __UpperCamelCase = ''' ################################################################################ !!!WARNING!!! ################################################################################ The "code_eval" metric executes untrusted model-generated code in Python. Although it is highly unlikely that model-generated code will do something overtly malicious in response to this test suite, model-generated code may act destructively due to a lack of model capability or alignment. Users are strongly encouraged to sandbox this evaluation suite so that it does not perform destructive actions on their host or network. For more information on how OpenAI sandboxes its code, see the paper "Evaluating Large Language Models Trained on Code" (https://arxiv.org/abs/2107.03374). Once you have read this disclaimer and taken appropriate precautions, set the environment variable HF_ALLOW_CODE_EVAL="1". Within Python you can to this with: >>> import os >>> os.environ["HF_ALLOW_CODE_EVAL"] = "1" ################################################################################\ ''' __UpperCamelCase = '''The MIT License Copyright (c) OpenAI (https://openai.com) Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase ( datasets.Metric ): def a_ ( self) -> List[str]: return datasets.MetricInfo( # This is the description that will appear on the metrics page. description=_DESCRIPTION, citation=_CITATION, inputs_description=_KWARGS_DESCRIPTION, features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('string')), 'references': datasets.Value('string'), }), homepage='https://github.com/openai/human-eval', codebase_urls=['https://github.com/openai/human-eval'], reference_urls=['https://github.com/openai/human-eval'], license=_LICENSE, ) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__=[1, 10, 100], lowerCAmelCase__=4, lowerCAmelCase__=3.0) -> Dict: if os.getenv('HF_ALLOW_CODE_EVAL', 0) != "1": raise ValueError(_WARNING) if os.name == "nt": raise NotImplementedError('This metric is currently not supported on Windows.') with ThreadPoolExecutor(max_workers=lowerCAmelCase__) as executor: snake_case_ = [] snake_case_ = Counter() snake_case_ = 0 snake_case_ = defaultdict(lowerCAmelCase__) for task_id, (candidates, test_case) in enumerate(zip(lowerCAmelCase__, lowerCAmelCase__)): for candidate in candidates: snake_case_ = candidate + '\n' + test_case snake_case_ = (test_program, timeout, task_id, completion_id[task_id]) snake_case_ = executor.submit(lowerCAmelCase__, *lowerCAmelCase__) futures.append(lowerCAmelCase__) completion_id[task_id] += 1 n_samples += 1 for future in as_completed(lowerCAmelCase__): snake_case_ = future.result() results[result["task_id"]].append((result['completion_id'], result)) snake_case_ , snake_case_ = [], [] for result in results.values(): result.sort() snake_case_ = [r[1]['passed'] for r in result] total.append(len(lowerCAmelCase__)) correct.append(sum(lowerCAmelCase__)) snake_case_ = np.array(lowerCAmelCase__) snake_case_ = np.array(lowerCAmelCase__) snake_case_ = k snake_case_ = {f'pass@{k}': estimate_pass_at_k(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).mean() for k in ks if (total >= k).all()} return pass_at_k, results def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Any: def estimator(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> float: if n - c < k: return 1.0 return 1.0 - np.prod(1.0 - k / np.arange(n - c + 1 , n + 1 ) ) if isinstance(UpperCAmelCase , UpperCAmelCase ): snake_case_ = itertools.repeat(UpperCAmelCase , len(UpperCAmelCase ) ) else: assert len(UpperCAmelCase ) == len(UpperCAmelCase ) snake_case_ = iter(UpperCAmelCase ) return np.array([estimator(int(UpperCAmelCase ) , int(UpperCAmelCase ) , UpperCAmelCase ) for n, c in zip(UpperCAmelCase , UpperCAmelCase )] )
69
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Union[str, Any] = VQModel( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=3 ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1000 ,) return CLIPTextModel(_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : Dict = self.dummy_uncond_unet _a : List[Any] = DDIMScheduler() _a : List[Any] = self.dummy_vq_model _a : str = LDMPipeline(unet=_a ,vqvae=_a ,scheduler=_a ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : List[str] = torch.manual_seed(0 ) _a : List[str] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Union[str, Any] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ,return_dict=_a )[0] _a : Tuple = image[0, -3:, -3:, -1] _a : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _a : int = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) _a : Any = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : List[str] = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Dict = ldm(generator=_a ,num_inference_steps=5 ,output_type='numpy' ).images _a : str = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _a : Optional[int] = np.array([0.4399, 0.4_4975, 0.4_6825, 0.474, 0.4359, 0.4581, 0.4_5095, 0.4341, 0.4447] ) _a : int = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
271
0
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool A__ : Dict ={ '''Acehnese Arabic''': '''ace_Arab''', '''Acehnese Latin''': '''ace_Latn''', '''Mesopotamian Arabic''': '''acm_Arab''', '''Ta\'izzi-Adeni Arabic''': '''acq_Arab''', '''Tunisian Arabic''': '''aeb_Arab''', '''Afrikaans''': '''afr_Latn''', '''South Levantine Arabic''': '''ajp_Arab''', '''Akan''': '''aka_Latn''', '''Amharic''': '''amh_Ethi''', '''North Levantine Arabic''': '''apc_Arab''', '''Modern Standard Arabic''': '''arb_Arab''', '''Modern Standard Arabic Romanized''': '''arb_Latn''', '''Najdi Arabic''': '''ars_Arab''', '''Moroccan Arabic''': '''ary_Arab''', '''Egyptian Arabic''': '''arz_Arab''', '''Assamese''': '''asm_Beng''', '''Asturian''': '''ast_Latn''', '''Awadhi''': '''awa_Deva''', '''Central Aymara''': '''ayr_Latn''', '''South Azerbaijani''': '''azb_Arab''', '''North Azerbaijani''': '''azj_Latn''', '''Bashkir''': '''bak_Cyrl''', '''Bambara''': '''bam_Latn''', '''Balinese''': '''ban_Latn''', '''Belarusian''': '''bel_Cyrl''', '''Bemba''': '''bem_Latn''', '''Bengali''': '''ben_Beng''', '''Bhojpuri''': '''bho_Deva''', '''Banjar Arabic''': '''bjn_Arab''', '''Banjar Latin''': '''bjn_Latn''', '''Standard Tibetan''': '''bod_Tibt''', '''Bosnian''': '''bos_Latn''', '''Buginese''': '''bug_Latn''', '''Bulgarian''': '''bul_Cyrl''', '''Catalan''': '''cat_Latn''', '''Cebuano''': '''ceb_Latn''', '''Czech''': '''ces_Latn''', '''Chokwe''': '''cjk_Latn''', '''Central Kurdish''': '''ckb_Arab''', '''Crimean Tatar''': '''crh_Latn''', '''Welsh''': '''cym_Latn''', '''Danish''': '''dan_Latn''', '''German''': '''deu_Latn''', '''Southwestern Dinka''': '''dik_Latn''', '''Dyula''': '''dyu_Latn''', '''Dzongkha''': '''dzo_Tibt''', '''Greek''': '''ell_Grek''', '''English''': '''eng_Latn''', '''Esperanto''': '''epo_Latn''', '''Estonian''': '''est_Latn''', '''Basque''': '''eus_Latn''', '''Ewe''': '''ewe_Latn''', '''Faroese''': '''fao_Latn''', '''Fijian''': '''fij_Latn''', '''Finnish''': '''fin_Latn''', '''Fon''': '''fon_Latn''', '''French''': '''fra_Latn''', '''Friulian''': '''fur_Latn''', '''Nigerian Fulfulde''': '''fuv_Latn''', '''Scottish Gaelic''': '''gla_Latn''', '''Irish''': '''gle_Latn''', '''Galician''': '''glg_Latn''', '''Guarani''': '''grn_Latn''', '''Gujarati''': '''guj_Gujr''', '''Haitian Creole''': '''hat_Latn''', '''Hausa''': '''hau_Latn''', '''Hebrew''': '''heb_Hebr''', '''Hindi''': '''hin_Deva''', '''Chhattisgarhi''': '''hne_Deva''', '''Croatian''': '''hrv_Latn''', '''Hungarian''': '''hun_Latn''', '''Armenian''': '''hye_Armn''', '''Igbo''': '''ibo_Latn''', '''Ilocano''': '''ilo_Latn''', '''Indonesian''': '''ind_Latn''', '''Icelandic''': '''isl_Latn''', '''Italian''': '''ita_Latn''', '''Javanese''': '''jav_Latn''', '''Japanese''': '''jpn_Jpan''', '''Kabyle''': '''kab_Latn''', '''Jingpho''': '''kac_Latn''', '''Kamba''': '''kam_Latn''', '''Kannada''': '''kan_Knda''', '''Kashmiri Arabic''': '''kas_Arab''', '''Kashmiri Devanagari''': '''kas_Deva''', '''Georgian''': '''kat_Geor''', '''Central Kanuri Arabic''': '''knc_Arab''', '''Central Kanuri Latin''': '''knc_Latn''', '''Kazakh''': '''kaz_Cyrl''', '''Kabiyè''': '''kbp_Latn''', '''Kabuverdianu''': '''kea_Latn''', '''Khmer''': '''khm_Khmr''', '''Kikuyu''': '''kik_Latn''', '''Kinyarwanda''': '''kin_Latn''', '''Kyrgyz''': '''kir_Cyrl''', '''Kimbundu''': '''kmb_Latn''', '''Northern Kurdish''': '''kmr_Latn''', '''Kikongo''': '''kon_Latn''', '''Korean''': '''kor_Hang''', '''Lao''': '''lao_Laoo''', '''Ligurian''': '''lij_Latn''', '''Limburgish''': '''lim_Latn''', '''Lingala''': '''lin_Latn''', '''Lithuanian''': '''lit_Latn''', '''Lombard''': '''lmo_Latn''', '''Latgalian''': '''ltg_Latn''', '''Luxembourgish''': '''ltz_Latn''', '''Luba-Kasai''': '''lua_Latn''', '''Ganda''': '''lug_Latn''', '''Luo''': '''luo_Latn''', '''Mizo''': '''lus_Latn''', '''Standard Latvian''': '''lvs_Latn''', '''Magahi''': '''mag_Deva''', '''Maithili''': '''mai_Deva''', '''Malayalam''': '''mal_Mlym''', '''Marathi''': '''mar_Deva''', '''Minangkabau Arabic ''': '''min_Arab''', '''Minangkabau Latin''': '''min_Latn''', '''Macedonian''': '''mkd_Cyrl''', '''Plateau Malagasy''': '''plt_Latn''', '''Maltese''': '''mlt_Latn''', '''Meitei Bengali''': '''mni_Beng''', '''Halh Mongolian''': '''khk_Cyrl''', '''Mossi''': '''mos_Latn''', '''Maori''': '''mri_Latn''', '''Burmese''': '''mya_Mymr''', '''Dutch''': '''nld_Latn''', '''Norwegian Nynorsk''': '''nno_Latn''', '''Norwegian Bokmål''': '''nob_Latn''', '''Nepali''': '''npi_Deva''', '''Northern Sotho''': '''nso_Latn''', '''Nuer''': '''nus_Latn''', '''Nyanja''': '''nya_Latn''', '''Occitan''': '''oci_Latn''', '''West Central Oromo''': '''gaz_Latn''', '''Odia''': '''ory_Orya''', '''Pangasinan''': '''pag_Latn''', '''Eastern Panjabi''': '''pan_Guru''', '''Papiamento''': '''pap_Latn''', '''Western Persian''': '''pes_Arab''', '''Polish''': '''pol_Latn''', '''Portuguese''': '''por_Latn''', '''Dari''': '''prs_Arab''', '''Southern Pashto''': '''pbt_Arab''', '''Ayacucho Quechua''': '''quy_Latn''', '''Romanian''': '''ron_Latn''', '''Rundi''': '''run_Latn''', '''Russian''': '''rus_Cyrl''', '''Sango''': '''sag_Latn''', '''Sanskrit''': '''san_Deva''', '''Santali''': '''sat_Olck''', '''Sicilian''': '''scn_Latn''', '''Shan''': '''shn_Mymr''', '''Sinhala''': '''sin_Sinh''', '''Slovak''': '''slk_Latn''', '''Slovenian''': '''slv_Latn''', '''Samoan''': '''smo_Latn''', '''Shona''': '''sna_Latn''', '''Sindhi''': '''snd_Arab''', '''Somali''': '''som_Latn''', '''Southern Sotho''': '''sot_Latn''', '''Spanish''': '''spa_Latn''', '''Tosk Albanian''': '''als_Latn''', '''Sardinian''': '''srd_Latn''', '''Serbian''': '''srp_Cyrl''', '''Swati''': '''ssw_Latn''', '''Sundanese''': '''sun_Latn''', '''Swedish''': '''swe_Latn''', '''Swahili''': '''swh_Latn''', '''Silesian''': '''szl_Latn''', '''Tamil''': '''tam_Taml''', '''Tatar''': '''tat_Cyrl''', '''Telugu''': '''tel_Telu''', '''Tajik''': '''tgk_Cyrl''', '''Tagalog''': '''tgl_Latn''', '''Thai''': '''tha_Thai''', '''Tigrinya''': '''tir_Ethi''', '''Tamasheq Latin''': '''taq_Latn''', '''Tamasheq Tifinagh''': '''taq_Tfng''', '''Tok Pisin''': '''tpi_Latn''', '''Tswana''': '''tsn_Latn''', '''Tsonga''': '''tso_Latn''', '''Turkmen''': '''tuk_Latn''', '''Tumbuka''': '''tum_Latn''', '''Turkish''': '''tur_Latn''', '''Twi''': '''twi_Latn''', '''Central Atlas Tamazight''': '''tzm_Tfng''', '''Uyghur''': '''uig_Arab''', '''Ukrainian''': '''ukr_Cyrl''', '''Umbundu''': '''umb_Latn''', '''Urdu''': '''urd_Arab''', '''Northern Uzbek''': '''uzn_Latn''', '''Venetian''': '''vec_Latn''', '''Vietnamese''': '''vie_Latn''', '''Waray''': '''war_Latn''', '''Wolof''': '''wol_Latn''', '''Xhosa''': '''xho_Latn''', '''Eastern Yiddish''': '''ydd_Hebr''', '''Yoruba''': '''yor_Latn''', '''Yue Chinese''': '''yue_Hant''', '''Chinese Simplified''': '''zho_Hans''', '''Chinese Traditional''': '''zho_Hant''', '''Standard Malay''': '''zsm_Latn''', '''Zulu''': '''zul_Latn''', } class UpperCAmelCase ( snake_case_ ): _lowercase: Tuple = '''facebook/nllb-200-distilled-600M''' _lowercase: Optional[int] = ( '''This is a tool that translates text from a language to another. It takes three inputs: `text`, which should ''' '''be the text to translate, `src_lang`, which should be the language of the text to translate and `tgt_lang`, ''' '''which should be the language for the desired ouput language. Both `src_lang` and `tgt_lang` are written in ''' '''plain English, such as \'Romanian\', or \'Albanian\'. It returns the text translated in `tgt_lang`.''' ) _lowercase: int = '''translator''' _lowercase: int = AutoTokenizer _lowercase: str = AutoModelForSeqaSeqLM _lowercase: str = LANGUAGE_CODES _lowercase: Union[str, Any] = ['''text''', '''text''', '''text'''] _lowercase: Union[str, Any] = ['''text'''] def lowercase__ ( self : Dict , __snake_case : Optional[Any] , __snake_case : Dict , __snake_case : Optional[int] ) -> Any: if src_lang not in self.lang_to_code: raise ValueError(f"{src_lang} is not a supported language." ) if tgt_lang not in self.lang_to_code: raise ValueError(f"{tgt_lang} is not a supported language." ) _lowerCAmelCase = self.lang_to_code[src_lang] _lowerCAmelCase = self.lang_to_code[tgt_lang] return self.pre_processor._build_translation_inputs( __snake_case , return_tensors="""pt""" , src_lang=__snake_case , tgt_lang=__snake_case ) def lowercase__ ( self : Optional[Any] , __snake_case : Any ) -> List[str]: return self.model.generate(**__snake_case ) def lowercase__ ( self : Union[str, Any] , __snake_case : List[Any] ) -> Optional[int]: return self.post_processor.decode(outputs[0].tolist() , skip_special_tokens=__snake_case )
70
'''simple docstring''' import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : int ,*_a : Optional[int] ,**_a : str ): '''simple docstring''' warnings.warn( 'The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use BeitImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL A_ :Any = logging.get_logger(__name__) class __A ( a ): """simple docstring""" UpperCamelCase__ : int =["""pixel_values"""] def __init__( self , lowerCamelCase__ = True , lowerCamelCase__ = None , lowerCamelCase__ = PIL.Image.BICUBIC , lowerCamelCase__ = True , lowerCamelCase__ = None , lowerCamelCase__ = 1 / 255 , lowerCamelCase__ = True , lowerCamelCase__ = True , lowerCamelCase__ = None , lowerCamelCase__ = None , **lowerCamelCase__ , ): """simple docstring""" super().__init__(**lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =size if size is not None else {'height': 256, 'width': 256} __UpperCamelCase : List[str] =get_size_dict(lowerCamelCase__ ) __UpperCamelCase : Tuple =crop_size if crop_size is not None else {'height': 224, 'width': 224} __UpperCamelCase : Optional[int] =get_size_dict(lowerCamelCase__ , param_name='crop_size' ) __UpperCamelCase : Tuple =do_resize __UpperCamelCase : Any =size __UpperCamelCase : Dict =resample __UpperCamelCase : Tuple =do_center_crop __UpperCamelCase : int =crop_size __UpperCamelCase : Optional[int] =do_rescale __UpperCamelCase : List[Any] =rescale_factor __UpperCamelCase : List[Any] =do_normalize __UpperCamelCase : Union[str, Any] =image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __UpperCamelCase : Optional[Any] =image_std if image_std is not None else IMAGENET_STANDARD_STD def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = PIL.Image.BICUBIC , lowerCamelCase__ = None , **lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : int =get_size_dict(lowerCamelCase__ ) if "height" not in size or "width" not in size: raise ValueError(f'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}' ) return resize( lowerCamelCase__ , size=(size['height'], size['width']) , resample=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , **lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Optional[int] =get_size_dict(lowerCamelCase__ ) if "height" not in size or "width" not in size: raise ValueError(f'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}' ) return center_crop(lowerCamelCase__ , size=(size['height'], size['width']) , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , **lowerCamelCase__ , ): """simple docstring""" return rescale(lowerCamelCase__ , scale=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , **lowerCamelCase__ , ): """simple docstring""" return normalize(lowerCamelCase__ , mean=lowerCamelCase__ , std=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__=None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = ChannelDimension.FIRST , **lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Optional[Any] =do_resize if do_resize is not None else self.do_resize __UpperCamelCase : Union[str, Any] =resample if resample is not None else self.resample __UpperCamelCase : Optional[Any] =do_center_crop if do_center_crop is not None else self.do_center_crop __UpperCamelCase : Any =do_rescale if do_rescale is not None else self.do_rescale __UpperCamelCase : Dict =rescale_factor if rescale_factor is not None else self.rescale_factor __UpperCamelCase : Optional[Any] =do_normalize if do_normalize is not None else self.do_normalize __UpperCamelCase : Optional[Any] =image_mean if image_mean is not None else self.image_mean __UpperCamelCase : Any =image_std if image_std is not None else self.image_std __UpperCamelCase : Union[str, Any] =size if size is not None else self.size __UpperCamelCase : List[str] =get_size_dict(lowerCamelCase__ ) __UpperCamelCase : Tuple =crop_size if crop_size is not None else self.crop_size __UpperCamelCase : Tuple =get_size_dict(lowerCamelCase__ , param_name='crop_size' ) __UpperCamelCase : List[Any] =make_list_of_images(lowerCamelCase__ ) if not valid_images(lowerCamelCase__ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. __UpperCamelCase : Optional[int] =[to_numpy_array(lowerCamelCase__ ) for image in images] if do_resize: __UpperCamelCase : Optional[Any] =[self.resize(image=lowerCamelCase__ , size=lowerCamelCase__ , resample=lowerCamelCase__ ) for image in images] if do_center_crop: __UpperCamelCase : int =[self.center_crop(image=lowerCamelCase__ , size=lowerCamelCase__ ) for image in images] if do_rescale: __UpperCamelCase : int =[self.rescale(image=lowerCamelCase__ , scale=lowerCamelCase__ ) for image in images] if do_normalize: __UpperCamelCase : Optional[int] =[self.normalize(image=lowerCamelCase__ , mean=lowerCamelCase__ , std=lowerCamelCase__ ) for image in images] __UpperCamelCase : Optional[Any] =[to_channel_dimension_format(lowerCamelCase__ , lowerCamelCase__ ) for image in images] __UpperCamelCase : List[str] ={'pixel_values': images} return BatchFeature(data=lowerCamelCase__ , tensor_type=lowerCamelCase__ )
71
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { """linear""": get_linear_schedule_with_warmup, """cosine""": get_cosine_schedule_with_warmup, """cosine_w_restarts""": get_cosine_with_hard_restarts_schedule_with_warmup, """polynomial""": get_polynomial_decay_schedule_with_warmup, """constant""": get_constant_schedule, """constant_w_warmup""": get_constant_schedule_with_warmup, } class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Optional[int] ,_a : Optional[Any]=None ,_a : Dict=None ,*_a : int ,**_a : str ): '''simple docstring''' super().__init__(*_a ,**_a ) if config is None: assert isinstance(self.model ,_a ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F""" {self.model.__class__}""" ) _a : List[Any] = self.model.config else: _a : Optional[int] = config _a : List[str] = data_args _a : List[Any] = self.config.tgt_vocab_size if isinstance(self.config ,_a ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ' padding..' ) if self.args.label_smoothing == 0: _a : List[str] = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss _a : Tuple = label_smoothed_nll_loss def __lowercase ( self : List[str] ,_a : int ): '''simple docstring''' if self.optimizer is None: _a : Union[str, Any] = ['bias', 'LayerNorm.weight'] _a : Tuple = [ { 'params': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], 'weight_decay': self.args.weight_decay, }, { 'params': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], 'weight_decay': 0.0, }, ] _a : Optional[int] = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: _a : Any = Adafactor _a : Dict = {'scale_parameter': False, 'relative_step': False} else: _a : Union[str, Any] = AdamW _a : str = { 'betas': (self.args.adam_betaa, self.args.adam_betaa), 'eps': self.args.adam_epsilon, } _a : Union[str, Any] = self.args.learning_rate if self.sharded_ddp: _a : str = OSS( params=_a ,optim=_a ,**_a ,) else: _a : Tuple = optimizer_cls(_a ,**_a ) if self.lr_scheduler is None: _a : List[Any] = self._get_lr_scheduler(_a ) else: # ignoring --lr_scheduler logger.warning('scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.' ) def __lowercase ( self : List[Any] ,_a : List[Any] ): '''simple docstring''' _a : str = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": _a : int = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": _a : List[str] = schedule_func(self.optimizer ,num_warmup_steps=self.args.warmup_steps ) else: _a : Optional[int] = schedule_func( self.optimizer ,num_warmup_steps=self.args.warmup_steps ,num_training_steps=_a ) return scheduler def __lowercase ( self : Tuple ): '''simple docstring''' if isinstance(self.train_dataset ,torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size ,distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) ,) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def __lowercase ( self : Dict ,_a : Dict ,_a : Any ,_a : Dict ): '''simple docstring''' if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Union[str, Any] = self.loss_fn(logits.view(-1 ,logits.shape[-1] ) ,labels.view(-1 ) ) else: # compute usual loss via models _a, _a : Union[str, Any] = model(**_a ,labels=_a ,use_cache=_a )[:2] else: # compute label smoothed loss _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Any = torch.nn.functional.log_softmax(_a ,dim=-1 ) _a, _a : List[str] = self.loss_fn(_a ,_a ,self.args.label_smoothing ,ignore_index=self.config.pad_token_id ) return loss, logits def __lowercase ( self : Optional[int] ,_a : Union[str, Any] ,_a : List[Any] ): '''simple docstring''' _a : Optional[int] = inputs.pop('labels' ) _a, _a : int = self._compute_loss(_a ,_a ,_a ) return loss def __lowercase ( self : Optional[Any] ,_a : nn.Module ,_a : Dict[str, Union[torch.Tensor, Any]] ,_a : bool ,_a : Optional[List[str]] = None ,): '''simple docstring''' _a : int = self._prepare_inputs(_a ) _a : Any = { 'max_length': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, 'num_beams': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: _a : int = self.model.generate( inputs['input_ids'] ,attention_mask=inputs['attention_mask'] ,**_a ,) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: _a : int = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) _a : Union[str, Any] = inputs.pop('labels' ) with torch.no_grad(): # compute loss on predict data _a, _a : Optional[int] = self._compute_loss(_a ,_a ,_a ) _a : Optional[Any] = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) _a : Optional[Any] = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: _a : Dict = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) return (loss, logits, labels) def __lowercase ( self : str ,_a : Tuple ,_a : Tuple ): '''simple docstring''' _a : List[Any] = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( 'Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be' F""" padded to `max_length`={max_length}""" ) _a : int = pad_token_id * torch.ones( (tensor.shape[0], max_length) ,dtype=tensor.dtype ,device=tensor.device ) _a : Union[str, Any] = tensor return padded_tensor
271
0
"""simple docstring""" lowerCAmelCase__ = '''0.21.0''' from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
72
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser __lowerCAmelCase = re.compile(r"""\s+""") def UpperCAmelCase_ (__a : Any ): """simple docstring""" return {"hash": hashlib.mda(re.sub(__a , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : List[str] = [len(__a ) for line in example['content'].splitlines()] return {"line_mean": np.mean(__a ), "line_max": max(__a )} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Union[str, Any] = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase_ (__a : Optional[int] , __a : Any ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def UpperCAmelCase_ (__a : int , __a : Union[str, Any]=5 ): """simple docstring""" _a : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] _a : List[str] = example['content'].splitlines() for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase_ (__a : List[str] , __a : Dict=5 , __a : Tuple=0.05 ): """simple docstring""" _a : Optional[int] = ['unit tests', 'test file', 'configuration file'] _a : int = example['content'].splitlines() _a : int = 0 _a : Dict = 0 # first test for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test _a : int = example['content'].count('\n' ) _a : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" _a : List[str] = ['def ', 'class ', 'for ', 'while '] _a : str = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase_ (__a : int , __a : Any=4 ): """simple docstring""" _a : List[str] = example['content'].splitlines() _a : Dict = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Optional[Any] = tokenizer(example['content'] , truncation=__a )['input_ids'] _a : Optional[int] = len(example['content'] ) / len(__a ) return {"ratio": ratio} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Dict = {} results.update(get_hash(__a ) ) results.update(line_stats(__a ) ) results.update(alpha_stats(__a ) ) results.update(char_token_ratio(__a ) ) results.update(is_autogenerated(__a ) ) results.update(is_config_or_test(__a ) ) results.update(has_no_keywords(__a ) ) results.update(has_few_assignments(__a ) ) return results def UpperCAmelCase_ (__a : Any , __a : Any , __a : str ): """simple docstring""" if not check_uniques(__a , __a ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase_ (__a : Union[str, Any] ): """simple docstring""" with open(__a , 'rb' ) as f_in: with gzip.open(str(__a ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(__a , __a ) os.unlink(__a ) # Settings __lowerCAmelCase = HfArgumentParser(PreprocessingArguments) __lowerCAmelCase = parser.parse_args() if args.num_workers is None: __lowerCAmelCase = multiprocessing.cpu_count() __lowerCAmelCase = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset __lowerCAmelCase = time.time() __lowerCAmelCase = load_dataset(args.dataset_name, split="""train""") print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing __lowerCAmelCase = time.time() __lowerCAmelCase = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes __lowerCAmelCase = set(ds.unique("""hash""")) __lowerCAmelCase = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics __lowerCAmelCase = time.time() __lowerCAmelCase = ds.filter(filter, fn_kwargs={"""uniques""": uniques, """args""": args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: __lowerCAmelCase = time.time() __lowerCAmelCase , __lowerCAmelCase = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file __lowerCAmelCase = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / """duplicate_clusters.json""", """w""") as f: json.dump(duplicate_clusters, f) __lowerCAmelCase = output_dir / """data""" data_dir.mkdir(exist_ok=True) __lowerCAmelCase = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): __lowerCAmelCase = str(data_dir / f'''file-{file_number+1:012}.json''') __lowerCAmelCase = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
271
0
import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING a =logging.get_logger(__name__) a ={ """microsoft/conditional-detr-resnet-50""": ( """https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json""" ), } class A_ ( SCREAMING_SNAKE_CASE ): _UpperCAmelCase : List[str] = '''conditional_detr''' _UpperCAmelCase : int = ['''past_key_values'''] _UpperCAmelCase : Optional[int] = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', } def __init__( self : Dict ,SCREAMING_SNAKE_CASE__ : int=True ,SCREAMING_SNAKE_CASE__ : List[str]=None ,SCREAMING_SNAKE_CASE__ : List[str]=3 ,SCREAMING_SNAKE_CASE__ : int=3_0_0 ,SCREAMING_SNAKE_CASE__ : str=6 ,SCREAMING_SNAKE_CASE__ : Dict=2_0_4_8 ,SCREAMING_SNAKE_CASE__ : Union[str, Any]=8 ,SCREAMING_SNAKE_CASE__ : int=6 ,SCREAMING_SNAKE_CASE__ : Optional[int]=2_0_4_8 ,SCREAMING_SNAKE_CASE__ : List[Any]=8 ,SCREAMING_SNAKE_CASE__ : int=0.0 ,SCREAMING_SNAKE_CASE__ : Tuple=0.0 ,SCREAMING_SNAKE_CASE__ : Union[str, Any]=True ,SCREAMING_SNAKE_CASE__ : Optional[Any]="relu" ,SCREAMING_SNAKE_CASE__ : List[Any]=2_5_6 ,SCREAMING_SNAKE_CASE__ : List[Any]=0.1 ,SCREAMING_SNAKE_CASE__ : Any=0.0 ,SCREAMING_SNAKE_CASE__ : List[Any]=0.0 ,SCREAMING_SNAKE_CASE__ : str=0.02 ,SCREAMING_SNAKE_CASE__ : Tuple=1.0 ,SCREAMING_SNAKE_CASE__ : str=False ,SCREAMING_SNAKE_CASE__ : Dict="sine" ,SCREAMING_SNAKE_CASE__ : int="resnet50" ,SCREAMING_SNAKE_CASE__ : Union[str, Any]=True ,SCREAMING_SNAKE_CASE__ : str=False ,SCREAMING_SNAKE_CASE__ : Tuple=2 ,SCREAMING_SNAKE_CASE__ : Optional[int]=5 ,SCREAMING_SNAKE_CASE__ : int=2 ,SCREAMING_SNAKE_CASE__ : List[str]=1 ,SCREAMING_SNAKE_CASE__ : int=1 ,SCREAMING_SNAKE_CASE__ : str=2 ,SCREAMING_SNAKE_CASE__ : Optional[int]=5 ,SCREAMING_SNAKE_CASE__ : int=2 ,SCREAMING_SNAKE_CASE__ : Dict=0.25 ,**SCREAMING_SNAKE_CASE__ : Union[str, Any] ,): if backbone_config is not None and use_timm_backbone: raise ValueError('You can\'t specify both `backbone_config` and `use_timm_backbone`.') if not use_timm_backbone: if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.') __lowerCamelCase : str = CONFIG_MAPPING['resnet'](out_features=['stage4']) elif isinstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__): __lowerCamelCase : int = backbone_config.get('model_type') __lowerCamelCase : Optional[int] = CONFIG_MAPPING[backbone_model_type] __lowerCamelCase : int = config_class.from_dict(SCREAMING_SNAKE_CASE__) __lowerCamelCase : Optional[Any] = use_timm_backbone __lowerCamelCase : Dict = backbone_config __lowerCamelCase : int = num_channels __lowerCamelCase : Union[str, Any] = num_queries __lowerCamelCase : List[Any] = d_model __lowerCamelCase : str = encoder_ffn_dim __lowerCamelCase : Union[str, Any] = encoder_layers __lowerCamelCase : Union[str, Any] = encoder_attention_heads __lowerCamelCase : Union[str, Any] = decoder_ffn_dim __lowerCamelCase : Optional[Any] = decoder_layers __lowerCamelCase : int = decoder_attention_heads __lowerCamelCase : Optional[Any] = dropout __lowerCamelCase : Optional[Any] = attention_dropout __lowerCamelCase : Any = activation_dropout __lowerCamelCase : int = activation_function __lowerCamelCase : Dict = init_std __lowerCamelCase : int = init_xavier_std __lowerCamelCase : Any = encoder_layerdrop __lowerCamelCase : str = decoder_layerdrop __lowerCamelCase : Dict = encoder_layers __lowerCamelCase : List[str] = auxiliary_loss __lowerCamelCase : Optional[int] = position_embedding_type __lowerCamelCase : List[str] = backbone __lowerCamelCase : Dict = use_pretrained_backbone __lowerCamelCase : Union[str, Any] = dilation # Hungarian matcher __lowerCamelCase : Dict = class_cost __lowerCamelCase : Dict = bbox_cost __lowerCamelCase : Any = giou_cost # Loss coefficients __lowerCamelCase : List[str] = mask_loss_coefficient __lowerCamelCase : Optional[int] = dice_loss_coefficient __lowerCamelCase : Any = cls_loss_coefficient __lowerCamelCase : str = bbox_loss_coefficient __lowerCamelCase : str = giou_loss_coefficient __lowerCamelCase : Optional[Any] = focal_alpha super().__init__(is_encoder_decoder=SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__) @property def lowerCAmelCase ( self : Union[str, Any]): return self.encoder_attention_heads @property def lowerCAmelCase ( self : int): return self.d_model def lowerCAmelCase ( self : str): __lowerCamelCase : Optional[int] = copy.deepcopy(self.__dict__) if self.backbone_config is not None: __lowerCamelCase : str = self.backbone_config.to_dict() __lowerCamelCase : Optional[int] = self.__class__.model_type return output class A_ ( SCREAMING_SNAKE_CASE ): _UpperCAmelCase : int = version.parse('''1.11''' ) @property def lowerCAmelCase ( self : Optional[int]): return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ('pixel_mask', {0: 'batch'}), ]) @property def lowerCAmelCase ( self : Optional[Any]): return 1E-5 @property def lowerCAmelCase ( self : str): return 1_2
73
'''simple docstring''' import argparse from typing import List import evaluate import numpy as np import torch from datasets import DatasetDict, load_dataset # New Code # # We'll be using StratifiedKFold for this example from sklearn.model_selection import StratifiedKFold from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to perform Cross Validation, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __lowerCAmelCase = 1_6 __lowerCAmelCase = 3_2 def UpperCAmelCase_ (__a : Accelerator , __a : DatasetDict , __a : List[int] , __a : List[int] , __a : int = 1_6 ): """simple docstring""" _a : Union[str, Any] = AutoTokenizer.from_pretrained('bert-base-cased' ) _a : str = DatasetDict( { 'train': dataset['train'].select(__a ), 'validation': dataset['train'].select(__a ), 'test': dataset['validation'], } ) def tokenize_function(__a : List[Any] ): # max_length=None => use the model max length (it's actually the default) _a : Optional[int] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=__a , max_length=__a ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _a : List[str] = datasets.map( __a , batched=__a , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _a : List[Any] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(__a : int ): # On TPU it's best to pad everything to the same length or training will be very slow. _a : Dict = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _a : Tuple = 1_6 elif accelerator.mixed_precision != "no": _a : List[Any] = 8 else: _a : List[Any] = None return tokenizer.pad( __a , padding='longest' , max_length=__a , pad_to_multiple_of=__a , return_tensors='pt' , ) # Instantiate dataloaders. _a : Any = DataLoader( tokenized_datasets['train'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[int] = DataLoader( tokenized_datasets['validation'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[Any] = DataLoader( tokenized_datasets['test'] , shuffle=__a , collate_fn=__a , batch_size=__a ) return train_dataloader, eval_dataloader, test_dataloader def UpperCAmelCase_ (__a : Any , __a : Union[str, Any] ): """simple docstring""" _a : Dict = [] # Download the dataset _a : Tuple = load_dataset('glue' , 'mrpc' ) # Create our splits _a : Union[str, Any] = StratifiedKFold(n_splits=int(args.num_folds ) ) # Initialize accelerator _a : Any = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _a : Optional[Any] = config['lr'] _a : Optional[int] = int(config['num_epochs'] ) _a : Dict = int(config['seed'] ) _a : Dict = int(config['batch_size'] ) _a : Optional[int] = evaluate.load('glue' , 'mrpc' ) # If the batch size is too big we use gradient accumulation _a : List[Any] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: _a : Any = batch_size // MAX_GPU_BATCH_SIZE _a : List[str] = MAX_GPU_BATCH_SIZE set_seed(__a ) # New Code # # Create our folds: _a : int = kfold.split(np.zeros(datasets['train'].num_rows ) , datasets['train']['label'] ) _a : Any = [] # Iterate over them for i, (train_idxs, valid_idxs) in enumerate(__a ): _a, _a, _a : Optional[Any] = get_fold_dataloaders( __a , __a , __a , __a , ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _a : Dict = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=__a ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _a : List[Any] = model.to(accelerator.device ) # Instantiate optimizer _a : List[str] = AdamW(params=model.parameters() , lr=__a ) # Instantiate scheduler _a : List[Any] = get_linear_schedule_with_warmup( optimizer=__a , num_warmup_steps=1_0_0 , num_training_steps=(len(__a ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _a, _a, _a, _a, _a : Union[str, Any] = accelerator.prepare( __a , __a , __a , __a , __a ) # Now we train the model for epoch in range(__a ): model.train() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _a : Dict = model(**__a ) _a : int = outputs.loss _a : Any = loss / gradient_accumulation_steps accelerator.backward(__a ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Union[str, Any] = model(**__a ) _a : Tuple = outputs.logits.argmax(dim=-1 ) _a, _a : Any = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=__a , references=__a , ) _a : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , __a ) # New Code # # We also run predictions on the test set at the very end _a : Any = [] for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Tuple = model(**__a ) _a : Dict = outputs.logits _a, _a : Optional[int] = accelerator.gather_for_metrics((predictions, batch['labels']) ) fold_predictions.append(predictions.cpu() ) if i == 0: # We need all of the test predictions test_references.append(references.cpu() ) # Use accelerator.print to print only on the main process. test_predictions.append(torch.cat(__a , dim=0 ) ) # We now need to release all our memory and get rid of the current model, optimizer, etc accelerator.free_memory() # New Code # # Finally we check the accuracy of our folded results: _a : Dict = torch.cat(__a , dim=0 ) _a : Any = torch.stack(__a , dim=0 ).sum(dim=0 ).div(int(args.num_folds ) ).argmax(dim=-1 ) _a : str = metric.compute(predictions=__a , references=__a ) accelerator.print('Average test metrics from all folds:' , __a ) def UpperCAmelCase_ (): """simple docstring""" _a : Any = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=__a , default=__a , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) # New Code # parser.add_argument('--num_folds' , type=__a , default=3 , help='The number of splits to perform across the dataset' ) _a : Any = parser.parse_args() _a : int = {'lr': 2e-5, 'num_epochs': 3, 'seed': 4_2, 'batch_size': 1_6} training_function(__a , __a ) if __name__ == "__main__": main()
271
0
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import KarrasVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class lowerCAmelCase_ ( _lowercase ): '''simple docstring''' _lowerCamelCase: UNetaDModel _lowerCamelCase: KarrasVeScheduler def __init__( self : List[Any] ,A_ : UNetaDModel ,A_ : KarrasVeScheduler ) -> Dict: super().__init__() self.register_modules(unet=A_ ,scheduler=A_ ) @torch.no_grad() def __call__( self : Optional[Any] ,A_ : int = 1 ,A_ : int = 50 ,A_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,A_ : Optional[str] = "pil" ,A_ : bool = True ,**A_ : Optional[Any] ,) -> Union[Tuple, ImagePipelineOutput]: A = self.unet.config.sample_size A = (batch_size, 3, img_size, img_size) A = self.unet # sample x_0 ~ N(0, sigma_0^2 * I) A = randn_tensor(A_ ,generator=A_ ,device=self.device ) * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(A_ ) for t in self.progress_bar(self.scheduler.timesteps ): # here sigma_t == t_i from the paper A = self.scheduler.schedule[t] A = self.scheduler.schedule[t - 1] if t > 0 else 0 # 1. Select temporarily increased noise level sigma_hat # 2. Add new noise to move from sample_i to sample_hat A , A = self.scheduler.add_noise_to_input(A_ ,A_ ,generator=A_ ) # 3. Predict the noise residual given the noise magnitude `sigma_hat` # The model inputs and output are adjusted by following eq. (213) in [1]. A = (sigma_hat / 2) * model((sample_hat + 1) / 2 ,sigma_hat / 2 ).sample # 4. Evaluate dx/dt at sigma_hat # 5. Take Euler step from sigma to sigma_prev A = self.scheduler.step(A_ ,A_ ,A_ ,A_ ) if sigma_prev != 0: # 6. Apply 2nd order correction # The model inputs and output are adjusted by following eq. (213) in [1]. A = (sigma_prev / 2) * model((step_output.prev_sample + 1) / 2 ,sigma_prev / 2 ).sample A = self.scheduler.step_correct( A_ ,A_ ,A_ ,A_ ,step_output.prev_sample ,step_output['derivative'] ,) A = step_output.prev_sample A = (sample / 2 + 0.5).clamp(0 ,1 ) A = sample.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": A = self.numpy_to_pil(A_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=A_ )
74
'''simple docstring''' from __future__ import annotations __lowerCAmelCase = [-1_0, -5, 0, 5, 5.1, 1_1, 1_3, 2_1, 3, 4, -2_1, -1_0, -5, -1, 0] __lowerCAmelCase = [-5, 0, 5, 5.1, 1_1, 1_3, 2_1, -1, 4, -1, -1_0, -5, -1, 0, -1] def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Optional[int] = [] _a : int = len(__a ) for i in range(__a ): _a : float = -1 for j in range(i + 1 , __a ): if arr[i] < arr[j]: _a : Any = arr[j] break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Tuple = [] for i, outer in enumerate(__a ): _a : float = -1 for inner in arr[i + 1 :]: if outer < inner: _a : Dict = inner break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : int = len(__a ) _a : list[float] = [] _a : list[float] = [-1] * arr_size for index in reversed(range(__a ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: _a : Dict = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) __lowerCAmelCase = ( """from __main__ import arr, next_greatest_element_slow, """ """next_greatest_element_fast, next_greatest_element""" ) print( """next_greatest_element_slow():""", timeit("""next_greatest_element_slow(arr)""", setup=setup), ) print( """next_greatest_element_fast():""", timeit("""next_greatest_element_fast(arr)""", setup=setup), ) print( """ next_greatest_element():""", timeit("""next_greatest_element(arr)""", setup=setup), )
271
0
'''simple docstring''' import argparse import collections import json from pathlib import Path import requests import torch import yaml from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTImageProcessor, MobileViTVaConfig, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, ) from transformers.utils import logging logging.set_verbosity_info() a_ : Optional[Any] = logging.get_logger(__name__) def a_ ( __snake_case : int ) -> Optional[int]: """simple docstring""" print('''Loading config file...''' ) def flatten_yaml_as_dict(__snake_case : int , __snake_case : Any="" , __snake_case : List[str]="." ): lowerCamelCase_ =[] for k, v in d.items(): lowerCamelCase_ =parent_key + sep + k if parent_key else k if isinstance(__snake_case , collections.abc.MutableMapping ): items.extend(flatten_yaml_as_dict(__snake_case , __snake_case , sep=__snake_case ).items() ) else: items.append((new_key, v) ) return dict(__snake_case ) lowerCamelCase_ =argparse.Namespace() with open(__snake_case , '''r''' ) as yaml_file: try: lowerCamelCase_ =yaml.load(__snake_case , Loader=yaml.FullLoader ) lowerCamelCase_ =flatten_yaml_as_dict(__snake_case ) for k, v in flat_cfg.items(): setattr(__snake_case , __snake_case , __snake_case ) except yaml.YAMLError as exc: logger.error('''Error while loading config file: {}. Error message: {}'''.format(__snake_case , str(__snake_case ) ) ) return config def a_ ( __snake_case : Tuple , __snake_case : List[str] ) -> Tuple: """simple docstring""" lowerCamelCase_ =MobileViTVaConfig() lowerCamelCase_ =False # dataset if task_name.startswith('''imagenet1k_''' ): lowerCamelCase_ =1000 if int(task_name.strip().split('''_''' )[-1] ) == 384: lowerCamelCase_ =384 else: lowerCamelCase_ =256 lowerCamelCase_ ='''imagenet-1k-id2label.json''' elif task_name.startswith('''imagenet21k_to_1k_''' ): lowerCamelCase_ =2_1000 if int(task_name.strip().split('''_''' )[-1] ) == 384: lowerCamelCase_ =384 else: lowerCamelCase_ =256 lowerCamelCase_ ='''imagenet-22k-id2label.json''' elif task_name.startswith('''ade20k_''' ): lowerCamelCase_ =151 lowerCamelCase_ =512 lowerCamelCase_ ='''ade20k-id2label.json''' lowerCamelCase_ =True elif task_name.startswith('''voc_''' ): lowerCamelCase_ =21 lowerCamelCase_ =512 lowerCamelCase_ ='''pascal-voc-id2label.json''' lowerCamelCase_ =True # orig_config lowerCamelCase_ =load_orig_config_file(__snake_case ) assert getattr(__snake_case , '''model.classification.name''' , -1 ) == "mobilevit_v2", "Invalid model" lowerCamelCase_ =getattr(__snake_case , '''model.classification.mitv2.width_multiplier''' , 1.0 ) assert ( getattr(__snake_case , '''model.classification.mitv2.attn_norm_layer''' , -1 ) == "layer_norm_2d" ), "Norm layers other than layer_norm_2d is not supported" lowerCamelCase_ =getattr(__snake_case , '''model.classification.activation.name''' , '''swish''' ) # config.image_size == getattr(orig_config, 'sampler.bs.crop_size_width', 256) if is_segmentation_model: lowerCamelCase_ =getattr(__snake_case , '''model.segmentation.output_stride''' , 16 ) if "_deeplabv3" in task_name: lowerCamelCase_ =getattr(__snake_case , '''model.segmentation.deeplabv3.aspp_rates''' , [12, 24, 36] ) lowerCamelCase_ =getattr(__snake_case , '''model.segmentation.deeplabv3.aspp_out_channels''' , 512 ) lowerCamelCase_ =getattr(__snake_case , '''model.segmentation.deeplabv3.aspp_dropout''' , 0.1 ) # id2label lowerCamelCase_ ='''huggingface/label-files''' lowerCamelCase_ =json.load(open(hf_hub_download(__snake_case , __snake_case , repo_type='''dataset''' ) , '''r''' ) ) lowerCamelCase_ ={int(__snake_case ): v for k, v in idalabel.items()} lowerCamelCase_ =idalabel lowerCamelCase_ ={v: k for k, v in idalabel.items()} return config def a_ ( __snake_case : Optional[Any] , __snake_case : Optional[int] , __snake_case : List[str] ) -> Optional[int]: """simple docstring""" lowerCamelCase_ =dct.pop(__snake_case ) lowerCamelCase_ =val def a_ ( __snake_case : List[str] , __snake_case : Optional[int]=False ) -> Optional[int]: """simple docstring""" if base_model: lowerCamelCase_ ='''''' else: lowerCamelCase_ ='''mobilevitv2.''' lowerCamelCase_ =[] for k in state_dict.keys(): if k[:8] == "encoder.": lowerCamelCase_ =k[8:] else: lowerCamelCase_ =k if ".block." in k: lowerCamelCase_ =k_new.replace('''.block.''' , '''.''' ) if ".conv." in k: lowerCamelCase_ =k_new.replace('''.conv.''' , '''.convolution.''' ) if ".norm." in k: lowerCamelCase_ =k_new.replace('''.norm.''' , '''.normalization.''' ) if "conv_1." in k: lowerCamelCase_ =k_new.replace('''conv_1.''' , F'''{model_prefix}conv_stem.''' ) for i in [1, 2]: if F'''layer_{i}.''' in k: lowerCamelCase_ =k_new.replace(F'''layer_{i}.''' , F'''{model_prefix}encoder.layer.{i-1}.layer.''' ) if ".exp_1x1." in k: lowerCamelCase_ =k_new.replace('''.exp_1x1.''' , '''.expand_1x1.''' ) if ".red_1x1." in k: lowerCamelCase_ =k_new.replace('''.red_1x1.''' , '''.reduce_1x1.''' ) for i in [3, 4, 5]: if F'''layer_{i}.0.''' in k: lowerCamelCase_ =k_new.replace(F'''layer_{i}.0.''' , F'''{model_prefix}encoder.layer.{i-1}.downsampling_layer.''' ) if F'''layer_{i}.1.local_rep.0.''' in k: lowerCamelCase_ =k_new.replace(F'''layer_{i}.1.local_rep.0.''' , F'''{model_prefix}encoder.layer.{i-1}.conv_kxk.''' ) if F'''layer_{i}.1.local_rep.1.''' in k: lowerCamelCase_ =k_new.replace(F'''layer_{i}.1.local_rep.1.''' , F'''{model_prefix}encoder.layer.{i-1}.conv_1x1.''' ) for i in [3, 4, 5]: if i == 3: lowerCamelCase_ =[0, 1] elif i == 4: lowerCamelCase_ =[0, 1, 2, 3] elif i == 5: lowerCamelCase_ =[0, 1, 2] for j in j_in: if F'''layer_{i}.1.global_rep.{j}.''' in k: lowerCamelCase_ =k_new.replace( F'''layer_{i}.1.global_rep.{j}.''' , F'''{model_prefix}encoder.layer.{i-1}.transformer.layer.{j}.''' ) if F'''layer_{i}.1.global_rep.{j+1}.''' in k: lowerCamelCase_ =k_new.replace( F'''layer_{i}.1.global_rep.{j+1}.''' , F'''{model_prefix}encoder.layer.{i-1}.layernorm.''' ) if F'''layer_{i}.1.conv_proj.''' in k: lowerCamelCase_ =k_new.replace(F'''layer_{i}.1.conv_proj.''' , F'''{model_prefix}encoder.layer.{i-1}.conv_projection.''' ) if "pre_norm_attn.0." in k: lowerCamelCase_ =k_new.replace('''pre_norm_attn.0.''' , '''layernorm_before.''' ) if "pre_norm_attn.1." in k: lowerCamelCase_ =k_new.replace('''pre_norm_attn.1.''' , '''attention.''' ) if "pre_norm_ffn.0." in k: lowerCamelCase_ =k_new.replace('''pre_norm_ffn.0.''' , '''layernorm_after.''' ) if "pre_norm_ffn.1." in k: lowerCamelCase_ =k_new.replace('''pre_norm_ffn.1.''' , '''ffn.conv1.''' ) if "pre_norm_ffn.3." in k: lowerCamelCase_ =k_new.replace('''pre_norm_ffn.3.''' , '''ffn.conv2.''' ) if "classifier.1." in k: lowerCamelCase_ =k_new.replace('''classifier.1.''' , '''classifier.''' ) if "seg_head." in k: lowerCamelCase_ =k_new.replace('''seg_head.''' , '''segmentation_head.''' ) if ".aspp_layer." in k: lowerCamelCase_ =k_new.replace('''.aspp_layer.''' , '''.''' ) if ".aspp_pool." in k: lowerCamelCase_ =k_new.replace('''.aspp_pool.''' , '''.''' ) rename_keys.append((k, k_new) ) return rename_keys def a_ ( __snake_case : Any ) -> Dict: """simple docstring""" lowerCamelCase_ =[] for k in state_dict.keys(): if k.startswith('''seg_head.aux_head.''' ): keys_to_ignore.append(__snake_case ) for k in keys_to_ignore: state_dict.pop(__snake_case , __snake_case ) def a_ ( ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ ='''http://images.cocodataset.org/val2017/000000039769.jpg''' # url = "https://cdn.britannica.com/86/141086-050-9D7C75EE/Gulfstream-G450-business-jet-passengers.jpg" lowerCamelCase_ =Image.open(requests.get(__snake_case , stream=__snake_case ).raw ) return im @torch.no_grad() def a_ ( __snake_case : Dict , __snake_case : str , __snake_case : Optional[int] , __snake_case : Union[str, Any] ) -> str: """simple docstring""" lowerCamelCase_ =get_mobilevitva_config(__snake_case , __snake_case ) # load original state_dict lowerCamelCase_ =torch.load(__snake_case , map_location='''cpu''' ) # load huggingface model if task_name.startswith('''ade20k_''' ) or task_name.startswith('''voc_''' ): lowerCamelCase_ =MobileViTVaForSemanticSegmentation(__snake_case ).eval() lowerCamelCase_ =False else: lowerCamelCase_ =MobileViTVaForImageClassification(__snake_case ).eval() lowerCamelCase_ =False # remove and rename some keys of load the original model lowerCamelCase_ =checkpoint remove_unused_keys(__snake_case ) lowerCamelCase_ =create_rename_keys(__snake_case , base_model=__snake_case ) for rename_key_src, rename_key_dest in rename_keys: rename_key(__snake_case , __snake_case , __snake_case ) # load modified state_dict model.load_state_dict(__snake_case ) # Check outputs on an image, prepared by MobileViTImageProcessor lowerCamelCase_ =MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 32 ) lowerCamelCase_ =image_processor(images=prepare_img() , return_tensors='''pt''' ) lowerCamelCase_ =model(**__snake_case ) # verify classification model if task_name.startswith('''imagenet''' ): lowerCamelCase_ =outputs.logits lowerCamelCase_ =logits.argmax(-1 ).item() print('''Predicted class:''' , model.config.idalabel[predicted_class_idx] ) if task_name.startswith('''imagenet1k_256''' ) and config.width_multiplier == 1.0: # expected_logits for base variant lowerCamelCase_ =torch.tensor([-1.63_36e00, -7.32_04e-02, -5.18_83e-01] ) assert torch.allclose(logits[0, :3] , __snake_case , atol=1e-4 ) Path(__snake_case ).mkdir(exist_ok=__snake_case ) print(F'''Saving model {task_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__snake_case ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__snake_case ) if __name__ == "__main__": a_ : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--task""", default="""imagenet1k_256""", type=str, help=( """Name of the task for which the MobileViTV2 model you'd like to convert is trained on . """ """ Classification (ImageNet-1k) - MobileViTV2 (256x256) : imagenet1k_256 - MobileViTV2 (Trained on 256x256 and Finetuned on 384x384) : imagenet1k_384 - MobileViTV2 (Trained on ImageNet-21k and Finetuned on ImageNet-1k 256x256) : imagenet21k_to_1k_256 - MobileViTV2 (Trained on ImageNet-21k, Finetuned on ImageNet-1k 256x256, and Finetuned on ImageNet-1k 384x384) : imagenet21k_to_1k_384 Segmentation - ADE20K Dataset : ade20k_deeplabv3 - Pascal VOC 2012 Dataset: voc_deeplabv3 """ ), choices=[ """imagenet1k_256""", """imagenet1k_384""", """imagenet21k_to_1k_256""", """imagenet21k_to_1k_384""", """ade20k_deeplabv3""", """voc_deeplabv3""", ], ) parser.add_argument( """--orig_checkpoint_path""", required=True, type=str, help="""Path to the original state dict (.pt file).""" ) parser.add_argument("""--orig_config_path""", required=True, type=str, help="""Path to the original config file.""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, type=str, help="""Path to the output PyTorch model directory.""" ) a_ : Dict = parser.parse_args() convert_mobilevitva_checkpoint( args.task, args.orig_checkpoint_path, args.orig_config_path, args.pytorch_dump_folder_path )
75
'''simple docstring''' import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home __lowerCAmelCase = HUGGINGFACE_HUB_CACHE __lowerCAmelCase = """config.json""" __lowerCAmelCase = """diffusion_pytorch_model.bin""" __lowerCAmelCase = """diffusion_flax_model.msgpack""" __lowerCAmelCase = """model.onnx""" __lowerCAmelCase = """diffusion_pytorch_model.safetensors""" __lowerCAmelCase = """weights.pb""" __lowerCAmelCase = """https://huggingface.co""" __lowerCAmelCase = default_cache_path __lowerCAmelCase = """diffusers_modules""" __lowerCAmelCase = os.getenv("""HF_MODULES_CACHE""", os.path.join(hf_cache_home, """modules""")) __lowerCAmelCase = ["""fp16""", """non-ema"""] __lowerCAmelCase = """.self_attn"""
271
0
# Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from packaging import version from .. import __version__ from .constants import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD from .doc import ( add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, copy_func, replace_return_docstrings, ) from .generic import ( ContextManagers, ExplicitEnum, ModelOutput, PaddingStrategy, TensorType, add_model_info_to_auto_map, cached_property, can_return_loss, expand_dims, find_labels, flatten_dict, infer_framework, is_jax_tensor, is_numpy_array, is_tensor, is_tf_symbolic_tensor, is_tf_tensor, is_torch_device, is_torch_dtype, is_torch_tensor, reshape, squeeze, strtobool, tensor_size, to_numpy, to_py_obj, transpose, working_or_temp_dir, ) from .hub import ( CLOUDFRONT_DISTRIB_PREFIX, DISABLE_TELEMETRY, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, EntryNotFoundError, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, cached_file, default_cache_path, define_sagemaker_information, download_url, extract_commit_hash, get_cached_models, get_file_from_repo, get_full_repo_name, has_file, http_user_agent, is_offline_mode, is_remote_url, move_cache, send_example_telemetry, try_to_load_from_cache, ) from .import_utils import ( ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, TORCH_FX_REQUIRED_VERSION, USE_JAX, USE_TF, USE_TORCH, DummyObject, OptionalDependencyNotAvailable, _LazyModule, ccl_version, direct_transformers_import, get_torch_version, is_accelerate_available, is_apex_available, is_bitsandbytes_available, is_bsa_available, is_coloredlogs_available, is_cython_available, is_datasets_available, is_decord_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_jieba_available, is_jumanpp_available, is_kenlm_available, is_keras_nlp_available, is_librosa_available, is_natten_available, is_ninja_available, is_onnx_available, is_openai_available, is_optimum_available, is_pandas_available, is_peft_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytest_available, is_pytorch_quantization_available, is_rjieba_available, is_sacremoses_available, is_safetensors_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_sudachi_available, is_tensorflow_probability_available, is_tensorflow_text_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_bfaa_cpu_available, is_torch_bfaa_gpu_available, is_torch_compile_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_neuroncore_available, is_torch_tensorrt_fx_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_torchdistx_available, is_torchdynamo_available, is_torchvision_available, is_training_run_on_sagemaker, is_vision_available, requires_backends, torch_only_method, ) a_ = 'pytorch_model.bin' a_ = 'pytorch_model.bin.index.json' a_ = 'adapter_config.json' a_ = 'adapter_model.bin' a_ = 'adapter_model.safetensors' a_ = 'tf_model.h5' a_ = 'tf_model.h5.index.json' a_ = 'model.ckpt' a_ = 'flax_model.msgpack' a_ = 'flax_model.msgpack.index.json' a_ = 'model.safetensors' a_ = 'model.safetensors.index.json' a_ = 'config.json' a_ = 'preprocessor_config.json' a_ = FEATURE_EXTRACTOR_NAME a_ = 'generation_config.json' a_ = 'modelcard.json' a_ = '▁' a_ = SENTENCEPIECE_UNDERLINE # Kept for backward compatibility a_ = [ [[0, 1, 0, 1], [1, 0, 0, 1]] ] * 2 # Needs to have 0s and 1s only since XLM uses it for langs too. a_ = [[7, 6, 0, 0, 1], [1, 2, 3, 0, 0], [0, 0, 0, 4, 5]] a_ = [[1, 1, 1, 1, 1], [1, 1, 1, 0, 0], [0, 0, 0, 1, 1]] def lowerCamelCase__ ( _a): if version.parse(_a) < version.parse(_a): if "dev" in min_version: SCREAMING_SNAKE_CASE : Optional[Any] = ( "This example requires a source install from HuggingFace Transformers (see " "`https://huggingface.co/docs/transformers/installation#install-from-source`)," ) else: SCREAMING_SNAKE_CASE : List[Any] = f"This example requires a minimum version of {min_version}," error_message += f" but the version found is {__version__}.\n" raise ImportError( error_message + "Check out https://github.com/huggingface/transformers/tree/main/examples#important-note for the examples corresponding to other " "versions of HuggingFace Transformers.")
76
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class UpperCAmelCase__ : """simple docstring""" def __init__( self : int ,_a : Any ,_a : Optional[int]=2 ,_a : Optional[Any]=True ,_a : Dict=False ,_a : Dict=10 ,_a : Any=3 ,_a : str=32 * 8 ,_a : Optional[int]=32 * 8 ,_a : int=4 ,_a : str=64 ,): '''simple docstring''' _a : Dict = parent _a : Union[str, Any] = batch_size _a : Tuple = is_training _a : List[str] = use_auxiliary_loss _a : Optional[Any] = num_queries _a : str = num_channels _a : List[str] = min_size _a : int = max_size _a : Optional[int] = num_labels _a : List[str] = hidden_dim _a : int = hidden_dim def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Tuple = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( _a ) _a : Optional[Any] = torch.ones([self.batch_size, self.min_size, self.max_size] ,device=_a ) _a : Union[str, Any] = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] ,device=_a ) > 0.5 ).float() _a : Tuple = (torch.rand((self.batch_size, self.num_labels) ,device=_a ) > 0.5).long() _a : Dict = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : int = MaskaFormerConfig( hidden_size=self.hidden_dim ,) _a : str = self.num_queries _a : Union[str, Any] = self.num_labels _a : Tuple = [1, 1, 1, 1] _a : Dict = self.num_channels _a : str = 64 _a : Tuple = 128 _a : Optional[Any] = self.hidden_dim _a : Union[str, Any] = self.hidden_dim _a : List[Any] = self.hidden_dim return config def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a, _a, _a, _a, _a : Optional[Any] = self.prepare_config_and_inputs() _a : str = {'pixel_values': pixel_values, 'pixel_mask': pixel_mask} return config, inputs_dict def __lowercase ( self : List[str] ,_a : Optional[Any] ,_a : str ): '''simple docstring''' _a : str = output.encoder_hidden_states _a : Any = output.pixel_decoder_hidden_states _a : Optional[Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,config.decoder_layers ) def __lowercase ( self : List[str] ,_a : str ,_a : List[Any] ,_a : Any ,_a : Union[str, Any]=False ): '''simple docstring''' with torch.no_grad(): _a : str = MaskaFormerModel(config=_a ) model.to(_a ) model.eval() _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[Any] = model(_a ,output_hidden_states=_a ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape ,(self.batch_size, self.num_queries, self.hidden_dim) ,) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(_a ,_a ) def __lowercase ( self : Tuple ,_a : List[Any] ,_a : Union[str, Any] ,_a : Tuple ,_a : List[str] ,_a : Any ): '''simple docstring''' _a : int = MaskaFormerForUniversalSegmentation(config=_a ) model.to(_a ) model.eval() def comm_check_on_output(_a : Any ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape ,(self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) ,) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape ,(self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[int] = model(_a ) comm_check_on_output(_a ) _a : List[str] = model( pixel_values=_a ,pixel_mask=_a ,mask_labels=_a ,class_labels=_a ) comm_check_on_output(_a ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape ,torch.Size([1] ) ) @require_torch class UpperCAmelCase__ ( lowercase__ , lowercase__ , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[int] = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () __UpperCAmelCase : Dict = {'''feature-extraction''': MaskaFormerModel} if is_torch_available() else {} __UpperCAmelCase : Dict = False __UpperCAmelCase : Tuple = False __UpperCAmelCase : Dict = False __UpperCAmelCase : List[Any] = False def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Union[str, Any] = MaskaFormerModelTester(self ) _a : Dict = ConfigTester(self ,config_class=_a ,has_text_modality=_a ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : Optional[int] ): '''simple docstring''' _a, _a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*_a ) @unittest.skip(reason='Mask2Former does not use inputs_embeds' ) def __lowercase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not have a get_input_embeddings method' ) def __lowercase ( self : str ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former is not a generative model' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not use token embeddings' ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason='Mask2Former has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def __lowercase ( self : Dict ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass def __lowercase ( self : int ): '''simple docstring''' _a, _a : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Union[str, Any] = model_class(_a ) _a : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a : Optional[Any] = [*signature.parameters.keys()] _a : List[Any] = ['pixel_values'] self.assertListEqual(arg_names[:1] ,_a ) @slow def __lowercase ( self : List[str] ): '''simple docstring''' for model_name in ["facebook/mask2former-swin-small-coco-instance"]: _a : Dict = MaskaFormerModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def __lowercase ( self : List[Any] ): '''simple docstring''' _a : int = (self.model_tester.min_size,) * 2 _a : Any = { 'pixel_values': torch.randn((2, 3, *size) ,device=_a ), 'mask_labels': torch.randn((2, 10, *size) ,device=_a ), 'class_labels': torch.zeros(2 ,10 ,device=_a ).long(), } _a : List[Any] = self.model_tester.get_config() _a : int = MaskaFormerForUniversalSegmentation(_a ).to(_a ) _a : str = model(**_a ) self.assertTrue(outputs.loss is not None ) def __lowercase ( self : List[str] ): '''simple docstring''' _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : int ): '''simple docstring''' _a, _a : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Any = model_class(_a ).to(_a ) _a : Optional[int] = model(**_a ,output_attentions=_a ) self.assertTrue(outputs.attentions is not None ) def __lowercase ( self : Tuple ): '''simple docstring''' if not self.model_tester.is_training: return _a : List[str] = self.all_model_classes[1] _a, _a, _a, _a, _a : List[str] = self.model_tester.prepare_config_and_inputs() _a : Any = model_class(_a ) model.to(_a ) model.train() _a : Union[str, Any] = model(_a ,mask_labels=_a ,class_labels=_a ).loss loss.backward() def __lowercase ( self : int ): '''simple docstring''' _a : int = self.all_model_classes[1] _a, _a, _a, _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs() _a : str = True _a : str = True _a : List[str] = model_class(_a ).to(_a ) model.train() _a : Optional[int] = model(_a ,mask_labels=_a ,class_labels=_a ) _a : Tuple = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _a : str = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() _a : Dict = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _a : List[str] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_a ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __lowerCAmelCase = 1e-4 def UpperCAmelCase_ (): """simple docstring""" _a : int = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_vision @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return "facebook/mask2former-swin-small-coco-instance" @cached_property def __lowercase ( self : Any ): '''simple docstring''' return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def __lowercase ( self : Any ): '''simple docstring''' _a : List[str] = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(_a ) _a : int = self.default_image_processor _a : Tuple = prepare_img() _a : Any = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Union[str, Any] = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[Any] = model(**_a ) _a : List[Any] = torch.tensor( [[-0.2790, -1.0717, -1.1668], [-0.5128, -0.3128, -0.4987], [-0.5832, 0.1971, -0.0197]] ).to(_a ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : str = torch.tensor( [[0.8973, 1.1847, 1.1776], [1.1934, 1.5040, 1.5128], [1.1153, 1.4486, 1.4951]] ).to(_a ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : Any = torch.tensor( [[2.1152, 1.7000, -0.8603], [1.5808, 1.8004, -0.9353], [1.6043, 1.7495, -0.5999]] ).to(_a ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Optional[Any] = self.default_image_processor _a : List[Any] = prepare_img() _a : str = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Any = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[int] = model(**_a ) # masks_queries_logits _a : Dict = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape ,(1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) _a : Dict = [ [-8.7839, -9.0056, -8.8121], [-7.4104, -7.0313, -6.5401], [-6.6105, -6.3427, -6.4675], ] _a : Optional[Any] = torch.tensor(_a ).to(_a ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] ,_a ,atol=_a ) ) # class_queries_logits _a : str = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape ,(1, model.config.num_queries, model.config.num_labels + 1) ) _a : str = torch.tensor( [ [1.8324, -8.0835, -4.1922], [0.8450, -9.0050, -3.6053], [0.3045, -7.7293, -3.0275], ] ).to(_a ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : Any = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Tuple = self.default_image_processor _a : Tuple = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] ,segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] ,return_tensors='pt' ,) _a : str = inputs['pixel_values'].to(_a ) _a : str = [el.to(_a ) for el in inputs['mask_labels']] _a : Dict = [el.to(_a ) for el in inputs['class_labels']] with torch.no_grad(): _a : List[str] = model(**_a ) self.assertTrue(outputs.loss is not None )
271
0
"""simple docstring""" import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class UpperCAmelCase_ ( _a): lowerCamelCase__ : BigBirdConfig lowerCamelCase__ : jnp.dtype = jnp.floataa lowerCamelCase__ : bool = True def _UpperCAmelCase ( self ) -> Optional[int]: super().setup() lowercase__ : Optional[int] = nn.Dense(5 , dtype=self.dtype ) def __call__( self , *a , **a ) -> Union[str, Any]: lowercase__ : List[str] = super().__call__(*a , **a ) lowercase__ : str = self.cls(outputs[2] ) return outputs[:2] + (cls_out,) class UpperCAmelCase_ ( _a): lowerCamelCase__ : Optional[int] = FlaxBigBirdForNaturalQuestionsModule def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str ): '''simple docstring''' def cross_entropy(_lowerCAmelCase : int , _lowerCAmelCase : Tuple , _lowerCAmelCase : str=None ): lowercase__ : List[str] = logits.shape[-1] lowercase__ : Any = (labels[..., None] == jnp.arange(_lowerCAmelCase )[None]).astype('f4' ) lowercase__ : Any = jax.nn.log_softmax(_lowerCAmelCase , axis=-1 ) lowercase__ : str = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: lowercase__ : List[Any] = reduction(_lowerCAmelCase ) return loss lowercase__ : int = partial(_lowerCAmelCase , reduction=jnp.mean ) lowercase__ : Tuple = cross_entropy(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : int = cross_entropy(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[Any] = cross_entropy(_lowerCAmelCase , _lowerCAmelCase ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class UpperCAmelCase_ : lowerCamelCase__ : str = "google/bigbird-roberta-base" lowerCamelCase__ : int = 3_0_0_0 lowerCamelCase__ : int = 1_0_5_0_0 lowerCamelCase__ : int = 1_2_8 lowerCamelCase__ : int = 3 lowerCamelCase__ : int = 1 lowerCamelCase__ : int = 5 # tx_args lowerCamelCase__ : float = 3E-5 lowerCamelCase__ : float = 0.0 lowerCamelCase__ : int = 2_0_0_0_0 lowerCamelCase__ : float = 0.0095 lowerCamelCase__ : str = "bigbird-roberta-natural-questions" lowerCamelCase__ : str = "training-expt" lowerCamelCase__ : str = "data/nq-training.jsonl" lowerCamelCase__ : str = "data/nq-validation.jsonl" def _UpperCAmelCase ( self ) -> Any: os.makedirs(self.base_dir , exist_ok=a ) lowercase__ : Tuple = os.path.join(self.base_dir , self.save_dir ) lowercase__ : int = self.batch_size_per_device * jax.device_count() @dataclass class UpperCAmelCase_ : lowerCamelCase__ : int lowerCamelCase__ : int = 4_0_9_6 # no dynamic padding on TPUs def __call__( self , a ) -> List[str]: lowercase__ : Any = self.collate_fn(a ) lowercase__ : Dict = jax.tree_util.tree_map(a , a ) return batch def _UpperCAmelCase ( self , a ) -> Union[str, Any]: lowercase__ , lowercase__ : Any = self.fetch_inputs(features['input_ids'] ) lowercase__ : Optional[Any] = { 'input_ids': jnp.array(a , dtype=jnp.intaa ), 'attention_mask': jnp.array(a , dtype=jnp.intaa ), 'start_labels': jnp.array(features['start_token'] , dtype=jnp.intaa ), 'end_labels': jnp.array(features['end_token'] , dtype=jnp.intaa ), 'pooled_labels': jnp.array(features['category'] , dtype=jnp.intaa ), } return batch def _UpperCAmelCase ( self , a ) -> List[str]: lowercase__ : List[Any] = [self._fetch_inputs(a ) for ids in input_ids] return zip(*a ) def _UpperCAmelCase ( self , a ) -> Tuple: lowercase__ : Any = [1 for _ in range(len(a ) )] while len(a ) < self.max_length: input_ids.append(self.pad_id ) attention_mask.append(0 ) return input_ids, attention_mask def a_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any]=None ): '''simple docstring''' if seed is not None: lowercase__ : Tuple = dataset.shuffle(seed=_lowerCAmelCase ) for i in range(len(_lowerCAmelCase ) // batch_size ): lowercase__ : Optional[int] = dataset[i * batch_size : (i + 1) * batch_size] yield dict(_lowerCAmelCase ) @partial(jax.pmap , axis_name='batch' ) def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Union[str, Any] , **_lowerCAmelCase : Optional[int] ): '''simple docstring''' def loss_fn(_lowerCAmelCase : Union[str, Any] ): lowercase__ : Any = model_inputs.pop('start_labels' ) lowercase__ : List[str] = model_inputs.pop('end_labels' ) lowercase__ : Any = model_inputs.pop('pooled_labels' ) lowercase__ : Dict = state.apply_fn(**_lowerCAmelCase , params=_lowerCAmelCase , dropout_rng=_lowerCAmelCase , train=_lowerCAmelCase ) lowercase__ , lowercase__ , lowercase__ : List[Any] = outputs return state.loss_fn( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) lowercase__ , lowercase__ : Dict = jax.random.split(_lowerCAmelCase ) lowercase__ : Dict = jax.value_and_grad(_lowerCAmelCase ) lowercase__ , lowercase__ : Dict = grad_fn(state.params ) lowercase__ : Optional[Any] = jax.lax.pmean({'loss': loss} , axis_name='batch' ) lowercase__ : Tuple = jax.lax.pmean(_lowerCAmelCase , 'batch' ) lowercase__ : Tuple = state.apply_gradients(grads=_lowerCAmelCase ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name='batch' ) def a_ ( _lowerCAmelCase : Tuple , **_lowerCAmelCase : List[str] ): '''simple docstring''' lowercase__ : List[str] = model_inputs.pop('start_labels' ) lowercase__ : int = model_inputs.pop('end_labels' ) lowercase__ : str = model_inputs.pop('pooled_labels' ) lowercase__ : Optional[int] = state.apply_fn(**_lowerCAmelCase , params=state.params , train=_lowerCAmelCase ) lowercase__ , lowercase__ , lowercase__ : List[str] = outputs lowercase__ : Optional[Any] = state.loss_fn(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : int = jax.lax.pmean({'loss': loss} , axis_name='batch' ) return metrics class UpperCAmelCase_ ( train_state.TrainState): lowerCamelCase__ : Callable = struct.field(pytree_node=_a) @dataclass class UpperCAmelCase_ : lowerCamelCase__ : Args lowerCamelCase__ : Callable lowerCamelCase__ : Callable lowerCamelCase__ : Callable lowerCamelCase__ : Callable lowerCamelCase__ : wandb lowerCamelCase__ : Callable = None def _UpperCAmelCase ( self , a , a , a , a=None ) -> Dict: lowercase__ : Optional[int] = model.params lowercase__ : Tuple = TrainState.create( apply_fn=model.__call__ , params=a , tx=a , loss_fn=a , ) if ckpt_dir is not None: lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ : Optional[Any] = restore_checkpoint(a , a ) lowercase__ : List[str] = { 'lr': args.lr, 'init_lr': args.init_lr, 'warmup_steps': args.warmup_steps, 'num_train_steps': num_train_steps, 'weight_decay': args.weight_decay, } lowercase__ , lowercase__ : str = build_tx(**a ) lowercase__ : Tuple = train_state.TrainState( step=a , apply_fn=model.__call__ , params=a , tx=a , opt_state=a , ) lowercase__ : Optional[int] = args lowercase__ : Any = data_collator lowercase__ : Union[str, Any] = lr lowercase__ : List[Any] = params lowercase__ : Tuple = jax_utils.replicate(a ) return state def _UpperCAmelCase ( self , a , a , a ) -> Optional[int]: lowercase__ : str = self.args lowercase__ : str = len(a ) // args.batch_size lowercase__ : List[Any] = jax.random.PRNGKey(0 ) lowercase__ : Tuple = jax.random.split(a , jax.device_count() ) for epoch in range(args.max_epochs ): lowercase__ : Any = jnp.array(0 , dtype=jnp.floataa ) lowercase__ : List[Any] = get_batched_dataset(a , args.batch_size , seed=a ) lowercase__ : int = 0 for batch in tqdm(a , total=a , desc=f"""Running EPOCH-{epoch}""" ): lowercase__ : Dict = self.data_collator(a ) lowercase__ , lowercase__ , lowercase__ : List[Any] = self.train_step_fn(a , a , **a ) running_loss += jax_utils.unreplicate(metrics['loss'] ) i += 1 if i % args.logging_steps == 0: lowercase__ : int = jax_utils.unreplicate(state.step ) lowercase__ : Dict = running_loss.item() / i lowercase__ : List[Any] = self.scheduler_fn(state_step - 1 ) lowercase__ : Dict = self.evaluate(a , a ) lowercase__ : Tuple = { 'step': state_step.item(), 'eval_loss': eval_loss.item(), 'tr_loss': tr_loss, 'lr': lr.item(), } tqdm.write(str(a ) ) self.logger.log(a , commit=a ) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f"""-e{epoch}-s{i}""" , state=a ) def _UpperCAmelCase ( self , a , a ) -> List[Any]: lowercase__ : Any = get_batched_dataset(a , self.args.batch_size ) lowercase__ : Optional[int] = len(a ) // self.args.batch_size lowercase__ : Union[str, Any] = jnp.array(0 , dtype=jnp.floataa ) lowercase__ : Optional[int] = 0 for batch in tqdm(a , total=a , desc='Evaluating ... ' ): lowercase__ : int = self.data_collator(a ) lowercase__ : Optional[int] = self.val_step_fn(a , **a ) running_loss += jax_utils.unreplicate(metrics['loss'] ) i += 1 return running_loss / i def _UpperCAmelCase ( self , a , a ) -> List[Any]: lowercase__ : int = jax_utils.unreplicate(a ) print(f"""SAVING CHECKPOINT IN {save_dir}""" , end=' ... ' ) self.model_save_fn(a , params=state.params ) with open(os.path.join(a , 'opt_state.msgpack' ) , 'wb' ) as f: f.write(to_bytes(state.opt_state ) ) joblib.dump(self.args , os.path.join(a , 'args.joblib' ) ) joblib.dump(self.data_collator , os.path.join(a , 'data_collator.joblib' ) ) with open(os.path.join(a , 'training_state.json' ) , 'w' ) as f: json.dump({'step': state.step.item()} , a ) print('DONE' ) def a_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : Any ): '''simple docstring''' print(f"""RESTORING CHECKPOINT FROM {save_dir}""" , end=' ... ' ) with open(os.path.join(_lowerCAmelCase , 'flax_model.msgpack' ) , 'rb' ) as f: lowercase__ : Tuple = from_bytes(state.params , f.read() ) with open(os.path.join(_lowerCAmelCase , 'opt_state.msgpack' ) , 'rb' ) as f: lowercase__ : List[str] = from_bytes(state.opt_state , f.read() ) lowercase__ : int = joblib.load(os.path.join(_lowerCAmelCase , 'args.joblib' ) ) lowercase__ : Optional[Any] = joblib.load(os.path.join(_lowerCAmelCase , 'data_collator.joblib' ) ) with open(os.path.join(_lowerCAmelCase , 'training_state.json' ) , 'r' ) as f: lowercase__ : Union[str, Any] = json.load(_lowerCAmelCase ) lowercase__ : Tuple = training_state['step'] print('DONE' ) return params, opt_state, step, args, data_collator def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Dict , _lowerCAmelCase : Any , _lowerCAmelCase : Any ): '''simple docstring''' lowercase__ : Union[str, Any] = num_train_steps - warmup_steps lowercase__ : int = optax.linear_schedule(init_value=_lowerCAmelCase , end_value=_lowerCAmelCase , transition_steps=_lowerCAmelCase ) lowercase__ : Optional[Any] = optax.linear_schedule(init_value=_lowerCAmelCase , end_value=1E-7 , transition_steps=_lowerCAmelCase ) lowercase__ : int = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : int ): '''simple docstring''' def weight_decay_mask(_lowerCAmelCase : Optional[int] ): lowercase__ : Tuple = traverse_util.flatten_dict(_lowerCAmelCase ) lowercase__ : Any = {k: (v[-1] != 'bias' and v[-2:] != ('LayerNorm', 'scale')) for k, v in params.items()} return traverse_util.unflatten_dict(_lowerCAmelCase ) lowercase__ : List[str] = scheduler_fn(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[Any] = optax.adamw(learning_rate=_lowerCAmelCase , weight_decay=_lowerCAmelCase , mask=_lowerCAmelCase ) return tx, lr
77
'''simple docstring''' import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def UpperCAmelCase_ (__a : List[Any] ): """simple docstring""" if ( (cp >= 0x4E_00 and cp <= 0x9F_FF) or (cp >= 0x34_00 and cp <= 0x4D_BF) # or (cp >= 0x2_00_00 and cp <= 0x2_A6_DF) # or (cp >= 0x2_A7_00 and cp <= 0x2_B7_3F) # or (cp >= 0x2_B7_40 and cp <= 0x2_B8_1F) # or (cp >= 0x2_B8_20 and cp <= 0x2_CE_AF) # or (cp >= 0xF9_00 and cp <= 0xFA_FF) or (cp >= 0x2_F8_00 and cp <= 0x2_FA_1F) # ): # return True return False def UpperCAmelCase_ (__a : str ): """simple docstring""" for char in word: _a : Union[str, Any] = ord(__a ) if not _is_chinese_char(__a ): return 0 return 1 def UpperCAmelCase_ (__a : List[str] ): """simple docstring""" _a : Dict = set() for token in tokens: _a : str = len(__a ) > 1 and is_chinese(__a ) if chinese_word: word_set.add(__a ) _a : Optional[Any] = list(__a ) return word_list def UpperCAmelCase_ (__a : List[str] , __a : set() ): """simple docstring""" if not chinese_word_set: return bert_tokens _a : Optional[Any] = max([len(__a ) for w in chinese_word_set] ) _a : Optional[int] = bert_tokens _a, _a : Any = 0, len(__a ) while start < end: _a : Tuple = True if is_chinese(bert_word[start] ): _a : Union[str, Any] = min(end - start , __a ) for i in range(__a , 1 , -1 ): _a : Optional[Any] = ''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): _a : Any = '##' + bert_word[j] _a : Union[str, Any] = start + i _a : int = False break if single_word: start += 1 return bert_word def UpperCAmelCase_ (__a : List[str] , __a : LTP , __a : BertTokenizer ): """simple docstring""" _a : int = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : Union[str, Any] = ltp_tokenizer.seg(lines[i : i + 1_0_0] )[0] _a : Optional[Any] = [get_chinese_word(__a ) for r in res] ltp_res.extend(__a ) assert len(__a ) == len(__a ) _a : str = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : List[str] = bert_tokenizer(lines[i : i + 1_0_0] , add_special_tokens=__a , truncation=__a , max_length=5_1_2 ) bert_res.extend(res['input_ids'] ) assert len(__a ) == len(__a ) _a : List[str] = [] for input_ids, chinese_word in zip(__a , __a ): _a : int = [] for id in input_ids: _a : Optional[int] = bert_tokenizer._convert_id_to_token(__a ) input_tokens.append(__a ) _a : List[str] = add_sub_symbol(__a , __a ) _a : Tuple = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(__a ): if token[:2] == "##": _a : str = token[2:] # save chinese tokens' pos if len(__a ) == 1 and _is_chinese_char(ord(__a ) ): ref_id.append(__a ) ref_ids.append(__a ) assert len(__a ) == len(__a ) return ref_ids def UpperCAmelCase_ (__a : Optional[Any] ): """simple docstring""" with open(args.file_name , 'r' , encoding='utf-8' ) as f: _a : Dict = f.readlines() _a : int = [line.strip() for line in data if len(__a ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' _a : int = LTP(args.ltp ) # faster in GPU device _a : Tuple = BertTokenizer.from_pretrained(args.bert ) _a : int = prepare_ref(__a , __a , __a ) with open(args.save_path , 'w' , encoding='utf-8' ) as f: _a : Optional[Any] = [json.dumps(__a ) + '\n' for ref in ref_ids] f.writelines(__a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser(description="""prepare_chinese_ref""") parser.add_argument( """--file_name""", type=str, default="""./resources/chinese-demo.txt""", help="""file need process, same as training data in lm""", ) parser.add_argument( """--ltp""", type=str, default="""./resources/ltp""", help="""resources for LTP tokenizer, usually a path""" ) parser.add_argument("""--bert""", type=str, default="""./resources/robert""", help="""resources for Bert tokenizer""") parser.add_argument("""--save_path""", type=str, default="""./resources/ref.txt""", help="""path to save res""") __lowerCAmelCase = parser.parse_args() main(args)
271
0
"""simple docstring""" import json import logging import os import socket import git import numpy as np import torch logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - PID: %(process)d - %(message)s""", datefmt="""%m/%d/%Y %H:%M:%S""", level=logging.INFO, ) snake_case_ = logging.getLogger(__name__) def _lowerCAmelCase ( lowercase_ ): UpperCAmelCase = git.Repo(search_parent_directories=lowercase_ ) UpperCAmelCase = { 'repo_id': str(lowercase_ ), 'repo_sha': str(repo.head.object.hexsha ), 'repo_branch': str(repo.active_branch ), } with open(os.path.join(lowercase_ , 'git_log.json' ) , 'w' ) as f: json.dump(lowercase_ , lowercase_ , indent=4 ) def _lowerCAmelCase ( lowercase_ ): if params.n_gpu <= 0: UpperCAmelCase = 0 UpperCAmelCase = -1 UpperCAmelCase = True UpperCAmelCase = False return assert torch.cuda.is_available() logger.info('Initializing GPUs' ) if params.n_gpu > 1: assert params.local_rank != -1 UpperCAmelCase = int(os.environ['WORLD_SIZE'] ) UpperCAmelCase = int(os.environ['N_GPU_NODE'] ) UpperCAmelCase = int(os.environ['RANK'] ) # number of nodes / node ID UpperCAmelCase = params.world_size // params.n_gpu_per_node UpperCAmelCase = params.global_rank // params.n_gpu_per_node UpperCAmelCase = True assert params.n_nodes == int(os.environ['N_NODES'] ) assert params.node_id == int(os.environ['NODE_RANK'] ) # local job (single GPU) else: assert params.local_rank == -1 UpperCAmelCase = 1 UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 1 UpperCAmelCase = 1 UpperCAmelCase = False # sanity checks assert params.n_nodes >= 1 assert 0 <= params.node_id < params.n_nodes assert 0 <= params.local_rank <= params.global_rank < params.world_size assert params.world_size == params.n_nodes * params.n_gpu_per_node # define whether this is the master process / if we are in multi-node distributed mode UpperCAmelCase = params.node_id == 0 and params.local_rank == 0 UpperCAmelCase = params.n_nodes > 1 # summary UpperCAmelCase = F"""--- Global rank: {params.global_rank} - """ logger.info(PREFIX + 'Number of nodes: %i' % params.n_nodes ) logger.info(PREFIX + 'Node ID : %i' % params.node_id ) logger.info(PREFIX + 'Local rank : %i' % params.local_rank ) logger.info(PREFIX + 'World size : %i' % params.world_size ) logger.info(PREFIX + 'GPUs per node : %i' % params.n_gpu_per_node ) logger.info(PREFIX + 'Master : %s' % str(params.is_master ) ) logger.info(PREFIX + 'Multi-node : %s' % str(params.multi_node ) ) logger.info(PREFIX + 'Multi-GPU : %s' % str(params.multi_gpu ) ) logger.info(PREFIX + 'Hostname : %s' % socket.gethostname() ) # set GPU device torch.cuda.set_device(params.local_rank ) # initialize multi-GPU if params.multi_gpu: logger.info('Initializing PyTorch distributed' ) torch.distributed.init_process_group( init_method='env://' , backend='nccl' , ) def _lowerCAmelCase ( lowercase_ ): np.random.seed(args.seed ) torch.manual_seed(args.seed ) if args.n_gpu > 0: torch.cuda.manual_seed_all(args.seed )
78
'''simple docstring''' import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Tuple ,*_a : List[str] ,**_a : Any ): '''simple docstring''' warnings.warn( 'The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use VideoMAEImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
'''simple docstring''' from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeling_utils import PreTrainedModel from ...utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging from .configuration_mobilenet_va import MobileNetVaConfig lowerCamelCase_ = logging.get_logger(__name__) # General docstring lowerCamelCase_ = '''MobileNetV1Config''' # Base docstring lowerCamelCase_ = '''google/mobilenet_v1_1.0_224''' lowerCamelCase_ = [1, 10_24, 7, 7] # Image classification docstring lowerCamelCase_ = '''google/mobilenet_v1_1.0_224''' lowerCamelCase_ = '''tabby, tabby cat''' lowerCamelCase_ = [ '''google/mobilenet_v1_1.0_224''', '''google/mobilenet_v1_0.75_192''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 ] def __lowercase ( __lowercase , __lowercase , __lowercase=None ) -> List[str]: '''simple docstring''' _A = {} if isinstance(__lowercase , __lowercase ): _A = model.mobilenet_va else: _A = model _A = "MobilenetV1/Conv2d_0/" _A = backbone.conv_stem.convolution.weight _A = backbone.conv_stem.normalization.bias _A = backbone.conv_stem.normalization.weight _A = backbone.conv_stem.normalization.running_mean _A = backbone.conv_stem.normalization.running_var for i in range(13 ): _A = i + 1 _A = i * 2 _A = backbone.layer[pt_index] _A = F'''MobilenetV1/Conv2d_{tf_index}_depthwise/''' _A = pointer.convolution.weight _A = pointer.normalization.bias _A = pointer.normalization.weight _A = pointer.normalization.running_mean _A = pointer.normalization.running_var _A = backbone.layer[pt_index + 1] _A = F'''MobilenetV1/Conv2d_{tf_index}_pointwise/''' _A = pointer.convolution.weight _A = pointer.normalization.bias _A = pointer.normalization.weight _A = pointer.normalization.running_mean _A = pointer.normalization.running_var if isinstance(__lowercase , __lowercase ): _A = "MobilenetV1/Logits/Conv2d_1c_1x1/" _A = model.classifier.weight _A = model.classifier.bias return tf_to_pt_map def __lowercase ( __lowercase , __lowercase , __lowercase ) -> Dict: '''simple docstring''' try: import numpy as np import tensorflow as tf except ImportError: logger.error( "Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see " "https://www.tensorflow.org/install/ for installation instructions." ) raise # Load weights from TF model _A = tf.train.list_variables(__lowercase ) _A = {} for name, shape in init_vars: logger.info(F'''Loading TF weight {name} with shape {shape}''' ) _A = tf.train.load_variable(__lowercase , __lowercase ) _A = array # Build TF to PyTorch weights loading map _A = _build_tf_to_pytorch_map(__lowercase , __lowercase , __lowercase ) for name, pointer in tf_to_pt_map.items(): logger.info(F'''Importing {name}''' ) if name not in tf_weights: logger.info(F'''{name} not in tf pre-trained weights, skipping''' ) continue _A = tf_weights[name] if "depthwise_weights" in name: logger.info("Transposing depthwise" ) _A = np.transpose(__lowercase , (2, 3, 0, 1) ) elif "weights" in name: logger.info("Transposing" ) if len(pointer.shape ) == 2: # copying into linear layer _A = array.squeeze().transpose() else: _A = np.transpose(__lowercase , (3, 2, 0, 1) ) if pointer.shape != array.shape: raise ValueError(F'''Pointer shape {pointer.shape} and array shape {array.shape} mismatched''' ) logger.info(F'''Initialize PyTorch weight {name} {array.shape}''' ) _A = torch.from_numpy(__lowercase ) tf_weights.pop(__lowercase , __lowercase ) tf_weights.pop(name + "/RMSProp" , __lowercase ) tf_weights.pop(name + "/RMSProp_1" , __lowercase ) tf_weights.pop(name + "/ExponentialMovingAverage" , __lowercase ) logger.info(F'''Weights not copied to PyTorch model: {', '.join(tf_weights.keys() )}''' ) return model def __lowercase ( __lowercase , __lowercase ) -> torch.Tensor: '''simple docstring''' _A , _A = features.shape[-2:] _A , _A = conv_layer.stride _A , _A = conv_layer.kernel_size if in_height % stride_height == 0: _A = max(kernel_height - stride_height , 0 ) else: _A = max(kernel_height - (in_height % stride_height) , 0 ) if in_width % stride_width == 0: _A = max(kernel_width - stride_width , 0 ) else: _A = max(kernel_width - (in_width % stride_width) , 0 ) _A = pad_along_width // 2 _A = pad_along_width - pad_left _A = pad_along_height // 2 _A = pad_along_height - pad_top _A = (pad_left, pad_right, pad_top, pad_bottom) return nn.functional.pad(__lowercase , __lowercase , "constant" , 0.0 ) class _UpperCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : int , __UpperCAmelCase : MobileNetVaConfig , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : Optional[int] = 1 , __UpperCAmelCase : Optional[int] = 1 , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[bool] = True , __UpperCAmelCase : Optional[bool or str] = True , ): '''simple docstring''' super().__init__() _A = config if in_channels % groups != 0: raise ValueError(f'''Input channels ({in_channels}) are not divisible by {groups} groups.''' ) if out_channels % groups != 0: raise ValueError(f'''Output channels ({out_channels}) are not divisible by {groups} groups.''' ) _A = 0 if config.tf_padding else int((kernel_size - 1) / 2 ) _A = nn.Convad( in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , kernel_size=__UpperCAmelCase , stride=__UpperCAmelCase , padding=__UpperCAmelCase , groups=__UpperCAmelCase , bias=__UpperCAmelCase , padding_mode="zeros" , ) if use_normalization: _A = nn.BatchNormad( num_features=__UpperCAmelCase , eps=config.layer_norm_eps , momentum=0.9997 , affine=__UpperCAmelCase , track_running_stats=__UpperCAmelCase , ) else: _A = None if use_activation: if isinstance(__UpperCAmelCase , __UpperCAmelCase ): _A = ACTaFN[use_activation] elif isinstance(config.hidden_act , __UpperCAmelCase ): _A = ACTaFN[config.hidden_act] else: _A = config.hidden_act else: _A = None def lowerCAmelCase ( self : str , __UpperCAmelCase : torch.Tensor ): '''simple docstring''' if self.config.tf_padding: _A = apply_tf_padding(__UpperCAmelCase , self.convolution ) _A = self.convolution(__UpperCAmelCase ) if self.normalization is not None: _A = self.normalization(__UpperCAmelCase ) if self.activation is not None: _A = self.activation(__UpperCAmelCase ) return features class _UpperCAmelCase ( snake_case_ ): """simple docstring""" snake_case = MobileNetVaConfig snake_case = load_tf_weights_in_mobilenet_va snake_case = '''mobilenet_v1''' snake_case = '''pixel_values''' snake_case = False def lowerCAmelCase ( self : List[str] , __UpperCAmelCase : Union[nn.Linear, nn.Convad] ): '''simple docstring''' if isinstance(__UpperCAmelCase , (nn.Linear, nn.Convad) ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() elif isinstance(__UpperCAmelCase , nn.BatchNormad ): module.bias.data.zero_() module.weight.data.fill_(1.0 ) lowerCamelCase_ = r''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`MobileNetV1Config`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' lowerCamelCase_ = r''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`MobileNetV1ImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( '''The bare MobileNetV1 model outputting raw hidden-states without any specific head on top.''' , snake_case_ , ) class _UpperCAmelCase ( snake_case_ ): """simple docstring""" def __init__( self : List[Any] , __UpperCAmelCase : MobileNetVaConfig , __UpperCAmelCase : bool = True ): '''simple docstring''' super().__init__(__UpperCAmelCase ) _A = config _A = 32 _A = max(int(depth * config.depth_multiplier ) , config.min_depth ) _A = MobileNetVaConvLayer( __UpperCAmelCase , in_channels=config.num_channels , out_channels=__UpperCAmelCase , kernel_size=3 , stride=2 , ) _A = [1, 2, 1, 2, 1, 2, 1, 1, 1, 1, 1, 2, 1] _A = nn.ModuleList() for i in range(13 ): _A = out_channels if strides[i] == 2 or i == 0: depth *= 2 _A = max(int(depth * config.depth_multiplier ) , config.min_depth ) self.layer.append( MobileNetVaConvLayer( __UpperCAmelCase , in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , kernel_size=3 , stride=strides[i] , groups=__UpperCAmelCase , ) ) self.layer.append( MobileNetVaConvLayer( __UpperCAmelCase , in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , kernel_size=1 , ) ) _A = nn.AdaptiveAvgPoolad((1, 1) ) if add_pooling_layer else None # Initialize weights and apply final processing self.post_init() def lowerCAmelCase ( self : List[str] , __UpperCAmelCase : List[str] ): '''simple docstring''' raise NotImplementedError @add_start_docstrings_to_model_forward(__UpperCAmelCase ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=__UpperCAmelCase , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def lowerCAmelCase ( self : List[str] , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None , ): '''simple docstring''' _A = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _A = return_dict if return_dict is not None else self.config.use_return_dict if pixel_values is None: raise ValueError("You have to specify pixel_values" ) _A = self.conv_stem(__UpperCAmelCase ) _A = () if output_hidden_states else None for i, layer_module in enumerate(self.layer ): _A = layer_module(__UpperCAmelCase ) if output_hidden_states: _A = all_hidden_states + (hidden_states,) _A = hidden_states if self.pooler is not None: _A = torch.flatten(self.pooler(__UpperCAmelCase ) , start_dim=1 ) else: _A = None if not return_dict: return tuple(v for v in [last_hidden_state, pooled_output, all_hidden_states] if v is not None ) return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=__UpperCAmelCase , pooler_output=__UpperCAmelCase , hidden_states=__UpperCAmelCase , ) @add_start_docstrings( ''' MobileNetV1 model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. ''' , snake_case_ , ) class _UpperCAmelCase ( snake_case_ ): """simple docstring""" def __init__( self : List[str] , __UpperCAmelCase : MobileNetVaConfig ): '''simple docstring''' super().__init__(__UpperCAmelCase ) _A = config.num_labels _A = MobileNetVaModel(__UpperCAmelCase ) _A = self.mobilenet_va.layer[-1].convolution.out_channels # Classifier head _A = nn.Dropout(config.classifier_dropout_prob , inplace=__UpperCAmelCase ) _A = nn.Linear(__UpperCAmelCase , config.num_labels ) if config.num_labels > 0 else nn.Identity() # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(__UpperCAmelCase ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=__UpperCAmelCase , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def lowerCAmelCase ( self : Union[str, Any] , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[bool] = None , ): '''simple docstring''' _A = return_dict if return_dict is not None else self.config.use_return_dict _A = self.mobilenet_va(__UpperCAmelCase , output_hidden_states=__UpperCAmelCase , return_dict=__UpperCAmelCase ) _A = outputs.pooler_output if return_dict else outputs[1] _A = self.classifier(self.dropout(__UpperCAmelCase ) ) _A = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: _A = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): _A = "single_label_classification" else: _A = "multi_label_classification" if self.config.problem_type == "regression": _A = MSELoss() if self.num_labels == 1: _A = loss_fct(logits.squeeze() , labels.squeeze() ) else: _A = loss_fct(__UpperCAmelCase , __UpperCAmelCase ) elif self.config.problem_type == "single_label_classification": _A = CrossEntropyLoss() _A = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": _A = BCEWithLogitsLoss() _A = loss_fct(__UpperCAmelCase , __UpperCAmelCase ) if not return_dict: _A = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return ImageClassifierOutputWithNoAttention( loss=__UpperCAmelCase , logits=__UpperCAmelCase , hidden_states=outputs.hidden_states , )
79
'''simple docstring''' from __future__ import annotations from random import choice def UpperCAmelCase_ (__a : str ): """simple docstring""" return choice(__a ) def UpperCAmelCase_ (__a : list[int] , __a : int ): """simple docstring""" _a : Dict = random_pivot(__a ) # partition based on pivot # linear time _a : Optional[int] = [e for e in lst if e < pivot] _a : List[str] = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(__a ) == k - 1: return pivot # pivot is in elements bigger than k elif len(__a ) < k - 1: return kth_number(__a , k - len(__a ) - 1 ) # pivot is in elements smaller than k else: return kth_number(__a , __a ) if __name__ == "__main__": import doctest doctest.testmod()
271
0
'''simple docstring''' import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class lowercase_ ( unittest.TestCase ): def __a ( self ): debug_launcher(test_script.main ) def __a ( self ): debug_launcher(test_ops.main )
80
'''simple docstring''' class UpperCAmelCase__ : """simple docstring""" def __init__( self : Dict ): '''simple docstring''' _a : Dict = {} def __lowercase ( self : Union[str, Any] ): '''simple docstring''' print(self.vertex ) for i in self.vertex: print(_a ,' -> ' ,' -> '.join([str(_a ) for j in self.vertex[i]] ) ) def __lowercase ( self : Dict ,_a : int ,_a : int ): '''simple docstring''' if from_vertex in self.vertex: self.vertex[from_vertex].append(_a ) else: # else make a new vertex _a : int = [to_vertex] def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Tuple = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(_a ,_a ) def __lowercase ( self : Union[str, Any] ,_a : int ,_a : list ): '''simple docstring''' _a : List[Any] = True print(_a ,end=' ' ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(_a ,_a ) if __name__ == "__main__": __lowerCAmelCase = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print("""DFS:""") g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
271
0
"""simple docstring""" import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class __A ( unittest.TestCase ): """simple docstring""" @slow def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: a =AutoImageProcessor.from_pretrained('''microsoft/dit-base-finetuned-rvlcdip''' ) a =AutoModelForImageClassification.from_pretrained('''microsoft/dit-base-finetuned-rvlcdip''' ) model.to(__A ) from datasets import load_dataset a =load_dataset('''nielsr/rvlcdip-demo''' ) a =dataset['''train'''][0]['''image'''].convert('''RGB''' ) a =image_processor(__A , return_tensors='''pt''' ).to(__A ) # forward pass with torch.no_grad(): a =model(**__A ) a =outputs.logits a =torch.Size((1, 16) ) self.assertEqual(logits.shape , __A ) a =torch.tensor( [-0.4_158, -0.4_092, -0.4_347] , device=__A , dtype=torch.float , ) self.assertTrue(torch.allclose(logits[0, :3] , __A , atol=1E-4 ) )
81
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = """▁""" __lowerCAmelCase = {"""vocab_file""": """sentencepiece.bpe.model""", """monolingual_vocab_file""": """dict.txt"""} __lowerCAmelCase = { """vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model""", }, """monolingual_vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt""", }, } __lowerCAmelCase = {"""vinai/bartpho-syllable""": 1_0_2_4} class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Optional[Any] = VOCAB_FILES_NAMES __UpperCAmelCase : Dict = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ['''input_ids''', '''attention_mask'''] def __init__( self : str ,_a : str ,_a : Any ,_a : Any="<s>" ,_a : Dict="</s>" ,_a : int="</s>" ,_a : Union[str, Any]="<s>" ,_a : List[Any]="<unk>" ,_a : Optional[Any]="<pad>" ,_a : List[str]="<mask>" ,_a : Optional[Dict[str, Any]] = None ,**_a : int ,): '''simple docstring''' _a : Any = AddedToken(_a ,lstrip=_a ,rstrip=_a ) if isinstance(_a ,_a ) else mask_token _a : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_a ,eos_token=_a ,unk_token=_a ,sep_token=_a ,cls_token=_a ,pad_token=_a ,mask_token=_a ,sp_model_kwargs=self.sp_model_kwargs ,**_a ,) _a : Optional[int] = vocab_file _a : Union[str, Any] = monolingual_vocab_file _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_a ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility _a : Union[str, Any] = {} _a : int = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(_a ) not in self.fairseq_tokens_to_ids: _a : int = cnt cnt += 1 with open(_a ,'r' ,encoding='utf-8' ) as f: for line in f.readlines(): _a : str = line.strip().split()[0] _a : Tuple = len(self.fairseq_tokens_to_ids ) if str(_a ) not in self.fairseq_tokens_to_ids: _a : List[str] = len(self.fairseq_tokens_to_ids ) _a : Tuple = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Union[str, Any] ): '''simple docstring''' _a : int = self.__dict__.copy() _a : str = None _a : Optional[Any] = self.sp_model.serialized_model_proto() return state def __setstate__( self : Tuple ,_a : Tuple ): '''simple docstring''' _a : Tuple = d # for backward compatibility if not hasattr(self ,'sp_model_kwargs' ): _a : List[str] = {} _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def __lowercase ( self : Dict ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _a : Dict = [self.cls_token_id] _a : int = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __lowercase ( self : int ,_a : List[int] ,_a : Optional[List[int]] = None ,_a : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a ,token_ids_a=_a ,already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def __lowercase ( self : Tuple ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' _a : List[str] = [self.sep_token_id] _a : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __lowercase ( self : Dict ): '''simple docstring''' return len(self.fairseq_ids_to_tokens ) def __lowercase ( self : Dict ): '''simple docstring''' _a : List[str] = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __lowercase ( self : Tuple ,_a : str ): '''simple docstring''' return self.sp_model.encode(_a ,out_type=_a ) def __lowercase ( self : Union[str, Any] ,_a : Union[str, Any] ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def __lowercase ( self : Any ,_a : int ): '''simple docstring''' return self.fairseq_ids_to_tokens[index] def __lowercase ( self : Tuple ,_a : Union[str, Any] ): '''simple docstring''' _a : str = ''.join(_a ).replace(_a ,' ' ).strip() return out_string def __lowercase ( self : Union[str, Any] ,_a : str ,_a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(_a ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['monolingual_vocab_file'] ,) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,_a ) elif not os.path.isfile(self.vocab_file ): with open(_a ,'wb' ) as fi: _a : List[Any] = self.sp_model.serialized_model_proto() fi.write(_a ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( _a ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file ,_a ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(_a ,'w' ,encoding='utf-8' ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(F"""{str(_a )} \n""" ) return out_vocab_file, out_monolingual_vocab_file
271
0
A__ = [0, 2, 4, 6, 8] A__ = [1, 3, 5, 7, 9] def _UpperCAmelCase ( snake_case , snake_case , snake_case , snake_case ): """simple docstring""" if remaining_length == 0: if digits[0] == 0 or digits[-1] == 0: return 0 for i in range(length // 2 - 1 , -1 , -1 ): remainder += digits[i] + digits[length - i - 1] if remainder % 2 == 0: return 0 remainder //= 10 return 1 if remaining_length == 1: if remainder % 2 == 0: return 0 _lowerCAmelCase = 0 for digit in range(10 ): _lowerCAmelCase = digit result += reversible_numbers( 0 , (remainder + 2 * digit) // 10 , snake_case , snake_case ) return result _lowerCAmelCase = 0 for digita in range(10 ): _lowerCAmelCase = digita if (remainder + digita) % 2 == 0: _lowerCAmelCase = ODD_DIGITS else: _lowerCAmelCase = EVEN_DIGITS for digita in other_parity_digits: _lowerCAmelCase = digita result += reversible_numbers( remaining_length - 2 , (remainder + digita + digita) // 10 , snake_case , snake_case , ) return result def _UpperCAmelCase ( snake_case = 9 ): """simple docstring""" _lowerCAmelCase = 0 for length in range(1 , max_power + 1 ): result += reversible_numbers(snake_case , 0 , [0] * length , snake_case ) return result if __name__ == "__main__": print(f"{solution() = }")
82
'''simple docstring''' import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = None __UpperCAmelCase : List[Any] = None @property def __lowercase ( self : Dict ): '''simple docstring''' return self.feat_extract_tester.prepare_feat_extract_dict() def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_a ,'feature_size' ) ) self.assertTrue(hasattr(_a ,'sampling_rate' ) ) self.assertTrue(hasattr(_a ,'padding_value' ) ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_tester.prepare_inputs_for_common() _a : str = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_a ) == len(_a ) for x, y in zip(_a ,processed_features[input_name] ) ) ) _a : Any = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Union[str, Any] = BatchFeature({input_name: speech_inputs} ,tensor_type='np' ) _a : Union[str, Any] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[int] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def __lowercase ( self : Any ): '''simple docstring''' _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : str = BatchFeature({input_name: speech_inputs} ,tensor_type='pt' ) _a : str = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : str = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : int = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = feat_extract.model_input_names[0] _a : int = BatchFeature({input_name: speech_inputs} ,tensor_type='tf' ) _a : Optional[int] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[Any] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def __lowercase ( self : Dict ,_a : Any=False ): '''simple docstring''' def _inputs_have_equal_length(_a : Tuple ): _a : Tuple = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : Optional[Any] ,_a : Union[str, Any] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : int = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Tuple = BatchFeature({input_name: speech_inputs} ) _a : str = self.feat_extract_tester.seq_length_diff _a : Dict = self.feat_extract_tester.max_seq_length + pad_diff _a : Dict = self.feat_extract_tester.min_seq_length _a : Optional[Any] = self.feat_extract_tester.batch_size _a : Tuple = self.feat_extract_tester.feature_size # test padding for List[int] + numpy _a : int = feat_extract.pad(_a ,padding=_a ) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad(_a ,padding='longest' ) _a : Any = input_a[input_name] _a : Optional[Any] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[-1] ) ) _a : List[str] = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) _a : str = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' )[input_name] _a : int = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,return_tensors='np' ) _a : Optional[int] = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy _a : Tuple = feat_extract.pad(_a ,pad_to_multiple_of=10 ) _a : List[str] = input_a[input_name] _a : str = feat_extract.pad(_a ,padding='longest' ,pad_to_multiple_of=10 ) _a : Tuple = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ) _a : Any = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ,return_tensors='np' ,) _a : Dict = input_a[input_name] self.assertTrue(all(len(_a ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) _a : List[str] = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_a ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] ,(batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct _a : Any = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1E-3 ) def __lowercase ( self : List[Any] ,_a : Optional[int]=False ): '''simple docstring''' def _inputs_have_equal_length(_a : List[str] ): _a : Union[str, Any] = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : List[str] ,_a : List[str] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : str = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Any = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) # truncate to smallest _a : Union[str, Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,truncation=_a ) _a : str = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ) _a : Tuple = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to smallest with np _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ,truncation=_a ,) _a : Any = input_a[input_name] _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ) _a : int = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to middle _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ,return_tensors='np' ,) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ) _a : Tuple = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,return_tensors='np' ) _a : Dict = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' ,truncation=_a )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy _a : Optional[Any] = 12 _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,truncation=_a ,) _a : Tuple = input_a[input_name] _a : str = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,) _a : List[Any] = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of _a : List[Any] = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: _a : Union[str, Any] = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Dict ): '''simple docstring''' self._check_truncation(numpify=_a ) def __lowercase ( self : str ): '''simple docstring''' self._check_truncation(numpify=_a ) @require_torch def __lowercase ( self : Dict ): '''simple docstring''' _a : Any = self.feature_extraction_class(**self.feat_extract_dict ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Optional[int] = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='pt' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : List[str] = self.feature_extraction_class(**self.feat_extract_dict ) _a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : Dict = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : Any = feat_extract.pad(_a ,padding='longest' ,return_tensors='tf' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : str = self.feat_extract_dict _a : List[Any] = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Tuple = [len(_a ) for x in speech_inputs] _a : int = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : str = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual(list(processed.attention_mask.shape ) ,list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() ,_a ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_dict _a : Tuple = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : Dict = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = [len(_a ) for x in speech_inputs] _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Any = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = min(_a ) _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,truncation=_a ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual( list(processed_pad.attention_mask.shape ) ,[processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() ,[max_length for x in speech_inputs] )
271
0
'''simple docstring''' from __future__ import annotations import math from collections.abc import Callable def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = 1_0_0 , ): _UpperCamelCase : Union[str, Any] = x_start _UpperCamelCase : List[Any] = fnc(UpperCAmelCase_ ) _UpperCamelCase : int = 0.0 for _ in range(UpperCAmelCase_ ): # Approximates curve as a sequence of linear lines and sums their length _UpperCamelCase : Union[str, Any] = (x_end - x_start) / steps + xa _UpperCamelCase : Any = fnc(UpperCAmelCase_ ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step _UpperCamelCase : Optional[int] = xa _UpperCamelCase : List[str] = fxa return length if __name__ == "__main__": def A__ ( UpperCAmelCase_ ): return math.sin(1_0 * x ) print('f(x) = sin(10 * x)') print('The length of the curve from x = -10 to x = 10 is:') snake_case_ : List[Any] = 10 while i <= 100000: print(F"""With {i} steps: {line_length(f, -10, 10, i)}""") i *= 10
83
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import KarrasVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : UNetaDModel __UpperCAmelCase : KarrasVeScheduler def __init__( self : Union[str, Any] ,_a : UNetaDModel ,_a : KarrasVeScheduler ): '''simple docstring''' super().__init__() self.register_modules(unet=_a ,scheduler=_a ) @torch.no_grad() def __call__( self : List[Any] ,_a : int = 1 ,_a : int = 50 ,_a : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,_a : Optional[str] = "pil" ,_a : bool = True ,**_a : List[Any] ,): '''simple docstring''' _a : Any = self.unet.config.sample_size _a : Optional[int] = (batch_size, 3, img_size, img_size) _a : Dict = self.unet # sample x_0 ~ N(0, sigma_0^2 * I) _a : Dict = randn_tensor(_a ,generator=_a ,device=self.device ) * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(_a ) for t in self.progress_bar(self.scheduler.timesteps ): # here sigma_t == t_i from the paper _a : Optional[int] = self.scheduler.schedule[t] _a : List[str] = self.scheduler.schedule[t - 1] if t > 0 else 0 # 1. Select temporarily increased noise level sigma_hat # 2. Add new noise to move from sample_i to sample_hat _a, _a : List[Any] = self.scheduler.add_noise_to_input(_a ,_a ,generator=_a ) # 3. Predict the noise residual given the noise magnitude `sigma_hat` # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_hat / 2) * model((sample_hat + 1) / 2 ,sigma_hat / 2 ).sample # 4. Evaluate dx/dt at sigma_hat # 5. Take Euler step from sigma to sigma_prev _a : Tuple = self.scheduler.step(_a ,_a ,_a ,_a ) if sigma_prev != 0: # 6. Apply 2nd order correction # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_prev / 2) * model((step_output.prev_sample + 1) / 2 ,sigma_prev / 2 ).sample _a : Optional[Any] = self.scheduler.step_correct( _a ,_a ,_a ,_a ,step_output.prev_sample ,step_output['derivative'] ,) _a : Dict = step_output.prev_sample _a : Tuple = (sample / 2 + 0.5).clamp(0 ,1 ) _a : Optional[Any] = sample.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": _a : List[str] = self.numpy_to_pil(_a ) if not return_dict: return (image,) return ImagePipelineOutput(images=_a )
271
0
"""simple docstring""" from typing import List from .keymap import KEYMAP, get_character def _snake_case ( lowercase__ : str ) -> Optional[Any]: '''simple docstring''' def decorator(lowercase__ : Any ): lowerCAmelCase_ :Optional[Any] = getattr(lowercase__ , """handle_key""" , [] ) handle += [key] setattr(lowercase__ , """handle_key""" , lowercase__ ) return func return decorator def _snake_case ( *lowercase__ : List[str] ) -> List[str]: '''simple docstring''' def decorator(lowercase__ : List[str] ): lowerCAmelCase_ :List[str] = getattr(lowercase__ , """handle_key""" , [] ) handle += keys setattr(lowercase__ , """handle_key""" , lowercase__ ) return func return decorator class _SCREAMING_SNAKE_CASE ( A__ ): def __new__( cls , __A , __A , __A ) -> Union[str, Any]: lowerCAmelCase_ :Tuple = super().__new__(cls , __A , __A , __A ) if not hasattr(__A , """key_handler""" ): setattr(__A , """key_handler""" , {} ) setattr(__A , """handle_input""" , KeyHandler.handle_input ) for value in attrs.values(): lowerCAmelCase_ :str = getattr(__A , """handle_key""" , [] ) for key in handled_keys: lowerCAmelCase_ :str = value return new_cls @staticmethod def __lowerCAmelCase ( cls ) -> str: lowerCAmelCase_ :Optional[int] = get_character() if char != KEYMAP["undefined"]: lowerCAmelCase_ :Optional[Any] = ord(__A ) lowerCAmelCase_ :int = cls.key_handler.get(__A ) if handler: lowerCAmelCase_ :Any = char return handler(cls ) else: return None def _snake_case ( cls : Dict ) -> Tuple: '''simple docstring''' return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy() )
84
'''simple docstring''' import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging __lowerCAmelCase = ( """https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py""" ) __lowerCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def UpperCAmelCase_ (): """simple docstring""" _a : Optional[int] = 'https://pypi.org/pypi/diffusers/json' _a : int = json.loads(request.urlopen(__a ).read() )['releases'].keys() return sorted(__a , key=lambda __a : version.Version(__a ) ) def UpperCAmelCase_ (): """simple docstring""" if HF_MODULES_CACHE in sys.path: return sys.path.append(__a ) os.makedirs(__a , exist_ok=__a ) _a : str = Path(__a ) / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : Union[str, os.PathLike] ): """simple docstring""" init_hf_modules() _a : Dict = Path(__a ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(__a , exist_ok=__a ) _a : Optional[int] = dynamic_module_path / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : str ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : int = f.read() # Imports of the form `import .xxx` _a : Tuple = re.findall('^\s*import\s+\.(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall('^\s*from\s+\.(\S+)\s+import' , __a , flags=re.MULTILINE ) # Unique-ify return list(set(__a ) ) def UpperCAmelCase_ (__a : Any ): """simple docstring""" _a : Optional[int] = False _a : Optional[int] = [module_file] _a : List[str] = [] # Let's recurse through all relative imports while not no_change: _a : str = [] for f in files_to_check: new_imports.extend(get_relative_imports(__a ) ) _a : Union[str, Any] = Path(__a ).parent _a : str = [str(module_path / m ) for m in new_imports] _a : Tuple = [f for f in new_import_files if f not in all_relative_imports] _a : Dict = [f"""{f}.py""" for f in new_import_files] _a : List[str] = len(__a ) == 0 all_relative_imports.extend(__a ) return all_relative_imports def UpperCAmelCase_ (__a : Tuple ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : Dict = f.read() # Imports of the form `import xxx` _a : Optional[int] = re.findall('^\s*import\s+(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall('^\s*from\s+(\S+)\s+import' , __a , flags=re.MULTILINE ) # Only keep the top-level module _a : List[str] = [imp.split('.' )[0] for imp in imports if not imp.startswith('.' )] # Unique-ify and test we got them all _a : Optional[int] = list(set(__a ) ) _a : List[str] = [] for imp in imports: try: importlib.import_module(__a ) except ImportError: missing_packages.append(__a ) if len(__a ) > 0: raise ImportError( 'This modeling file requires the following packages that were not found in your environment: ' f"""{', '.join(__a )}. Run `pip install {' '.join(__a )}`""" ) return get_relative_imports(__a ) def UpperCAmelCase_ (__a : Any , __a : str ): """simple docstring""" _a : Any = module_path.replace(os.path.sep , '.' ) _a : Union[str, Any] = importlib.import_module(__a ) if class_name is None: return find_pipeline_class(__a ) return getattr(__a , __a ) def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" from ..pipelines import DiffusionPipeline _a : List[str] = dict(inspect.getmembers(__a , inspect.isclass ) ) _a : str = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls , __a ) and cls.__module__.split('.' )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( f"""Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:""" f""" {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in""" f""" {loaded_module}.""" ) _a : Any = cls return pipeline_class def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , ): """simple docstring""" _a : str = str(__a ) _a : Optional[Any] = os.path.join(__a , __a ) if os.path.isfile(__a ): _a : Tuple = module_file_or_url _a : Optional[Any] = 'local' elif pretrained_model_name_or_path.count('/' ) == 0: _a : int = get_diffusers_versions() # cut ".dev0" _a : Any = 'v' + '.'.join(__version__.split('.' )[:3] ) # retrieve github version that matches if revision is None: _a : Any = latest_version if latest_version[1:] in available_versions else 'main' logger.info(f"""Defaulting to latest_version: {revision}.""" ) elif revision in available_versions: _a : Any = f"""v{revision}""" elif revision == "main": _a : Optional[int] = revision else: raise ValueError( f"""`custom_revision`: {revision} does not exist. Please make sure to choose one of""" f""" {', '.join(available_versions + ['main'] )}.""" ) # community pipeline on GitHub _a : Tuple = COMMUNITY_PIPELINES_URL.format(revision=__a , pipeline=__a ) try: _a : Any = cached_download( __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = 'git' _a : Any = pretrained_model_name_or_path + '.py' except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise else: try: # Load from URL or cache if already cached _a : Optional[Any] = hf_hub_download( __a , __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = os.path.join('local' , '--'.join(pretrained_model_name_or_path.split('/' ) ) ) except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise # Check we have all the requirements in our environment _a : Optional[int] = check_imports(__a ) # Now we move the module inside our cached dynamic modules. _a : Optional[Any] = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(__a ) _a : Any = Path(__a ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(__a , submodule_path / module_file ) for module_needed in modules_needed: _a : Dict = f"""{module_needed}.py""" shutil.copy(os.path.join(__a , __a ) , submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(__a , __a ): _a : Optional[Any] = use_auth_token elif use_auth_token is True: _a : List[Any] = HfFolder.get_token() else: _a : Dict = None _a : int = model_info(__a , revision=__a , token=__a ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. _a : Optional[int] = submodule_path / commit_hash _a : str = full_submodule + os.path.sep + commit_hash create_dynamic_module(__a ) if not (submodule_path / module_file).exists(): shutil.copy(__a , submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( __a , f"""{module_needed}.py""" , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return os.path.join(__a , __a ) def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[str] = None , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , **__a : str , ): """simple docstring""" _a : Dict = get_cached_module_file( __a , __a , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return get_class_in_module(__a , final_module.replace('.py' , '' ) )
271
0
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() _SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) def UpperCamelCase_( snake_case : Union[str, Any] , snake_case : Optional[Any]=False ): '''simple docstring''' snake_case_ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'blocks.{i}.norm1.weight', f'vit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((f'blocks.{i}.norm1.bias', f'vit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((f'blocks.{i}.attn.proj.weight', f'vit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((f'blocks.{i}.attn.proj.bias', f'vit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((f'blocks.{i}.norm2.weight', f'vit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((f'blocks.{i}.norm2.bias', f'vit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc1.weight', f'vit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc1.bias', f'vit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc2.weight', f'vit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc2.bias', f'vit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "vit.embeddings.cls_token"), ("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "vit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" snake_case_ = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def UpperCamelCase_( snake_case : Any , snake_case : Tuple , snake_case : str=False ): '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: snake_case_ = "" else: snake_case_ = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) snake_case_ = state_dict.pop(f'blocks.{i}.attn.qkv.weight' ) snake_case_ = state_dict.pop(f'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict snake_case_ = in_proj_weight[ : config.hidden_size, : ] snake_case_ = in_proj_bias[: config.hidden_size] snake_case_ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] snake_case_ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] snake_case_ = in_proj_weight[ -config.hidden_size :, : ] snake_case_ = in_proj_bias[-config.hidden_size :] def UpperCamelCase_( snake_case : Tuple ): '''simple docstring''' snake_case_ = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(snake_case , snake_case ) def UpperCamelCase_( snake_case : List[str] , snake_case : Union[str, Any] , snake_case : int ): '''simple docstring''' snake_case_ = dct.pop(snake_case ) snake_case_ = val def UpperCamelCase_( ): '''simple docstring''' snake_case_ = "http://images.cocodataset.org/val2017/000000039769.jpg" snake_case_ = Image.open(requests.get(snake_case , stream=snake_case ).raw ) return im @torch.no_grad() def UpperCamelCase_( snake_case : Optional[Any] , snake_case : Any , snake_case : str=True ): '''simple docstring''' snake_case_ = ViTConfig() # patch_size if model_name[-1] == "8": snake_case_ = 8 # set labels if required if not base_model: snake_case_ = 1_0_0_0 snake_case_ = "huggingface/label-files" snake_case_ = "imagenet-1k-id2label.json" snake_case_ = json.load(open(hf_hub_download(snake_case , snake_case , repo_type="dataset" ) , "r" ) ) snake_case_ = {int(snake_case ): v for k, v in idalabel.items()} snake_case_ = idalabel snake_case_ = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: snake_case_ = 3_8_4 snake_case_ = 1_5_3_6 snake_case_ = 1_2 snake_case_ = 6 # load original model from torch hub snake_case_ = torch.hub.load("facebookresearch/dino:main" , snake_case ) original_model.eval() # load state_dict of original model, remove and rename some keys snake_case_ = original_model.state_dict() if base_model: remove_classification_head_(snake_case ) snake_case_ = create_rename_keys(snake_case , base_model=snake_case ) for src, dest in rename_keys: rename_key(snake_case , snake_case , snake_case ) read_in_q_k_v(snake_case , snake_case , snake_case ) # load HuggingFace model if base_model: snake_case_ = ViTModel(snake_case , add_pooling_layer=snake_case ).eval() else: snake_case_ = ViTForImageClassification(snake_case ).eval() model.load_state_dict(snake_case ) # Check outputs on an image, prepared by ViTImageProcessor snake_case_ = ViTImageProcessor() snake_case_ = image_processor(images=prepare_img() , return_tensors="pt" ) snake_case_ = encoding["pixel_values"] snake_case_ = model(snake_case ) if base_model: snake_case_ = original_model(snake_case ) assert torch.allclose(snake_case , outputs.last_hidden_state[:, 0, :] , atol=1e-1 ) else: snake_case_ = original_model(snake_case ) assert logits.shape == outputs.logits.shape assert torch.allclose(snake_case , outputs.logits , atol=1e-3 ) Path(snake_case ).mkdir(exist_ok=snake_case ) print(f'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(snake_case ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(snake_case ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="dino_vitb16", type=str, help="Name of the model trained with DINO you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--base_model", action="store_true", help="Whether to only convert the base model (no projection head weights).", ) parser.set_defaults(base_model=True) _SCREAMING_SNAKE_CASE : Optional[Any] = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
85
'''simple docstring''' def UpperCAmelCase_ (__a : list , __a : list , __a : int ): """simple docstring""" _a : Optional[Any] = len(__a ) _a : int = [[0] * n for i in range(__a )] for i in range(__a ): _a : Tuple = y_points[i] for i in range(2 , __a ): for j in range(__a , __a ): _a : Tuple = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
271
0
"""simple docstring""" import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 lowerCamelCase__ = data_utils.TransfoXLTokenizer lowerCamelCase__ = data_utils.TransfoXLCorpus lowerCamelCase__ = data_utils lowerCamelCase__ = data_utils def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(_UpperCamelCase , 'rb' ) as fp: __lowerCAmelCase : Dict = pickle.load(_UpperCamelCase , encoding='latin1' ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) __lowerCAmelCase : Union[str, Any] = pytorch_dump_folder_path + '/' + VOCAB_FILES_NAMES['pretrained_vocab_file'] print(F"Save vocabulary to {pytorch_vocab_dump_path}" ) __lowerCAmelCase : Optional[int] = corpus.vocab.__dict__ torch.save(_UpperCamelCase , _UpperCamelCase ) __lowerCAmelCase : Tuple = corpus.__dict__ corpus_dict_no_vocab.pop('vocab' , _UpperCamelCase ) __lowerCAmelCase : Tuple = pytorch_dump_folder_path + '/' + CORPUS_NAME print(F"Save dataset to {pytorch_dataset_dump_path}" ) torch.save(_UpperCamelCase , _UpperCamelCase ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model __lowerCAmelCase : str = os.path.abspath(_UpperCamelCase ) __lowerCAmelCase : Any = os.path.abspath(_UpperCamelCase ) print(F"Converting Transformer XL checkpoint from {tf_path} with config at {config_path}." ) # Initialise PyTorch model if transfo_xl_config_file == "": __lowerCAmelCase : List[Any] = TransfoXLConfig() else: __lowerCAmelCase : List[Any] = TransfoXLConfig.from_json_file(_UpperCamelCase ) print(F"Building PyTorch model from configuration: {config}" ) __lowerCAmelCase : Union[str, Any] = TransfoXLLMHeadModel(_UpperCamelCase ) __lowerCAmelCase : Tuple = load_tf_weights_in_transfo_xl(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) # Save pytorch-model __lowerCAmelCase : str = os.path.join(_UpperCamelCase , _UpperCamelCase ) __lowerCAmelCase : Union[str, Any] = os.path.join(_UpperCamelCase , _UpperCamelCase ) print(F"Save PyTorch model to {os.path.abspath(_UpperCamelCase )}" ) torch.save(model.state_dict() , _UpperCamelCase ) print(F"Save configuration file to {os.path.abspath(_UpperCamelCase )}" ) with open(_UpperCamelCase , 'w' , encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the folder to store the PyTorch model or dataset/vocab.""", ) parser.add_argument( """--tf_checkpoint_path""", default="""""", type=str, help="""An optional path to a TensorFlow checkpoint path to be converted.""", ) parser.add_argument( """--transfo_xl_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained BERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--transfo_xl_dataset_file""", default="""""", type=str, help="""An optional dataset file to be converted in a vocabulary.""", ) lowerCamelCase__ = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
86
'''simple docstring''' import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = inspect.getfile(accelerate.test_utils ) __UpperCAmelCase : List[str] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_cli.py'''] ) __UpperCAmelCase : Dict = ['''accelerate''', '''launch'''] __UpperCAmelCase : Dict = Path.home() / '''.cache/huggingface/accelerate''' __UpperCAmelCase : Dict = '''default_config.yaml''' __UpperCAmelCase : Optional[Any] = config_folder / config_file __UpperCAmelCase : Dict = config_folder / '''_default_config.yaml''' __UpperCAmelCase : Any = Path('''tests/test_configs''' ) @classmethod def __lowercase ( cls : int ): '''simple docstring''' if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def __lowercase ( cls : List[Any] ): '''simple docstring''' if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Dict = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' for config in sorted(self.test_config_path.glob('**/*.yaml' ) ): with self.subTest(config_file=_a ): execute_subprocess_async( self.base_cmd + ['--config_file', str(_a ), self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[int] ): '''simple docstring''' execute_subprocess_async(['accelerate', 'test'] ,env=os.environ.copy() ) class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = '''test-tpu''' __UpperCAmelCase : Any = '''us-central1-a''' __UpperCAmelCase : List[Any] = '''ls''' __UpperCAmelCase : Any = ['''accelerate''', '''tpu-config'''] __UpperCAmelCase : Dict = '''cd /usr/share''' __UpperCAmelCase : Any = '''tests/test_samples/test_command_file.sh''' __UpperCAmelCase : List[Any] = '''Running gcloud compute tpus tpu-vm ssh''' def __lowercase ( self : Dict ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Any = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--debug'] ,return_stdout=_a ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : int ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : str ): '''simple docstring''' _a : List[str] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--command', 'echo "Hello World"', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo \"Hello World\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Any = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command_file', self.command_file, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Union[str, Any] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command_file', self.command_file, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Any ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--accelerate_version', '12.0.0', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,)
271
0
import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase = '''▁''' UpperCamelCase = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class snake_case_ ( __A ,unittest.TestCase ): __A : Optional[int] = BigBirdTokenizer __A : Any = BigBirdTokenizerFast __A : Dict = True __A : Optional[Any] = True def __UpperCamelCase ( self : Union[str, Any] ) -> Optional[Any]: super().setUp() lowercase__ : Union[str, Any] = self.tokenizer_class(lowercase_ , keep_accents=lowercase_ ) tokenizer.save_pretrained(self.tmpdirname ) def __UpperCamelCase ( self : Tuple ) -> List[Any]: lowercase__ : List[Any] = "<s>" lowercase__ : Optional[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase_ ) , lowercase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase_ ) , lowercase_ ) def __UpperCamelCase ( self : Optional[Any] ) -> int: lowercase__ : Union[str, Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "[MASK]" ) self.assertEqual(len(lowercase_ ) , 10_04 ) def __UpperCamelCase ( self : List[Any] ) -> str: self.assertEqual(self.get_tokenizer().vocab_size , 10_00 ) def __UpperCamelCase ( self : Optional[int] ) -> Tuple: if not self.test_rust_tokenizer: return lowercase__ : Dict = self.get_tokenizer() lowercase__ : Any = self.get_rust_tokenizer() lowercase__ : str = "I was born in 92000, and this is falsé." lowercase__ : Optional[int] = tokenizer.tokenize(lowercase_ ) lowercase__ : Any = rust_tokenizer.tokenize(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) lowercase__ : List[str] = tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) lowercase__ : List[str] = rust_tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) lowercase__ : str = self.get_rust_tokenizer() lowercase__ : Union[str, Any] = tokenizer.encode(lowercase_ ) lowercase__ : Optional[Any] = rust_tokenizer.encode(lowercase_ ) self.assertListEqual(lowercase_ , lowercase_ ) def __UpperCamelCase ( self : Dict ) -> str: lowercase__ : Optional[Any] = BigBirdTokenizer(lowercase_ , keep_accents=lowercase_ ) lowercase__ : List[str] = tokenizer.tokenize("This is a test" ) self.assertListEqual(lowercase_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowercase_ ) , [2_85, 46, 10, 1_70, 3_82] , ) lowercase__ : Dict = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( lowercase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) lowercase__ : Tuple = tokenizer.convert_tokens_to_ids(lowercase_ ) self.assertListEqual( lowercase_ , [8, 21, 84, 55, 24, 19, 7, 0, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) lowercase__ : Tuple = tokenizer.convert_ids_to_tokens(lowercase_ ) self.assertListEqual( lowercase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def __UpperCamelCase ( self : List[Any] ) -> Optional[int]: return BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) @slow def __UpperCamelCase ( self : Optional[int] ) -> int: lowercase__ : Optional[Any] = "Hello World!" lowercase__ : Optional[int] = [65, 1_85_36, 22_60, 1_01, 66] self.assertListEqual(lowercase_ , self.big_tokenizer.encode(lowercase_ ) ) @slow def __UpperCamelCase ( self : Tuple ) -> Optional[Any]: lowercase__ : Tuple = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) # fmt: off lowercase__ : Optional[Any] = [65, 8_71, 4_19, 3_58, 9_46, 9_91, 25_21, 4_52, 3_58, 13_57, 3_87, 77_51, 35_36, 1_12, 9_85, 4_56, 1_26, 8_65, 9_38, 54_00, 57_34, 4_58, 13_68, 4_67, 7_86, 24_62, 52_46, 11_59, 6_33, 8_65, 45_19, 4_57, 5_82, 8_52, 25_57, 4_27, 9_16, 5_08, 4_05, 3_43_24, 4_97, 3_91, 4_08, 1_13_42, 12_44, 3_85, 1_00, 9_38, 9_85, 4_56, 5_74, 3_62, 1_25_97, 32_00, 31_29, 11_72, 66] # noqa: E231 # fmt: on self.assertListEqual(lowercase_ , self.big_tokenizer.encode(lowercase_ ) ) @require_torch @slow def __UpperCamelCase ( self : List[Any] ) -> int: import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence lowercase__ : Optional[int] = list(self.big_tokenizer.get_vocab().keys() )[:10] lowercase__ : str = " ".join(lowercase_ ) lowercase__ : List[str] = self.big_tokenizer.encode_plus(lowercase_ , return_tensors="pt" , return_token_type_ids=lowercase_ ) lowercase__ : int = self.big_tokenizer.batch_encode_plus( [sequence + " " + sequence] , return_tensors="pt" , return_token_type_ids=lowercase_ ) lowercase__ : Optional[int] = BigBirdConfig(attention_type="original_full" ) lowercase__ : List[str] = BigBirdModel(lowercase_ ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**lowercase_ ) model(**lowercase_ ) @slow def __UpperCamelCase ( self : str ) -> Optional[int]: lowercase__ : List[Any] = BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) lowercase__ : str = tokenizer.decode(tokenizer("Paris is the [MASK]." ).input_ids ) self.assertTrue(decoded_text == "[CLS] Paris is the[MASK].[SEP]" ) @slow def __UpperCamelCase ( self : Optional[int] ) -> Optional[Any]: # fmt: off lowercase__ : str = {"input_ids": [[65, 3_92_86, 4_58, 3_63_35, 20_01, 4_56, 1_30_73, 1_32_66, 4_55, 1_13, 77_46, 17_41, 1_11_57, 3_91, 1_30_73, 1_32_66, 4_55, 1_13, 39_67, 3_54_12, 1_13, 49_36, 1_09, 38_70, 23_77, 1_13, 3_00_84, 4_57_20, 4_58, 1_34, 1_74_96, 1_12, 5_03, 1_16_72, 1_13, 1_18, 1_12, 56_65, 1_33_47, 3_86_87, 1_12, 14_96, 3_13_89, 1_12, 32_68, 4_72_64, 1_34, 9_62, 1_12, 1_63_77, 80_35, 2_31_30, 4_30, 1_21_69, 1_55_18, 2_85_92, 4_58, 1_46, 4_16_97, 1_09, 3_91, 1_21_69, 1_55_18, 1_66_89, 4_58, 1_46, 4_13_58, 1_09, 4_52, 7_26, 40_34, 1_11, 7_63, 3_54_12, 50_82, 3_88, 19_03, 1_11, 90_51, 3_91, 28_70, 4_89_18, 19_00, 11_23, 5_50, 9_98, 1_12, 95_86, 1_59_85, 4_55, 3_91, 4_10, 2_29_55, 3_76_36, 1_14, 66], [65, 4_48, 1_74_96, 4_19, 36_63, 3_85, 7_63, 1_13, 2_75_33, 28_70, 32_83, 1_30_43, 16_39, 2_47_13, 5_23, 6_56, 2_40_13, 1_85_50, 25_21, 5_17, 2_70_14, 2_12_44, 4_20, 12_12, 14_65, 3_91, 9_27, 48_33, 3_88, 5_78, 1_17_86, 1_14, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 4_84, 21_69, 76_87, 2_19_32, 1_81_46, 7_26, 3_63, 1_70_32, 33_91, 1_14, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowercase_ , model_name="google/bigbird-roberta-base" , revision="215c99f1600e06f83acce68422f2035b2b5c3510" , )
87
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __lowerCAmelCase = TypeVar("""T""") class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Tuple ,_a : T ): '''simple docstring''' _a : List[str] = data _a : Node[T] | None = None def __str__( self : Dict ): '''simple docstring''' return F"""{self.data}""" class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Optional[int] ): '''simple docstring''' _a : Node[T] | None = None def __iter__( self : str ): '''simple docstring''' _a : Tuple = self.top while node: yield node.data _a : int = node.next def __str__( self : str ): '''simple docstring''' return "->".join([str(_a ) for item in self] ) def __len__( self : Optional[Any] ): '''simple docstring''' return len(tuple(iter(self ) ) ) def __lowercase ( self : str ): '''simple docstring''' return self.top is None def __lowercase ( self : List[Any] ,_a : T ): '''simple docstring''' _a : int = Node(_a ) if not self.is_empty(): _a : Optional[Any] = self.top _a : List[str] = node def __lowercase ( self : Tuple ): '''simple docstring''' if self.is_empty(): raise IndexError('pop from empty stack' ) assert isinstance(self.top ,_a ) _a : List[Any] = self.top _a : int = self.top.next return pop_node.data def __lowercase ( self : List[str] ): '''simple docstring''' if self.is_empty(): raise IndexError('peek from empty stack' ) assert self.top is not None return self.top.data def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = None if __name__ == "__main__": from doctest import testmod testmod()
271
0
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase : List[Any] = logging.get_logger(__name__) __lowerCAmelCase : Optional[int] = { 'xlnet-base-cased': 'https://huggingface.co/xlnet-base-cased/resolve/main/config.json', 'xlnet-large-cased': 'https://huggingface.co/xlnet-large-cased/resolve/main/config.json', } class UpperCAmelCase_ ( _A ): '''simple docstring''' a__ = """xlnet""" a__ = ["""mems"""] a__ = { """n_token""": """vocab_size""", # Backward compatibility """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : int , UpperCamelCase__ : List[str]=3_2000 , UpperCamelCase__ : Optional[int]=1024 , UpperCamelCase__ : Union[str, Any]=24 , UpperCamelCase__ : Optional[int]=16 , UpperCamelCase__ : List[str]=4096 , UpperCamelCase__ : Union[str, Any]="gelu" , UpperCamelCase__ : Tuple=True , UpperCamelCase__ : Tuple="bi" , UpperCamelCase__ : List[str]=0.02 , UpperCamelCase__ : List[str]=1E-12 , UpperCamelCase__ : Tuple=0.1 , UpperCamelCase__ : Tuple=512 , UpperCamelCase__ : str=None , UpperCamelCase__ : Dict=True , UpperCamelCase__ : Optional[int]=False , UpperCamelCase__ : List[str]=False , UpperCamelCase__ : Any=-1 , UpperCamelCase__ : Dict=False , UpperCamelCase__ : int="last" , UpperCamelCase__ : Optional[Any]=True , UpperCamelCase__ : List[Any]="tanh" , UpperCamelCase__ : Tuple=0.1 , UpperCamelCase__ : List[str]=5 , UpperCamelCase__ : Tuple=5 , UpperCamelCase__ : str=5 , UpperCamelCase__ : Optional[int]=1 , UpperCamelCase__ : List[str]=2 , **UpperCamelCase__ : List[Any] , ) -> Any: """simple docstring""" __magic_name__ = vocab_size __magic_name__ = d_model __magic_name__ = n_layer __magic_name__ = n_head if d_model % n_head != 0: raise ValueError(F'''\'d_model % n_head\' ({d_model % n_head}) should be equal to 0''' ) if "d_head" in kwargs: if kwargs["d_head"] != d_model // n_head: raise ValueError( F'''`d_head` ({kwargs['d_head']}) should be equal to `d_model // n_head` ({d_model // n_head})''' ) __magic_name__ = d_model // n_head __magic_name__ = ff_activation __magic_name__ = d_inner __magic_name__ = untie_r __magic_name__ = attn_type __magic_name__ = initializer_range __magic_name__ = layer_norm_eps __magic_name__ = dropout __magic_name__ = mem_len __magic_name__ = reuse_len __magic_name__ = bi_data __magic_name__ = clamp_len __magic_name__ = same_length __magic_name__ = summary_type __magic_name__ = summary_use_proj __magic_name__ = summary_activation __magic_name__ = summary_last_dropout __magic_name__ = start_n_top __magic_name__ = end_n_top __magic_name__ = bos_token_id __magic_name__ = pad_token_id __magic_name__ = eos_token_id if "use_cache" in kwargs: warnings.warn( """The `use_cache` argument is deprecated and will be removed in a future version, use `use_mems_eval`""" """ instead.""" , UpperCamelCase__ , ) __magic_name__ = kwargs["""use_cache"""] __magic_name__ = use_mems_eval __magic_name__ = use_mems_train super().__init__(pad_token_id=UpperCamelCase__ , bos_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , **UpperCamelCase__ ) @property def _lowercase ( self : int ) -> Tuple: """simple docstring""" logger.info(F'''The model {self.model_type} is one of the few models that has no sequence length limit.''' ) return -1 @max_position_embeddings.setter def _lowercase ( self : Any , UpperCamelCase__ : Optional[Any] ) -> Dict: """simple docstring""" raise NotImplementedError( F'''The model {self.model_type} is one of the few models that has no sequence length limit.''' )
88
'''simple docstring''' import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : List[str] ): '''simple docstring''' torch.manual_seed(0 ) _a : int = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : str = self.dummy_uncond_unet _a : int = PNDMScheduler() _a : str = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Any = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ,return_dict=_a )[0] _a : List[Any] = image[0, -3:, -3:, -1] _a : Any = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : List[Any] = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[str] = 'google/ddpm-cifar10-32' _a : str = UNetaDModel.from_pretrained(_a ) _a : Union[str, Any] = PNDMScheduler() _a : Tuple = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : str = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,output_type='numpy' ).images _a : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : Tuple = np.array([0.1564, 0.1_4645, 0.1406, 0.1_4715, 0.1_2425, 0.1_4045, 0.1_3115, 0.1_2175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
271
0
'''simple docstring''' from __future__ import annotations import math import random from typing import Any class __magic_name__ : def __init__( self : Dict ): _a : list[Any] = [] _a : int = 0 _a : int = 0 def __lowercase ( self : Optional[Any] ): return self.head == self.tail def __lowercase ( self : List[Any] ,_UpperCAmelCase : Any ): self.data.append(_UpperCAmelCase ) _a : Dict = self.tail + 1 def __lowercase ( self : int ): _a : str = self.data[self.head] _a : List[Any] = self.head + 1 return ret def __lowercase ( self : Dict ): return self.tail - self.head def __lowercase ( self : List[Any] ): print(self.data ) print('**************' ) print(self.data[self.head : self.tail] ) class __magic_name__ : def __init__( self : List[str] ,_UpperCAmelCase : Any ): _a : Dict = data _a : MyNode | None = None _a : MyNode | None = None _a : int = 1 def __lowercase ( self : Dict ): return self.data def __lowercase ( self : Tuple ): return self.left def __lowercase ( self : Optional[Any] ): return self.right def __lowercase ( self : Optional[Any] ): return self.height def __lowercase ( self : str ,_UpperCAmelCase : Any ): _a : str = data def __lowercase ( self : str ,_UpperCAmelCase : MyNode | None ): _a : Tuple = node def __lowercase ( self : Optional[Any] ,_UpperCAmelCase : MyNode | None ): _a : Tuple = node def __lowercase ( self : Dict ,_UpperCAmelCase : int ): _a : Dict = height def __lowerCamelCase ( lowerCAmelCase_ ) -> int: if node is None: return 0 return node.get_height() def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> int: if a > b: return a return b def __lowerCamelCase ( lowerCAmelCase_ ) -> MyNode: print('left rotation node:' , node.get_data() ) _a : str = node.get_left() assert ret is not None node.set_left(ret.get_right() ) ret.set_right(lowerCAmelCase_ ) _a : Any = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(lowerCAmelCase_ ) _a : str = my_max(get_height(ret.get_right() ) , get_height(ret.get_left() ) ) + 1 ret.set_height(lowerCAmelCase_ ) return ret def __lowerCamelCase ( lowerCAmelCase_ ) -> MyNode: print('right rotation node:' , node.get_data() ) _a : Optional[Any] = node.get_right() assert ret is not None node.set_right(ret.get_left() ) ret.set_left(lowerCAmelCase_ ) _a : Tuple = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(lowerCAmelCase_ ) _a : List[str] = my_max(get_height(ret.get_right() ) , get_height(ret.get_left() ) ) + 1 ret.set_height(lowerCAmelCase_ ) return ret def __lowerCamelCase ( lowerCAmelCase_ ) -> MyNode: _a : Union[str, Any] = node.get_left() assert left_child is not None node.set_left(left_rotation(lowerCAmelCase_ ) ) return right_rotation(lowerCAmelCase_ ) def __lowerCamelCase ( lowerCAmelCase_ ) -> MyNode: _a : str = node.get_right() assert right_child is not None node.set_right(right_rotation(lowerCAmelCase_ ) ) return left_rotation(lowerCAmelCase_ ) def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> MyNode | None: if node is None: return MyNode(lowerCAmelCase_ ) if data < node.get_data(): node.set_left(insert_node(node.get_left() , lowerCAmelCase_ ) ) if ( get_height(node.get_left() ) - get_height(node.get_right() ) == 2 ): # an unbalance detected _a : Any = node.get_left() assert left_child is not None if ( data < left_child.get_data() ): # new node is the left child of the left child _a : Union[str, Any] = right_rotation(lowerCAmelCase_ ) else: _a : List[str] = lr_rotation(lowerCAmelCase_ ) else: node.set_right(insert_node(node.get_right() , lowerCAmelCase_ ) ) if get_height(node.get_right() ) - get_height(node.get_left() ) == 2: _a : str = node.get_right() assert right_child is not None if data < right_child.get_data(): _a : Any = rl_rotation(lowerCAmelCase_ ) else: _a : str = left_rotation(lowerCAmelCase_ ) _a : str = my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(lowerCAmelCase_ ) return node def __lowerCamelCase ( lowerCAmelCase_ ) -> Any: while True: _a : Tuple = root.get_right() if right_child is None: break _a : Any = right_child return root.get_data() def __lowerCamelCase ( lowerCAmelCase_ ) -> Any: while True: _a : Any = root.get_left() if left_child is None: break _a : Optional[int] = left_child return root.get_data() def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> MyNode | None: _a : Optional[int] = root.get_left() _a : Optional[int] = root.get_right() if root.get_data() == data: if left_child is not None and right_child is not None: _a : int = get_left_most(lowerCAmelCase_ ) root.set_data(lowerCAmelCase_ ) root.set_right(del_node(lowerCAmelCase_ , lowerCAmelCase_ ) ) elif left_child is not None: _a : Optional[int] = left_child elif right_child is not None: _a : Optional[int] = right_child else: return None elif root.get_data() > data: if left_child is None: print('No such data' ) return root else: root.set_left(del_node(lowerCAmelCase_ , lowerCAmelCase_ ) ) else: # root.get_data() < data if right_child is None: return root else: root.set_right(del_node(lowerCAmelCase_ , lowerCAmelCase_ ) ) if get_height(lowerCAmelCase_ ) - get_height(lowerCAmelCase_ ) == 2: assert right_child is not None if get_height(right_child.get_right() ) > get_height(right_child.get_left() ): _a : Union[str, Any] = left_rotation(lowerCAmelCase_ ) else: _a : Dict = rl_rotation(lowerCAmelCase_ ) elif get_height(lowerCAmelCase_ ) - get_height(lowerCAmelCase_ ) == -2: assert left_child is not None if get_height(left_child.get_left() ) > get_height(left_child.get_right() ): _a : Any = right_rotation(lowerCAmelCase_ ) else: _a : int = lr_rotation(lowerCAmelCase_ ) _a : Tuple = my_max(get_height(root.get_right() ) , get_height(root.get_left() ) ) + 1 root.set_height(lowerCAmelCase_ ) return root class __magic_name__ : def __init__( self : Tuple ): _a : MyNode | None = None def __lowercase ( self : List[Any] ): return get_height(self.root ) def __lowercase ( self : Tuple ,_UpperCAmelCase : Any ): print('insert:' + str(_UpperCAmelCase ) ) _a : List[Any] = insert_node(self.root ,_UpperCAmelCase ) def __lowercase ( self : Any ,_UpperCAmelCase : Any ): print('delete:' + str(_UpperCAmelCase ) ) if self.root is None: print('Tree is empty!' ) return _a : Union[str, Any] = del_node(self.root ,_UpperCAmelCase ) def __str__( self : int ,): # a level traversale, gives a more intuitive look on the tree _a : Optional[Any] = '' _a : Tuple = MyQueue() q.push(self.root ) _a : Dict = self.get_height() if layer == 0: return output _a : int = 0 while not q.is_empty(): _a : List[str] = q.pop() _a : Optional[int] = ' ' * int(math.pow(2 ,layer - 1 ) ) output += space if node is None: output += "*" q.push(_UpperCAmelCase ) q.push(_UpperCAmelCase ) else: output += str(node.get_data() ) q.push(node.get_left() ) q.push(node.get_right() ) output += space _a : Dict = cnt + 1 for i in range(100 ): if cnt == math.pow(2 ,_UpperCAmelCase ) - 1: _a : List[Any] = layer - 1 if layer == 0: output += "\n*************************************" return output output += "\n" break output += "\n*************************************" return output def __lowerCamelCase ( ) -> None: import doctest doctest.testmod() if __name__ == "__main__": _test() __lowerCAmelCase = AVLtree() __lowerCAmelCase = list(range(10)) random.shuffle(lst) for i in lst: t.insert(i) print(str(t)) random.shuffle(lst) for i in lst: t.del_node(i) print(str(t))
89
'''simple docstring''' import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __lowerCAmelCase = logging.getLogger() @unittest.skip('''Temporarily disable the doc tests.''' ) @require_torch @require_tf @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : str ,_a : Path ,_a : Union[str, None] = None ,_a : Union[List[str], None] = None ,_a : Union[str, List[str], None] = None ,_a : bool = True ,): '''simple docstring''' _a : Optional[int] = [file for file in os.listdir(_a ) if os.path.isfile(os.path.join(_a ,_a ) )] if identifier is not None: _a : List[str] = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(_a ,_a ): for n_ in n_identifier: _a : Tuple = [file for file in files if n_ not in file] else: _a : Optional[Any] = [file for file in files if n_identifier not in file] _a : List[str] = ignore_files or [] ignore_files.append('__init__.py' ) _a : Tuple = [file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' ,_a ) if only_modules: _a : Any = file.split('.' )[0] try: _a : List[str] = getattr(_a ,_a ) _a : int = doctest.DocTestSuite(_a ) _a : Any = unittest.TextTestRunner().run(_a ) self.assertIs(len(result.failures ) ,0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: _a : Union[str, Any] = doctest.testfile(str('..' / directory / file ) ,optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed ,0 ) def __lowercase ( self : Any ): '''simple docstring''' _a : int = Path('src/transformers' ) _a : List[Any] = 'modeling' _a : Optional[Any] = [ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(_a ,identifier=_a ,ignore_files=_a ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Optional[Any] = Path('src/transformers' ) _a : Optional[Any] = 'tokenization' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Dict = Path('src/transformers' ) _a : str = 'configuration' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : Any ): '''simple docstring''' _a : Tuple = Path('src/transformers' ) _a : List[Any] = ['configuration', 'modeling', 'tokenization'] self.analyze_directory(_a ,n_identifier=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = Path('docs/source' ) _a : List[str] = ['favicon.ico'] self.analyze_directory(_a ,ignore_files=_a ,only_modules=_a )
271
0
def lowerCamelCase_ ( UpperCamelCase__ : list[list[int | float]] ) -> int: """simple docstring""" __lowerCamelCase = len(UpperCamelCase__ ) __lowerCamelCase = len(matrix[0] ) __lowerCamelCase = min(UpperCamelCase__ , UpperCamelCase__ ) for row in range(UpperCamelCase__ ): # Check if diagonal element is not zero if matrix[row][row] != 0: # Eliminate all the elements below the diagonal for col in range(row + 1 , UpperCamelCase__ ): __lowerCamelCase = matrix[col][row] / matrix[row][row] for i in range(UpperCamelCase__ , UpperCamelCase__ ): matrix[col][i] -= multiplier * matrix[row][i] else: # Find a non-zero diagonal element to swap rows __lowerCamelCase = True for i in range(row + 1 , UpperCamelCase__ ): if matrix[i][row] != 0: __lowerCamelCase , __lowerCamelCase = matrix[i], matrix[row] __lowerCamelCase = False break if reduce: rank -= 1 for i in range(UpperCamelCase__ ): __lowerCamelCase = matrix[i][rank] # Reduce the row pointer by one to stay on the same row row -= 1 return rank if __name__ == "__main__": import doctest doctest.testmod()
90
'''simple docstring''' import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def UpperCAmelCase_ (__a : Optional[Any] , __a : str , __a : Optional[Any]=None ): """simple docstring""" assert torch_layer.weight.shape == weight.shape, f"""{torch_layer} layer.weight does not match""" _a : str = nn.Parameter(__a ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f"""{torch_layer} layer.bias does not match""" _a : Any = nn.Parameter(__a ) def UpperCAmelCase_ (__a : int , __a : Optional[Any] , __a : int ): """simple docstring""" _a : Tuple = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : Dict = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Optional[Any] , __a : Optional[int] , __a : List[str] ): """simple docstring""" _a : Dict = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : str = np.asarray(weights[2] ) _a : int = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Any , __a : Any , __a : Optional[Any] ): """simple docstring""" _a : List[str] = weights[0][0][0] _a : List[Any] = np.asarray(layer_norm_a[0] ) _a : List[str] = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # lsh weights + output _a : List[str] = weights[0][1] if len(__a ) < 4: set_layer_weights_in_torch_lsh(__a , torch_block.attention , __a ) else: set_layer_weights_in_torch_local(__a , torch_block.attention , __a ) # intermediate weighs _a : Optional[Any] = weights[2][0][1][2] # Chunked Feed Forward if len(__a ) == 4: _a : Union[str, Any] = intermediate_weights[2] # layernorm 2 _a : Any = np.asarray(intermediate_weights[0][0] ) _a : List[Any] = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # intermediate dense _a : Any = np.asarray(intermediate_weights[1][0] ) _a : Any = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) # intermediate out _a : Optional[int] = np.asarray(intermediate_weights[4][0] ) _a : int = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Dict , __a : Dict , __a : List[Any] ): """simple docstring""" _a : Optional[int] = torch_model.reformer # word embeds _a : Tuple = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(__a ) , ) if isinstance(weights[3] , __a ): _a : Any = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _a : List[Any] = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f"""{position_embeddings[emb_idx]} emb does not match""" _a : Any = nn.Parameter(torch.tensor(__a ) ) _a : List[str] = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( __a ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _a : Tuple = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(__a , __a , __a ) # output layer norm _a : Optional[Any] = np.asarray(weights[7][0] ) _a : int = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # output embeddings _a : List[str] = np.asarray(weights[9][0] ) _a : int = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Tuple , __a : Optional[Any] , __a : Dict ): """simple docstring""" _a : List[Any] = ReformerConfig.from_json_file(__a ) print(f"""Building PyTorch model from configuration: {config}""" ) _a : int = ReformerModelWithLMHead(__a ) with open(__a , 'rb' ) as f: _a : Optional[Any] = pickle.load(__a )['weights'] set_model_weights_in_torch(__a , __a , config.hidden_size ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , __a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--trax_model_pkl_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained Reformer model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __lowerCAmelCase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
271
0
"""simple docstring""" from math import isqrt, loga def _A (__a ) -> list[int]: """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , __a , __a ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = False return [i for i in range(2 , __a ) if is_prime[i]] def _A (__a = 80_08_00 , __a = 80_08_00 ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = degree * loga(__a ) SCREAMING_SNAKE_CASE_ : str = int(__a ) SCREAMING_SNAKE_CASE_ : Optional[int] = calculate_prime_numbers(__a ) SCREAMING_SNAKE_CASE_ : int = 0 SCREAMING_SNAKE_CASE_ : List[Any] = 0 SCREAMING_SNAKE_CASE_ : List[Any] = len(__a ) - 1 while left < right: while ( prime_numbers[right] * loga(prime_numbers[left] ) + prime_numbers[left] * loga(prime_numbers[right] ) > upper_bound ): right -= 1 hybrid_integers_count += right - left left += 1 return hybrid_integers_count if __name__ == "__main__": print(f'''{solution() = }''')
91
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Union[str, Any] = VQModel( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=3 ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1000 ,) return CLIPTextModel(_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : Dict = self.dummy_uncond_unet _a : List[Any] = DDIMScheduler() _a : List[Any] = self.dummy_vq_model _a : str = LDMPipeline(unet=_a ,vqvae=_a ,scheduler=_a ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : List[str] = torch.manual_seed(0 ) _a : List[str] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Union[str, Any] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ,return_dict=_a )[0] _a : Tuple = image[0, -3:, -3:, -1] _a : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _a : int = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) _a : Any = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : List[str] = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Dict = ldm(generator=_a ,num_inference_steps=5 ,output_type='numpy' ).images _a : str = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _a : Optional[int] = np.array([0.4399, 0.4_4975, 0.4_6825, 0.474, 0.4359, 0.4581, 0.4_5095, 0.4341, 0.4447] ) _a : int = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
271
0
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { """BAAI/AltCLIP""": """https://huggingface.co/BAAI/AltCLIP/resolve/main/config.json""", # See all AltCLIP models at https://huggingface.co/models?filter=altclip } class a__ ( snake_case__ ): _a : List[Any] = """altclip_text_model""" def __init__( self , _A=2_5_0_0_0_2 , _A=1_0_2_4 , _A=2_4 , _A=1_6 , _A=4_0_9_6 , _A="gelu" , _A=0.1 , _A=0.1 , _A=5_1_4 , _A=1 , _A=0.02 , _A=0.02 , _A=1E-0_5 , _A=1 , _A=0 , _A=2 , _A="absolute" , _A=True , _A=7_6_8 , **_A , ): """simple docstring""" super().__init__(pad_token_id=_A , bos_token_id=_A , eos_token_id=_A , **_A ) __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = hidden_act __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = type_vocab_size __lowerCAmelCase = initializer_range __lowerCAmelCase = initializer_factor __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = position_embedding_type __lowerCAmelCase = use_cache __lowerCAmelCase = project_dim class a__ ( snake_case__ ): _a : List[Any] = """altclip_vision_model""" def __init__( self , _A=7_6_8 , _A=3_0_7_2 , _A=5_1_2 , _A=1_2 , _A=1_2 , _A=3 , _A=2_2_4 , _A=3_2 , _A="quick_gelu" , _A=1E-5 , _A=0.0 , _A=0.02 , _A=1.0 , **_A , ): """simple docstring""" super().__init__(**_A ) __lowerCAmelCase = hidden_size __lowerCAmelCase = intermediate_size __lowerCAmelCase = projection_dim __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = num_channels __lowerCAmelCase = patch_size __lowerCAmelCase = image_size __lowerCAmelCase = initializer_range __lowerCAmelCase = initializer_factor __lowerCAmelCase = attention_dropout __lowerCAmelCase = layer_norm_eps __lowerCAmelCase = hidden_act @classmethod def __SCREAMING_SNAKE_CASE( cls , _A , **_A ): """simple docstring""" cls._set_token_in_kwargs(_A ) __lowerCAmelCase , __lowerCAmelCase = cls.get_config_dict(_A , **_A ) # get the vision config dict if we are loading from AltCLIPConfig if config_dict.get("model_type" ) == "altclip": __lowerCAmelCase = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict['model_type']} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(_A , **_A ) class a__ ( snake_case__ ): _a : Any = """altclip""" _a : Optional[int] = True def __init__( self , _A=None , _A=None , _A=7_6_8 , _A=2.65_92 , **_A ): """simple docstring""" __lowerCAmelCase = kwargs.pop("text_config_dict" , _A ) __lowerCAmelCase = kwargs.pop("vision_config_dict" , _A ) super().__init__(**_A ) # Instead of simply assigning `[text|vision]_config_dict` to `[text|vision]_config`, we use the values in # `[text|vision]_config_dict` to update the values in `[text|vision]_config`. The values should be same in most # cases, but we don't want to break anything regarding `_config_dict` that existed before commit `8827e1b2`. if text_config_dict is not None: if text_config is None: __lowerCAmelCase = {} # This is the complete result when using `text_config_dict`. __lowerCAmelCase = AltCLIPTextConfig(**_A ).to_dict() # Give a warning if the values exist in both `_text_config_dict` and `text_config` but being different. for key, value in _text_config_dict.items(): if key in text_config and value != text_config[key] and key not in ["transformers_version"]: # If specified in `text_config_dict` if key in text_config_dict: __lowerCAmelCase = ( f"""`{key}` is found in both `text_config_dict` and `text_config` but with different values. """ f"""The value `text_config_dict[\"{key}\"]` will be used instead.""" ) # If inferred from default argument values (just to be super careful) else: __lowerCAmelCase = ( f"""`text_config_dict` is provided which will be used to initialize `AltCLIPTextConfig`. The """ f"""value `text_config[\"{key}\"]` will be overriden.""" ) logger.warning(_A ) # Update all values in `text_config` with the ones in `_text_config_dict`. text_config.update(_text_config_dict ) if vision_config_dict is not None: if vision_config is None: __lowerCAmelCase = {} # This is the complete result when using `vision_config_dict`. __lowerCAmelCase = AltCLIPVisionConfig(**_A ).to_dict() # convert keys to string instead of integer if "id2label" in _vision_config_dict: __lowerCAmelCase = { str(_A ): value for key, value in _vision_config_dict["id2label"].items() } # Give a warning if the values exist in both `_vision_config_dict` and `vision_config` but being different. for key, value in _vision_config_dict.items(): if key in vision_config and value != vision_config[key] and key not in ["transformers_version"]: # If specified in `vision_config_dict` if key in vision_config_dict: __lowerCAmelCase = ( f"""`{key}` is found in both `vision_config_dict` and `vision_config` but with different """ f"""values. The value `vision_config_dict[\"{key}\"]` will be used instead.""" ) # If inferred from default argument values (just to be super careful) else: __lowerCAmelCase = ( f"""`vision_config_dict` is provided which will be used to initialize `AltCLIPVisionConfig`. """ f"""The value `vision_config[\"{key}\"]` will be overriden.""" ) logger.warning(_A ) # Update all values in `vision_config` with the ones in `_vision_config_dict`. vision_config.update(_vision_config_dict ) if text_config is None: __lowerCAmelCase = {} logger.info("`text_config` is `None`. Initializing the `AltCLIPTextConfig` with default values." ) if vision_config is None: __lowerCAmelCase = {} logger.info("`vision_config` is `None`. initializing the `AltCLIPVisionConfig` with default values." ) __lowerCAmelCase = AltCLIPTextConfig(**_A ) __lowerCAmelCase = AltCLIPVisionConfig(**_A ) __lowerCAmelCase = projection_dim __lowerCAmelCase = logit_scale_init_value __lowerCAmelCase = 1.0 @classmethod def __SCREAMING_SNAKE_CASE( cls , _A , _A , **_A ): """simple docstring""" return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **_A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = copy.deepcopy(self.__dict__ ) __lowerCAmelCase = self.text_config.to_dict() __lowerCAmelCase = self.vision_config.to_dict() __lowerCAmelCase = self.__class__.model_type return output
92
'''simple docstring''' import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : int ,*_a : Optional[int] ,**_a : str ): '''simple docstring''' warnings.warn( 'The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use BeitImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
'''simple docstring''' def snake_case_ ( __SCREAMING_SNAKE_CASE : int = 50000000 ): """simple docstring""" lowercase_ : Any = set() lowercase_ : List[Any] = int((limit - 24) ** (1 / 2) ) lowercase_ : str = set(range(3 , prime_square_limit + 1 , 2 ) ) primes.add(2 ) for p in range(3 , prime_square_limit + 1 , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , prime_square_limit + 1 , __SCREAMING_SNAKE_CASE ) ) ) for primea in primes: lowercase_ : Dict = primea * primea for primea in primes: lowercase_ : Dict = primea * primea * primea if square + cube >= limit - 16: break for primea in primes: lowercase_ : Optional[int] = primea * primea * primea * primea lowercase_ : Any = square + cube + tetr if total >= limit: break ret.add(__SCREAMING_SNAKE_CASE ) return len(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(f"""{solution() = }""")
93
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { """linear""": get_linear_schedule_with_warmup, """cosine""": get_cosine_schedule_with_warmup, """cosine_w_restarts""": get_cosine_with_hard_restarts_schedule_with_warmup, """polynomial""": get_polynomial_decay_schedule_with_warmup, """constant""": get_constant_schedule, """constant_w_warmup""": get_constant_schedule_with_warmup, } class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Optional[int] ,_a : Optional[Any]=None ,_a : Dict=None ,*_a : int ,**_a : str ): '''simple docstring''' super().__init__(*_a ,**_a ) if config is None: assert isinstance(self.model ,_a ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F""" {self.model.__class__}""" ) _a : List[Any] = self.model.config else: _a : Optional[int] = config _a : List[str] = data_args _a : List[Any] = self.config.tgt_vocab_size if isinstance(self.config ,_a ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ' padding..' ) if self.args.label_smoothing == 0: _a : List[str] = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss _a : Tuple = label_smoothed_nll_loss def __lowercase ( self : List[str] ,_a : int ): '''simple docstring''' if self.optimizer is None: _a : Union[str, Any] = ['bias', 'LayerNorm.weight'] _a : Tuple = [ { 'params': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], 'weight_decay': self.args.weight_decay, }, { 'params': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], 'weight_decay': 0.0, }, ] _a : Optional[int] = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: _a : Any = Adafactor _a : Dict = {'scale_parameter': False, 'relative_step': False} else: _a : Union[str, Any] = AdamW _a : str = { 'betas': (self.args.adam_betaa, self.args.adam_betaa), 'eps': self.args.adam_epsilon, } _a : Union[str, Any] = self.args.learning_rate if self.sharded_ddp: _a : str = OSS( params=_a ,optim=_a ,**_a ,) else: _a : Tuple = optimizer_cls(_a ,**_a ) if self.lr_scheduler is None: _a : List[Any] = self._get_lr_scheduler(_a ) else: # ignoring --lr_scheduler logger.warning('scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.' ) def __lowercase ( self : List[Any] ,_a : List[Any] ): '''simple docstring''' _a : str = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": _a : int = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": _a : List[str] = schedule_func(self.optimizer ,num_warmup_steps=self.args.warmup_steps ) else: _a : Optional[int] = schedule_func( self.optimizer ,num_warmup_steps=self.args.warmup_steps ,num_training_steps=_a ) return scheduler def __lowercase ( self : Tuple ): '''simple docstring''' if isinstance(self.train_dataset ,torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size ,distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) ,) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def __lowercase ( self : Dict ,_a : Dict ,_a : Any ,_a : Dict ): '''simple docstring''' if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Union[str, Any] = self.loss_fn(logits.view(-1 ,logits.shape[-1] ) ,labels.view(-1 ) ) else: # compute usual loss via models _a, _a : Union[str, Any] = model(**_a ,labels=_a ,use_cache=_a )[:2] else: # compute label smoothed loss _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Any = torch.nn.functional.log_softmax(_a ,dim=-1 ) _a, _a : List[str] = self.loss_fn(_a ,_a ,self.args.label_smoothing ,ignore_index=self.config.pad_token_id ) return loss, logits def __lowercase ( self : Optional[int] ,_a : Union[str, Any] ,_a : List[Any] ): '''simple docstring''' _a : Optional[int] = inputs.pop('labels' ) _a, _a : int = self._compute_loss(_a ,_a ,_a ) return loss def __lowercase ( self : Optional[Any] ,_a : nn.Module ,_a : Dict[str, Union[torch.Tensor, Any]] ,_a : bool ,_a : Optional[List[str]] = None ,): '''simple docstring''' _a : int = self._prepare_inputs(_a ) _a : Any = { 'max_length': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, 'num_beams': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: _a : int = self.model.generate( inputs['input_ids'] ,attention_mask=inputs['attention_mask'] ,**_a ,) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: _a : int = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) _a : Union[str, Any] = inputs.pop('labels' ) with torch.no_grad(): # compute loss on predict data _a, _a : Optional[int] = self._compute_loss(_a ,_a ,_a ) _a : Optional[Any] = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) _a : Optional[Any] = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: _a : Dict = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) return (loss, logits, labels) def __lowercase ( self : str ,_a : Tuple ,_a : Tuple ): '''simple docstring''' _a : List[Any] = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( 'Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be' F""" padded to `max_length`={max_length}""" ) _a : int = pad_token_id * torch.ones( (tensor.shape[0], max_length) ,dtype=tensor.dtype ,device=tensor.device ) _a : Union[str, Any] = tensor return padded_tensor
271
0
import qiskit def __lowerCamelCase ( UpperCAmelCase_ : int = 2 ): """simple docstring""" a :Tuple = qubits # Using Aer's simulator a :Union[str, Any] = qiskit.Aer.get_backend('''aer_simulator''' ) # Creating a Quantum Circuit acting on the q register a :str = qiskit.QuantumCircuit(UpperCAmelCase_ , UpperCAmelCase_ ) # Adding a H gate on qubit 0 (now q0 in superposition) circuit.h(0 ) for i in range(1 , UpperCAmelCase_ ): # Adding CX (CNOT) gate circuit.cx(i - 1 , UpperCAmelCase_ ) # Mapping the quantum measurement to the classical bits circuit.measure(list(range(UpperCAmelCase_ ) ) , list(range(UpperCAmelCase_ ) ) ) # Now measuring any one qubit would affect other qubits to collapse # their super position and have same state as the measured one. # Executing the circuit on the simulator a :Union[str, Any] = qiskit.execute(UpperCAmelCase_ , UpperCAmelCase_ , shots=1000 ) return job.result().get_counts(UpperCAmelCase_ ) if __name__ == "__main__": print(F"""Total count for various states are: {quantum_entanglement(3)}""")
94
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser __lowerCAmelCase = re.compile(r"""\s+""") def UpperCAmelCase_ (__a : Any ): """simple docstring""" return {"hash": hashlib.mda(re.sub(__a , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : List[str] = [len(__a ) for line in example['content'].splitlines()] return {"line_mean": np.mean(__a ), "line_max": max(__a )} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Union[str, Any] = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase_ (__a : Optional[int] , __a : Any ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def UpperCAmelCase_ (__a : int , __a : Union[str, Any]=5 ): """simple docstring""" _a : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] _a : List[str] = example['content'].splitlines() for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase_ (__a : List[str] , __a : Dict=5 , __a : Tuple=0.05 ): """simple docstring""" _a : Optional[int] = ['unit tests', 'test file', 'configuration file'] _a : int = example['content'].splitlines() _a : int = 0 _a : Dict = 0 # first test for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test _a : int = example['content'].count('\n' ) _a : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" _a : List[str] = ['def ', 'class ', 'for ', 'while '] _a : str = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase_ (__a : int , __a : Any=4 ): """simple docstring""" _a : List[str] = example['content'].splitlines() _a : Dict = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Optional[Any] = tokenizer(example['content'] , truncation=__a )['input_ids'] _a : Optional[int] = len(example['content'] ) / len(__a ) return {"ratio": ratio} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Dict = {} results.update(get_hash(__a ) ) results.update(line_stats(__a ) ) results.update(alpha_stats(__a ) ) results.update(char_token_ratio(__a ) ) results.update(is_autogenerated(__a ) ) results.update(is_config_or_test(__a ) ) results.update(has_no_keywords(__a ) ) results.update(has_few_assignments(__a ) ) return results def UpperCAmelCase_ (__a : Any , __a : Any , __a : str ): """simple docstring""" if not check_uniques(__a , __a ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase_ (__a : Union[str, Any] ): """simple docstring""" with open(__a , 'rb' ) as f_in: with gzip.open(str(__a ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(__a , __a ) os.unlink(__a ) # Settings __lowerCAmelCase = HfArgumentParser(PreprocessingArguments) __lowerCAmelCase = parser.parse_args() if args.num_workers is None: __lowerCAmelCase = multiprocessing.cpu_count() __lowerCAmelCase = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset __lowerCAmelCase = time.time() __lowerCAmelCase = load_dataset(args.dataset_name, split="""train""") print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing __lowerCAmelCase = time.time() __lowerCAmelCase = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes __lowerCAmelCase = set(ds.unique("""hash""")) __lowerCAmelCase = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics __lowerCAmelCase = time.time() __lowerCAmelCase = ds.filter(filter, fn_kwargs={"""uniques""": uniques, """args""": args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: __lowerCAmelCase = time.time() __lowerCAmelCase , __lowerCAmelCase = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file __lowerCAmelCase = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / """duplicate_clusters.json""", """w""") as f: json.dump(duplicate_clusters, f) __lowerCAmelCase = output_dir / """data""" data_dir.mkdir(exist_ok=True) __lowerCAmelCase = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): __lowerCAmelCase = str(data_dir / f'''file-{file_number+1:012}.json''') __lowerCAmelCase = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
271
0
def _A ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int ): """simple docstring""" a__ : int =(num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def _A ( ): """simple docstring""" print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
95
'''simple docstring''' import argparse from typing import List import evaluate import numpy as np import torch from datasets import DatasetDict, load_dataset # New Code # # We'll be using StratifiedKFold for this example from sklearn.model_selection import StratifiedKFold from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to perform Cross Validation, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __lowerCAmelCase = 1_6 __lowerCAmelCase = 3_2 def UpperCAmelCase_ (__a : Accelerator , __a : DatasetDict , __a : List[int] , __a : List[int] , __a : int = 1_6 ): """simple docstring""" _a : Union[str, Any] = AutoTokenizer.from_pretrained('bert-base-cased' ) _a : str = DatasetDict( { 'train': dataset['train'].select(__a ), 'validation': dataset['train'].select(__a ), 'test': dataset['validation'], } ) def tokenize_function(__a : List[Any] ): # max_length=None => use the model max length (it's actually the default) _a : Optional[int] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=__a , max_length=__a ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _a : List[str] = datasets.map( __a , batched=__a , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _a : List[Any] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(__a : int ): # On TPU it's best to pad everything to the same length or training will be very slow. _a : Dict = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _a : Tuple = 1_6 elif accelerator.mixed_precision != "no": _a : List[Any] = 8 else: _a : List[Any] = None return tokenizer.pad( __a , padding='longest' , max_length=__a , pad_to_multiple_of=__a , return_tensors='pt' , ) # Instantiate dataloaders. _a : Any = DataLoader( tokenized_datasets['train'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[int] = DataLoader( tokenized_datasets['validation'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[Any] = DataLoader( tokenized_datasets['test'] , shuffle=__a , collate_fn=__a , batch_size=__a ) return train_dataloader, eval_dataloader, test_dataloader def UpperCAmelCase_ (__a : Any , __a : Union[str, Any] ): """simple docstring""" _a : Dict = [] # Download the dataset _a : Tuple = load_dataset('glue' , 'mrpc' ) # Create our splits _a : Union[str, Any] = StratifiedKFold(n_splits=int(args.num_folds ) ) # Initialize accelerator _a : Any = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _a : Optional[Any] = config['lr'] _a : Optional[int] = int(config['num_epochs'] ) _a : Dict = int(config['seed'] ) _a : Dict = int(config['batch_size'] ) _a : Optional[int] = evaluate.load('glue' , 'mrpc' ) # If the batch size is too big we use gradient accumulation _a : List[Any] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: _a : Any = batch_size // MAX_GPU_BATCH_SIZE _a : List[str] = MAX_GPU_BATCH_SIZE set_seed(__a ) # New Code # # Create our folds: _a : int = kfold.split(np.zeros(datasets['train'].num_rows ) , datasets['train']['label'] ) _a : Any = [] # Iterate over them for i, (train_idxs, valid_idxs) in enumerate(__a ): _a, _a, _a : Optional[Any] = get_fold_dataloaders( __a , __a , __a , __a , ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _a : Dict = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=__a ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _a : List[Any] = model.to(accelerator.device ) # Instantiate optimizer _a : List[str] = AdamW(params=model.parameters() , lr=__a ) # Instantiate scheduler _a : List[Any] = get_linear_schedule_with_warmup( optimizer=__a , num_warmup_steps=1_0_0 , num_training_steps=(len(__a ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _a, _a, _a, _a, _a : Union[str, Any] = accelerator.prepare( __a , __a , __a , __a , __a ) # Now we train the model for epoch in range(__a ): model.train() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _a : Dict = model(**__a ) _a : int = outputs.loss _a : Any = loss / gradient_accumulation_steps accelerator.backward(__a ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Union[str, Any] = model(**__a ) _a : Tuple = outputs.logits.argmax(dim=-1 ) _a, _a : Any = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=__a , references=__a , ) _a : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , __a ) # New Code # # We also run predictions on the test set at the very end _a : Any = [] for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Tuple = model(**__a ) _a : Dict = outputs.logits _a, _a : Optional[int] = accelerator.gather_for_metrics((predictions, batch['labels']) ) fold_predictions.append(predictions.cpu() ) if i == 0: # We need all of the test predictions test_references.append(references.cpu() ) # Use accelerator.print to print only on the main process. test_predictions.append(torch.cat(__a , dim=0 ) ) # We now need to release all our memory and get rid of the current model, optimizer, etc accelerator.free_memory() # New Code # # Finally we check the accuracy of our folded results: _a : Dict = torch.cat(__a , dim=0 ) _a : Any = torch.stack(__a , dim=0 ).sum(dim=0 ).div(int(args.num_folds ) ).argmax(dim=-1 ) _a : str = metric.compute(predictions=__a , references=__a ) accelerator.print('Average test metrics from all folds:' , __a ) def UpperCAmelCase_ (): """simple docstring""" _a : Any = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=__a , default=__a , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) # New Code # parser.add_argument('--num_folds' , type=__a , default=3 , help='The number of splits to perform across the dataset' ) _a : Any = parser.parse_args() _a : int = {'lr': 2e-5, 'num_epochs': 3, 'seed': 4_2, 'batch_size': 1_6} training_function(__a , __a ) if __name__ == "__main__": main()
271
0
"""simple docstring""" import gc import unittest import numpy as np import torch from diffusers import StableDiffusionKDiffusionPipeline from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def A_ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A_ ( self ): _lowerCamelCase : Tuple = StableDiffusionKDiffusionPipeline.from_pretrained('CompVis/stable-diffusion-v1-4' ) _lowerCamelCase : Any = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) sd_pipe.set_scheduler('sample_euler' ) _lowerCamelCase : Dict = 'A painting of a squirrel eating a burger' _lowerCamelCase : List[Any] = torch.manual_seed(0 ) _lowerCamelCase : Dict = sd_pipe([prompt] , generator=lowercase , guidance_scale=9.0 , num_inference_steps=20 , output_type='np' ) _lowerCamelCase : Union[str, Any] = output.images _lowerCamelCase : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _lowerCamelCase : Dict = np.array([0.04_47, 0.04_92, 0.04_68, 0.04_08, 0.03_83, 0.04_08, 0.03_54, 0.03_80, 0.03_39] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A_ ( self ): _lowerCamelCase : Any = StableDiffusionKDiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base' ) _lowerCamelCase : Any = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) sd_pipe.set_scheduler('sample_euler' ) _lowerCamelCase : Union[str, Any] = 'A painting of a squirrel eating a burger' _lowerCamelCase : List[str] = torch.manual_seed(0 ) _lowerCamelCase : Tuple = sd_pipe([prompt] , generator=lowercase , guidance_scale=9.0 , num_inference_steps=20 , output_type='np' ) _lowerCamelCase : Optional[int] = output.images _lowerCamelCase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _lowerCamelCase : List[str] = np.array([0.12_37, 0.13_20, 0.14_38, 0.13_59, 0.13_90, 0.11_32, 0.12_77, 0.11_75, 0.11_12] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-1 def A_ ( self ): _lowerCamelCase : Union[str, Any] = StableDiffusionKDiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base' ) _lowerCamelCase : int = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) sd_pipe.set_scheduler('sample_dpmpp_2m' ) _lowerCamelCase : int = 'A painting of a squirrel eating a burger' _lowerCamelCase : List[str] = torch.manual_seed(0 ) _lowerCamelCase : Any = sd_pipe( [prompt] , generator=lowercase , guidance_scale=7.5 , num_inference_steps=15 , output_type='np' , use_karras_sigmas=lowercase , ) _lowerCamelCase : List[Any] = output.images _lowerCamelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _lowerCamelCase : Any = np.array( [0.11_38_16_89, 0.12_11_29_21, 0.1_38_94_57, 0.12_54_96_06, 0.1_24_49_64, 0.10_83_15_17, 0.11_56_28_66, 0.10_86_78_16, 0.10_49_90_48] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
96
'''simple docstring''' from __future__ import annotations __lowerCAmelCase = [-1_0, -5, 0, 5, 5.1, 1_1, 1_3, 2_1, 3, 4, -2_1, -1_0, -5, -1, 0] __lowerCAmelCase = [-5, 0, 5, 5.1, 1_1, 1_3, 2_1, -1, 4, -1, -1_0, -5, -1, 0, -1] def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Optional[int] = [] _a : int = len(__a ) for i in range(__a ): _a : float = -1 for j in range(i + 1 , __a ): if arr[i] < arr[j]: _a : Any = arr[j] break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Tuple = [] for i, outer in enumerate(__a ): _a : float = -1 for inner in arr[i + 1 :]: if outer < inner: _a : Dict = inner break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : int = len(__a ) _a : list[float] = [] _a : list[float] = [-1] * arr_size for index in reversed(range(__a ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: _a : Dict = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) __lowerCAmelCase = ( """from __main__ import arr, next_greatest_element_slow, """ """next_greatest_element_fast, next_greatest_element""" ) print( """next_greatest_element_slow():""", timeit("""next_greatest_element_slow(arr)""", setup=setup), ) print( """next_greatest_element_fast():""", timeit("""next_greatest_element_fast(arr)""", setup=setup), ) print( """ next_greatest_element():""", timeit("""next_greatest_element(arr)""", setup=setup), )
271
0
'''simple docstring''' import collections import inspect import unittest from transformers import FocalNetConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, ) from transformers.models.focalnet.modeling_focalnet import FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase : """simple docstring""" def __init__( self , UpperCamelCase_ , UpperCamelCase_=13 , UpperCamelCase_=32 , UpperCamelCase_=2 , UpperCamelCase_=3 , UpperCamelCase_=16 , UpperCamelCase_=[32, 64, 128] , UpperCamelCase_=[1, 2, 1] , UpperCamelCase_=[2, 2, 4] , UpperCamelCase_=2 , UpperCamelCase_=2.0 , UpperCamelCase_=True , UpperCamelCase_=0.0 , UpperCamelCase_=0.0 , UpperCamelCase_=0.1 , UpperCamelCase_="gelu" , UpperCamelCase_=False , UpperCamelCase_=True , UpperCamelCase_=0.02 , UpperCamelCase_=1e-5 , UpperCamelCase_=True , UpperCamelCase_=None , UpperCamelCase_=True , UpperCamelCase_=10 , UpperCamelCase_=8 , UpperCamelCase_=["stage1", "stage2"] , UpperCamelCase_=[1, 2] , ): '''simple docstring''' UpperCamelCase__ :Tuple = parent UpperCamelCase__ :Any = batch_size UpperCamelCase__ :int = image_size UpperCamelCase__ :int = patch_size UpperCamelCase__ :int = num_channels UpperCamelCase__ :Union[str, Any] = embed_dim UpperCamelCase__ :int = hidden_sizes UpperCamelCase__ :Optional[int] = depths UpperCamelCase__ :Optional[Any] = num_heads UpperCamelCase__ :Optional[Any] = window_size UpperCamelCase__ :List[Any] = mlp_ratio UpperCamelCase__ :List[Any] = qkv_bias UpperCamelCase__ :List[Any] = hidden_dropout_prob UpperCamelCase__ :List[str] = attention_probs_dropout_prob UpperCamelCase__ :List[Any] = drop_path_rate UpperCamelCase__ :int = hidden_act UpperCamelCase__ :List[str] = use_absolute_embeddings UpperCamelCase__ :Optional[int] = patch_norm UpperCamelCase__ :int = layer_norm_eps UpperCamelCase__ :str = initializer_range UpperCamelCase__ :Any = is_training UpperCamelCase__ :Optional[int] = scope UpperCamelCase__ :Dict = use_labels UpperCamelCase__ :List[str] = type_sequence_label_size UpperCamelCase__ :Union[str, Any] = encoder_stride UpperCamelCase__ :List[Any] = out_features UpperCamelCase__ :int = out_indices def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase__ :int = None if self.use_labels: UpperCamelCase__ :List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase__ :Union[str, Any] = self.get_config() return config, pixel_values, labels def lowerCAmelCase__ ( self ): '''simple docstring''' return FocalNetConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , hidden_sizes=self.hidden_sizes , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :Tuple = FocalNetModel(config=UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() UpperCamelCase__ :int = model(UpperCamelCase_ ) UpperCamelCase__ :int = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) UpperCamelCase__ :List[Any] = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :Tuple = FocalNetBackbone(config=UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() UpperCamelCase__ :List[Any] = model(UpperCamelCase_ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size, 8, 8] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[:-1] ) # verify backbone works with out_features=None UpperCamelCase__ :Optional[Any] = None UpperCamelCase__ :Union[str, Any] = FocalNetBackbone(config=UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() UpperCamelCase__ :Optional[int] = model(UpperCamelCase_ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size * 2, 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :int = FocalNetForMaskedImageModeling(config=UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() UpperCamelCase__ :List[Any] = model(UpperCamelCase_ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCamelCase__ :Dict = 1 UpperCamelCase__ :List[Any] = FocalNetForMaskedImageModeling(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() UpperCamelCase__ :Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCamelCase__ :Optional[int] = model(UpperCamelCase_ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = self.type_sequence_label_size UpperCamelCase__ :List[Any] = FocalNetForImageClassification(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() UpperCamelCase__ :Tuple = model(UpperCamelCase_ , labels=UpperCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCamelCase__ :int = 1 UpperCamelCase__ :Dict = FocalNetForImageClassification(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() UpperCamelCase__ :Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCamelCase__ :Dict = model(UpperCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :str = self.prepare_config_and_inputs() UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ :Any = config_and_inputs UpperCamelCase__ :str = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase ( A__ , A__ , unittest.TestCase ): """simple docstring""" _a = ( ( FocalNetModel, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetBackbone, ) if is_torch_available() else () ) _a = ( {'feature-extraction': FocalNetModel, 'image-classification': FocalNetForImageClassification} if is_torch_available() else {} ) _a = False _a = False _a = False _a = False _a = False def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Dict = FocalNetModelTester(self ) UpperCamelCase__ :Optional[Any] = ConfigTester(self , config_class=UpperCamelCase_ , embed_dim=37 , has_text_modality=UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCAmelCase__ ( self ): '''simple docstring''' return def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase_ ) @unittest.skip(reason='''FocalNet does not use inputs_embeds''' ) def lowerCAmelCase__ ( self ): '''simple docstring''' pass @unittest.skip(reason='''FocalNet does not use feedforward chunking''' ) def lowerCAmelCase__ ( self ): '''simple docstring''' pass def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ , UpperCamelCase__ :Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: UpperCamelCase__ :Optional[Any] = model_class(UpperCamelCase_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCamelCase__ :Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCamelCase_ , nn.Linear ) ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ , UpperCamelCase__ :Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: UpperCamelCase__ :Tuple = model_class(UpperCamelCase_ ) UpperCamelCase__ :int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase__ :str = [*signature.parameters.keys()] UpperCamelCase__ :Union[str, Any] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , UpperCamelCase_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = model_class(UpperCamelCase_ ) model.to(UpperCamelCase_ ) model.eval() with torch.no_grad(): UpperCamelCase__ :Tuple = model(**self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_ ) ) UpperCamelCase__ :Optional[int] = outputs.hidden_states UpperCamelCase__ :int = getattr( self.model_tester , '''expected_num_hidden_layers''' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(UpperCamelCase_ ) , UpperCamelCase_ ) # FocalNet has a different seq_length UpperCamelCase__ :Optional[Any] = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) UpperCamelCase__ :int = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) UpperCamelCase__ :Optional[int] = outputs.reshaped_hidden_states self.assertEqual(len(UpperCamelCase_ ) , UpperCamelCase_ ) UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ :str = reshaped_hidden_states[0].shape UpperCamelCase__ :Union[str, Any] = ( reshaped_hidden_states[0].view(UpperCamelCase_ , UpperCamelCase_ , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ , UpperCamelCase__ :Dict = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase__ :Any = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes[:-1]: UpperCamelCase__ :Tuple = True self.check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase__ :str = True self.check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ , UpperCamelCase__ :Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase__ :Union[str, Any] = 3 UpperCamelCase__ :Dict = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) UpperCamelCase__ :Any = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) UpperCamelCase__ :Optional[Any] = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) UpperCamelCase__ :Optional[Any] = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes[:-1]: UpperCamelCase__ :List[Any] = True self.check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase__ :str = True self.check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , (padded_height, padded_width) ) @slow def lowerCAmelCase__ ( self ): '''simple docstring''' for model_name in FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase__ :int = FocalNetModel.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ , UpperCamelCase__ :Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase__ :List[str] = _config_zero_init(UpperCamelCase_ ) for model_class in self.all_model_classes: UpperCamelCase__ :Optional[Any] = model_class(config=UpperCamelCase_ ) for name, param in model.named_parameters(): if "embeddings" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class lowercase ( unittest.TestCase ): """simple docstring""" @cached_property def lowerCAmelCase__ ( self ): '''simple docstring''' return AutoImageProcessor.from_pretrained('''microsoft/focalnet-tiny''' ) if is_vision_available() else None @slow def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Dict = FocalNetForImageClassification.from_pretrained('''microsoft/focalnet-tiny''' ).to(UpperCamelCase_ ) UpperCamelCase__ :str = self.default_image_processor UpperCamelCase__ :Dict = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) UpperCamelCase__ :int = image_processor(images=UpperCamelCase_ , return_tensors='''pt''' ).to(UpperCamelCase_ ) # forward pass with torch.no_grad(): UpperCamelCase__ :Optional[int] = model(**UpperCamelCase_ ) # verify the logits UpperCamelCase__ :List[Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , UpperCamelCase_ ) UpperCamelCase__ :Optional[int] = torch.tensor([0.2166, -0.4368, 0.2191] ).to(UpperCamelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCamelCase_ , atol=1e-4 ) ) self.assertTrue(outputs.logits.argmax(dim=-1 ).item() , 281 ) @require_torch class lowercase ( A__ , unittest.TestCase ): """simple docstring""" _a = (FocalNetBackbone,) if is_torch_available() else () _a = FocalNetConfig _a = False def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = FocalNetModelTester(self )
97
'''simple docstring''' import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home __lowerCAmelCase = HUGGINGFACE_HUB_CACHE __lowerCAmelCase = """config.json""" __lowerCAmelCase = """diffusion_pytorch_model.bin""" __lowerCAmelCase = """diffusion_flax_model.msgpack""" __lowerCAmelCase = """model.onnx""" __lowerCAmelCase = """diffusion_pytorch_model.safetensors""" __lowerCAmelCase = """weights.pb""" __lowerCAmelCase = """https://huggingface.co""" __lowerCAmelCase = default_cache_path __lowerCAmelCase = """diffusers_modules""" __lowerCAmelCase = os.getenv("""HF_MODULES_CACHE""", os.path.join(hf_cache_home, """modules""")) __lowerCAmelCase = ["""fp16""", """non-ema"""] __lowerCAmelCase = """.self_attn"""
271
0
"""simple docstring""" import requests lowerCAmelCase__ : Optional[int] = 'YOUR API KEY' def a_ ( lowerCamelCase , lowerCamelCase = giphy_api_key ): UpperCAmelCase__ = '+'.join(query.split() ) UpperCAmelCase__ = f'''https://api.giphy.com/v1/gifs/search?q={formatted_query}&api_key={api_key}''' UpperCAmelCase__ = requests.get(lowerCamelCase ).json()['data'] return [gif["url"] for gif in gifs] if __name__ == "__main__": print('\n'.join(get_gifs('space ship')))
98
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class UpperCAmelCase__ : """simple docstring""" def __init__( self : int ,_a : Any ,_a : Optional[int]=2 ,_a : Optional[Any]=True ,_a : Dict=False ,_a : Dict=10 ,_a : Any=3 ,_a : str=32 * 8 ,_a : Optional[int]=32 * 8 ,_a : int=4 ,_a : str=64 ,): '''simple docstring''' _a : Dict = parent _a : Union[str, Any] = batch_size _a : Tuple = is_training _a : List[str] = use_auxiliary_loss _a : Optional[Any] = num_queries _a : str = num_channels _a : List[str] = min_size _a : int = max_size _a : Optional[int] = num_labels _a : List[str] = hidden_dim _a : int = hidden_dim def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Tuple = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( _a ) _a : Optional[Any] = torch.ones([self.batch_size, self.min_size, self.max_size] ,device=_a ) _a : Union[str, Any] = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] ,device=_a ) > 0.5 ).float() _a : Tuple = (torch.rand((self.batch_size, self.num_labels) ,device=_a ) > 0.5).long() _a : Dict = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : int = MaskaFormerConfig( hidden_size=self.hidden_dim ,) _a : str = self.num_queries _a : Union[str, Any] = self.num_labels _a : Tuple = [1, 1, 1, 1] _a : Dict = self.num_channels _a : str = 64 _a : Tuple = 128 _a : Optional[Any] = self.hidden_dim _a : Union[str, Any] = self.hidden_dim _a : List[Any] = self.hidden_dim return config def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a, _a, _a, _a, _a : Optional[Any] = self.prepare_config_and_inputs() _a : str = {'pixel_values': pixel_values, 'pixel_mask': pixel_mask} return config, inputs_dict def __lowercase ( self : List[str] ,_a : Optional[Any] ,_a : str ): '''simple docstring''' _a : str = output.encoder_hidden_states _a : Any = output.pixel_decoder_hidden_states _a : Optional[Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,config.decoder_layers ) def __lowercase ( self : List[str] ,_a : str ,_a : List[Any] ,_a : Any ,_a : Union[str, Any]=False ): '''simple docstring''' with torch.no_grad(): _a : str = MaskaFormerModel(config=_a ) model.to(_a ) model.eval() _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[Any] = model(_a ,output_hidden_states=_a ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape ,(self.batch_size, self.num_queries, self.hidden_dim) ,) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(_a ,_a ) def __lowercase ( self : Tuple ,_a : List[Any] ,_a : Union[str, Any] ,_a : Tuple ,_a : List[str] ,_a : Any ): '''simple docstring''' _a : int = MaskaFormerForUniversalSegmentation(config=_a ) model.to(_a ) model.eval() def comm_check_on_output(_a : Any ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape ,(self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) ,) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape ,(self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[int] = model(_a ) comm_check_on_output(_a ) _a : List[str] = model( pixel_values=_a ,pixel_mask=_a ,mask_labels=_a ,class_labels=_a ) comm_check_on_output(_a ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape ,torch.Size([1] ) ) @require_torch class UpperCAmelCase__ ( lowercase__ , lowercase__ , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[int] = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () __UpperCAmelCase : Dict = {'''feature-extraction''': MaskaFormerModel} if is_torch_available() else {} __UpperCAmelCase : Dict = False __UpperCAmelCase : Tuple = False __UpperCAmelCase : Dict = False __UpperCAmelCase : List[Any] = False def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Union[str, Any] = MaskaFormerModelTester(self ) _a : Dict = ConfigTester(self ,config_class=_a ,has_text_modality=_a ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : Optional[int] ): '''simple docstring''' _a, _a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*_a ) @unittest.skip(reason='Mask2Former does not use inputs_embeds' ) def __lowercase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not have a get_input_embeddings method' ) def __lowercase ( self : str ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former is not a generative model' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not use token embeddings' ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason='Mask2Former has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def __lowercase ( self : Dict ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass def __lowercase ( self : int ): '''simple docstring''' _a, _a : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Union[str, Any] = model_class(_a ) _a : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a : Optional[Any] = [*signature.parameters.keys()] _a : List[Any] = ['pixel_values'] self.assertListEqual(arg_names[:1] ,_a ) @slow def __lowercase ( self : List[str] ): '''simple docstring''' for model_name in ["facebook/mask2former-swin-small-coco-instance"]: _a : Dict = MaskaFormerModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def __lowercase ( self : List[Any] ): '''simple docstring''' _a : int = (self.model_tester.min_size,) * 2 _a : Any = { 'pixel_values': torch.randn((2, 3, *size) ,device=_a ), 'mask_labels': torch.randn((2, 10, *size) ,device=_a ), 'class_labels': torch.zeros(2 ,10 ,device=_a ).long(), } _a : List[Any] = self.model_tester.get_config() _a : int = MaskaFormerForUniversalSegmentation(_a ).to(_a ) _a : str = model(**_a ) self.assertTrue(outputs.loss is not None ) def __lowercase ( self : List[str] ): '''simple docstring''' _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : int ): '''simple docstring''' _a, _a : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Any = model_class(_a ).to(_a ) _a : Optional[int] = model(**_a ,output_attentions=_a ) self.assertTrue(outputs.attentions is not None ) def __lowercase ( self : Tuple ): '''simple docstring''' if not self.model_tester.is_training: return _a : List[str] = self.all_model_classes[1] _a, _a, _a, _a, _a : List[str] = self.model_tester.prepare_config_and_inputs() _a : Any = model_class(_a ) model.to(_a ) model.train() _a : Union[str, Any] = model(_a ,mask_labels=_a ,class_labels=_a ).loss loss.backward() def __lowercase ( self : int ): '''simple docstring''' _a : int = self.all_model_classes[1] _a, _a, _a, _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs() _a : str = True _a : str = True _a : List[str] = model_class(_a ).to(_a ) model.train() _a : Optional[int] = model(_a ,mask_labels=_a ,class_labels=_a ) _a : Tuple = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _a : str = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() _a : Dict = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _a : List[str] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_a ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __lowerCAmelCase = 1e-4 def UpperCAmelCase_ (): """simple docstring""" _a : int = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_vision @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return "facebook/mask2former-swin-small-coco-instance" @cached_property def __lowercase ( self : Any ): '''simple docstring''' return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def __lowercase ( self : Any ): '''simple docstring''' _a : List[str] = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(_a ) _a : int = self.default_image_processor _a : Tuple = prepare_img() _a : Any = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Union[str, Any] = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[Any] = model(**_a ) _a : List[Any] = torch.tensor( [[-0.2790, -1.0717, -1.1668], [-0.5128, -0.3128, -0.4987], [-0.5832, 0.1971, -0.0197]] ).to(_a ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : str = torch.tensor( [[0.8973, 1.1847, 1.1776], [1.1934, 1.5040, 1.5128], [1.1153, 1.4486, 1.4951]] ).to(_a ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : Any = torch.tensor( [[2.1152, 1.7000, -0.8603], [1.5808, 1.8004, -0.9353], [1.6043, 1.7495, -0.5999]] ).to(_a ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Optional[Any] = self.default_image_processor _a : List[Any] = prepare_img() _a : str = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Any = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[int] = model(**_a ) # masks_queries_logits _a : Dict = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape ,(1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) _a : Dict = [ [-8.7839, -9.0056, -8.8121], [-7.4104, -7.0313, -6.5401], [-6.6105, -6.3427, -6.4675], ] _a : Optional[Any] = torch.tensor(_a ).to(_a ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] ,_a ,atol=_a ) ) # class_queries_logits _a : str = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape ,(1, model.config.num_queries, model.config.num_labels + 1) ) _a : str = torch.tensor( [ [1.8324, -8.0835, -4.1922], [0.8450, -9.0050, -3.6053], [0.3045, -7.7293, -3.0275], ] ).to(_a ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : Any = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Tuple = self.default_image_processor _a : Tuple = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] ,segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] ,return_tensors='pt' ,) _a : str = inputs['pixel_values'].to(_a ) _a : str = [el.to(_a ) for el in inputs['mask_labels']] _a : Dict = [el.to(_a ) for el in inputs['class_labels']] with torch.no_grad(): _a : List[str] = model(**_a ) self.assertTrue(outputs.loss is not None )
271
0
import os from collections.abc import Iterator def A_ ( A__ = "." ) -> Iterator[str]: for dir_path, dir_names, filenames in os.walk(A__ ): a__ : Dict = [d for d in dir_names if d != 'scripts' and d[0] not in '._'] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(A__ )[1] in (".py", ".ipynb"): yield os.path.join(A__ , A__ ).lstrip('./' ) def A_ ( A__ ) -> List[str]: return F'{i * " "}*' if i else "\n##" def A_ ( A__ , A__ ) -> str: a__ : Any = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(A__ ) or old_parts[i] != new_part) and new_part: print(F'{md_prefix(A__ )} {new_part.replace("_" , " " ).title()}' ) return new_path def A_ ( A__ = "." ) -> None: a__ : Optional[Any] = '' for filepath in sorted(good_file_paths(A__ ) ): a__ , a__ : Union[str, Any] = os.path.split(A__ ) if filepath != old_path: a__ : Tuple = print_path(A__ , A__ ) a__ : Optional[int] = (filepath.count(os.sep ) + 1) if filepath else 0 a__ : Any = F'{filepath}/{filename}'.replace(' ' , '%20' ) a__ : List[Any] = os.path.splitext(filename.replace('_' , ' ' ).title() )[0] print(F'{md_prefix(A__ )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md(""".""")
99
'''simple docstring''' import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def UpperCAmelCase_ (__a : List[Any] ): """simple docstring""" if ( (cp >= 0x4E_00 and cp <= 0x9F_FF) or (cp >= 0x34_00 and cp <= 0x4D_BF) # or (cp >= 0x2_00_00 and cp <= 0x2_A6_DF) # or (cp >= 0x2_A7_00 and cp <= 0x2_B7_3F) # or (cp >= 0x2_B7_40 and cp <= 0x2_B8_1F) # or (cp >= 0x2_B8_20 and cp <= 0x2_CE_AF) # or (cp >= 0xF9_00 and cp <= 0xFA_FF) or (cp >= 0x2_F8_00 and cp <= 0x2_FA_1F) # ): # return True return False def UpperCAmelCase_ (__a : str ): """simple docstring""" for char in word: _a : Union[str, Any] = ord(__a ) if not _is_chinese_char(__a ): return 0 return 1 def UpperCAmelCase_ (__a : List[str] ): """simple docstring""" _a : Dict = set() for token in tokens: _a : str = len(__a ) > 1 and is_chinese(__a ) if chinese_word: word_set.add(__a ) _a : Optional[Any] = list(__a ) return word_list def UpperCAmelCase_ (__a : List[str] , __a : set() ): """simple docstring""" if not chinese_word_set: return bert_tokens _a : Optional[Any] = max([len(__a ) for w in chinese_word_set] ) _a : Optional[int] = bert_tokens _a, _a : Any = 0, len(__a ) while start < end: _a : Tuple = True if is_chinese(bert_word[start] ): _a : Union[str, Any] = min(end - start , __a ) for i in range(__a , 1 , -1 ): _a : Optional[Any] = ''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): _a : Any = '##' + bert_word[j] _a : Union[str, Any] = start + i _a : int = False break if single_word: start += 1 return bert_word def UpperCAmelCase_ (__a : List[str] , __a : LTP , __a : BertTokenizer ): """simple docstring""" _a : int = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : Union[str, Any] = ltp_tokenizer.seg(lines[i : i + 1_0_0] )[0] _a : Optional[Any] = [get_chinese_word(__a ) for r in res] ltp_res.extend(__a ) assert len(__a ) == len(__a ) _a : str = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : List[str] = bert_tokenizer(lines[i : i + 1_0_0] , add_special_tokens=__a , truncation=__a , max_length=5_1_2 ) bert_res.extend(res['input_ids'] ) assert len(__a ) == len(__a ) _a : List[str] = [] for input_ids, chinese_word in zip(__a , __a ): _a : int = [] for id in input_ids: _a : Optional[int] = bert_tokenizer._convert_id_to_token(__a ) input_tokens.append(__a ) _a : List[str] = add_sub_symbol(__a , __a ) _a : Tuple = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(__a ): if token[:2] == "##": _a : str = token[2:] # save chinese tokens' pos if len(__a ) == 1 and _is_chinese_char(ord(__a ) ): ref_id.append(__a ) ref_ids.append(__a ) assert len(__a ) == len(__a ) return ref_ids def UpperCAmelCase_ (__a : Optional[Any] ): """simple docstring""" with open(args.file_name , 'r' , encoding='utf-8' ) as f: _a : Dict = f.readlines() _a : int = [line.strip() for line in data if len(__a ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' _a : int = LTP(args.ltp ) # faster in GPU device _a : Tuple = BertTokenizer.from_pretrained(args.bert ) _a : int = prepare_ref(__a , __a , __a ) with open(args.save_path , 'w' , encoding='utf-8' ) as f: _a : Optional[Any] = [json.dumps(__a ) + '\n' for ref in ref_ids] f.writelines(__a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser(description="""prepare_chinese_ref""") parser.add_argument( """--file_name""", type=str, default="""./resources/chinese-demo.txt""", help="""file need process, same as training data in lm""", ) parser.add_argument( """--ltp""", type=str, default="""./resources/ltp""", help="""resources for LTP tokenizer, usually a path""" ) parser.add_argument("""--bert""", type=str, default="""./resources/robert""", help="""resources for Bert tokenizer""") parser.add_argument("""--save_path""", type=str, default="""./resources/ref.txt""", help="""path to save res""") __lowerCAmelCase = parser.parse_args() main(args)
271
0
"""simple docstring""" import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing the experiment tracking capability, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __magic_name__ = 16 __magic_name__ = 32 def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ = 16 ): __SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained("""bert-base-cased""" ) __SCREAMING_SNAKE_CASE = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(UpperCamelCase_ ): # max_length=None => use the model max length (it's actually the default) __SCREAMING_SNAKE_CASE = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=UpperCamelCase_ , max_length=UpperCamelCase_ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): __SCREAMING_SNAKE_CASE = datasets.map( UpperCamelCase_ , batched=UpperCamelCase_ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __SCREAMING_SNAKE_CASE = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(UpperCamelCase_ ): # On TPU it's best to pad everything to the same length or training will be very slow. __SCREAMING_SNAKE_CASE = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": __SCREAMING_SNAKE_CASE = 16 elif accelerator.mixed_precision != "no": __SCREAMING_SNAKE_CASE = 8 else: __SCREAMING_SNAKE_CASE = None return tokenizer.pad( UpperCamelCase_ , padding="""longest""" , max_length=UpperCamelCase_ , pad_to_multiple_of=UpperCamelCase_ , return_tensors="""pt""" , ) # Instantiate dataloaders. __SCREAMING_SNAKE_CASE = DataLoader( tokenized_datasets["""train"""] , shuffle=UpperCamelCase_ , collate_fn=UpperCamelCase_ , batch_size=UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = DataLoader( tokenized_datasets["""validation"""] , shuffle=UpperCamelCase_ , collate_fn=UpperCamelCase_ , batch_size=UpperCamelCase_ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("TESTING_MOCKED_DATALOADERS", None) == "1": from accelerate.test_utils.training import mocked_dataloaders __magic_name__ = mocked_dataloaders # noqa: F811 def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ ): # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""" , UpperCamelCase_ ) == "1": __SCREAMING_SNAKE_CASE = 2 # Initialize Accelerator # New Code # # We pass in "all" to `log_with` to grab all available trackers in the environment # Note: If using a custom `Tracker` class, should be passed in here such as: # >>> log_with = ["all", MyCustomTrackerClassInstance()] if args.with_tracking: __SCREAMING_SNAKE_CASE = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , log_with="""all""" , project_dir=args.project_dir ) else: __SCREAMING_SNAKE_CASE = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __SCREAMING_SNAKE_CASE = config["""lr"""] __SCREAMING_SNAKE_CASE = int(config["""num_epochs"""] ) __SCREAMING_SNAKE_CASE = int(config["""seed"""] ) __SCREAMING_SNAKE_CASE = int(config["""batch_size"""] ) set_seed(UpperCamelCase_ ) __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = get_dataloaders(UpperCamelCase_ , UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = evaluate.load("""glue""" , """mrpc""" ) # If the batch size is too big we use gradient accumulation __SCREAMING_SNAKE_CASE = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: __SCREAMING_SNAKE_CASE = batch_size // MAX_GPU_BATCH_SIZE __SCREAMING_SNAKE_CASE = MAX_GPU_BATCH_SIZE # Instantiate the model (we build the model here so that the seed also control new weights initialization) __SCREAMING_SNAKE_CASE = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""" , return_dict=UpperCamelCase_ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). __SCREAMING_SNAKE_CASE = model.to(accelerator.device ) # Instantiate optimizer __SCREAMING_SNAKE_CASE = AdamW(params=model.parameters() , lr=UpperCamelCase_ ) # Instantiate scheduler __SCREAMING_SNAKE_CASE = get_linear_schedule_with_warmup( optimizer=UpperCamelCase_ , num_warmup_steps=100 , num_training_steps=(len(UpperCamelCase_ ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = accelerator.prepare( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # New Code # # We need to initialize the trackers we use. Overall configurations can also be stored if args.with_tracking: __SCREAMING_SNAKE_CASE = os.path.split(UpperCamelCase_ )[-1].split(""".""" )[0] accelerator.init_trackers(UpperCamelCase_ , UpperCamelCase_ ) # Now we train the model for epoch in range(UpperCamelCase_ ): model.train() # New Code # # For our tracking example, we will log the total loss of each epoch if args.with_tracking: __SCREAMING_SNAKE_CASE = 0 for step, batch in enumerate(UpperCamelCase_ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) __SCREAMING_SNAKE_CASE = model(**UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = outputs.loss # New Code # if args.with_tracking: total_loss += loss.detach().float() __SCREAMING_SNAKE_CASE = loss / gradient_accumulation_steps accelerator.backward(UpperCamelCase_ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(UpperCamelCase_ ): # We could avoid this line since we set the accelerator with `device_placement=True` (the default). batch.to(accelerator.device ) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(**UpperCamelCase_ ) __SCREAMING_SNAKE_CASE = outputs.logits.argmax(dim=-1 ) __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = accelerator.gather_for_metrics((predictions, batch["""labels"""]) ) metric.add_batch( predictions=UpperCamelCase_ , references=UpperCamelCase_ , ) __SCREAMING_SNAKE_CASE = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"epoch {epoch}:" , UpperCamelCase_ ) # New Code # # To actually log, we call `Accelerator.log` # The values passed can be of `str`, `int`, `float` or `dict` of `str` to `float`/`int` if args.with_tracking: accelerator.log( { """accuracy""": eval_metric["""accuracy"""], """f1""": eval_metric["""f1"""], """train_loss""": total_loss.item() / len(UpperCamelCase_ ), """epoch""": epoch, } , step=UpperCamelCase_ , ) # New Code # # When a run is finished, you should call `accelerator.end_training()` # to close all of the open trackers if args.with_tracking: accelerator.end_training() def _lowerCAmelCase ( ): __SCREAMING_SNAKE_CASE = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""" , type=UpperCamelCase_ , default=UpperCamelCase_ , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) parser.add_argument( """--with_tracking""" , action="""store_true""" , help="""Whether to load in all available experiment trackers from the environment and use them for logging.""" , ) parser.add_argument( """--project_dir""" , type=UpperCamelCase_ , default="""logs""" , help="""Location on where to store experiment tracking logs` and relevent project information""" , ) __SCREAMING_SNAKE_CASE = parser.parse_args() __SCREAMING_SNAKE_CASE = {"""lr""": 2e-5, """num_epochs""": 3, """seed""": 42, """batch_size""": 16} training_function(UpperCamelCase_ , UpperCamelCase_ ) if __name__ == "__main__": main()
100
'''simple docstring''' import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Tuple ,*_a : List[str] ,**_a : Any ): '''simple docstring''' warnings.warn( 'The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use VideoMAEImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
# Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position lowercase__ :int = "2.13.1" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("3.7"): raise ImportWarning( "To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition." ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( "To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n" "If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`." ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip lowercase__ :Tuple = concatenate_datasets lowercase__ :List[str] = DownloadConfig lowercase__ :Optional[int] = DownloadManager lowercase__ :Optional[int] = DownloadMode lowercase__ :Any = DownloadConfig lowercase__ :str = DownloadMode lowercase__ :List[str] = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
101
'''simple docstring''' from __future__ import annotations from random import choice def UpperCAmelCase_ (__a : str ): """simple docstring""" return choice(__a ) def UpperCAmelCase_ (__a : list[int] , __a : int ): """simple docstring""" _a : Dict = random_pivot(__a ) # partition based on pivot # linear time _a : Optional[int] = [e for e in lst if e < pivot] _a : List[str] = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(__a ) == k - 1: return pivot # pivot is in elements bigger than k elif len(__a ) < k - 1: return kth_number(__a , k - len(__a ) - 1 ) # pivot is in elements smaller than k else: return kth_number(__a , __a ) if __name__ == "__main__": import doctest doctest.testmod()
271
0
"""simple docstring""" import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _UpperCAmelCase : '''simple docstring''' def __init__(self , a_ , a_=13 , a_=3 , a_=True , a_=True , a_=0.1 , a_=0.1 , a_=2_24 , a_=10_00 , a_=[3, 3, 6, 4] , a_=[48, 56, 1_12, 2_20] , ): '''simple docstring''' __snake_case : Any = parent __snake_case : Tuple = batch_size __snake_case : Dict = num_channels __snake_case : Dict = is_training __snake_case : List[Any] = use_labels __snake_case : Dict = hidden_dropout_prob __snake_case : Union[str, Any] = attention_probs_dropout_prob __snake_case : Optional[int] = num_labels __snake_case : str = image_size __snake_case : List[str] = layer_depths __snake_case : Dict = embed_dims def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : Optional[int] = None if self.use_labels: __snake_case : List[str] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Dict = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act='''gelu''' , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=a_ , layer_scale_init_value=1E-5 , ) def SCREAMING_SNAKE_CASE (self , a_ , a_ , a_ ): '''simple docstring''' __snake_case : str = SwiftFormerModel(config=a_ ) model.to(a_ ) model.eval() __snake_case : Union[str, Any] = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7) ) def SCREAMING_SNAKE_CASE (self , a_ , a_ , a_ ): '''simple docstring''' __snake_case : Dict = self.num_labels __snake_case : Tuple = SwiftFormerForImageClassification(a_ ) model.to(a_ ) model.eval() __snake_case : Any = model(a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) __snake_case : Optional[Any] = SwiftFormerForImageClassification(a_ ) model.to(a_ ) model.eval() __snake_case : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[str] = model(a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' ((__snake_case) , (__snake_case) , (__snake_case)) : Optional[int] = self.prepare_config_and_inputs() __snake_case : Optional[Any] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class _UpperCAmelCase ( __snake_case, __snake_case, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ =(SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () lowerCamelCase__ =( {'feature-extraction': SwiftFormerModel, 'image-classification': SwiftFormerForImageClassification} if is_torch_available() else {} ) lowerCamelCase__ =False lowerCamelCase__ =False lowerCamelCase__ =False lowerCamelCase__ =False lowerCamelCase__ =False def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Union[str, Any] = SwiftFormerModelTester(self ) __snake_case : Tuple = ConfigTester( self , config_class=a_ , has_text_modality=a_ , hidden_size=37 , num_attention_heads=12 , num_hidden_layers=12 , ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='''SwiftFormer does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' pass def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case , __snake_case : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : int = model_class(a_ ) __snake_case : List[str] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a_ , nn.Linear ) ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Dict = model_class(a_ ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : List[str] = [*signature.parameters.keys()] __snake_case : List[str] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , a_ ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a_ ) @slow def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : int = SwiftFormerModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) @unittest.skip(reason='''SwiftFormer does not output attentions''' ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' pass def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' def check_hidden_states_output(a_ , a_ , a_ ): __snake_case : List[str] = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): __snake_case : Union[str, Any] = model(**self._prepare_for_class(a_ , a_ ) ) __snake_case : Optional[Any] = outputs.hidden_states __snake_case : Union[str, Any] = 8 self.assertEqual(len(a_ ) , a_ ) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(a_ ) ): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ] ) , ) __snake_case , __snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Any = True check_hidden_states_output(a_ , a_ , a_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : int = True check_hidden_states_output(a_ , a_ , a_ ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' def _config_zero_init(a_ ): __snake_case : Optional[Any] = copy.deepcopy(a_ ) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(a_ , a_ , 1E-10 ) if isinstance(getattr(a_ , a_ , a_ ) , a_ ): __snake_case : Union[str, Any] = _config_zero_init(getattr(a_ , a_ ) ) setattr(a_ , a_ , a_ ) return configs_no_init __snake_case , __snake_case : List[str] = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : Dict = _config_zero_init(a_ ) for model_class in self.all_model_classes: __snake_case : Dict = model_class(config=a_ ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9) / 1E9).round().item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' pass def lowercase ( ) ->List[Any]: """simple docstring""" __snake_case : int = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' return ViTImageProcessor.from_pretrained('''MBZUAI/swiftformer-xs''' ) if is_vision_available() else None @slow def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[Any] = SwiftFormerForImageClassification.from_pretrained('''MBZUAI/swiftformer-xs''' ).to(a_ ) __snake_case : Optional[int] = self.default_image_processor __snake_case : Dict = prepare_img() __snake_case : Tuple = image_processor(images=a_ , return_tensors='''pt''' ).to(a_ ) # forward pass with torch.no_grad(): __snake_case : Any = model(**a_ ) # verify the logits __snake_case : int = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , a_ ) __snake_case : List[Any] = torch.tensor([[-2.1703E00, 2.1107E00, -2.0811E00]] ).to(a_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a_ , atol=1E-4 ) )
102
'''simple docstring''' class UpperCAmelCase__ : """simple docstring""" def __init__( self : Dict ): '''simple docstring''' _a : Dict = {} def __lowercase ( self : Union[str, Any] ): '''simple docstring''' print(self.vertex ) for i in self.vertex: print(_a ,' -> ' ,' -> '.join([str(_a ) for j in self.vertex[i]] ) ) def __lowercase ( self : Dict ,_a : int ,_a : int ): '''simple docstring''' if from_vertex in self.vertex: self.vertex[from_vertex].append(_a ) else: # else make a new vertex _a : int = [to_vertex] def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Tuple = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(_a ,_a ) def __lowercase ( self : Union[str, Any] ,_a : int ,_a : list ): '''simple docstring''' _a : List[Any] = True print(_a ,end=' ' ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(_a ,_a ) if __name__ == "__main__": __lowerCAmelCase = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print("""DFS:""") g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
271
0
import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class __snake_case : @staticmethod def UpperCAmelCase__ ( *A_ : Tuple , **A_ : Optional[Any]): pass def UpperCamelCase( __UpperCamelCase : Image ): lowerCAmelCase_ : Union[str, Any] = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class __snake_case ( unittest.TestCase ): _a = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def UpperCAmelCase__ ( self : Tuple , A_ : str , A_ : Tuple , A_ : str): lowerCAmelCase_ : str = DepthEstimationPipeline(model=A_ , image_processor=A_) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def UpperCAmelCase__ ( self : List[str] , A_ : Union[str, Any] , A_ : int): lowerCAmelCase_ : Dict = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''') self.assertEqual({'''predicted_depth''': ANY(torch.Tensor), '''depth''': ANY(Image.Image)} , A_) import datasets lowerCAmelCase_ : int = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' , '''image''' , split='''test''') lowerCAmelCase_ : Any = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png'''), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ]) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor), '''depth''': ANY(Image.Image)}, {'''predicted_depth''': ANY(torch.Tensor), '''depth''': ANY(Image.Image)}, {'''predicted_depth''': ANY(torch.Tensor), '''depth''': ANY(Image.Image)}, {'''predicted_depth''': ANY(torch.Tensor), '''depth''': ANY(Image.Image)}, {'''predicted_depth''': ANY(torch.Tensor), '''depth''': ANY(Image.Image)}, ] , A_ , ) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''') def UpperCAmelCase__ ( self : Optional[int]): pass @slow @require_torch def UpperCAmelCase__ ( self : List[str]): lowerCAmelCase_ : Any = '''Intel/dpt-large''' lowerCAmelCase_ : Optional[int] = pipeline('''depth-estimation''' , model=A_) lowerCAmelCase_ : List[str] = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''') lowerCAmelCase_ : List[Any] = hashimage(outputs['''depth''']) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item()) , 29.304) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item()) , 2.662) @require_torch def UpperCAmelCase__ ( self : List[Any]): # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''')
103
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = """▁""" __lowerCAmelCase = {"""vocab_file""": """sentencepiece.bpe.model""", """monolingual_vocab_file""": """dict.txt"""} __lowerCAmelCase = { """vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model""", }, """monolingual_vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt""", }, } __lowerCAmelCase = {"""vinai/bartpho-syllable""": 1_0_2_4} class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Optional[Any] = VOCAB_FILES_NAMES __UpperCAmelCase : Dict = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ['''input_ids''', '''attention_mask'''] def __init__( self : str ,_a : str ,_a : Any ,_a : Any="<s>" ,_a : Dict="</s>" ,_a : int="</s>" ,_a : Union[str, Any]="<s>" ,_a : List[Any]="<unk>" ,_a : Optional[Any]="<pad>" ,_a : List[str]="<mask>" ,_a : Optional[Dict[str, Any]] = None ,**_a : int ,): '''simple docstring''' _a : Any = AddedToken(_a ,lstrip=_a ,rstrip=_a ) if isinstance(_a ,_a ) else mask_token _a : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_a ,eos_token=_a ,unk_token=_a ,sep_token=_a ,cls_token=_a ,pad_token=_a ,mask_token=_a ,sp_model_kwargs=self.sp_model_kwargs ,**_a ,) _a : Optional[int] = vocab_file _a : Union[str, Any] = monolingual_vocab_file _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_a ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility _a : Union[str, Any] = {} _a : int = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(_a ) not in self.fairseq_tokens_to_ids: _a : int = cnt cnt += 1 with open(_a ,'r' ,encoding='utf-8' ) as f: for line in f.readlines(): _a : str = line.strip().split()[0] _a : Tuple = len(self.fairseq_tokens_to_ids ) if str(_a ) not in self.fairseq_tokens_to_ids: _a : List[str] = len(self.fairseq_tokens_to_ids ) _a : Tuple = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Union[str, Any] ): '''simple docstring''' _a : int = self.__dict__.copy() _a : str = None _a : Optional[Any] = self.sp_model.serialized_model_proto() return state def __setstate__( self : Tuple ,_a : Tuple ): '''simple docstring''' _a : Tuple = d # for backward compatibility if not hasattr(self ,'sp_model_kwargs' ): _a : List[str] = {} _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def __lowercase ( self : Dict ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _a : Dict = [self.cls_token_id] _a : int = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __lowercase ( self : int ,_a : List[int] ,_a : Optional[List[int]] = None ,_a : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a ,token_ids_a=_a ,already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def __lowercase ( self : Tuple ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' _a : List[str] = [self.sep_token_id] _a : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __lowercase ( self : Dict ): '''simple docstring''' return len(self.fairseq_ids_to_tokens ) def __lowercase ( self : Dict ): '''simple docstring''' _a : List[str] = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __lowercase ( self : Tuple ,_a : str ): '''simple docstring''' return self.sp_model.encode(_a ,out_type=_a ) def __lowercase ( self : Union[str, Any] ,_a : Union[str, Any] ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def __lowercase ( self : Any ,_a : int ): '''simple docstring''' return self.fairseq_ids_to_tokens[index] def __lowercase ( self : Tuple ,_a : Union[str, Any] ): '''simple docstring''' _a : str = ''.join(_a ).replace(_a ,' ' ).strip() return out_string def __lowercase ( self : Union[str, Any] ,_a : str ,_a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(_a ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['monolingual_vocab_file'] ,) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,_a ) elif not os.path.isfile(self.vocab_file ): with open(_a ,'wb' ) as fi: _a : List[Any] = self.sp_model.serialized_model_proto() fi.write(_a ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( _a ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file ,_a ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(_a ,'w' ,encoding='utf-8' ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(F"""{str(_a )} \n""" ) return out_vocab_file, out_monolingual_vocab_file
271
0
'''simple docstring''' import qiskit def _A ( A__ , A__ ): """simple docstring""" __lowercase = qiskit.Aer.get_backend('''aer_simulator''' ) # Create a Quantum Circuit acting on the q register __lowercase = qiskit.QuantumCircuit(A__ , A__ ) # Apply X (NOT) Gate to Qubits 0 & 1 circuit.x(0 ) circuit.x(1 ) # Map the quantum measurement to the classical bits circuit.measure([0, 1] , [0, 1] ) # Execute the circuit on the qasm simulator __lowercase = qiskit.execute(A__ , A__ , shots=1000 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(A__ ) if __name__ == "__main__": lowerCAmelCase__ = single_qubit_measure(2, 2) print(f'Total count for various states are: {counts}')
104
'''simple docstring''' import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = None __UpperCAmelCase : List[Any] = None @property def __lowercase ( self : Dict ): '''simple docstring''' return self.feat_extract_tester.prepare_feat_extract_dict() def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_a ,'feature_size' ) ) self.assertTrue(hasattr(_a ,'sampling_rate' ) ) self.assertTrue(hasattr(_a ,'padding_value' ) ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_tester.prepare_inputs_for_common() _a : str = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_a ) == len(_a ) for x, y in zip(_a ,processed_features[input_name] ) ) ) _a : Any = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Union[str, Any] = BatchFeature({input_name: speech_inputs} ,tensor_type='np' ) _a : Union[str, Any] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[int] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def __lowercase ( self : Any ): '''simple docstring''' _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : str = BatchFeature({input_name: speech_inputs} ,tensor_type='pt' ) _a : str = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : str = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : int = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = feat_extract.model_input_names[0] _a : int = BatchFeature({input_name: speech_inputs} ,tensor_type='tf' ) _a : Optional[int] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[Any] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def __lowercase ( self : Dict ,_a : Any=False ): '''simple docstring''' def _inputs_have_equal_length(_a : Tuple ): _a : Tuple = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : Optional[Any] ,_a : Union[str, Any] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : int = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Tuple = BatchFeature({input_name: speech_inputs} ) _a : str = self.feat_extract_tester.seq_length_diff _a : Dict = self.feat_extract_tester.max_seq_length + pad_diff _a : Dict = self.feat_extract_tester.min_seq_length _a : Optional[Any] = self.feat_extract_tester.batch_size _a : Tuple = self.feat_extract_tester.feature_size # test padding for List[int] + numpy _a : int = feat_extract.pad(_a ,padding=_a ) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad(_a ,padding='longest' ) _a : Any = input_a[input_name] _a : Optional[Any] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[-1] ) ) _a : List[str] = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) _a : str = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' )[input_name] _a : int = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,return_tensors='np' ) _a : Optional[int] = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy _a : Tuple = feat_extract.pad(_a ,pad_to_multiple_of=10 ) _a : List[str] = input_a[input_name] _a : str = feat_extract.pad(_a ,padding='longest' ,pad_to_multiple_of=10 ) _a : Tuple = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ) _a : Any = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ,return_tensors='np' ,) _a : Dict = input_a[input_name] self.assertTrue(all(len(_a ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) _a : List[str] = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_a ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] ,(batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct _a : Any = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1E-3 ) def __lowercase ( self : List[Any] ,_a : Optional[int]=False ): '''simple docstring''' def _inputs_have_equal_length(_a : List[str] ): _a : Union[str, Any] = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : List[str] ,_a : List[str] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : str = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Any = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) # truncate to smallest _a : Union[str, Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,truncation=_a ) _a : str = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ) _a : Tuple = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to smallest with np _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ,truncation=_a ,) _a : Any = input_a[input_name] _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ) _a : int = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to middle _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ,return_tensors='np' ,) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ) _a : Tuple = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,return_tensors='np' ) _a : Dict = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' ,truncation=_a )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy _a : Optional[Any] = 12 _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,truncation=_a ,) _a : Tuple = input_a[input_name] _a : str = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,) _a : List[Any] = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of _a : List[Any] = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: _a : Union[str, Any] = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Dict ): '''simple docstring''' self._check_truncation(numpify=_a ) def __lowercase ( self : str ): '''simple docstring''' self._check_truncation(numpify=_a ) @require_torch def __lowercase ( self : Dict ): '''simple docstring''' _a : Any = self.feature_extraction_class(**self.feat_extract_dict ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Optional[int] = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='pt' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : List[str] = self.feature_extraction_class(**self.feat_extract_dict ) _a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : Dict = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : Any = feat_extract.pad(_a ,padding='longest' ,return_tensors='tf' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : str = self.feat_extract_dict _a : List[Any] = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Tuple = [len(_a ) for x in speech_inputs] _a : int = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : str = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual(list(processed.attention_mask.shape ) ,list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() ,_a ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_dict _a : Tuple = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : Dict = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = [len(_a ) for x in speech_inputs] _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Any = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = min(_a ) _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,truncation=_a ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual( list(processed_pad.attention_mask.shape ) ,[processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() ,[max_length for x in speech_inputs] )
271
0
"""simple docstring""" import warnings from ...utils import is_sklearn_available, requires_backends if is_sklearn_available(): from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef a : Tuple = ( '''This metric will be removed from the library soon, metrics should be handled with the 🤗 Evaluate ''' '''library. You can have a look at this example script for pointers: ''' '''https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py''' ) def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : List[str] ) ->Dict: '''simple docstring''' warnings.warn(_lowercase , _lowercase ) requires_backends(_lowercase , "sklearn" ) return (preds == labels).mean() def _SCREAMING_SNAKE_CASE ( _lowercase : Tuple , _lowercase : Union[str, Any] ) ->Optional[int]: '''simple docstring''' warnings.warn(_lowercase , _lowercase ) requires_backends(_lowercase , "sklearn" ) a : Dict = simple_accuracy(_lowercase , _lowercase ) a : Optional[int] = fa_score(y_true=_lowercase , y_pred=_lowercase ) return { "acc": acc, "f1": fa, "acc_and_f1": (acc + fa) / 2, } def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : Tuple ) ->str: '''simple docstring''' warnings.warn(_lowercase , _lowercase ) requires_backends(_lowercase , "sklearn" ) a : Union[str, Any] = pearsonr(_lowercase , _lowercase )[0] a : List[Any] = spearmanr(_lowercase , _lowercase )[0] return { "pearson": pearson_corr, "spearmanr": spearman_corr, "corr": (pearson_corr + spearman_corr) / 2, } def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : str , _lowercase : Optional[int] ) ->str: '''simple docstring''' warnings.warn(_lowercase , _lowercase ) requires_backends(_lowercase , "sklearn" ) assert len(_lowercase ) == len(_lowercase ), F"""Predictions and labels have mismatched lengths {len(_lowercase )} and {len(_lowercase )}""" if task_name == "cola": return {"mcc": matthews_corrcoef(_lowercase , _lowercase )} elif task_name == "sst-2": return {"acc": simple_accuracy(_lowercase , _lowercase )} elif task_name == "mrpc": return acc_and_fa(_lowercase , _lowercase ) elif task_name == "sts-b": return pearson_and_spearman(_lowercase , _lowercase ) elif task_name == "qqp": return acc_and_fa(_lowercase , _lowercase ) elif task_name == "mnli": return {"mnli/acc": simple_accuracy(_lowercase , _lowercase )} elif task_name == "mnli-mm": return {"mnli-mm/acc": simple_accuracy(_lowercase , _lowercase )} elif task_name == "qnli": return {"acc": simple_accuracy(_lowercase , _lowercase )} elif task_name == "rte": return {"acc": simple_accuracy(_lowercase , _lowercase )} elif task_name == "wnli": return {"acc": simple_accuracy(_lowercase , _lowercase )} elif task_name == "hans": return {"acc": simple_accuracy(_lowercase , _lowercase )} else: raise KeyError(_lowercase ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : Tuple , _lowercase : Tuple ) ->Optional[int]: '''simple docstring''' warnings.warn(_lowercase , _lowercase ) requires_backends(_lowercase , "sklearn" ) if len(_lowercase ) != len(_lowercase ): raise ValueError(F"""Predictions and labels have mismatched lengths {len(_lowercase )} and {len(_lowercase )}""" ) if task_name == "xnli": return {"acc": simple_accuracy(_lowercase , _lowercase )} else: raise KeyError(_lowercase )
105
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import KarrasVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : UNetaDModel __UpperCAmelCase : KarrasVeScheduler def __init__( self : Union[str, Any] ,_a : UNetaDModel ,_a : KarrasVeScheduler ): '''simple docstring''' super().__init__() self.register_modules(unet=_a ,scheduler=_a ) @torch.no_grad() def __call__( self : List[Any] ,_a : int = 1 ,_a : int = 50 ,_a : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,_a : Optional[str] = "pil" ,_a : bool = True ,**_a : List[Any] ,): '''simple docstring''' _a : Any = self.unet.config.sample_size _a : Optional[int] = (batch_size, 3, img_size, img_size) _a : Dict = self.unet # sample x_0 ~ N(0, sigma_0^2 * I) _a : Dict = randn_tensor(_a ,generator=_a ,device=self.device ) * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(_a ) for t in self.progress_bar(self.scheduler.timesteps ): # here sigma_t == t_i from the paper _a : Optional[int] = self.scheduler.schedule[t] _a : List[str] = self.scheduler.schedule[t - 1] if t > 0 else 0 # 1. Select temporarily increased noise level sigma_hat # 2. Add new noise to move from sample_i to sample_hat _a, _a : List[Any] = self.scheduler.add_noise_to_input(_a ,_a ,generator=_a ) # 3. Predict the noise residual given the noise magnitude `sigma_hat` # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_hat / 2) * model((sample_hat + 1) / 2 ,sigma_hat / 2 ).sample # 4. Evaluate dx/dt at sigma_hat # 5. Take Euler step from sigma to sigma_prev _a : Tuple = self.scheduler.step(_a ,_a ,_a ,_a ) if sigma_prev != 0: # 6. Apply 2nd order correction # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_prev / 2) * model((step_output.prev_sample + 1) / 2 ,sigma_prev / 2 ).sample _a : Optional[Any] = self.scheduler.step_correct( _a ,_a ,_a ,_a ,step_output.prev_sample ,step_output['derivative'] ,) _a : Dict = step_output.prev_sample _a : Tuple = (sample / 2 + 0.5).clamp(0 ,1 ) _a : Optional[Any] = sample.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": _a : List[str] = self.numpy_to_pil(_a ) if not return_dict: return (image,) return ImagePipelineOutput(images=_a )
271
0
"""simple docstring""" from math import ceil def __SCREAMING_SNAKE_CASE ( A_ = 10_01 ): lowerCAmelCase__ : Union[str, Any] = 1 for i in range(1 , int(ceil(n / 2.0 ) ) ): lowerCAmelCase__ : int = 2 * i + 1 lowerCAmelCase__ : Any = 2 * i lowerCAmelCase__ : Optional[Any] = total + 4 * odd**2 - 6 * even return total if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution()) else: try: __UpperCamelCase : Optional[Any] = int(sys.argv[1]) print(solution(n)) except ValueError: print('''Invalid entry - please enter a number''')
106
'''simple docstring''' import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging __lowerCAmelCase = ( """https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py""" ) __lowerCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def UpperCAmelCase_ (): """simple docstring""" _a : Optional[int] = 'https://pypi.org/pypi/diffusers/json' _a : int = json.loads(request.urlopen(__a ).read() )['releases'].keys() return sorted(__a , key=lambda __a : version.Version(__a ) ) def UpperCAmelCase_ (): """simple docstring""" if HF_MODULES_CACHE in sys.path: return sys.path.append(__a ) os.makedirs(__a , exist_ok=__a ) _a : str = Path(__a ) / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : Union[str, os.PathLike] ): """simple docstring""" init_hf_modules() _a : Dict = Path(__a ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(__a , exist_ok=__a ) _a : Optional[int] = dynamic_module_path / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : str ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : int = f.read() # Imports of the form `import .xxx` _a : Tuple = re.findall('^\s*import\s+\.(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall('^\s*from\s+\.(\S+)\s+import' , __a , flags=re.MULTILINE ) # Unique-ify return list(set(__a ) ) def UpperCAmelCase_ (__a : Any ): """simple docstring""" _a : Optional[int] = False _a : Optional[int] = [module_file] _a : List[str] = [] # Let's recurse through all relative imports while not no_change: _a : str = [] for f in files_to_check: new_imports.extend(get_relative_imports(__a ) ) _a : Union[str, Any] = Path(__a ).parent _a : str = [str(module_path / m ) for m in new_imports] _a : Tuple = [f for f in new_import_files if f not in all_relative_imports] _a : Dict = [f"""{f}.py""" for f in new_import_files] _a : List[str] = len(__a ) == 0 all_relative_imports.extend(__a ) return all_relative_imports def UpperCAmelCase_ (__a : Tuple ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : Dict = f.read() # Imports of the form `import xxx` _a : Optional[int] = re.findall('^\s*import\s+(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall('^\s*from\s+(\S+)\s+import' , __a , flags=re.MULTILINE ) # Only keep the top-level module _a : List[str] = [imp.split('.' )[0] for imp in imports if not imp.startswith('.' )] # Unique-ify and test we got them all _a : Optional[int] = list(set(__a ) ) _a : List[str] = [] for imp in imports: try: importlib.import_module(__a ) except ImportError: missing_packages.append(__a ) if len(__a ) > 0: raise ImportError( 'This modeling file requires the following packages that were not found in your environment: ' f"""{', '.join(__a )}. Run `pip install {' '.join(__a )}`""" ) return get_relative_imports(__a ) def UpperCAmelCase_ (__a : Any , __a : str ): """simple docstring""" _a : Any = module_path.replace(os.path.sep , '.' ) _a : Union[str, Any] = importlib.import_module(__a ) if class_name is None: return find_pipeline_class(__a ) return getattr(__a , __a ) def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" from ..pipelines import DiffusionPipeline _a : List[str] = dict(inspect.getmembers(__a , inspect.isclass ) ) _a : str = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls , __a ) and cls.__module__.split('.' )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( f"""Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:""" f""" {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in""" f""" {loaded_module}.""" ) _a : Any = cls return pipeline_class def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , ): """simple docstring""" _a : str = str(__a ) _a : Optional[Any] = os.path.join(__a , __a ) if os.path.isfile(__a ): _a : Tuple = module_file_or_url _a : Optional[Any] = 'local' elif pretrained_model_name_or_path.count('/' ) == 0: _a : int = get_diffusers_versions() # cut ".dev0" _a : Any = 'v' + '.'.join(__version__.split('.' )[:3] ) # retrieve github version that matches if revision is None: _a : Any = latest_version if latest_version[1:] in available_versions else 'main' logger.info(f"""Defaulting to latest_version: {revision}.""" ) elif revision in available_versions: _a : Any = f"""v{revision}""" elif revision == "main": _a : Optional[int] = revision else: raise ValueError( f"""`custom_revision`: {revision} does not exist. Please make sure to choose one of""" f""" {', '.join(available_versions + ['main'] )}.""" ) # community pipeline on GitHub _a : Tuple = COMMUNITY_PIPELINES_URL.format(revision=__a , pipeline=__a ) try: _a : Any = cached_download( __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = 'git' _a : Any = pretrained_model_name_or_path + '.py' except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise else: try: # Load from URL or cache if already cached _a : Optional[Any] = hf_hub_download( __a , __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = os.path.join('local' , '--'.join(pretrained_model_name_or_path.split('/' ) ) ) except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise # Check we have all the requirements in our environment _a : Optional[int] = check_imports(__a ) # Now we move the module inside our cached dynamic modules. _a : Optional[Any] = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(__a ) _a : Any = Path(__a ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(__a , submodule_path / module_file ) for module_needed in modules_needed: _a : Dict = f"""{module_needed}.py""" shutil.copy(os.path.join(__a , __a ) , submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(__a , __a ): _a : Optional[Any] = use_auth_token elif use_auth_token is True: _a : List[Any] = HfFolder.get_token() else: _a : Dict = None _a : int = model_info(__a , revision=__a , token=__a ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. _a : Optional[int] = submodule_path / commit_hash _a : str = full_submodule + os.path.sep + commit_hash create_dynamic_module(__a ) if not (submodule_path / module_file).exists(): shutil.copy(__a , submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( __a , f"""{module_needed}.py""" , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return os.path.join(__a , __a ) def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[str] = None , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , **__a : str , ): """simple docstring""" _a : Dict = get_cached_module_file( __a , __a , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return get_class_in_module(__a , final_module.replace('.py' , '' ) )
271
0
import json import re from typing import TYPE_CHECKING, List, Optional, Tuple, Union import numpy as np from ...utils import is_tf_available, is_torch_available, logging if TYPE_CHECKING: if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_codegen import CodeGenTokenizer __lowerCAmelCase : Dict = logging.get_logger(__name__) __lowerCAmelCase : Optional[int] = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} __lowerCAmelCase : Tuple = { 'vocab_file': { 'Salesforce/codegen-350M-mono': 'https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/vocab.json', }, 'merges_file': { 'Salesforce/codegen-350M-mono': 'https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/merges.txt', }, 'tokenizer_file': { 'Salesforce/codegen-350M-mono': ( 'https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/tokenizer.json' ), }, } __lowerCAmelCase : Tuple = { 'Salesforce/codegen-350M-mono': 2048, } class snake_case__ (_UpperCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ : Any = ["""input_ids""", """attention_mask"""] SCREAMING_SNAKE_CASE_ : Dict = CodeGenTokenizer def __init__( self : str , __lowerCamelCase : Optional[int]=None , __lowerCamelCase : int=None , __lowerCamelCase : int=None , __lowerCamelCase : List[Any]="<|endoftext|>" , __lowerCamelCase : str="<|endoftext|>" , __lowerCamelCase : List[Any]="<|endoftext|>" , __lowerCamelCase : List[Any]=False , **__lowerCamelCase : Optional[int] , ) -> Optional[int]: super().__init__( __lowerCamelCase , __lowerCamelCase , tokenizer_file=__lowerCamelCase , unk_token=__lowerCamelCase , bos_token=__lowerCamelCase , eos_token=__lowerCamelCase , add_prefix_space=__lowerCamelCase , **__lowerCamelCase , ) if kwargs.pop("add_bos_token" , __lowerCamelCase ): a = kwargs.pop("name_or_path" , "" ) raise ValueError( "Currenty GPT2's fast tokenizer does NOT support adding a BOS token." "Instead you should use GPT2's slow tokenizer class `CodeGenTokenizer` as follows: \n" f"""`CodeGenTokenizer.from_pretrained('{model_id}')`\nor\n""" f"""`AutoTokenizer.from_pretrained('{model_id}', use_fast=False)`\n""" "This issue will be fixed soon, see: https://github.com/huggingface/tokenizers/pull/1005." " so that the fast tokenizer works correctly." ) a = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , __lowerCamelCase ) != add_prefix_space: a = getattr(__lowerCamelCase , pre_tok_state.pop("type" ) ) a = add_prefix_space a = pre_tok_class(**__lowerCamelCase ) a = add_prefix_space def __UpperCAmelCase ( self : Dict , *__lowerCamelCase : str , **__lowerCamelCase : Optional[int] ) -> BatchEncoding: a = kwargs.get("is_split_into_words" , __lowerCamelCase ) assert self.add_prefix_space or not is_split_into_words, ( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*__lowerCamelCase , **__lowerCamelCase ) def __UpperCAmelCase ( self : Union[str, Any] , *__lowerCamelCase : List[str] , **__lowerCamelCase : List[str] ) -> BatchEncoding: a = kwargs.get("is_split_into_words" , __lowerCamelCase ) assert self.add_prefix_space or not is_split_into_words, ( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ "to use it with pretokenized inputs." ) return super()._encode_plus(*__lowerCamelCase , **__lowerCamelCase ) def __UpperCAmelCase ( self : List[Any] , __lowerCamelCase : str , __lowerCamelCase : Optional[str] = None ) -> Tuple[str]: a = self._tokenizer.model.save(__lowerCamelCase , name=__lowerCamelCase ) return tuple(__lowerCamelCase ) def __UpperCAmelCase ( self : Optional[int] , __lowerCamelCase : Union[int, List[int], "np.ndarray", "torch.Tensor", "tf.Tensor"] , __lowerCamelCase : bool = False , __lowerCamelCase : bool = None , __lowerCamelCase : Optional[List[str]] = None , **__lowerCamelCase : int , ) -> str: a = super().decode( token_ids=__lowerCamelCase , skip_special_tokens=__lowerCamelCase , clean_up_tokenization_spaces=__lowerCamelCase , **__lowerCamelCase , ) if truncate_before_pattern is not None and len(__lowerCamelCase ) > 0: a = self.truncate(__lowerCamelCase , __lowerCamelCase ) return decoded_text def __UpperCAmelCase ( self : List[Any] , __lowerCamelCase : Any , __lowerCamelCase : List[str] ) -> int: def find_re(__lowerCamelCase : Union[str, Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Tuple ): a = pattern.search(__lowerCamelCase , __lowerCamelCase ) return m.start() if m else -1 a = [re.compile(__lowerCamelCase , re.MULTILINE ) for pattern in truncate_before_pattern] a = list(re.finditer("^print" , __lowerCamelCase , re.MULTILINE ) ) if len(__lowerCamelCase ) > 1: a = completion[: prints[1].start()] a = list(re.finditer("^def" , __lowerCamelCase , re.MULTILINE ) ) if len(__lowerCamelCase ) > 1: a = completion[: defs[1].start()] a = 0 a = [ pos for pos in [find_re(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) for terminal in terminals] if pos != -1 ] if len(__lowerCamelCase ) > 0: return completion[: min(__lowerCamelCase )] else: return completion
107
'''simple docstring''' def UpperCAmelCase_ (__a : list , __a : list , __a : int ): """simple docstring""" _a : Optional[Any] = len(__a ) _a : int = [[0] * n for i in range(__a )] for i in range(__a ): _a : Tuple = y_points[i] for i in range(2 , __a ): for j in range(__a , __a ): _a : Tuple = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
271
0
"""simple docstring""" import argparse import json import os from collections import OrderedDict import torch from transformers import LukeConfig, LukeForMaskedLM, MLukeTokenizer, XLMRobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def a__ ( SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : Optional[Any] , SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' with open(SCREAMING_SNAKE_CASE ) as metadata_file: lowerCAmelCase : Tuple = json.load(SCREAMING_SNAKE_CASE ) lowerCAmelCase : Optional[Any] = LukeConfig(use_entity_aware_attention=SCREAMING_SNAKE_CASE , **metadata["model_config"] ) # Load in the weights from the checkpoint_path lowerCAmelCase : Optional[int] = torch.load(SCREAMING_SNAKE_CASE , map_location="cpu" )["module"] # Load the entity vocab file lowerCAmelCase : List[str] = load_original_entity_vocab(SCREAMING_SNAKE_CASE ) # add an entry for [MASK2] lowerCAmelCase : Dict = max(entity_vocab.values() ) + 1 config.entity_vocab_size += 1 lowerCAmelCase : List[Any] = XLMRobertaTokenizer.from_pretrained(metadata["model_config"]["bert_model_name"] ) # Add special tokens to the token vocabulary for downstream tasks lowerCAmelCase : Union[str, Any] = AddedToken("<ent>" , lstrip=SCREAMING_SNAKE_CASE , rstrip=SCREAMING_SNAKE_CASE ) lowerCAmelCase : List[str] = AddedToken("<ent2>" , lstrip=SCREAMING_SNAKE_CASE , rstrip=SCREAMING_SNAKE_CASE ) tokenizer.add_special_tokens({"additional_special_tokens": [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(f"""Saving tokenizer to {pytorch_dump_folder_path}""" ) tokenizer.save_pretrained(SCREAMING_SNAKE_CASE ) with open(os.path.join(SCREAMING_SNAKE_CASE , "tokenizer_config.json" ) , "r" ) as f: lowerCAmelCase : int = json.load(SCREAMING_SNAKE_CASE ) lowerCAmelCase : Optional[int] = "MLukeTokenizer" with open(os.path.join(SCREAMING_SNAKE_CASE , "tokenizer_config.json" ) , "w" ) as f: json.dump(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) with open(os.path.join(SCREAMING_SNAKE_CASE , MLukeTokenizer.vocab_files_names["entity_vocab_file"] ) , "w" ) as f: json.dump(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) lowerCAmelCase : List[Any] = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE ) # Initialize the embeddings of the special tokens lowerCAmelCase : Union[str, Any] = tokenizer.convert_tokens_to_ids(["@"] )[0] lowerCAmelCase : Tuple = tokenizer.convert_tokens_to_ids(["#"] )[0] lowerCAmelCase : Optional[int] = state_dict["embeddings.word_embeddings.weight"] lowerCAmelCase : str = word_emb[ent_init_index].unsqueeze(0 ) lowerCAmelCase : int = word_emb[enta_init_index].unsqueeze(0 ) lowerCAmelCase : Optional[Any] = torch.cat([word_emb, ent_emb, enta_emb] ) # add special tokens for 'entity_predictions.bias' for bias_name in ["lm_head.decoder.bias", "lm_head.bias"]: lowerCAmelCase : Optional[Any] = state_dict[bias_name] lowerCAmelCase : List[str] = decoder_bias[ent_init_index].unsqueeze(0 ) lowerCAmelCase : Tuple = decoder_bias[enta_init_index].unsqueeze(0 ) lowerCAmelCase : Tuple = torch.cat([decoder_bias, ent_decoder_bias, enta_decoder_bias] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: lowerCAmelCase : int = f"""encoder.layer.{layer_index}.attention.self.""" lowerCAmelCase : List[str] = state_dict[prefix + matrix_name] lowerCAmelCase : Optional[int] = state_dict[prefix + matrix_name] lowerCAmelCase : Dict = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks lowerCAmelCase : Any = state_dict["entity_embeddings.entity_embeddings.weight"] lowerCAmelCase : List[str] = entity_emb[entity_vocab["[MASK]"]].unsqueeze(0 ) lowerCAmelCase : Optional[int] = torch.cat([entity_emb, entity_mask_emb] ) # add [MASK2] for 'entity_predictions.bias' lowerCAmelCase : Optional[Any] = state_dict["entity_predictions.bias"] lowerCAmelCase : Any = entity_prediction_bias[entity_vocab["[MASK]"]].unsqueeze(0 ) lowerCAmelCase : List[Any] = torch.cat([entity_prediction_bias, entity_mask_bias] ) lowerCAmelCase : Dict = LukeForMaskedLM(config=SCREAMING_SNAKE_CASE ).eval() state_dict.pop("entity_predictions.decoder.weight" ) state_dict.pop("lm_head.decoder.weight" ) state_dict.pop("lm_head.decoder.bias" ) lowerCAmelCase : str = OrderedDict() for key, value in state_dict.items(): if not (key.startswith("lm_head" ) or key.startswith("entity_predictions" )): lowerCAmelCase : str = state_dict[key] else: lowerCAmelCase : int = state_dict[key] lowerCAmelCase , lowerCAmelCase : Optional[int] = model.load_state_dict(SCREAMING_SNAKE_CASE , strict=SCREAMING_SNAKE_CASE ) if set(SCREAMING_SNAKE_CASE ) != {"luke.embeddings.position_ids"}: raise ValueError(f"""Unexpected unexpected_keys: {unexpected_keys}""" ) if set(SCREAMING_SNAKE_CASE ) != { "lm_head.decoder.weight", "lm_head.decoder.bias", "entity_predictions.decoder.weight", }: raise ValueError(f"""Unexpected missing_keys: {missing_keys}""" ) model.tie_weights() assert (model.luke.embeddings.word_embeddings.weight == model.lm_head.decoder.weight).all() assert (model.luke.entity_embeddings.entity_embeddings.weight == model.entity_predictions.decoder.weight).all() # Check outputs lowerCAmelCase : str = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE , task="entity_classification" ) lowerCAmelCase : str = "ISO 639-3 uses the code fas for the dialects spoken across Iran and アフガニスタン (Afghanistan)." lowerCAmelCase : Dict = (0, 9) lowerCAmelCase : Optional[int] = tokenizer(SCREAMING_SNAKE_CASE , entity_spans=[span] , return_tensors="pt" ) lowerCAmelCase : Optional[Any] = model(**SCREAMING_SNAKE_CASE ) # Verify word hidden states if model_size == "large": raise NotImplementedError else: # base lowerCAmelCase : Tuple = torch.Size((1, 3_3, 7_6_8) ) lowerCAmelCase : Dict = torch.tensor([[0.0_892, 0.0_596, -0.2_819], [0.0_134, 0.1_199, 0.0_573], [-0.0_169, 0.0_927, 0.0_644]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( f"""Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}""" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , SCREAMING_SNAKE_CASE , atol=1E-4 ): raise ValueError # Verify entity hidden states if model_size == "large": raise NotImplementedError else: # base lowerCAmelCase : Tuple = torch.Size((1, 1, 7_6_8) ) lowerCAmelCase : Optional[int] = torch.tensor([[-0.1_482, 0.0_609, 0.0_322]] ) if not (outputs.entity_last_hidden_state.shape == expected_shape): raise ValueError( f"""Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is""" f""" {expected_shape}""" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , SCREAMING_SNAKE_CASE , atol=1E-4 ): raise ValueError # Verify masked word/entity prediction lowerCAmelCase : Tuple = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE ) lowerCAmelCase : List[str] = "Tokyo is the capital of <mask>." lowerCAmelCase : Optional[int] = (2_4, 3_0) lowerCAmelCase : Any = tokenizer(SCREAMING_SNAKE_CASE , entity_spans=[span] , return_tensors="pt" ) lowerCAmelCase : Any = model(**SCREAMING_SNAKE_CASE ) lowerCAmelCase : Union[str, Any] = encoding["input_ids"][0].tolist() lowerCAmelCase : str = input_ids.index(tokenizer.convert_tokens_to_ids("<mask>" ) ) lowerCAmelCase : Tuple = outputs.logits[0][mask_position_id].argmax(dim=-1 ) assert "Japan" == tokenizer.decode(SCREAMING_SNAKE_CASE ) lowerCAmelCase : Any = outputs.entity_logits[0][0].argmax().item() lowerCAmelCase : List[str] = [ entity for entity, entity_id in tokenizer.entity_vocab.items() if entity_id == predicted_entity_id ] assert [e for e in multilingual_predicted_entities if e.startswith("en:" )][0] == "en:Japan" # Finally, save our PyTorch model and tokenizer print("Saving PyTorch model to {}".format(SCREAMING_SNAKE_CASE ) ) model.save_pretrained(SCREAMING_SNAKE_CASE ) def a__ ( SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' lowerCAmelCase : Optional[Any] = ["[MASK]", "[PAD]", "[UNK]"] lowerCAmelCase : Tuple = [json.loads(SCREAMING_SNAKE_CASE ) for line in open(SCREAMING_SNAKE_CASE )] lowerCAmelCase : Any = {} for entry in data: lowerCAmelCase : Tuple = entry["id"] for entity_name, language in entry["entities"]: if entity_name in SPECIAL_TOKENS: lowerCAmelCase : Tuple = entity_id break lowerCAmelCase : Tuple = f"""{language}:{entity_name}""" lowerCAmelCase : Tuple = entity_id return new_mapping if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Path to a pytorch_model.bin file.''') parser.add_argument( '''--metadata_path''', default=None, type=str, help='''Path to a metadata.json file, defining the configuration.''' ) parser.add_argument( '''--entity_vocab_path''', default=None, type=str, help='''Path to an entity_vocab.tsv file, containing the entity vocabulary.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to where to dump the output PyTorch model.''' ) parser.add_argument( '''--model_size''', default='''base''', type=str, choices=['''base''', '''large'''], help='''Size of the model to be converted.''' ) lowerCAmelCase__ = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
108
'''simple docstring''' import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = inspect.getfile(accelerate.test_utils ) __UpperCAmelCase : List[str] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_cli.py'''] ) __UpperCAmelCase : Dict = ['''accelerate''', '''launch'''] __UpperCAmelCase : Dict = Path.home() / '''.cache/huggingface/accelerate''' __UpperCAmelCase : Dict = '''default_config.yaml''' __UpperCAmelCase : Optional[Any] = config_folder / config_file __UpperCAmelCase : Dict = config_folder / '''_default_config.yaml''' __UpperCAmelCase : Any = Path('''tests/test_configs''' ) @classmethod def __lowercase ( cls : int ): '''simple docstring''' if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def __lowercase ( cls : List[Any] ): '''simple docstring''' if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Dict = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' for config in sorted(self.test_config_path.glob('**/*.yaml' ) ): with self.subTest(config_file=_a ): execute_subprocess_async( self.base_cmd + ['--config_file', str(_a ), self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[int] ): '''simple docstring''' execute_subprocess_async(['accelerate', 'test'] ,env=os.environ.copy() ) class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = '''test-tpu''' __UpperCAmelCase : Any = '''us-central1-a''' __UpperCAmelCase : List[Any] = '''ls''' __UpperCAmelCase : Any = ['''accelerate''', '''tpu-config'''] __UpperCAmelCase : Dict = '''cd /usr/share''' __UpperCAmelCase : Any = '''tests/test_samples/test_command_file.sh''' __UpperCAmelCase : List[Any] = '''Running gcloud compute tpus tpu-vm ssh''' def __lowercase ( self : Dict ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Any = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--debug'] ,return_stdout=_a ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : int ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : str ): '''simple docstring''' _a : List[str] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--command', 'echo "Hello World"', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo \"Hello World\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Any = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command_file', self.command_file, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Union[str, Any] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command_file', self.command_file, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Any ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--accelerate_version', '12.0.0', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,)
271
0
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=13 , _SCREAMING_SNAKE_CASE=7 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=99 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=5 , _SCREAMING_SNAKE_CASE=4 , _SCREAMING_SNAKE_CASE=37 , _SCREAMING_SNAKE_CASE="gelu" , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=512 , _SCREAMING_SNAKE_CASE=16 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=4 , _SCREAMING_SNAKE_CASE=None , ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase : Optional[Any] = parent UpperCAmelCase : Tuple = batch_size UpperCAmelCase : Optional[int] = seq_length UpperCAmelCase : int = is_training UpperCAmelCase : str = use_token_type_ids UpperCAmelCase : Optional[int] = use_labels UpperCAmelCase : List[Any] = vocab_size UpperCAmelCase : Optional[Any] = hidden_size UpperCAmelCase : Optional[Any] = num_hidden_layers UpperCAmelCase : int = num_attention_heads UpperCAmelCase : List[Any] = intermediate_size UpperCAmelCase : int = hidden_act UpperCAmelCase : Dict = hidden_dropout_prob UpperCAmelCase : List[Any] = attention_probs_dropout_prob UpperCAmelCase : Optional[int] = max_position_embeddings UpperCAmelCase : int = type_vocab_size UpperCAmelCase : Any = type_sequence_label_size UpperCAmelCase : Optional[Any] = initializer_range UpperCAmelCase : Union[str, Any] = num_labels UpperCAmelCase : Any = num_choices UpperCAmelCase : str = scope UpperCAmelCase : str = self.vocab_size - 1 def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: '''simple docstring''' UpperCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase : Any = None if self.use_token_type_ids: UpperCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase : Union[str, Any] = None UpperCAmelCase : Dict = None UpperCAmelCase : Tuple = None if self.use_labels: UpperCAmelCase : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase : Tuple = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase : Union[str, Any] = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) UpperCAmelCase : str = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE ) -> Any: '''simple docstring''' UpperCAmelCase : Tuple = OpenAIGPTModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCAmelCase : List[str] = model(_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , head_mask=_SCREAMING_SNAKE_CASE ) UpperCAmelCase : List[Any] = model(_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Tuple = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase : Optional[int] = OpenAIGPTLMHeadModel(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCAmelCase : Dict = model(_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE ) -> Dict: '''simple docstring''' UpperCAmelCase : Tuple = OpenAIGPTDoubleHeadsModel(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCAmelCase : Optional[int] = model(_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , *_SCREAMING_SNAKE_CASE ) -> int: '''simple docstring''' UpperCAmelCase : str = self.num_labels UpperCAmelCase : Optional[int] = OpenAIGPTForSequenceClassification(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() UpperCAmelCase : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase : Optional[int] = model(_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase : List[str] = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) : Any = config_and_inputs UpperCAmelCase : Dict = { """input_ids""": input_ids, """token_type_ids""": token_type_ids, """head_mask""": head_mask, } return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): __lowerCAmelCase : Union[str, Any] = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) __lowerCAmelCase : Union[str, Any] = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly __lowerCAmelCase : int = ( { 'feature-extraction': OpenAIGPTModel, 'text-classification': OpenAIGPTForSequenceClassification, 'text-generation': OpenAIGPTLMHeadModel, 'zero-shot': OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[str]: '''simple docstring''' if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def SCREAMING_SNAKE_CASE ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ) -> int: '''simple docstring''' UpperCAmelCase : List[str] = super()._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_labels=_SCREAMING_SNAKE_CASE ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": UpperCAmelCase : Union[str, Any] = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=_SCREAMING_SNAKE_CASE , ) UpperCAmelCase : str = inputs_dict["""labels"""] UpperCAmelCase : List[Any] = inputs_dict["""labels"""] UpperCAmelCase : Union[str, Any] = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=_SCREAMING_SNAKE_CASE , ) UpperCAmelCase : str = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_SCREAMING_SNAKE_CASE ) return inputs_dict def SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase : str = OpenAIGPTModelTester(self ) UpperCAmelCase : Any = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , n_embd=37 ) def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*_SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*_SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self ) -> List[str]: '''simple docstring''' UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*_SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: '''simple docstring''' UpperCAmelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*_SCREAMING_SNAKE_CASE ) @slow def SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: '''simple docstring''' for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase : Optional[Any] = OpenAIGPTModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase : Union[str, Any] = OpenAIGPTLMHeadModel.from_pretrained("""openai-gpt""" ) model.to(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Dict = torch.tensor([[481, 4735, 544]] , dtype=torch.long , device=_SCREAMING_SNAKE_CASE ) # the president is UpperCAmelCase : Optional[int] = [ 481, 4735, 544, 246, 963, 870, 762, 239, 244, 40477, 244, 249, 719, 881, 487, 544, 240, 244, 603, 481, ] # the president is a very good man. " \n " i\'m sure he is, " said the UpperCAmelCase : List[Any] = model.generate(_SCREAMING_SNAKE_CASE , do_sample=_SCREAMING_SNAKE_CASE ) self.assertListEqual(output_ids[0].tolist() , _SCREAMING_SNAKE_CASE )
109
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __lowerCAmelCase = TypeVar("""T""") class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Tuple ,_a : T ): '''simple docstring''' _a : List[str] = data _a : Node[T] | None = None def __str__( self : Dict ): '''simple docstring''' return F"""{self.data}""" class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Optional[int] ): '''simple docstring''' _a : Node[T] | None = None def __iter__( self : str ): '''simple docstring''' _a : Tuple = self.top while node: yield node.data _a : int = node.next def __str__( self : str ): '''simple docstring''' return "->".join([str(_a ) for item in self] ) def __len__( self : Optional[Any] ): '''simple docstring''' return len(tuple(iter(self ) ) ) def __lowercase ( self : str ): '''simple docstring''' return self.top is None def __lowercase ( self : List[Any] ,_a : T ): '''simple docstring''' _a : int = Node(_a ) if not self.is_empty(): _a : Optional[Any] = self.top _a : List[str] = node def __lowercase ( self : Tuple ): '''simple docstring''' if self.is_empty(): raise IndexError('pop from empty stack' ) assert isinstance(self.top ,_a ) _a : List[Any] = self.top _a : int = self.top.next return pop_node.data def __lowercase ( self : List[str] ): '''simple docstring''' if self.is_empty(): raise IndexError('peek from empty stack' ) assert self.top is not None return self.top.data def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = None if __name__ == "__main__": from doctest import testmod testmod()
271
0
import heapq as hq import math from collections.abc import Iterator class UpperCAmelCase_ : '''simple docstring''' def __init__( self , __A ): """simple docstring""" lowerCamelCase : Union[str, Any] = str(id_ ) lowerCamelCase : List[Any] = None lowerCamelCase : int = None lowerCamelCase : Union[str, Any] = [] lowerCamelCase : Any = {} # {vertex:distance} def __lt__( self , __A ): """simple docstring""" return self.key < other.key def __repr__( self ): """simple docstring""" return self.id def _snake_case ( self , __A ): """simple docstring""" self.neighbors.append(_a ) def _snake_case ( self , __A , __A ): """simple docstring""" lowerCamelCase : Dict = weight def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , __a ) graph[b - 1].add_edge(graph[a - 1] , __a ) def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : List[str] = [] for u in graph: lowerCamelCase : Any = math.inf lowerCamelCase : Optional[Any] = None lowerCamelCase : Any = 0 lowerCamelCase : Tuple = graph[:] while q: lowerCamelCase : Any = min(__a ) q.remove(__a ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): lowerCamelCase : List[Any] = u lowerCamelCase : Any = u.edges[v.id] for i in range(1 , len(__a ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' for u in graph: lowerCamelCase : Any = math.inf lowerCamelCase : List[Any] = None lowerCamelCase : Union[str, Any] = 0 lowerCamelCase : Tuple = list(__a ) hq.heapify(__a ) while h: lowerCamelCase : Any = hq.heappop(__a ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): lowerCamelCase : int = u lowerCamelCase : str = u.edges[v.id] hq.heapify(__a ) for i in range(1 , len(__a ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def lowercase_( ): '''simple docstring''' if __name__ == "__main__": import doctest doctest.testmod()
283
'''simple docstring''' import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : List[str] ): '''simple docstring''' torch.manual_seed(0 ) _a : int = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : str = self.dummy_uncond_unet _a : int = PNDMScheduler() _a : str = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Any = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ,return_dict=_a )[0] _a : List[Any] = image[0, -3:, -3:, -1] _a : Any = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : List[Any] = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[str] = 'google/ddpm-cifar10-32' _a : str = UNetaDModel.from_pretrained(_a ) _a : Union[str, Any] = PNDMScheduler() _a : Tuple = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : str = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,output_type='numpy' ).images _a : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : Tuple = np.array([0.1564, 0.1_4645, 0.1406, 0.1_4715, 0.1_2425, 0.1_4045, 0.1_3115, 0.1_2175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
271
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _A = logging.get_logger(__name__) _A = '''▁''' _A = {'''vocab_file''': '''sentencepiece.bpe.model''', '''monolingual_vocab_file''': '''dict.txt'''} _A = { '''vocab_file''': { '''vinai/bartpho-syllable''': '''https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model''', }, '''monolingual_vocab_file''': { '''vinai/bartpho-syllable''': '''https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt''', }, } _A = {'''vinai/bartpho-syllable''': 1_024} class lowercase_ ( lowercase__ ): A__ : Optional[Any] = VOCAB_FILES_NAMES A__ : Dict = PRETRAINED_VOCAB_FILES_MAP A__ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Dict = ['''input_ids''', '''attention_mask'''] def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase="<s>" , __UpperCamelCase="</s>" , __UpperCamelCase="</s>" , __UpperCamelCase="<s>" , __UpperCamelCase="<unk>" , __UpperCamelCase="<pad>" , __UpperCamelCase="<mask>" , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" UpperCamelCase_ = AddedToken(_a , lstrip=_a , rstrip=_a ) if isinstance(_a , _a ) else mask_token UpperCamelCase_ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_a , eos_token=_a , unk_token=_a , sep_token=_a , cls_token=_a , pad_token=_a , mask_token=_a , sp_model_kwargs=self.sp_model_kwargs , **_a , ) UpperCamelCase_ = vocab_file UpperCamelCase_ = monolingual_vocab_file UpperCamelCase_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_a ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility UpperCamelCase_ = {} UpperCamelCase_ = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(_a ) not in self.fairseq_tokens_to_ids: UpperCamelCase_ = cnt cnt += 1 with open(_a , """r""" , encoding="""utf-8""" ) as f: for line in f.readlines(): UpperCamelCase_ = line.strip().split()[0] UpperCamelCase_ = len(self.fairseq_tokens_to_ids ) if str(_a ) not in self.fairseq_tokens_to_ids: UpperCamelCase_ = len(self.fairseq_tokens_to_ids ) UpperCamelCase_ = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self ): """simple docstring""" UpperCamelCase_ = self.__dict__.copy() UpperCamelCase_ = None UpperCamelCase_ = self.sp_model.serialized_model_proto() return state def __setstate__( self , __UpperCamelCase ): """simple docstring""" UpperCamelCase_ = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): UpperCamelCase_ = {} UpperCamelCase_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCamelCase_ = [self.cls_token_id] UpperCamelCase_ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a , token_ids_a=_a , already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase = None ): """simple docstring""" UpperCamelCase_ = [self.sep_token_id] UpperCamelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def lowerCamelCase_ ( self ): """simple docstring""" return len(self.fairseq_ids_to_tokens ) def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowerCamelCase_ ( self , __UpperCamelCase ): """simple docstring""" return self.sp_model.encode(_a , out_type=_a ) def lowerCamelCase_ ( self , __UpperCamelCase ): """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def lowerCamelCase_ ( self , __UpperCamelCase ): """simple docstring""" return self.fairseq_ids_to_tokens[index] def lowerCamelCase_ ( self , __UpperCamelCase ): """simple docstring""" UpperCamelCase_ = ''.join(_a ).replace(_a , """ """ ).strip() return out_string def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase = None ): """simple docstring""" if not os.path.isdir(_a ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCamelCase_ = os.path.join( _a , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCamelCase_ = os.path.join( _a , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""monolingual_vocab_file"""] , ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _a ) elif not os.path.isfile(self.vocab_file ): with open(_a , """wb""" ) as fi: UpperCamelCase_ = self.sp_model.serialized_model_proto() fi.write(_a ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( _a ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file , _a ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(_a , """w""" , encoding="""utf-8""" ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(f'''{str(_a )} \n''' ) return out_vocab_file, out_monolingual_vocab_file
122
'''simple docstring''' import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __lowerCAmelCase = logging.getLogger() @unittest.skip('''Temporarily disable the doc tests.''' ) @require_torch @require_tf @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : str ,_a : Path ,_a : Union[str, None] = None ,_a : Union[List[str], None] = None ,_a : Union[str, List[str], None] = None ,_a : bool = True ,): '''simple docstring''' _a : Optional[int] = [file for file in os.listdir(_a ) if os.path.isfile(os.path.join(_a ,_a ) )] if identifier is not None: _a : List[str] = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(_a ,_a ): for n_ in n_identifier: _a : Tuple = [file for file in files if n_ not in file] else: _a : Optional[Any] = [file for file in files if n_identifier not in file] _a : List[str] = ignore_files or [] ignore_files.append('__init__.py' ) _a : Tuple = [file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' ,_a ) if only_modules: _a : Any = file.split('.' )[0] try: _a : List[str] = getattr(_a ,_a ) _a : int = doctest.DocTestSuite(_a ) _a : Any = unittest.TextTestRunner().run(_a ) self.assertIs(len(result.failures ) ,0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: _a : Union[str, Any] = doctest.testfile(str('..' / directory / file ) ,optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed ,0 ) def __lowercase ( self : Any ): '''simple docstring''' _a : int = Path('src/transformers' ) _a : List[Any] = 'modeling' _a : Optional[Any] = [ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(_a ,identifier=_a ,ignore_files=_a ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Optional[Any] = Path('src/transformers' ) _a : Optional[Any] = 'tokenization' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Dict = Path('src/transformers' ) _a : str = 'configuration' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : Any ): '''simple docstring''' _a : Tuple = Path('src/transformers' ) _a : List[Any] = ['configuration', 'modeling', 'tokenization'] self.analyze_directory(_a ,n_identifier=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = Path('docs/source' ) _a : List[str] = ['favicon.ico'] self.analyze_directory(_a ,ignore_files=_a ,only_modules=_a )
271
0
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. A ={'LayoutLMv2Config', 'LayoutLMv3Config'} @is_pipeline_test class _a ( unittest.TestCase ): __a : int = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING __a : Any = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: __a : Any = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: __a : Tuple = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def A ( self : Optional[int] , lowercase : Optional[Any] , lowercase : Optional[Any] , lowercase : List[str] ): '''simple docstring''' UpperCAmelCase = ZeroShotClassificationPipeline( model=_a , tokenizer=_a , candidate_labels=['''polics''', '''health'''] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def A ( self : List[Any] , lowercase : Optional[int] , lowercase : Optional[Any] ): '''simple docstring''' UpperCAmelCase = classifier('''Who are you voting for in 2020?''' , candidate_labels='''politics''' ) self.assertEqual(_a , {'''sequence''': ANY(_a ), '''labels''': [ANY(_a )], '''scores''': [ANY(_a )]} ) # No kwarg UpperCAmelCase = classifier('''Who are you voting for in 2020?''' , ['''politics'''] ) self.assertEqual(_a , {'''sequence''': ANY(_a ), '''labels''': [ANY(_a )], '''scores''': [ANY(_a )]} ) UpperCAmelCase = classifier('''Who are you voting for in 2020?''' , candidate_labels=['''politics'''] ) self.assertEqual(_a , {'''sequence''': ANY(_a ), '''labels''': [ANY(_a )], '''scores''': [ANY(_a )]} ) UpperCAmelCase = classifier('''Who are you voting for in 2020?''' , candidate_labels='''politics, public health''' ) self.assertEqual( _a , {'''sequence''': ANY(_a ), '''labels''': [ANY(_a ), ANY(_a )], '''scores''': [ANY(_a ), ANY(_a )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs['''scores'''] ) ) , 1.0 ) UpperCAmelCase = classifier('''Who are you voting for in 2020?''' , candidate_labels=['''politics''', '''public health'''] ) self.assertEqual( _a , {'''sequence''': ANY(_a ), '''labels''': [ANY(_a ), ANY(_a )], '''scores''': [ANY(_a ), ANY(_a )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs['''scores'''] ) ) , 1.0 ) UpperCAmelCase = classifier( '''Who are you voting for in 2020?''' , candidate_labels='''politics''' , hypothesis_template='''This text is about {}''' ) self.assertEqual(_a , {'''sequence''': ANY(_a ), '''labels''': [ANY(_a )], '''scores''': [ANY(_a )]} ) # https://github.com/huggingface/transformers/issues/13846 UpperCAmelCase = classifier(['''I am happy'''] , ['''positive''', '''negative'''] ) self.assertEqual( _a , [ {'''sequence''': ANY(_a ), '''labels''': [ANY(_a ), ANY(_a )], '''scores''': [ANY(_a ), ANY(_a )]} for i in range(1 ) ] , ) UpperCAmelCase = classifier(['''I am happy''', '''I am sad'''] , ['''positive''', '''negative'''] ) self.assertEqual( _a , [ {'''sequence''': ANY(_a ), '''labels''': [ANY(_a ), ANY(_a )], '''scores''': [ANY(_a ), ANY(_a )]} for i in range(2 ) ] , ) with self.assertRaises(_a ): classifier('''''' , candidate_labels='''politics''' ) with self.assertRaises(_a ): classifier(_a , candidate_labels='''politics''' ) with self.assertRaises(_a ): classifier('''Who are you voting for in 2020?''' , candidate_labels='''''' ) with self.assertRaises(_a ): classifier('''Who are you voting for in 2020?''' , candidate_labels=_a ) with self.assertRaises(_a ): classifier( '''Who are you voting for in 2020?''' , candidate_labels='''politics''' , hypothesis_template='''Not formatting template''' , ) with self.assertRaises(_a ): classifier( '''Who are you voting for in 2020?''' , candidate_labels='''politics''' , hypothesis_template=_a , ) self.run_entailment_id(_a ) def A ( self : str , lowercase : Pipeline ): '''simple docstring''' UpperCAmelCase = zero_shot_classifier.model.config UpperCAmelCase = config.labelaid UpperCAmelCase = zero_shot_classifier.entailment_id UpperCAmelCase = {'LABEL_0': 0, 'LABEL_1': 1, 'LABEL_2': 2} self.assertEqual(zero_shot_classifier.entailment_id , -1 ) UpperCAmelCase = {'entailment': 0, 'neutral': 1, 'contradiction': 2} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) UpperCAmelCase = {'ENTAIL': 0, 'NON-ENTAIL': 1} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) UpperCAmelCase = {'ENTAIL': 2, 'NEUTRAL': 1, 'CONTR': 0} self.assertEqual(zero_shot_classifier.entailment_id , 2 ) UpperCAmelCase = original_labelaid self.assertEqual(_a , zero_shot_classifier.entailment_id ) @require_torch def A ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase = pipeline( '''zero-shot-classification''' , model='''sshleifer/tiny-distilbert-base-cased-distilled-squad''' , framework='''pt''' , ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( '''Who are you voting for in 2020?''' * 100 , candidate_labels=['''politics''', '''public health''', '''science'''] ) @require_torch def A ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase = pipeline( '''zero-shot-classification''' , model='''sshleifer/tiny-distilbert-base-cased-distilled-squad''' , framework='''pt''' , ) UpperCAmelCase = zero_shot_classifier( '''Who are you voting for in 2020?''' , candidate_labels=['''politics''', '''public health''', '''science'''] ) self.assertEqual( nested_simplify(_a ) , { '''sequence''': '''Who are you voting for in 2020?''', '''labels''': ['''science''', '''public health''', '''politics'''], '''scores''': [0.333, 0.333, 0.333], } , ) @require_tf def A ( self : Dict ): '''simple docstring''' UpperCAmelCase = pipeline( '''zero-shot-classification''' , model='''sshleifer/tiny-distilbert-base-cased-distilled-squad''' , framework='''tf''' , ) UpperCAmelCase = zero_shot_classifier( '''Who are you voting for in 2020?''' , candidate_labels=['''politics''', '''public health''', '''science'''] ) self.assertEqual( nested_simplify(_a ) , { '''sequence''': '''Who are you voting for in 2020?''', '''labels''': ['''science''', '''public health''', '''politics'''], '''scores''': [0.333, 0.333, 0.333], } , ) @slow @require_torch def A ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase = pipeline('''zero-shot-classification''' , model='''roberta-large-mnli''' , framework='''pt''' ) UpperCAmelCase = zero_shot_classifier( '''Who are you voting for in 2020?''' , candidate_labels=['''politics''', '''public health''', '''science'''] ) self.assertEqual( nested_simplify(_a ) , { '''sequence''': '''Who are you voting for in 2020?''', '''labels''': ['''politics''', '''public health''', '''science'''], '''scores''': [0.976, 0.015, 0.009], } , ) UpperCAmelCase = zero_shot_classifier( '''The dominant sequence transduction models are based on complex recurrent or convolutional neural networks''' ''' in an encoder-decoder configuration. The best performing models also connect the encoder and decoder''' ''' through an attention mechanism. We propose a new simple network architecture, the Transformer, based''' ''' solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two''' ''' machine translation tasks show these models to be superior in quality while being more parallelizable''' ''' and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014''' ''' English-to-German translation task, improving over the existing best results, including ensembles by''' ''' over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new''' ''' single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small''' ''' fraction of the training costs of the best models from the literature. We show that the Transformer''' ''' generalizes well to other tasks by applying it successfully to English constituency parsing both with''' ''' large and limited training data.''' , candidate_labels=['''machine learning''', '''statistics''', '''translation''', '''vision'''] , multi_label=_a , ) self.assertEqual( nested_simplify(_a ) , { '''sequence''': ( '''The dominant sequence transduction models are based on complex recurrent or convolutional neural''' ''' networks in an encoder-decoder configuration. The best performing models also connect the''' ''' encoder and decoder through an attention mechanism. We propose a new simple network''' ''' architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence''' ''' and convolutions entirely. Experiments on two machine translation tasks show these models to be''' ''' superior in quality while being more parallelizable and requiring significantly less time to''' ''' train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,''' ''' improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014''' ''' English-to-French translation task, our model establishes a new single-model state-of-the-art''' ''' BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training''' ''' costs of the best models from the literature. We show that the Transformer generalizes well to''' ''' other tasks by applying it successfully to English constituency parsing both with large and''' ''' limited training data.''' ), '''labels''': ['''translation''', '''machine learning''', '''vision''', '''statistics'''], '''scores''': [0.817, 0.713, 0.018, 0.018], } , ) @slow @require_tf def A ( self : List[str] ): '''simple docstring''' UpperCAmelCase = pipeline('''zero-shot-classification''' , model='''roberta-large-mnli''' , framework='''tf''' ) UpperCAmelCase = zero_shot_classifier( '''Who are you voting for in 2020?''' , candidate_labels=['''politics''', '''public health''', '''science'''] ) self.assertEqual( nested_simplify(_a ) , { '''sequence''': '''Who are you voting for in 2020?''', '''labels''': ['''politics''', '''public health''', '''science'''], '''scores''': [0.976, 0.015, 0.009], } , ) UpperCAmelCase = zero_shot_classifier( '''The dominant sequence transduction models are based on complex recurrent or convolutional neural networks''' ''' in an encoder-decoder configuration. The best performing models also connect the encoder and decoder''' ''' through an attention mechanism. We propose a new simple network architecture, the Transformer, based''' ''' solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two''' ''' machine translation tasks show these models to be superior in quality while being more parallelizable''' ''' and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014''' ''' English-to-German translation task, improving over the existing best results, including ensembles by''' ''' over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new''' ''' single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small''' ''' fraction of the training costs of the best models from the literature. We show that the Transformer''' ''' generalizes well to other tasks by applying it successfully to English constituency parsing both with''' ''' large and limited training data.''' , candidate_labels=['''machine learning''', '''statistics''', '''translation''', '''vision'''] , multi_label=_a , ) self.assertEqual( nested_simplify(_a ) , { '''sequence''': ( '''The dominant sequence transduction models are based on complex recurrent or convolutional neural''' ''' networks in an encoder-decoder configuration. The best performing models also connect the''' ''' encoder and decoder through an attention mechanism. We propose a new simple network''' ''' architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence''' ''' and convolutions entirely. Experiments on two machine translation tasks show these models to be''' ''' superior in quality while being more parallelizable and requiring significantly less time to''' ''' train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,''' ''' improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014''' ''' English-to-French translation task, our model establishes a new single-model state-of-the-art''' ''' BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training''' ''' costs of the best models from the literature. We show that the Transformer generalizes well to''' ''' other tasks by applying it successfully to English constituency parsing both with large and''' ''' limited training data.''' ), '''labels''': ['''translation''', '''machine learning''', '''vision''', '''statistics'''], '''scores''': [0.817, 0.713, 0.018, 0.018], } , )
34
'''simple docstring''' import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def UpperCAmelCase_ (__a : Optional[Any] , __a : str , __a : Optional[Any]=None ): """simple docstring""" assert torch_layer.weight.shape == weight.shape, f"""{torch_layer} layer.weight does not match""" _a : str = nn.Parameter(__a ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f"""{torch_layer} layer.bias does not match""" _a : Any = nn.Parameter(__a ) def UpperCAmelCase_ (__a : int , __a : Optional[Any] , __a : int ): """simple docstring""" _a : Tuple = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : Dict = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Optional[Any] , __a : Optional[int] , __a : List[str] ): """simple docstring""" _a : Dict = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : str = np.asarray(weights[2] ) _a : int = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Any , __a : Any , __a : Optional[Any] ): """simple docstring""" _a : List[str] = weights[0][0][0] _a : List[Any] = np.asarray(layer_norm_a[0] ) _a : List[str] = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # lsh weights + output _a : List[str] = weights[0][1] if len(__a ) < 4: set_layer_weights_in_torch_lsh(__a , torch_block.attention , __a ) else: set_layer_weights_in_torch_local(__a , torch_block.attention , __a ) # intermediate weighs _a : Optional[Any] = weights[2][0][1][2] # Chunked Feed Forward if len(__a ) == 4: _a : Union[str, Any] = intermediate_weights[2] # layernorm 2 _a : Any = np.asarray(intermediate_weights[0][0] ) _a : List[Any] = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # intermediate dense _a : Any = np.asarray(intermediate_weights[1][0] ) _a : Any = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) # intermediate out _a : Optional[int] = np.asarray(intermediate_weights[4][0] ) _a : int = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Dict , __a : Dict , __a : List[Any] ): """simple docstring""" _a : Optional[int] = torch_model.reformer # word embeds _a : Tuple = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(__a ) , ) if isinstance(weights[3] , __a ): _a : Any = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _a : List[Any] = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f"""{position_embeddings[emb_idx]} emb does not match""" _a : Any = nn.Parameter(torch.tensor(__a ) ) _a : List[str] = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( __a ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _a : Tuple = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(__a , __a , __a ) # output layer norm _a : Optional[Any] = np.asarray(weights[7][0] ) _a : int = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # output embeddings _a : List[str] = np.asarray(weights[9][0] ) _a : int = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Tuple , __a : Optional[Any] , __a : Dict ): """simple docstring""" _a : List[Any] = ReformerConfig.from_json_file(__a ) print(f"""Building PyTorch model from configuration: {config}""" ) _a : int = ReformerModelWithLMHead(__a ) with open(__a , 'rb' ) as f: _a : Optional[Any] = pickle.load(__a )['weights'] set_model_weights_in_torch(__a , __a , config.hidden_size ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , __a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--trax_model_pkl_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained Reformer model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __lowerCAmelCase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
271
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available __UpperCamelCase : Optional[Any] = {"""tokenization_herbert""": ["""HerbertTokenizer"""]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Union[str, Any] = ["""HerbertTokenizerFast"""] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys __UpperCamelCase : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
307
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Union[str, Any] = VQModel( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=3 ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1000 ,) return CLIPTextModel(_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : Dict = self.dummy_uncond_unet _a : List[Any] = DDIMScheduler() _a : List[Any] = self.dummy_vq_model _a : str = LDMPipeline(unet=_a ,vqvae=_a ,scheduler=_a ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : List[str] = torch.manual_seed(0 ) _a : List[str] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Union[str, Any] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ,return_dict=_a )[0] _a : Tuple = image[0, -3:, -3:, -1] _a : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _a : int = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) _a : Any = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : List[str] = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Dict = ldm(generator=_a ,num_inference_steps=5 ,output_type='numpy' ).images _a : str = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _a : Optional[int] = np.array([0.4399, 0.4_4975, 0.4_6825, 0.474, 0.4359, 0.4581, 0.4_5095, 0.4341, 0.4447] ) _a : int = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
271
0
from __future__ import annotations def lowerCAmelCase_ (lowerCAmelCase__: list[list[int]] ): """simple docstring""" for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(__a ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 , len(__a ) ): for j in range(1 , len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] , matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
147
'''simple docstring''' import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : int ,*_a : Optional[int] ,**_a : str ): '''simple docstring''' warnings.warn( 'The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use BeitImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _snake_case : Optional[Any] = parse(importlib.metadata.version('torch')) def a_ ( lowerCAmelCase_ : Union[str, Version], lowerCAmelCase_ : str, lowerCAmelCase_ : str ): if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(F"""`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}""" ) __lowerCAmelCase = STR_OPERATION_TO_FUNC[operation] if isinstance(__a, __a ): __lowerCAmelCase = parse(importlib.metadata.version(__a ) ) return operation(__a, parse(__a ) ) def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : str ): return compare_versions(__a, __a, __a )
284
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { """linear""": get_linear_schedule_with_warmup, """cosine""": get_cosine_schedule_with_warmup, """cosine_w_restarts""": get_cosine_with_hard_restarts_schedule_with_warmup, """polynomial""": get_polynomial_decay_schedule_with_warmup, """constant""": get_constant_schedule, """constant_w_warmup""": get_constant_schedule_with_warmup, } class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Optional[int] ,_a : Optional[Any]=None ,_a : Dict=None ,*_a : int ,**_a : str ): '''simple docstring''' super().__init__(*_a ,**_a ) if config is None: assert isinstance(self.model ,_a ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F""" {self.model.__class__}""" ) _a : List[Any] = self.model.config else: _a : Optional[int] = config _a : List[str] = data_args _a : List[Any] = self.config.tgt_vocab_size if isinstance(self.config ,_a ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ' padding..' ) if self.args.label_smoothing == 0: _a : List[str] = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss _a : Tuple = label_smoothed_nll_loss def __lowercase ( self : List[str] ,_a : int ): '''simple docstring''' if self.optimizer is None: _a : Union[str, Any] = ['bias', 'LayerNorm.weight'] _a : Tuple = [ { 'params': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], 'weight_decay': self.args.weight_decay, }, { 'params': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], 'weight_decay': 0.0, }, ] _a : Optional[int] = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: _a : Any = Adafactor _a : Dict = {'scale_parameter': False, 'relative_step': False} else: _a : Union[str, Any] = AdamW _a : str = { 'betas': (self.args.adam_betaa, self.args.adam_betaa), 'eps': self.args.adam_epsilon, } _a : Union[str, Any] = self.args.learning_rate if self.sharded_ddp: _a : str = OSS( params=_a ,optim=_a ,**_a ,) else: _a : Tuple = optimizer_cls(_a ,**_a ) if self.lr_scheduler is None: _a : List[Any] = self._get_lr_scheduler(_a ) else: # ignoring --lr_scheduler logger.warning('scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.' ) def __lowercase ( self : List[Any] ,_a : List[Any] ): '''simple docstring''' _a : str = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": _a : int = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": _a : List[str] = schedule_func(self.optimizer ,num_warmup_steps=self.args.warmup_steps ) else: _a : Optional[int] = schedule_func( self.optimizer ,num_warmup_steps=self.args.warmup_steps ,num_training_steps=_a ) return scheduler def __lowercase ( self : Tuple ): '''simple docstring''' if isinstance(self.train_dataset ,torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size ,distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) ,) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def __lowercase ( self : Dict ,_a : Dict ,_a : Any ,_a : Dict ): '''simple docstring''' if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Union[str, Any] = self.loss_fn(logits.view(-1 ,logits.shape[-1] ) ,labels.view(-1 ) ) else: # compute usual loss via models _a, _a : Union[str, Any] = model(**_a ,labels=_a ,use_cache=_a )[:2] else: # compute label smoothed loss _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Any = torch.nn.functional.log_softmax(_a ,dim=-1 ) _a, _a : List[str] = self.loss_fn(_a ,_a ,self.args.label_smoothing ,ignore_index=self.config.pad_token_id ) return loss, logits def __lowercase ( self : Optional[int] ,_a : Union[str, Any] ,_a : List[Any] ): '''simple docstring''' _a : Optional[int] = inputs.pop('labels' ) _a, _a : int = self._compute_loss(_a ,_a ,_a ) return loss def __lowercase ( self : Optional[Any] ,_a : nn.Module ,_a : Dict[str, Union[torch.Tensor, Any]] ,_a : bool ,_a : Optional[List[str]] = None ,): '''simple docstring''' _a : int = self._prepare_inputs(_a ) _a : Any = { 'max_length': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, 'num_beams': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: _a : int = self.model.generate( inputs['input_ids'] ,attention_mask=inputs['attention_mask'] ,**_a ,) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: _a : int = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) _a : Union[str, Any] = inputs.pop('labels' ) with torch.no_grad(): # compute loss on predict data _a, _a : Optional[int] = self._compute_loss(_a ,_a ,_a ) _a : Optional[Any] = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) _a : Optional[Any] = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: _a : Dict = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) return (loss, logits, labels) def __lowercase ( self : str ,_a : Tuple ,_a : Tuple ): '''simple docstring''' _a : List[Any] = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( 'Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be' F""" padded to `max_length`={max_length}""" ) _a : int = pad_token_id * torch.ones( (tensor.shape[0], max_length) ,dtype=tensor.dtype ,device=tensor.device ) _a : Union[str, Any] = tensor return padded_tensor
271
0
import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() lowerCamelCase = logging.get_logger('''transformers.models.encodec''') lowerCamelCase = { '''quantizer.vq.layers.*._codebook.inited''': '''quantizer.layers.*.codebook.inited''', '''quantizer.vq.layers.*._codebook.cluster_size''': '''quantizer.layers.*.codebook.cluster_size''', '''quantizer.vq.layers.*._codebook.embed''': '''quantizer.layers.*.codebook.embed''', '''quantizer.vq.layers.*._codebook.embed_avg''': '''quantizer.layers.*.codebook.embed_avg''', } lowerCamelCase = { '''encoder.model.0.conv.conv''': '''encoder.layers.0.conv''', '''encoder.model.1.block.1.conv.conv''': '''encoder.layers.1.block.1.conv''', '''encoder.model.1.block.3.conv.conv''': '''encoder.layers.1.block.3.conv''', '''encoder.model.1.shortcut.conv.conv''': '''encoder.layers.1.shortcut.conv''', '''encoder.model.3.conv.conv''': '''encoder.layers.3.conv''', '''encoder.model.4.block.1.conv.conv''': '''encoder.layers.4.block.1.conv''', '''encoder.model.4.block.3.conv.conv''': '''encoder.layers.4.block.3.conv''', '''encoder.model.4.shortcut.conv.conv''': '''encoder.layers.4.shortcut.conv''', '''encoder.model.6.conv.conv''': '''encoder.layers.6.conv''', '''encoder.model.7.block.1.conv.conv''': '''encoder.layers.7.block.1.conv''', '''encoder.model.7.block.3.conv.conv''': '''encoder.layers.7.block.3.conv''', '''encoder.model.7.shortcut.conv.conv''': '''encoder.layers.7.shortcut.conv''', '''encoder.model.9.conv.conv''': '''encoder.layers.9.conv''', '''encoder.model.10.block.1.conv.conv''': '''encoder.layers.10.block.1.conv''', '''encoder.model.10.block.3.conv.conv''': '''encoder.layers.10.block.3.conv''', '''encoder.model.10.shortcut.conv.conv''': '''encoder.layers.10.shortcut.conv''', '''encoder.model.12.conv.conv''': '''encoder.layers.12.conv''', '''encoder.model.13.lstm''': '''encoder.layers.13.lstm''', '''encoder.model.15.conv.conv''': '''encoder.layers.15.conv''', } lowerCamelCase = { '''encoder.model.0.conv.norm''': '''encoder.layers.0.norm''', '''encoder.model.1.block.1.conv.norm''': '''encoder.layers.1.block.1.norm''', '''encoder.model.1.block.3.conv.norm''': '''encoder.layers.1.block.3.norm''', '''encoder.model.1.shortcut.conv.norm''': '''encoder.layers.1.shortcut.norm''', '''encoder.model.3.conv.norm''': '''encoder.layers.3.norm''', '''encoder.model.4.block.1.conv.norm''': '''encoder.layers.4.block.1.norm''', '''encoder.model.4.block.3.conv.norm''': '''encoder.layers.4.block.3.norm''', '''encoder.model.4.shortcut.conv.norm''': '''encoder.layers.4.shortcut.norm''', '''encoder.model.6.conv.norm''': '''encoder.layers.6.norm''', '''encoder.model.7.block.1.conv.norm''': '''encoder.layers.7.block.1.norm''', '''encoder.model.7.block.3.conv.norm''': '''encoder.layers.7.block.3.norm''', '''encoder.model.7.shortcut.conv.norm''': '''encoder.layers.7.shortcut.norm''', '''encoder.model.9.conv.norm''': '''encoder.layers.9.norm''', '''encoder.model.10.block.1.conv.norm''': '''encoder.layers.10.block.1.norm''', '''encoder.model.10.block.3.conv.norm''': '''encoder.layers.10.block.3.norm''', '''encoder.model.10.shortcut.conv.norm''': '''encoder.layers.10.shortcut.norm''', '''encoder.model.12.conv.norm''': '''encoder.layers.12.norm''', '''encoder.model.15.conv.norm''': '''encoder.layers.15.norm''', } lowerCamelCase = { '''decoder.model.0.conv.conv''': '''decoder.layers.0.conv''', '''decoder.model.1.lstm''': '''decoder.layers.1.lstm''', '''decoder.model.3.convtr.convtr''': '''decoder.layers.3.conv''', '''decoder.model.4.block.1.conv.conv''': '''decoder.layers.4.block.1.conv''', '''decoder.model.4.block.3.conv.conv''': '''decoder.layers.4.block.3.conv''', '''decoder.model.4.shortcut.conv.conv''': '''decoder.layers.4.shortcut.conv''', '''decoder.model.6.convtr.convtr''': '''decoder.layers.6.conv''', '''decoder.model.7.block.1.conv.conv''': '''decoder.layers.7.block.1.conv''', '''decoder.model.7.block.3.conv.conv''': '''decoder.layers.7.block.3.conv''', '''decoder.model.7.shortcut.conv.conv''': '''decoder.layers.7.shortcut.conv''', '''decoder.model.9.convtr.convtr''': '''decoder.layers.9.conv''', '''decoder.model.10.block.1.conv.conv''': '''decoder.layers.10.block.1.conv''', '''decoder.model.10.block.3.conv.conv''': '''decoder.layers.10.block.3.conv''', '''decoder.model.10.shortcut.conv.conv''': '''decoder.layers.10.shortcut.conv''', '''decoder.model.12.convtr.convtr''': '''decoder.layers.12.conv''', '''decoder.model.13.block.1.conv.conv''': '''decoder.layers.13.block.1.conv''', '''decoder.model.13.block.3.conv.conv''': '''decoder.layers.13.block.3.conv''', '''decoder.model.13.shortcut.conv.conv''': '''decoder.layers.13.shortcut.conv''', '''decoder.model.15.conv.conv''': '''decoder.layers.15.conv''', } lowerCamelCase = { '''decoder.model.0.conv.norm''': '''decoder.layers.0.norm''', '''decoder.model.3.convtr.norm''': '''decoder.layers.3.norm''', '''decoder.model.4.block.1.conv.norm''': '''decoder.layers.4.block.1.norm''', '''decoder.model.4.block.3.conv.norm''': '''decoder.layers.4.block.3.norm''', '''decoder.model.4.shortcut.conv.norm''': '''decoder.layers.4.shortcut.norm''', '''decoder.model.6.convtr.norm''': '''decoder.layers.6.norm''', '''decoder.model.7.block.1.conv.norm''': '''decoder.layers.7.block.1.norm''', '''decoder.model.7.block.3.conv.norm''': '''decoder.layers.7.block.3.norm''', '''decoder.model.7.shortcut.conv.norm''': '''decoder.layers.7.shortcut.norm''', '''decoder.model.9.convtr.norm''': '''decoder.layers.9.norm''', '''decoder.model.10.block.1.conv.norm''': '''decoder.layers.10.block.1.norm''', '''decoder.model.10.block.3.conv.norm''': '''decoder.layers.10.block.3.norm''', '''decoder.model.10.shortcut.conv.norm''': '''decoder.layers.10.shortcut.norm''', '''decoder.model.12.convtr.norm''': '''decoder.layers.12.norm''', '''decoder.model.13.block.1.conv.norm''': '''decoder.layers.13.block.1.norm''', '''decoder.model.13.block.3.conv.norm''': '''decoder.layers.13.block.3.norm''', '''decoder.model.13.shortcut.conv.norm''': '''decoder.layers.13.shortcut.norm''', '''decoder.model.15.conv.norm''': '''decoder.layers.15.norm''', } lowerCamelCase = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } lowerCamelCase = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } lowerCamelCase = [] lowerCamelCase = [] def lowerCamelCase_ ( _a , _a , _a , _a , _a ): """simple docstring""" for attribute in key.split('''.''' ): lowerCAmelCase__ : List[Any] = getattr(__a , __a ) if weight_type is not None: lowerCAmelCase__ : List[Any] = getattr(__a , __a ).shape else: lowerCAmelCase__ : List[str] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": lowerCAmelCase__ : Tuple = value elif weight_type == "weight_g": lowerCAmelCase__ : Dict = value elif weight_type == "weight_v": lowerCAmelCase__ : str = value elif weight_type == "bias": lowerCAmelCase__ : Dict = value elif weight_type == "running_mean": lowerCAmelCase__ : str = value elif weight_type == "running_var": lowerCAmelCase__ : str = value elif weight_type == "num_batches_tracked": lowerCAmelCase__ : Optional[Any] = value elif weight_type == "weight_ih_l0": lowerCAmelCase__ : List[Any] = value elif weight_type == "weight_hh_l0": lowerCAmelCase__ : int = value elif weight_type == "bias_ih_l0": lowerCAmelCase__ : Optional[int] = value elif weight_type == "bias_hh_l0": lowerCAmelCase__ : Dict = value elif weight_type == "weight_ih_l1": lowerCAmelCase__ : str = value elif weight_type == "weight_hh_l1": lowerCAmelCase__ : Dict = value elif weight_type == "bias_ih_l1": lowerCAmelCase__ : Union[str, Any] = value elif weight_type == "bias_hh_l1": lowerCAmelCase__ : Tuple = value else: lowerCAmelCase__ : Dict = value logger.info(f'{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.' ) def lowerCamelCase_ ( _a , _a ): """simple docstring""" for key in ignore_keys: if key.endswith('''.*''' ): if name.startswith(key[:-1] ): return True elif ".*." in key: lowerCAmelCase__ : int = key.split('''.*.''' ) if prefix in name and suffix in name: return True elif key in name: return True return False def lowerCamelCase_ ( _a , _a , _a ): """simple docstring""" lowerCAmelCase__ : Optional[int] = [] if model_name == "encodec_24khz" or "encodec_32khz": lowerCAmelCase__ : Tuple = MAPPING_24K elif model_name == "encodec_48khz": lowerCAmelCase__ : Any = MAPPING_48K else: raise ValueError(f'Unsupported model: {model_name}' ) for name, value in orig_dict.items(): if should_ignore(__a , __a ): logger.info(f'{name} was ignored' ) continue lowerCAmelCase__ : Tuple = False for key, mapped_key in MAPPING.items(): if "*" in key: lowerCAmelCase__ : Union[str, Any] = key.split('''.*.''' ) if prefix in name and suffix in name: lowerCAmelCase__ : Union[str, Any] = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith('''embed''' ) and name.endswith('''embed_avg''' ): continue lowerCAmelCase__ : int = True if "*" in mapped_key: lowerCAmelCase__ : int = name.split(__a )[0].split('''.''' )[-2] lowerCAmelCase__ : Dict = mapped_key.replace('''*''' , __a ) if "weight_g" in name: lowerCAmelCase__ : List[Any] = 'weight_g' elif "weight_v" in name: lowerCAmelCase__ : List[str] = 'weight_v' elif "weight_ih_l0" in name: lowerCAmelCase__ : Tuple = 'weight_ih_l0' elif "weight_hh_l0" in name: lowerCAmelCase__ : Optional[Any] = 'weight_hh_l0' elif "bias_ih_l0" in name: lowerCAmelCase__ : List[str] = 'bias_ih_l0' elif "bias_hh_l0" in name: lowerCAmelCase__ : Union[str, Any] = 'bias_hh_l0' elif "weight_ih_l1" in name: lowerCAmelCase__ : Dict = 'weight_ih_l1' elif "weight_hh_l1" in name: lowerCAmelCase__ : Optional[int] = 'weight_hh_l1' elif "bias_ih_l1" in name: lowerCAmelCase__ : Optional[Any] = 'bias_ih_l1' elif "bias_hh_l1" in name: lowerCAmelCase__ : Dict = 'bias_hh_l1' elif "bias" in name: lowerCAmelCase__ : List[Any] = 'bias' elif "weight" in name: lowerCAmelCase__ : Dict = 'weight' elif "running_mean" in name: lowerCAmelCase__ : Optional[Any] = 'running_mean' elif "running_var" in name: lowerCAmelCase__ : List[Any] = 'running_var' elif "num_batches_tracked" in name: lowerCAmelCase__ : Tuple = 'num_batches_tracked' else: lowerCAmelCase__ : List[str] = None set_recursively(__a , __a , __a , __a , __a ) continue if not is_used: unused_weights.append(__a ) logger.warning(f'Unused weights: {unused_weights}' ) @torch.no_grad() def lowerCamelCase_ ( _a , _a , _a , _a=None , _a=None , ): """simple docstring""" if config_path is not None: lowerCAmelCase__ : str = EncodecConfig.from_pretrained(__a ) else: lowerCAmelCase__ : Optional[int] = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": lowerCAmelCase__ : str = [8, 5, 4, 4] lowerCAmelCase__ : Tuple = [2.2] lowerCAmelCase__ : Tuple = 64 lowerCAmelCase__ : Optional[Any] = 32_000 lowerCAmelCase__ : List[str] = 2_048 lowerCAmelCase__ : Optional[int] = False lowerCAmelCase__ : Optional[int] = False lowerCAmelCase__ : Dict = False elif model_name == "encodec_48khz": lowerCAmelCase__ : int = [8, 5, 4, 2] lowerCAmelCase__ : Dict = [3.0, 6.0, 12.0, 24.0] lowerCAmelCase__ : Union[str, Any] = 48_000 lowerCAmelCase__ : Union[str, Any] = 2 lowerCAmelCase__ : Tuple = False lowerCAmelCase__ : str = 'time_group_norm' lowerCAmelCase__ : Any = True lowerCAmelCase__ : Tuple = 1.0 lowerCAmelCase__ : Tuple = 0.01 else: raise ValueError(f'Unknown model name: {model_name}' ) lowerCAmelCase__ : Union[str, Any] = EncodecModel(__a ) lowerCAmelCase__ : Tuple = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(__a ) lowerCAmelCase__ : Any = torch.load(__a ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights lowerCAmelCase__ : List[Any] = original_checkpoint['best_state'] recursively_load_weights(__a , __a , __a ) model.save_pretrained(__a ) if repo_id: print('''Pushing to the hub...''' ) feature_extractor.push_to_hub(__a ) model.push_to_hub(__a ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '''--model''', default='''encodec_24khz''', type=str, help='''The model to convert. Should be one of \'encodec_24khz\', \'encodec_32khz\', \'encodec_48khz\'.''', ) parser.add_argument('''--checkpoint_path''', required=True, default=None, type=str, help='''Path to original checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, default=None, type=str, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) lowerCamelCase = parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
131
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser __lowerCAmelCase = re.compile(r"""\s+""") def UpperCAmelCase_ (__a : Any ): """simple docstring""" return {"hash": hashlib.mda(re.sub(__a , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : List[str] = [len(__a ) for line in example['content'].splitlines()] return {"line_mean": np.mean(__a ), "line_max": max(__a )} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Union[str, Any] = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase_ (__a : Optional[int] , __a : Any ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def UpperCAmelCase_ (__a : int , __a : Union[str, Any]=5 ): """simple docstring""" _a : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] _a : List[str] = example['content'].splitlines() for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase_ (__a : List[str] , __a : Dict=5 , __a : Tuple=0.05 ): """simple docstring""" _a : Optional[int] = ['unit tests', 'test file', 'configuration file'] _a : int = example['content'].splitlines() _a : int = 0 _a : Dict = 0 # first test for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test _a : int = example['content'].count('\n' ) _a : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" _a : List[str] = ['def ', 'class ', 'for ', 'while '] _a : str = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase_ (__a : int , __a : Any=4 ): """simple docstring""" _a : List[str] = example['content'].splitlines() _a : Dict = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Optional[Any] = tokenizer(example['content'] , truncation=__a )['input_ids'] _a : Optional[int] = len(example['content'] ) / len(__a ) return {"ratio": ratio} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Dict = {} results.update(get_hash(__a ) ) results.update(line_stats(__a ) ) results.update(alpha_stats(__a ) ) results.update(char_token_ratio(__a ) ) results.update(is_autogenerated(__a ) ) results.update(is_config_or_test(__a ) ) results.update(has_no_keywords(__a ) ) results.update(has_few_assignments(__a ) ) return results def UpperCAmelCase_ (__a : Any , __a : Any , __a : str ): """simple docstring""" if not check_uniques(__a , __a ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase_ (__a : Union[str, Any] ): """simple docstring""" with open(__a , 'rb' ) as f_in: with gzip.open(str(__a ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(__a , __a ) os.unlink(__a ) # Settings __lowerCAmelCase = HfArgumentParser(PreprocessingArguments) __lowerCAmelCase = parser.parse_args() if args.num_workers is None: __lowerCAmelCase = multiprocessing.cpu_count() __lowerCAmelCase = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset __lowerCAmelCase = time.time() __lowerCAmelCase = load_dataset(args.dataset_name, split="""train""") print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing __lowerCAmelCase = time.time() __lowerCAmelCase = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes __lowerCAmelCase = set(ds.unique("""hash""")) __lowerCAmelCase = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics __lowerCAmelCase = time.time() __lowerCAmelCase = ds.filter(filter, fn_kwargs={"""uniques""": uniques, """args""": args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: __lowerCAmelCase = time.time() __lowerCAmelCase , __lowerCAmelCase = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file __lowerCAmelCase = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / """duplicate_clusters.json""", """w""") as f: json.dump(duplicate_clusters, f) __lowerCAmelCase = output_dir / """data""" data_dir.mkdir(exist_ok=True) __lowerCAmelCase = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): __lowerCAmelCase = str(data_dir / f'''file-{file_number+1:012}.json''') __lowerCAmelCase = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
271
0
"""simple docstring""" import webbrowser from sys import argv from urllib.parse import parse_qs, quote import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": lowerCAmelCase__ = '''%20'''.join(argv[1:]) if len(argv) > 1 else quote(str(input('''Search: '''))) print('''Googling.....''') lowerCAmelCase__ = F"https://www.google.com/search?q={query}&num=100" lowerCAmelCase__ = requests.get( url, headers={'''User-Agent''': str(UserAgent().random)}, ) try: lowerCAmelCase__ = ( BeautifulSoup(res.text, '''html.parser''') .find('''div''', attrs={'''class''': '''yuRUbf'''}) .find('''a''') .get('''href''') ) except AttributeError: lowerCAmelCase__ = parse_qs( BeautifulSoup(res.text, '''html.parser''') .find('''div''', attrs={'''class''': '''kCrYT'''}) .find('''a''') .get('''href''') )['''url'''][0] webbrowser.open(link)
108
'''simple docstring''' import argparse from typing import List import evaluate import numpy as np import torch from datasets import DatasetDict, load_dataset # New Code # # We'll be using StratifiedKFold for this example from sklearn.model_selection import StratifiedKFold from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to perform Cross Validation, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __lowerCAmelCase = 1_6 __lowerCAmelCase = 3_2 def UpperCAmelCase_ (__a : Accelerator , __a : DatasetDict , __a : List[int] , __a : List[int] , __a : int = 1_6 ): """simple docstring""" _a : Union[str, Any] = AutoTokenizer.from_pretrained('bert-base-cased' ) _a : str = DatasetDict( { 'train': dataset['train'].select(__a ), 'validation': dataset['train'].select(__a ), 'test': dataset['validation'], } ) def tokenize_function(__a : List[Any] ): # max_length=None => use the model max length (it's actually the default) _a : Optional[int] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=__a , max_length=__a ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _a : List[str] = datasets.map( __a , batched=__a , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _a : List[Any] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(__a : int ): # On TPU it's best to pad everything to the same length or training will be very slow. _a : Dict = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _a : Tuple = 1_6 elif accelerator.mixed_precision != "no": _a : List[Any] = 8 else: _a : List[Any] = None return tokenizer.pad( __a , padding='longest' , max_length=__a , pad_to_multiple_of=__a , return_tensors='pt' , ) # Instantiate dataloaders. _a : Any = DataLoader( tokenized_datasets['train'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[int] = DataLoader( tokenized_datasets['validation'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[Any] = DataLoader( tokenized_datasets['test'] , shuffle=__a , collate_fn=__a , batch_size=__a ) return train_dataloader, eval_dataloader, test_dataloader def UpperCAmelCase_ (__a : Any , __a : Union[str, Any] ): """simple docstring""" _a : Dict = [] # Download the dataset _a : Tuple = load_dataset('glue' , 'mrpc' ) # Create our splits _a : Union[str, Any] = StratifiedKFold(n_splits=int(args.num_folds ) ) # Initialize accelerator _a : Any = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _a : Optional[Any] = config['lr'] _a : Optional[int] = int(config['num_epochs'] ) _a : Dict = int(config['seed'] ) _a : Dict = int(config['batch_size'] ) _a : Optional[int] = evaluate.load('glue' , 'mrpc' ) # If the batch size is too big we use gradient accumulation _a : List[Any] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: _a : Any = batch_size // MAX_GPU_BATCH_SIZE _a : List[str] = MAX_GPU_BATCH_SIZE set_seed(__a ) # New Code # # Create our folds: _a : int = kfold.split(np.zeros(datasets['train'].num_rows ) , datasets['train']['label'] ) _a : Any = [] # Iterate over them for i, (train_idxs, valid_idxs) in enumerate(__a ): _a, _a, _a : Optional[Any] = get_fold_dataloaders( __a , __a , __a , __a , ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _a : Dict = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=__a ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _a : List[Any] = model.to(accelerator.device ) # Instantiate optimizer _a : List[str] = AdamW(params=model.parameters() , lr=__a ) # Instantiate scheduler _a : List[Any] = get_linear_schedule_with_warmup( optimizer=__a , num_warmup_steps=1_0_0 , num_training_steps=(len(__a ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _a, _a, _a, _a, _a : Union[str, Any] = accelerator.prepare( __a , __a , __a , __a , __a ) # Now we train the model for epoch in range(__a ): model.train() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _a : Dict = model(**__a ) _a : int = outputs.loss _a : Any = loss / gradient_accumulation_steps accelerator.backward(__a ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Union[str, Any] = model(**__a ) _a : Tuple = outputs.logits.argmax(dim=-1 ) _a, _a : Any = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=__a , references=__a , ) _a : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , __a ) # New Code # # We also run predictions on the test set at the very end _a : Any = [] for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Tuple = model(**__a ) _a : Dict = outputs.logits _a, _a : Optional[int] = accelerator.gather_for_metrics((predictions, batch['labels']) ) fold_predictions.append(predictions.cpu() ) if i == 0: # We need all of the test predictions test_references.append(references.cpu() ) # Use accelerator.print to print only on the main process. test_predictions.append(torch.cat(__a , dim=0 ) ) # We now need to release all our memory and get rid of the current model, optimizer, etc accelerator.free_memory() # New Code # # Finally we check the accuracy of our folded results: _a : Dict = torch.cat(__a , dim=0 ) _a : Any = torch.stack(__a , dim=0 ).sum(dim=0 ).div(int(args.num_folds ) ).argmax(dim=-1 ) _a : str = metric.compute(predictions=__a , references=__a ) accelerator.print('Average test metrics from all folds:' , __a ) def UpperCAmelCase_ (): """simple docstring""" _a : Any = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=__a , default=__a , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) # New Code # parser.add_argument('--num_folds' , type=__a , default=3 , help='The number of splits to perform across the dataset' ) _a : Any = parser.parse_args() _a : int = {'lr': 2e-5, 'num_epochs': 3, 'seed': 4_2, 'batch_size': 1_6} training_function(__a , __a ) if __name__ == "__main__": main()
271
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule SCREAMING_SNAKE_CASE : List[str] = {"""tokenization_wav2vec2_phoneme""": ["""Wav2Vec2PhonemeCTCTokenizer"""]} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys SCREAMING_SNAKE_CASE : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
102
'''simple docstring''' from __future__ import annotations __lowerCAmelCase = [-1_0, -5, 0, 5, 5.1, 1_1, 1_3, 2_1, 3, 4, -2_1, -1_0, -5, -1, 0] __lowerCAmelCase = [-5, 0, 5, 5.1, 1_1, 1_3, 2_1, -1, 4, -1, -1_0, -5, -1, 0, -1] def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Optional[int] = [] _a : int = len(__a ) for i in range(__a ): _a : float = -1 for j in range(i + 1 , __a ): if arr[i] < arr[j]: _a : Any = arr[j] break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Tuple = [] for i, outer in enumerate(__a ): _a : float = -1 for inner in arr[i + 1 :]: if outer < inner: _a : Dict = inner break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : int = len(__a ) _a : list[float] = [] _a : list[float] = [-1] * arr_size for index in reversed(range(__a ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: _a : Dict = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) __lowerCAmelCase = ( """from __main__ import arr, next_greatest_element_slow, """ """next_greatest_element_fast, next_greatest_element""" ) print( """next_greatest_element_slow():""", timeit("""next_greatest_element_slow(arr)""", setup=setup), ) print( """next_greatest_element_fast():""", timeit("""next_greatest_element_fast(arr)""", setup=setup), ) print( """ next_greatest_element():""", timeit("""next_greatest_element(arr)""", setup=setup), )
271
0
"""simple docstring""" import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class a__ : def __init__( self , _a , _a=sys.maxsize ): lowercase : List[str] = 'bilinear' lowercase : List[Any] = max_size lowercase : Dict = short_edge_length def __call__( self , _a ): lowercase : List[str] = [] for img in imgs: lowercase : Union[str, Any] = img.shape[:2] # later: provide list and randomly choose index for resize lowercase : Optional[int] = np.random.randint(self.short_edge_length[0] , self.short_edge_length[1] + 1 ) if size == 0: return img lowercase : Any = size * 1.0 / min(_a , _a ) if h < w: lowercase : Optional[Any] = size, scale * w else: lowercase : Union[str, Any] = scale * h, size if max(_a , _a ) > self.max_size: lowercase : Any = self.max_size * 1.0 / max(_a , _a ) lowercase : Any = newh * scale lowercase : Dict = neww * scale lowercase : str = int(neww + 0.5 ) lowercase : Optional[Any] = int(newh + 0.5 ) if img.dtype == np.uinta: lowercase : Any = Image.fromarray(_a ) lowercase : List[Any] = pil_image.resize((neww, newh) , PILImageResampling.BILINEAR ) lowercase : Tuple = np.asarray(_a ) else: lowercase : Dict = img.permute(2 , 0 , 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw lowercase : Any = nn.functional.interpolate( _a , (newh, neww) , mode=self.interp_method , align_corners=_a ).squeeze(0 ) img_augs.append(_a ) return img_augs class a__ : def __init__( self , _a ): lowercase : Dict = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST] , cfg.INPUT.MAX_SIZE_TEST ) lowercase : Any = cfg.INPUT.FORMAT lowercase : List[Any] = cfg.SIZE_DIVISIBILITY lowercase : str = cfg.PAD_VALUE lowercase : List[str] = cfg.INPUT.MAX_SIZE_TEST lowercase : Union[str, Any] = cfg.MODEL.DEVICE lowercase : Optional[Any] = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) lowercase : Optional[int] = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) lowercase : Optional[Any] = lambda _a : (x - self.pixel_mean) / self.pixel_std def __magic_name__ ( self , _a ): lowercase : List[Any] = tuple(max(_a ) for s in zip(*[img.shape for img in images] ) ) lowercase : List[Any] = [im.shape[-2:] for im in images] lowercase : Dict = [ nn.functional.pad( _a , [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]] , value=self.pad_value , ) for size, im in zip(_a , _a ) ] return torch.stack(_a ), torch.tensor(_a ) def __call__( self , _a , _a=False ): with torch.no_grad(): if not isinstance(_a , _a ): lowercase : Optional[Any] = [images] if single_image: assert len(_a ) == 1 for i in range(len(_a ) ): if isinstance(images[i] , torch.Tensor ): images.insert(_a , images.pop(_a ).to(self.device ).float() ) elif not isinstance(images[i] , torch.Tensor ): images.insert( _a , torch.as_tensor(img_tensorize(images.pop(_a ) , input_format=self.input_format ) ) .to(self.device ) .float() , ) # resize smallest edge lowercase : List[Any] = torch.tensor([im.shape[:2] for im in images] ) lowercase : Any = self.aug(_a ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic lowercase : Any = [self.normalizer(_a ) for x in images] # now pad them to do the following operations lowercase : Any = self.pad(_a ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad lowercase : List[str] = torch.true_divide(_a , _a ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def __magic_name__ ( __snake_case : Tuple , __snake_case : List[Any] ) -> Any: boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def __magic_name__ ( __snake_case : Any , __snake_case : Tuple[int, int] ) -> List[str]: assert torch.isfinite(__a ).all(), "Box tensor contains infinite or NaN!" lowercase : List[str] = box_size tensor[:, 0].clamp_(min=0 , max=__a ) tensor[:, 1].clamp_(min=0 , max=__a ) tensor[:, 2].clamp_(min=0 , max=__a ) tensor[:, 3].clamp_(min=0 , max=__a )
202
'''simple docstring''' import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home __lowerCAmelCase = HUGGINGFACE_HUB_CACHE __lowerCAmelCase = """config.json""" __lowerCAmelCase = """diffusion_pytorch_model.bin""" __lowerCAmelCase = """diffusion_flax_model.msgpack""" __lowerCAmelCase = """model.onnx""" __lowerCAmelCase = """diffusion_pytorch_model.safetensors""" __lowerCAmelCase = """weights.pb""" __lowerCAmelCase = """https://huggingface.co""" __lowerCAmelCase = default_cache_path __lowerCAmelCase = """diffusers_modules""" __lowerCAmelCase = os.getenv("""HF_MODULES_CACHE""", os.path.join(hf_cache_home, """modules""")) __lowerCAmelCase = ["""fp16""", """non-ema"""] __lowerCAmelCase = """.self_attn"""
271
0
"""simple docstring""" def UpperCamelCase_ ( lowerCAmelCase__ : str ) -> Union[str, Any]: """simple docstring""" if not all(x.isalpha() for x in string ): raise ValueError('String must only contain alphabetic characters.' ) lowerCAmelCase_ : Dict = sorted(string.lower() ) return len(__a ) == len(set(__a ) ) if __name__ == "__main__": lowercase__ : Tuple = input("""Enter a string """).strip() lowercase__ : List[str] = is_isogram(input_str) print(f'{input_str} is {"an" if isogram else "not an"} isogram.')
224
'''simple docstring''' import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class UpperCAmelCase__ : """simple docstring""" def __init__( self : int ,_a : Any ,_a : Optional[int]=2 ,_a : Optional[Any]=True ,_a : Dict=False ,_a : Dict=10 ,_a : Any=3 ,_a : str=32 * 8 ,_a : Optional[int]=32 * 8 ,_a : int=4 ,_a : str=64 ,): '''simple docstring''' _a : Dict = parent _a : Union[str, Any] = batch_size _a : Tuple = is_training _a : List[str] = use_auxiliary_loss _a : Optional[Any] = num_queries _a : str = num_channels _a : List[str] = min_size _a : int = max_size _a : Optional[int] = num_labels _a : List[str] = hidden_dim _a : int = hidden_dim def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Tuple = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( _a ) _a : Optional[Any] = torch.ones([self.batch_size, self.min_size, self.max_size] ,device=_a ) _a : Union[str, Any] = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] ,device=_a ) > 0.5 ).float() _a : Tuple = (torch.rand((self.batch_size, self.num_labels) ,device=_a ) > 0.5).long() _a : Dict = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : int = MaskaFormerConfig( hidden_size=self.hidden_dim ,) _a : str = self.num_queries _a : Union[str, Any] = self.num_labels _a : Tuple = [1, 1, 1, 1] _a : Dict = self.num_channels _a : str = 64 _a : Tuple = 128 _a : Optional[Any] = self.hidden_dim _a : Union[str, Any] = self.hidden_dim _a : List[Any] = self.hidden_dim return config def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a, _a, _a, _a, _a : Optional[Any] = self.prepare_config_and_inputs() _a : str = {'pixel_values': pixel_values, 'pixel_mask': pixel_mask} return config, inputs_dict def __lowercase ( self : List[str] ,_a : Optional[Any] ,_a : str ): '''simple docstring''' _a : str = output.encoder_hidden_states _a : Any = output.pixel_decoder_hidden_states _a : Optional[Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) ,config.decoder_layers ) def __lowercase ( self : List[str] ,_a : str ,_a : List[Any] ,_a : Any ,_a : Union[str, Any]=False ): '''simple docstring''' with torch.no_grad(): _a : str = MaskaFormerModel(config=_a ) model.to(_a ) model.eval() _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[Any] = model(_a ,output_hidden_states=_a ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape ,(self.batch_size, self.num_queries, self.hidden_dim) ,) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(_a ,_a ) def __lowercase ( self : Tuple ,_a : List[Any] ,_a : Union[str, Any] ,_a : Tuple ,_a : List[str] ,_a : Any ): '''simple docstring''' _a : int = MaskaFormerForUniversalSegmentation(config=_a ) model.to(_a ) model.eval() def comm_check_on_output(_a : Any ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape ,(self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) ,) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape ,(self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _a : Any = model(pixel_values=_a ,pixel_mask=_a ) _a : Optional[int] = model(_a ) comm_check_on_output(_a ) _a : List[str] = model( pixel_values=_a ,pixel_mask=_a ,mask_labels=_a ,class_labels=_a ) comm_check_on_output(_a ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape ,torch.Size([1] ) ) @require_torch class UpperCAmelCase__ ( lowercase__ , lowercase__ , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[int] = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () __UpperCAmelCase : Dict = {'''feature-extraction''': MaskaFormerModel} if is_torch_available() else {} __UpperCAmelCase : Dict = False __UpperCAmelCase : Tuple = False __UpperCAmelCase : Dict = False __UpperCAmelCase : List[Any] = False def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Union[str, Any] = MaskaFormerModelTester(self ) _a : Dict = ConfigTester(self ,config_class=_a ,has_text_modality=_a ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' self.config_tester.run_common_tests() def __lowercase ( self : Optional[int] ): '''simple docstring''' _a, _a : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*_a ) @unittest.skip(reason='Mask2Former does not use inputs_embeds' ) def __lowercase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not have a get_input_embeddings method' ) def __lowercase ( self : str ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former is not a generative model' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass @unittest.skip(reason='Mask2Former does not use token embeddings' ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip( reason='Mask2Former has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def __lowercase ( self : Dict ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowercase ( self : List[Any] ): '''simple docstring''' pass def __lowercase ( self : int ): '''simple docstring''' _a, _a : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Union[str, Any] = model_class(_a ) _a : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a : Optional[Any] = [*signature.parameters.keys()] _a : List[Any] = ['pixel_values'] self.assertListEqual(arg_names[:1] ,_a ) @slow def __lowercase ( self : List[str] ): '''simple docstring''' for model_name in ["facebook/mask2former-swin-small-coco-instance"]: _a : Dict = MaskaFormerModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def __lowercase ( self : List[Any] ): '''simple docstring''' _a : int = (self.model_tester.min_size,) * 2 _a : Any = { 'pixel_values': torch.randn((2, 3, *size) ,device=_a ), 'mask_labels': torch.randn((2, 10, *size) ,device=_a ), 'class_labels': torch.zeros(2 ,10 ,device=_a ).long(), } _a : List[Any] = self.model_tester.get_config() _a : int = MaskaFormerForUniversalSegmentation(_a ).to(_a ) _a : str = model(**_a ) self.assertTrue(outputs.loss is not None ) def __lowercase ( self : List[str] ): '''simple docstring''' _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a ,**_a ,output_hidden_states=_a ) def __lowercase ( self : int ): '''simple docstring''' _a, _a : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a : Any = model_class(_a ).to(_a ) _a : Optional[int] = model(**_a ,output_attentions=_a ) self.assertTrue(outputs.attentions is not None ) def __lowercase ( self : Tuple ): '''simple docstring''' if not self.model_tester.is_training: return _a : List[str] = self.all_model_classes[1] _a, _a, _a, _a, _a : List[str] = self.model_tester.prepare_config_and_inputs() _a : Any = model_class(_a ) model.to(_a ) model.train() _a : Union[str, Any] = model(_a ,mask_labels=_a ,class_labels=_a ).loss loss.backward() def __lowercase ( self : int ): '''simple docstring''' _a : int = self.all_model_classes[1] _a, _a, _a, _a, _a : List[Any] = self.model_tester.prepare_config_and_inputs() _a : str = True _a : str = True _a : List[str] = model_class(_a ).to(_a ) model.train() _a : Optional[int] = model(_a ,mask_labels=_a ,class_labels=_a ) _a : Tuple = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _a : str = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() _a : Dict = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _a : List[str] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_a ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) __lowerCAmelCase = 1e-4 def UpperCAmelCase_ (): """simple docstring""" _a : int = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_vision @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' return "facebook/mask2former-swin-small-coco-instance" @cached_property def __lowercase ( self : Any ): '''simple docstring''' return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def __lowercase ( self : Any ): '''simple docstring''' _a : List[str] = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(_a ) _a : int = self.default_image_processor _a : Tuple = prepare_img() _a : Any = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Union[str, Any] = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[Any] = model(**_a ) _a : List[Any] = torch.tensor( [[-0.2790, -1.0717, -1.1668], [-0.5128, -0.3128, -0.4987], [-0.5832, 0.1971, -0.0197]] ).to(_a ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : str = torch.tensor( [[0.8973, 1.1847, 1.1776], [1.1934, 1.5040, 1.5128], [1.1153, 1.4486, 1.4951]] ).to(_a ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] ,_a ,atol=_a ) ) _a : Any = torch.tensor( [[2.1152, 1.7000, -0.8603], [1.5808, 1.8004, -0.9353], [1.6043, 1.7495, -0.5999]] ).to(_a ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Optional[Any] = self.default_image_processor _a : List[Any] = prepare_img() _a : str = image_processor(_a ,return_tensors='pt' ).to(_a ) _a : Any = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a ,(1, 3, 384, 384) ) with torch.no_grad(): _a : Optional[int] = model(**_a ) # masks_queries_logits _a : Dict = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape ,(1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) _a : Dict = [ [-8.7839, -9.0056, -8.8121], [-7.4104, -7.0313, -6.5401], [-6.6105, -6.3427, -6.4675], ] _a : Optional[Any] = torch.tensor(_a ).to(_a ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] ,_a ,atol=_a ) ) # class_queries_logits _a : str = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape ,(1, model.config.num_queries, model.config.num_labels + 1) ) _a : str = torch.tensor( [ [1.8324, -8.0835, -4.1922], [0.8450, -9.0050, -3.6053], [0.3045, -7.7293, -3.0275], ] ).to(_a ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] ,_a ,atol=_a ) ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : Any = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() _a : Tuple = self.default_image_processor _a : Tuple = image_processor( [np.zeros((3, 800, 1333) ), np.zeros((3, 800, 1333) )] ,segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] ,return_tensors='pt' ,) _a : str = inputs['pixel_values'].to(_a ) _a : str = [el.to(_a ) for el in inputs['mask_labels']] _a : Dict = [el.to(_a ) for el in inputs['class_labels']] with torch.no_grad(): _a : List[str] = model(**_a ) self.assertTrue(outputs.loss is not None )
271
0
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor _snake_case = logging.get_logger(__name__) class UpperCAmelCase_ ( lowercase__ ): '''simple docstring''' def __init__( self , *__A , **__A ): """simple docstring""" warnings.warn( "The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use LayoutLMv2ImageProcessor instead." , _a , ) super().__init__(*_a , **_a )
283
'''simple docstring''' import argparse import json from typing import List from ltp import LTP from transformers import BertTokenizer def UpperCAmelCase_ (__a : List[Any] ): """simple docstring""" if ( (cp >= 0x4E_00 and cp <= 0x9F_FF) or (cp >= 0x34_00 and cp <= 0x4D_BF) # or (cp >= 0x2_00_00 and cp <= 0x2_A6_DF) # or (cp >= 0x2_A7_00 and cp <= 0x2_B7_3F) # or (cp >= 0x2_B7_40 and cp <= 0x2_B8_1F) # or (cp >= 0x2_B8_20 and cp <= 0x2_CE_AF) # or (cp >= 0xF9_00 and cp <= 0xFA_FF) or (cp >= 0x2_F8_00 and cp <= 0x2_FA_1F) # ): # return True return False def UpperCAmelCase_ (__a : str ): """simple docstring""" for char in word: _a : Union[str, Any] = ord(__a ) if not _is_chinese_char(__a ): return 0 return 1 def UpperCAmelCase_ (__a : List[str] ): """simple docstring""" _a : Dict = set() for token in tokens: _a : str = len(__a ) > 1 and is_chinese(__a ) if chinese_word: word_set.add(__a ) _a : Optional[Any] = list(__a ) return word_list def UpperCAmelCase_ (__a : List[str] , __a : set() ): """simple docstring""" if not chinese_word_set: return bert_tokens _a : Optional[Any] = max([len(__a ) for w in chinese_word_set] ) _a : Optional[int] = bert_tokens _a, _a : Any = 0, len(__a ) while start < end: _a : Tuple = True if is_chinese(bert_word[start] ): _a : Union[str, Any] = min(end - start , __a ) for i in range(__a , 1 , -1 ): _a : Optional[Any] = ''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): _a : Any = '##' + bert_word[j] _a : Union[str, Any] = start + i _a : int = False break if single_word: start += 1 return bert_word def UpperCAmelCase_ (__a : List[str] , __a : LTP , __a : BertTokenizer ): """simple docstring""" _a : int = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : Union[str, Any] = ltp_tokenizer.seg(lines[i : i + 1_0_0] )[0] _a : Optional[Any] = [get_chinese_word(__a ) for r in res] ltp_res.extend(__a ) assert len(__a ) == len(__a ) _a : str = [] for i in range(0 , len(__a ) , 1_0_0 ): _a : List[str] = bert_tokenizer(lines[i : i + 1_0_0] , add_special_tokens=__a , truncation=__a , max_length=5_1_2 ) bert_res.extend(res['input_ids'] ) assert len(__a ) == len(__a ) _a : List[str] = [] for input_ids, chinese_word in zip(__a , __a ): _a : int = [] for id in input_ids: _a : Optional[int] = bert_tokenizer._convert_id_to_token(__a ) input_tokens.append(__a ) _a : List[str] = add_sub_symbol(__a , __a ) _a : Tuple = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(__a ): if token[:2] == "##": _a : str = token[2:] # save chinese tokens' pos if len(__a ) == 1 and _is_chinese_char(ord(__a ) ): ref_id.append(__a ) ref_ids.append(__a ) assert len(__a ) == len(__a ) return ref_ids def UpperCAmelCase_ (__a : Optional[Any] ): """simple docstring""" with open(args.file_name , 'r' , encoding='utf-8' ) as f: _a : Dict = f.readlines() _a : int = [line.strip() for line in data if len(__a ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' _a : int = LTP(args.ltp ) # faster in GPU device _a : Tuple = BertTokenizer.from_pretrained(args.bert ) _a : int = prepare_ref(__a , __a , __a ) with open(args.save_path , 'w' , encoding='utf-8' ) as f: _a : Optional[Any] = [json.dumps(__a ) + '\n' for ref in ref_ids] f.writelines(__a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser(description="""prepare_chinese_ref""") parser.add_argument( """--file_name""", type=str, default="""./resources/chinese-demo.txt""", help="""file need process, same as training data in lm""", ) parser.add_argument( """--ltp""", type=str, default="""./resources/ltp""", help="""resources for LTP tokenizer, usually a path""" ) parser.add_argument("""--bert""", type=str, default="""./resources/robert""", help="""resources for Bert tokenizer""") parser.add_argument("""--save_path""", type=str, default="""./resources/ref.txt""", help="""path to save res""") __lowerCAmelCase = parser.parse_args() main(args)
271
0
from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_outputs import ( BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import logging from .configuration_regnet import RegNetConfig _A = logging.get_logger(__name__) # General docstring _A = '''RegNetConfig''' # Base docstring _A = '''facebook/regnet-y-040''' _A = [1, 1_088, 7, 7] # Image classification docstring _A = '''facebook/regnet-y-040''' _A = '''tabby, tabby cat''' _A = [ '''facebook/regnet-y-040''', # See all regnet models at https://huggingface.co/models?filter=regnet ] class lowercase_ ( nn.Module ): def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 3 , __UpperCamelCase = 1 , __UpperCamelCase = 1 , __UpperCamelCase = "relu" , ): """simple docstring""" super().__init__() UpperCamelCase_ = nn.Convad( _a , _a , kernel_size=_a , stride=_a , padding=kernel_size // 2 , groups=_a , bias=_a , ) UpperCamelCase_ = nn.BatchNormad(_a ) UpperCamelCase_ = ACTaFN[activation] if activation is not None else nn.Identity() def lowerCamelCase_ ( self , __UpperCamelCase ): """simple docstring""" UpperCamelCase_ = self.convolution(_a ) UpperCamelCase_ = self.normalization(_a ) UpperCamelCase_ = self.activation(_a ) return hidden_state class lowercase_ ( nn.Module ): def __init__( self , __UpperCamelCase ): """simple docstring""" super().__init__() UpperCamelCase_ = RegNetConvLayer( config.num_channels , config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act ) UpperCamelCase_ = config.num_channels def lowerCamelCase_ ( self , __UpperCamelCase ): """simple docstring""" UpperCamelCase_ = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( """Make sure that the channel dimension of the pixel values match with the one set in the configuration.""" ) UpperCamelCase_ = self.embedder(_a ) return hidden_state class lowercase_ ( nn.Module ): def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 2 ): """simple docstring""" super().__init__() UpperCamelCase_ = nn.Convad(_a , _a , kernel_size=1 , stride=_a , bias=_a ) UpperCamelCase_ = nn.BatchNormad(_a ) def lowerCamelCase_ ( self , __UpperCamelCase ): """simple docstring""" UpperCamelCase_ = self.convolution(_a ) UpperCamelCase_ = self.normalization(_a ) return hidden_state class lowercase_ ( nn.Module ): def __init__( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" super().__init__() UpperCamelCase_ = nn.AdaptiveAvgPoolad((1, 1) ) UpperCamelCase_ = nn.Sequential( nn.Convad(_a , _a , kernel_size=1 ) , nn.ReLU() , nn.Convad(_a , _a , kernel_size=1 ) , nn.Sigmoid() , ) def lowerCamelCase_ ( self , __UpperCamelCase ): """simple docstring""" UpperCamelCase_ = self.pooler(_a ) UpperCamelCase_ = self.attention(_a ) UpperCamelCase_ = hidden_state * attention return hidden_state class lowercase_ ( nn.Module ): def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 1 ): """simple docstring""" super().__init__() UpperCamelCase_ = in_channels != out_channels or stride != 1 UpperCamelCase_ = max(1 , out_channels // config.groups_width ) UpperCamelCase_ = ( RegNetShortCut(_a , _a , stride=_a ) if should_apply_shortcut else nn.Identity() ) UpperCamelCase_ = nn.Sequential( RegNetConvLayer(_a , _a , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(_a , _a , stride=_a , groups=_a , activation=config.hidden_act ) , RegNetConvLayer(_a , _a , kernel_size=1 , activation=_a ) , ) UpperCamelCase_ = ACTaFN[config.hidden_act] def lowerCamelCase_ ( self , __UpperCamelCase ): """simple docstring""" UpperCamelCase_ = hidden_state UpperCamelCase_ = self.layer(_a ) UpperCamelCase_ = self.shortcut(_a ) hidden_state += residual UpperCamelCase_ = self.activation(_a ) return hidden_state class lowercase_ ( nn.Module ): def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 1 ): """simple docstring""" super().__init__() UpperCamelCase_ = in_channels != out_channels or stride != 1 UpperCamelCase_ = max(1 , out_channels // config.groups_width ) UpperCamelCase_ = ( RegNetShortCut(_a , _a , stride=_a ) if should_apply_shortcut else nn.Identity() ) UpperCamelCase_ = nn.Sequential( RegNetConvLayer(_a , _a , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(_a , _a , stride=_a , groups=_a , activation=config.hidden_act ) , RegNetSELayer(_a , reduced_channels=int(round(in_channels / 4 ) ) ) , RegNetConvLayer(_a , _a , kernel_size=1 , activation=_a ) , ) UpperCamelCase_ = ACTaFN[config.hidden_act] def lowerCamelCase_ ( self , __UpperCamelCase ): """simple docstring""" UpperCamelCase_ = hidden_state UpperCamelCase_ = self.layer(_a ) UpperCamelCase_ = self.shortcut(_a ) hidden_state += residual UpperCamelCase_ = self.activation(_a ) return hidden_state class lowercase_ ( nn.Module ): def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 2 , __UpperCamelCase = 2 , ): """simple docstring""" super().__init__() UpperCamelCase_ = RegNetXLayer if config.layer_type == 'x' else RegNetYLayer UpperCamelCase_ = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer( _a , _a , _a , stride=_a , ) , *[layer(_a , _a , _a ) for _ in range(depth - 1 )] , ) def lowerCamelCase_ ( self , __UpperCamelCase ): """simple docstring""" UpperCamelCase_ = self.layers(_a ) return hidden_state class lowercase_ ( nn.Module ): def __init__( self , __UpperCamelCase ): """simple docstring""" super().__init__() UpperCamelCase_ = nn.ModuleList([] ) # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( RegNetStage( _a , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , ) ) UpperCamelCase_ = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(_a , config.depths[1:] ): self.stages.append(RegNetStage(_a , _a , _a , depth=_a ) ) def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase = False , __UpperCamelCase = True ): """simple docstring""" UpperCamelCase_ = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: UpperCamelCase_ = hidden_states + (hidden_state,) UpperCamelCase_ = stage_module(_a ) if output_hidden_states: UpperCamelCase_ = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=_a , hidden_states=_a ) class lowercase_ ( lowercase__ ): A__ : int = RegNetConfig A__ : Any = '''regnet''' A__ : Optional[Any] = '''pixel_values''' A__ : List[str] = True def lowerCamelCase_ ( self , __UpperCamelCase ): """simple docstring""" if isinstance(_a , nn.Convad ): nn.init.kaiming_normal_(module.weight , mode="""fan_out""" , nonlinearity="""relu""" ) elif isinstance(_a , (nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight , 1 ) nn.init.constant_(module.bias , 0 ) def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase=False ): """simple docstring""" if isinstance(_a , _a ): UpperCamelCase_ = value _A = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' _A = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConvNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~file_utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( """The bare RegNet model outputting raw features without any specific head on top.""" , lowercase__ , ) # Copied from transformers.models.resnet.modeling_resnet.ResNetModel with RESNET->REGNET,ResNet->RegNet class lowercase_ ( lowercase__ ): def __init__( self , __UpperCamelCase ): """simple docstring""" super().__init__(_a ) UpperCamelCase_ = config UpperCamelCase_ = RegNetEmbeddings(_a ) UpperCamelCase_ = RegNetEncoder(_a ) UpperCamelCase_ = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_a ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=_a , config_class=_CONFIG_FOR_DOC , modality="""vision""" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None ): """simple docstring""" UpperCamelCase_ = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) UpperCamelCase_ = return_dict if return_dict is not None else self.config.use_return_dict UpperCamelCase_ = self.embedder(_a ) UpperCamelCase_ = self.encoder( _a , output_hidden_states=_a , return_dict=_a ) UpperCamelCase_ = encoder_outputs[0] UpperCamelCase_ = self.pooler(_a ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=_a , pooler_output=_a , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( """ RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. """ , lowercase__ , ) # Copied from transformers.models.resnet.modeling_resnet.ResNetForImageClassification with RESNET->REGNET,ResNet->RegNet,resnet->regnet class lowercase_ ( lowercase__ ): def __init__( self , __UpperCamelCase ): """simple docstring""" super().__init__(_a ) UpperCamelCase_ = config.num_labels UpperCamelCase_ = RegNetModel(_a ) # classification head UpperCamelCase_ = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_a ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=_a , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def lowerCamelCase_ ( self , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , ): """simple docstring""" UpperCamelCase_ = return_dict if return_dict is not None else self.config.use_return_dict UpperCamelCase_ = self.regnet(_a , output_hidden_states=_a , return_dict=_a ) UpperCamelCase_ = outputs.pooler_output if return_dict else outputs[1] UpperCamelCase_ = self.classifier(_a ) UpperCamelCase_ = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: UpperCamelCase_ = 'regression' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): UpperCamelCase_ = 'single_label_classification' else: UpperCamelCase_ = 'multi_label_classification' if self.config.problem_type == "regression": UpperCamelCase_ = MSELoss() if self.num_labels == 1: UpperCamelCase_ = loss_fct(logits.squeeze() , labels.squeeze() ) else: UpperCamelCase_ = loss_fct(_a , _a ) elif self.config.problem_type == "single_label_classification": UpperCamelCase_ = CrossEntropyLoss() UpperCamelCase_ = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": UpperCamelCase_ = BCEWithLogitsLoss() UpperCamelCase_ = loss_fct(_a , _a ) if not return_dict: UpperCamelCase_ = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=_a , logits=_a , hidden_states=outputs.hidden_states )
122
'''simple docstring''' import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Tuple ,*_a : List[str] ,**_a : Any ): '''simple docstring''' warnings.warn( 'The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use VideoMAEImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
'''simple docstring''' import argparse import torch from transformers import RemBertConfig, RemBertModel, load_tf_weights_in_rembert from transformers.utils import logging logging.set_verbosity_info() def snake_case_ (_a : Tuple , _a : List[str] , _a : Union[str, Any] ): UpperCAmelCase = RemBertConfig.from_json_file(__a ) print('''Building PyTorch model from configuration: {}'''.format(str(__a ) ) ) UpperCAmelCase = RemBertModel(__a ) # Load weights from tf checkpoint load_tf_weights_in_rembert(__a , __a , __a ) # Save pytorch-model print('''Save PyTorch model to {}'''.format(__a ) ) torch.save(model.state_dict() , __a ) if __name__ == "__main__": A =argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--rembert_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained RemBERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) A =parser.parse_args() convert_rembert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.rembert_config_file, args.pytorch_dump_path)
34
'''simple docstring''' from __future__ import annotations from random import choice def UpperCAmelCase_ (__a : str ): """simple docstring""" return choice(__a ) def UpperCAmelCase_ (__a : list[int] , __a : int ): """simple docstring""" _a : Dict = random_pivot(__a ) # partition based on pivot # linear time _a : Optional[int] = [e for e in lst if e < pivot] _a : List[str] = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(__a ) == k - 1: return pivot # pivot is in elements bigger than k elif len(__a ) < k - 1: return kth_number(__a , k - len(__a ) - 1 ) # pivot is in elements smaller than k else: return kth_number(__a , __a ) if __name__ == "__main__": import doctest doctest.testmod()
271
0
import tempfile import torch from diffusers import IPNDMScheduler from .test_schedulers import SchedulerCommonTest class __SCREAMING_SNAKE_CASE( lowercase__ ): _UpperCAmelCase = (IPNDMScheduler,) _UpperCAmelCase = (('''num_inference_steps''', 5_0),) def lowerCAmelCase_ ( self: Any , **UpperCamelCase: Union[str, Any] ) -> Any: snake_case__ = {'num_train_timesteps': 10_00} config.update(**_a ) return config def lowerCAmelCase_ ( self: Dict , UpperCamelCase: Dict=0 , **UpperCamelCase: List[str] ) -> Tuple: snake_case__ = dict(self.forward_default_kwargs ) snake_case__ = kwargs.pop('num_inference_steps' , _a ) snake_case__ = self.dummy_sample snake_case__ = 0.1 * sample snake_case__ = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: snake_case__ = self.get_scheduler_config(**_a ) snake_case__ = scheduler_class(**_a ) scheduler.set_timesteps(_a ) # copy over dummy past residuals snake_case__ = dummy_past_residuals[:] if time_step is None: snake_case__ = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_a ) snake_case__ = scheduler_class.from_pretrained(_a ) new_scheduler.set_timesteps(_a ) # copy over dummy past residuals snake_case__ = dummy_past_residuals[:] snake_case__ = scheduler.step(_a , _a , _a , **_a ).prev_sample snake_case__ = new_scheduler.step(_a , _a , _a , **_a ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" snake_case__ = scheduler.step(_a , _a , _a , **_a ).prev_sample snake_case__ = new_scheduler.step(_a , _a , _a , **_a ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def lowerCAmelCase_ ( self: Optional[Any] ) -> Dict: pass def lowerCAmelCase_ ( self: List[Any] , UpperCamelCase: Tuple=0 , **UpperCamelCase: Dict ) -> Optional[int]: snake_case__ = dict(self.forward_default_kwargs ) snake_case__ = kwargs.pop('num_inference_steps' , _a ) snake_case__ = self.dummy_sample snake_case__ = 0.1 * sample snake_case__ = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) scheduler.set_timesteps(_a ) # copy over dummy past residuals (must be after setting timesteps) snake_case__ = dummy_past_residuals[:] if time_step is None: snake_case__ = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_a ) snake_case__ = scheduler_class.from_pretrained(_a ) # copy over dummy past residuals new_scheduler.set_timesteps(_a ) # copy over dummy past residual (must be after setting timesteps) snake_case__ = dummy_past_residuals[:] snake_case__ = scheduler.step(_a , _a , _a , **_a ).prev_sample snake_case__ = new_scheduler.step(_a , _a , _a , **_a ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" snake_case__ = scheduler.step(_a , _a , _a , **_a ).prev_sample snake_case__ = new_scheduler.step(_a , _a , _a , **_a ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def lowerCAmelCase_ ( self: Dict , **UpperCamelCase: List[Any] ) -> Dict: snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config(**_a ) snake_case__ = scheduler_class(**_a ) snake_case__ = 10 snake_case__ = self.dummy_model() snake_case__ = self.dummy_sample_deter scheduler.set_timesteps(_a ) for i, t in enumerate(scheduler.timesteps ): snake_case__ = model(_a , _a ) snake_case__ = scheduler.step(_a , _a , _a ).prev_sample for i, t in enumerate(scheduler.timesteps ): snake_case__ = model(_a , _a ) snake_case__ = scheduler.step(_a , _a , _a ).prev_sample return sample def lowerCAmelCase_ ( self: Optional[Any] ) -> Tuple: snake_case__ = dict(self.forward_default_kwargs ) snake_case__ = kwargs.pop('num_inference_steps' , _a ) for scheduler_class in self.scheduler_classes: snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = self.dummy_sample snake_case__ = 0.1 * sample if num_inference_steps is not None and hasattr(_a , 'set_timesteps' ): scheduler.set_timesteps(_a ) elif num_inference_steps is not None and not hasattr(_a , 'set_timesteps' ): snake_case__ = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) snake_case__ = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] snake_case__ = dummy_past_residuals[:] snake_case__ = scheduler.timesteps[5] snake_case__ = scheduler.timesteps[6] snake_case__ = scheduler.step(_a , _a , _a , **_a ).prev_sample snake_case__ = scheduler.step(_a , _a , _a , **_a ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) snake_case__ = scheduler.step(_a , _a , _a , **_a ).prev_sample snake_case__ = scheduler.step(_a , _a , _a , **_a ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def lowerCAmelCase_ ( self: int ) -> Union[str, Any]: for timesteps in [1_00, 10_00]: self.check_over_configs(num_train_timesteps=_a , time_step=_a ) def lowerCAmelCase_ ( self: Optional[int] ) -> List[Any]: for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 1_00] ): self.check_over_forward(num_inference_steps=_a , time_step=_a ) def lowerCAmelCase_ ( self: Dict ) -> Optional[Any]: snake_case__ = self.full_loop() snake_case__ = torch.mean(torch.abs(_a ) ) assert abs(result_mean.item() - 2_54_05_29 ) < 10
307
'''simple docstring''' class UpperCAmelCase__ : """simple docstring""" def __init__( self : Dict ): '''simple docstring''' _a : Dict = {} def __lowercase ( self : Union[str, Any] ): '''simple docstring''' print(self.vertex ) for i in self.vertex: print(_a ,' -> ' ,' -> '.join([str(_a ) for j in self.vertex[i]] ) ) def __lowercase ( self : Dict ,_a : int ,_a : int ): '''simple docstring''' if from_vertex in self.vertex: self.vertex[from_vertex].append(_a ) else: # else make a new vertex _a : int = [to_vertex] def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Tuple = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(_a ,_a ) def __lowercase ( self : Union[str, Any] ,_a : int ,_a : list ): '''simple docstring''' _a : List[Any] = True print(_a ,end=' ' ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(_a ,_a ) if __name__ == "__main__": __lowerCAmelCase = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print("""DFS:""") g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
271
0
from pathlib import PurePosixPath from typing import Optional import fsspec from fsspec import AbstractFileSystem from huggingface_hub.hf_api import DatasetInfo from ..utils.file_utils import get_authentication_headers_for_url from ..utils.hub import hf_hub_url class _a ( lowercase__ ): A = '''''' A = '''hf-legacy''' # "hf://"" is reserved for hffs def __init__(self, SCREAMING_SNAKE_CASE_ = None, SCREAMING_SNAKE_CASE_ = None, **SCREAMING_SNAKE_CASE_, ) -> Any: super().__init__(self, **_a ) UpperCAmelCase_: Dict = repo_info UpperCAmelCase_: Tuple = token UpperCAmelCase_: Any = None def __snake_case (self ) -> int: if self.dir_cache is None: UpperCAmelCase_: int = {} for hf_file in self.repo_info.siblings: # TODO(QL): add sizes UpperCAmelCase_: str = { 'name': hf_file.rfilename, 'size': None, 'type': 'file', } self.dir_cache.update( { str(_a ): {"""name""": str(_a ), """size""": None, """type""": """directory"""} for d in list(PurePosixPath(hf_file.rfilename ).parents )[:-1] } ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ = "rb", **SCREAMING_SNAKE_CASE_, ) -> Any: if not isinstance(self.repo_info, _a ): raise NotImplementedError(f'Open is only implemented for dataset repositories, but got {self.repo_info}' ) UpperCAmelCase_: Dict = hf_hub_url(self.repo_info.id, _a, revision=self.repo_info.sha ) return fsspec.open( _a, mode=_a, headers=get_authentication_headers_for_url(_a, use_auth_token=self.token ), client_kwargs={"""trust_env""": True}, ).open() def __snake_case (self, SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) -> Tuple: self._get_dirs() UpperCAmelCase_: Any = self._strip_protocol(_a ) if path in self.dir_cache: return self.dir_cache[path] else: raise FileNotFoundError(_a ) def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_=False, **SCREAMING_SNAKE_CASE_ ) -> Optional[Any]: self._get_dirs() UpperCAmelCase_: List[str] = PurePosixPath(path.strip("""/""" ) ) UpperCAmelCase_: Dict = {} for p, f in self.dir_cache.items(): UpperCAmelCase_: int = PurePosixPath(p.strip("""/""" ) ) UpperCAmelCase_: Union[str, Any] = p.parent if root == path: UpperCAmelCase_: str = f UpperCAmelCase_: str = list(paths.values() ) if detail: return out else: return sorted(f["""name"""] for f in out )
147
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = """▁""" __lowerCAmelCase = {"""vocab_file""": """sentencepiece.bpe.model""", """monolingual_vocab_file""": """dict.txt"""} __lowerCAmelCase = { """vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model""", }, """monolingual_vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt""", }, } __lowerCAmelCase = {"""vinai/bartpho-syllable""": 1_0_2_4} class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Optional[Any] = VOCAB_FILES_NAMES __UpperCAmelCase : Dict = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ['''input_ids''', '''attention_mask'''] def __init__( self : str ,_a : str ,_a : Any ,_a : Any="<s>" ,_a : Dict="</s>" ,_a : int="</s>" ,_a : Union[str, Any]="<s>" ,_a : List[Any]="<unk>" ,_a : Optional[Any]="<pad>" ,_a : List[str]="<mask>" ,_a : Optional[Dict[str, Any]] = None ,**_a : int ,): '''simple docstring''' _a : Any = AddedToken(_a ,lstrip=_a ,rstrip=_a ) if isinstance(_a ,_a ) else mask_token _a : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_a ,eos_token=_a ,unk_token=_a ,sep_token=_a ,cls_token=_a ,pad_token=_a ,mask_token=_a ,sp_model_kwargs=self.sp_model_kwargs ,**_a ,) _a : Optional[int] = vocab_file _a : Union[str, Any] = monolingual_vocab_file _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_a ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility _a : Union[str, Any] = {} _a : int = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(_a ) not in self.fairseq_tokens_to_ids: _a : int = cnt cnt += 1 with open(_a ,'r' ,encoding='utf-8' ) as f: for line in f.readlines(): _a : str = line.strip().split()[0] _a : Tuple = len(self.fairseq_tokens_to_ids ) if str(_a ) not in self.fairseq_tokens_to_ids: _a : List[str] = len(self.fairseq_tokens_to_ids ) _a : Tuple = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Union[str, Any] ): '''simple docstring''' _a : int = self.__dict__.copy() _a : str = None _a : Optional[Any] = self.sp_model.serialized_model_proto() return state def __setstate__( self : Tuple ,_a : Tuple ): '''simple docstring''' _a : Tuple = d # for backward compatibility if not hasattr(self ,'sp_model_kwargs' ): _a : List[str] = {} _a : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def __lowercase ( self : Dict ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _a : Dict = [self.cls_token_id] _a : int = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __lowercase ( self : int ,_a : List[int] ,_a : Optional[List[int]] = None ,_a : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_a ,token_ids_a=_a ,already_has_special_tokens=_a ) if token_ids_a is None: return [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1, 1] + ([0] * len(_a )) + [1] def __lowercase ( self : Tuple ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' _a : List[str] = [self.sep_token_id] _a : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __lowercase ( self : Dict ): '''simple docstring''' return len(self.fairseq_ids_to_tokens ) def __lowercase ( self : Dict ): '''simple docstring''' _a : List[str] = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __lowercase ( self : Tuple ,_a : str ): '''simple docstring''' return self.sp_model.encode(_a ,out_type=_a ) def __lowercase ( self : Union[str, Any] ,_a : Union[str, Any] ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def __lowercase ( self : Any ,_a : int ): '''simple docstring''' return self.fairseq_ids_to_tokens[index] def __lowercase ( self : Tuple ,_a : Union[str, Any] ): '''simple docstring''' _a : str = ''.join(_a ).replace(_a ,' ' ).strip() return out_string def __lowercase ( self : Union[str, Any] ,_a : str ,_a : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(_a ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _a : int = os.path.join( _a ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['monolingual_vocab_file'] ,) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,_a ) elif not os.path.isfile(self.vocab_file ): with open(_a ,'wb' ) as fi: _a : List[Any] = self.sp_model.serialized_model_proto() fi.write(_a ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( _a ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file ,_a ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(_a ,'w' ,encoding='utf-8' ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(F"""{str(_a )} \n""" ) return out_vocab_file, out_monolingual_vocab_file
271
0
import os from argparse import ArgumentParser, Namespace from ..data import SingleSentenceClassificationProcessor as Processor from ..pipelines import TextClassificationPipeline from ..utils import is_tf_available, is_torch_available, logging from . import BaseTransformersCLICommand if not is_tf_available() and not is_torch_available(): raise RuntimeError('At least one of PyTorch or TensorFlow 2.0+ should be installed to use CLI training') # TF training parameters _snake_case : Optional[int] = False _snake_case : Union[str, Any] = False def a_ ( lowerCAmelCase_ : Namespace ): return TrainCommand(__a ) class _UpperCAmelCase ( lowercase__ ): """simple docstring""" @staticmethod def lowercase ( lowerCAmelCase_ : ArgumentParser ) -> Optional[int]: __lowerCAmelCase = parser.add_parser('train' , help='CLI tool to train a model on a task.' ) train_parser.add_argument( '--train_data' , type=_a , required=_a , help='path to train (and optionally evaluation) dataset as a csv with tab separated labels and sentences.' , ) train_parser.add_argument( '--column_label' , type=_a , default=0 , help='Column of the dataset csv file with example labels.' ) train_parser.add_argument( '--column_text' , type=_a , default=1 , help='Column of the dataset csv file with example texts.' ) train_parser.add_argument( '--column_id' , type=_a , default=2 , help='Column of the dataset csv file with example ids.' ) train_parser.add_argument( '--skip_first_row' , action='store_true' , help='Skip the first row of the csv file (headers).' ) train_parser.add_argument('--validation_data' , type=_a , default='' , help='path to validation dataset.' ) train_parser.add_argument( '--validation_split' , type=_a , default=0.1 , help='if validation dataset is not provided, fraction of train dataset to use as validation dataset.' , ) train_parser.add_argument('--output' , type=_a , default='./' , help='path to saved the trained model.' ) train_parser.add_argument( '--task' , type=_a , default='text_classification' , help='Task to train the model on.' ) train_parser.add_argument( '--model' , type=_a , default='bert-base-uncased' , help='Model\'s name or path to stored model.' ) train_parser.add_argument('--train_batch_size' , type=_a , default=3_2 , help='Batch size for training.' ) train_parser.add_argument('--valid_batch_size' , type=_a , default=6_4 , help='Batch size for validation.' ) train_parser.add_argument('--learning_rate' , type=_a , default=3e-5 , help='Learning rate.' ) train_parser.add_argument('--adam_epsilon' , type=_a , default=1e-08 , help='Epsilon for Adam optimizer.' ) train_parser.set_defaults(func=_a ) def __init__( self : int , lowerCAmelCase_ : Namespace ) -> Union[str, Any]: __lowerCAmelCase = logging.get_logger('transformers-cli/training' ) __lowerCAmelCase = 'tf' if is_tf_available() else 'torch' os.makedirs(args.output , exist_ok=_a ) __lowerCAmelCase = args.output __lowerCAmelCase = args.column_label __lowerCAmelCase = args.column_text __lowerCAmelCase = args.column_id self.logger.info(f"""Loading {args.task} pipeline for {args.model}""" ) if args.task == "text_classification": __lowerCAmelCase = TextClassificationPipeline.from_pretrained(args.model ) elif args.task == "token_classification": raise NotImplementedError elif args.task == "question_answering": raise NotImplementedError self.logger.info(f"""Loading dataset from {args.train_data}""" ) __lowerCAmelCase = Processor.create_from_csv( args.train_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) __lowerCAmelCase = None if args.validation_data: self.logger.info(f"""Loading validation dataset from {args.validation_data}""" ) __lowerCAmelCase = Processor.create_from_csv( args.validation_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) __lowerCAmelCase = args.validation_split __lowerCAmelCase = args.train_batch_size __lowerCAmelCase = args.valid_batch_size __lowerCAmelCase = args.learning_rate __lowerCAmelCase = args.adam_epsilon def lowercase ( self : Any ) -> Tuple: if self.framework == "tf": return self.run_tf() return self.run_torch() def lowercase ( self : Tuple ) -> Any: raise NotImplementedError def lowercase ( self : List[str] ) -> Tuple: self.pipeline.fit( self.train_dataset , validation_data=self.valid_dataset , validation_split=self.validation_split , learning_rate=self.learning_rate , adam_epsilon=self.adam_epsilon , train_batch_size=self.train_batch_size , valid_batch_size=self.valid_batch_size , ) # Save trained pipeline self.pipeline.save_pretrained(self.output )
284
'''simple docstring''' import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = None __UpperCAmelCase : List[Any] = None @property def __lowercase ( self : Dict ): '''simple docstring''' return self.feat_extract_tester.prepare_feat_extract_dict() def __lowercase ( self : str ): '''simple docstring''' _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(_a ,'feature_size' ) ) self.assertTrue(hasattr(_a ,'sampling_rate' ) ) self.assertTrue(hasattr(_a ,'padding_value' ) ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_tester.prepare_inputs_for_common() _a : str = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(_a ) == len(_a ) for x, y in zip(_a ,processed_features[input_name] ) ) ) _a : Any = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Union[str, Any] = BatchFeature({input_name: speech_inputs} ,tensor_type='np' ) _a : Union[str, Any] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[int] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_torch def __lowercase ( self : Any ): '''simple docstring''' _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : int = feat_extract.model_input_names[0] _a : str = BatchFeature({input_name: speech_inputs} ,tensor_type='pt' ) _a : str = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : str = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : int = self.feat_extract_tester.prepare_inputs_for_common(equal_length=_a ) _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = feat_extract.model_input_names[0] _a : int = BatchFeature({input_name: speech_inputs} ,tensor_type='tf' ) _a : Optional[int] = processed_features[input_name] if len(batch_features_input.shape ) < 3: _a : Optional[Any] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.feature_size) ) def __lowercase ( self : Dict ,_a : Any=False ): '''simple docstring''' def _inputs_have_equal_length(_a : Tuple ): _a : Tuple = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : Optional[Any] ,_a : Union[str, Any] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : int = self.feature_extraction_class(**self.feat_extract_dict ) _a : Tuple = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Tuple = BatchFeature({input_name: speech_inputs} ) _a : str = self.feat_extract_tester.seq_length_diff _a : Dict = self.feat_extract_tester.max_seq_length + pad_diff _a : Dict = self.feat_extract_tester.min_seq_length _a : Optional[Any] = self.feat_extract_tester.batch_size _a : Tuple = self.feat_extract_tester.feature_size # test padding for List[int] + numpy _a : int = feat_extract.pad(_a ,padding=_a ) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad(_a ,padding='longest' ) _a : Any = input_a[input_name] _a : Optional[Any] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[-1] ) ) _a : List[str] = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) _a : str = input_a[input_name] # max_length parameter has to be provided when setting `padding="max_length"` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' )[input_name] _a : int = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,return_tensors='np' ) _a : Optional[int] = input_a[input_name] self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) self.assertTrue(len(input_a[0] ) == pad_min_length ) self.assertTrue(len(input_a[1] ) == pad_min_length + pad_diff ) self.assertTrue(input_a.shape[:2] == (batch_size, len(input_a[0] )) ) self.assertTrue(input_a.shape[:2] == (batch_size, pad_max_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == input_a.shape[2] == feature_size ) # test padding for `pad_to_multiple_of` for List[int] + numpy _a : Tuple = feat_extract.pad(_a ,pad_to_multiple_of=10 ) _a : List[str] = input_a[input_name] _a : str = feat_extract.pad(_a ,padding='longest' ,pad_to_multiple_of=10 ) _a : Tuple = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ) _a : Any = input_a[input_name] _a : Optional[int] = feat_extract.pad( _a ,padding='max_length' ,pad_to_multiple_of=10 ,max_length=_a ,return_tensors='np' ,) _a : Dict = input_a[input_name] self.assertTrue(all(len(_a ) % 10 == 0 for x in input_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) _a : List[str] = pad_max_length if pad_max_length % 10 == 0 else (pad_max_length // 10 + 1) * 10 self.assertTrue(all(len(_a ) == expected_mult_pad_length for x in input_a ) ) self.assertEqual(input_a.shape[:2] ,(batch_size, expected_mult_pad_length) ) if feature_size > 1: self.assertTrue(input_a.shape[2] == feature_size ) # Check padding value is correct _a : Any = (np.ones(self.feat_extract_tester.feature_size ) * feat_extract.padding_value).sum() self.assertTrue( abs(np.asarray(input_a[0] )[pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[1] )[pad_min_length + pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - pad_diff) ) < 1E-3 ) self.assertTrue( abs( np.asarray(input_a[2] )[pad_min_length + 2 * pad_diff :].sum() - padding_vector_sum * (pad_max_length - pad_min_length - 2 * pad_diff) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (pad_max_length - pad_min_length) ) < 1E-3 ) self.assertTrue( abs(input_a[0, pad_min_length:].sum() - padding_vector_sum * (expected_mult_pad_length - pad_min_length) ) < 1E-3 ) def __lowercase ( self : List[Any] ,_a : Optional[int]=False ): '''simple docstring''' def _inputs_have_equal_length(_a : List[str] ): _a : Union[str, Any] = len(input[0] ) for input_slice in input[1:]: if len(_a ) != length: return False return True def _inputs_are_equal(_a : List[str] ,_a : List[str] ): if len(_a ) != len(_a ): return False for input_slice_a, input_slice_a in zip(_a ,_a ): if not np.allclose(np.asarray(_a ) ,np.asarray(_a ) ,atol=1E-3 ): return False return True _a : Dict = self.feature_extraction_class(**self.feat_extract_dict ) _a : str = self.feat_extract_tester.prepare_inputs_for_common(numpify=_a ) _a : Any = feat_extract.model_input_names[0] _a : List[Any] = BatchFeature({input_name: speech_inputs} ) # truncate to smallest _a : Union[str, Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,truncation=_a ) _a : str = input_a[input_name] _a : List[str] = feat_extract.pad(_a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ) _a : Tuple = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to smallest with np _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ,truncation=_a ,) _a : Any = input_a[input_name] _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,return_tensors='np' ) _a : int = input_a[input_name] self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(input_a.shape[1] == len(speech_inputs[0] ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) # truncate to middle _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ,return_tensors='np' ,) _a : List[Any] = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,truncation=_a ) _a : Tuple = input_a[input_name] _a : Tuple = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[1] ) ,return_tensors='np' ) _a : Dict = input_a[input_name] self.assertTrue(input_a.shape[1] == len(speech_inputs[1] ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertTrue(_inputs_are_equal(_a ,_a ) ) # since truncation forces padding to be smaller than longest input # function can't return `np.ndarray`, but has to return list self.assertFalse(_inputs_have_equal_length(_a ) ) self.assertTrue(len(input_a[-1] ) == len(speech_inputs[-1] ) ) # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # padding has to be max_length when setting `truncation=True` with self.assertRaises(_a ): feat_extract.pad(_a ,padding='longest' ,truncation=_a )[input_name] # max_length parameter has to be provided when setting `truncation=True` and padding="max_length" with self.assertRaises(_a ): feat_extract.pad(_a ,padding='max_length' ,truncation=_a )[input_name] # test truncation for `pad_to_multiple_of` for List[int] + numpy _a : Optional[Any] = 12 _a : List[Any] = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,truncation=_a ,) _a : Tuple = input_a[input_name] _a : str = feat_extract.pad( _a ,padding='max_length' ,max_length=len(speech_inputs[0] ) ,pad_to_multiple_of=_a ,) _a : List[Any] = input_a[input_name] # retrieve expected_length as multiple of pad_to_multiple_of _a : List[Any] = len(speech_inputs[0] ) if expected_length % pad_to_multiple_of != 0: _a : Union[str, Any] = ((len(speech_inputs[0] ) // pad_to_multiple_of) + 1) * pad_to_multiple_of self.assertTrue(len(input_a[0] ) == expected_length ) self.assertTrue(_inputs_have_equal_length(_a ) ) self.assertFalse(_inputs_have_equal_length(_a ) ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' self._check_padding(numpify=_a ) def __lowercase ( self : Dict ): '''simple docstring''' self._check_truncation(numpify=_a ) def __lowercase ( self : str ): '''simple docstring''' self._check_truncation(numpify=_a ) @require_torch def __lowercase ( self : Dict ): '''simple docstring''' _a : Any = self.feature_extraction_class(**self.feat_extract_dict ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Optional[int] = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : List[str] = feat_extract.pad(_a ,padding='longest' ,return_tensors='pt' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) @require_tf def __lowercase ( self : int ): '''simple docstring''' _a : List[str] = self.feature_extraction_class(**self.feat_extract_dict ) _a : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : Dict = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' )[input_name] _a : Any = feat_extract.pad(_a ,padding='longest' ,return_tensors='tf' )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_tf.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : str = self.feat_extract_dict _a : List[Any] = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : List[Any] = self.feat_extract_tester.prepare_inputs_for_common() _a : Tuple = [len(_a ) for x in speech_inputs] _a : int = feat_extract.model_input_names[0] _a : Optional[Any] = BatchFeature({input_name: speech_inputs} ) _a : str = feat_extract.pad(_a ,padding='longest' ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual(list(processed.attention_mask.shape ) ,list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() ,_a ) def __lowercase ( self : int ): '''simple docstring''' _a : Any = self.feat_extract_dict _a : Tuple = True _a : Optional[int] = self.feature_extraction_class(**_a ) _a : Dict = self.feat_extract_tester.prepare_inputs_for_common() _a : Dict = [len(_a ) for x in speech_inputs] _a : Union[str, Any] = feat_extract.model_input_names[0] _a : Any = BatchFeature({input_name: speech_inputs} ) _a : List[Any] = min(_a ) _a : Dict = feat_extract.pad( _a ,padding='max_length' ,max_length=_a ,truncation=_a ,return_tensors='np' ) self.assertIn('attention_mask' ,_a ) self.assertListEqual( list(processed_pad.attention_mask.shape ) ,[processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() ,[max_length for x in speech_inputs] )
271
0
from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def lowerCamelCase_ ( _a ): """simple docstring""" lowerCAmelCase__ : int = int(number**0.5 ) return number == sq * sq def lowerCamelCase_ ( _a , _a , _a , _a , _a , _a ): """simple docstring""" lowerCAmelCase__ : int = x_num * y_den * z_den + y_num * x_den * z_den + z_num * x_den * y_den lowerCAmelCase__ : int = x_den * y_den * z_den lowerCAmelCase__ : int = gcd(__a , __a ) top //= hcf bottom //= hcf return top, bottom def lowerCamelCase_ ( _a = 35 ): """simple docstring""" lowerCAmelCase__ : set = set() lowerCAmelCase__ : int lowerCAmelCase__ : Fraction = Fraction(0 ) lowerCAmelCase__ : tuple[int, int] for x_num in range(1 , order + 1 ): for x_den in range(x_num + 1 , order + 1 ): for y_num in range(1 , order + 1 ): for y_den in range(y_num + 1 , order + 1 ): # n=1 lowerCAmelCase__ : Any = x_num * y_den + x_den * y_num lowerCAmelCase__ : int = x_den * y_den lowerCAmelCase__ : List[Any] = gcd(__a , __a ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: lowerCAmelCase__ : Tuple = add_three( __a , __a , __a , __a , __a , __a ) unique_s.add(__a ) # n=2 lowerCAmelCase__ : int = ( x_num * x_num * y_den * y_den + x_den * x_den * y_num * y_num ) lowerCAmelCase__ : Optional[int] = x_den * x_den * y_den * y_den if is_sq(__a ) and is_sq(__a ): lowerCAmelCase__ : Tuple = int(sqrt(__a ) ) lowerCAmelCase__ : Optional[int] = int(sqrt(__a ) ) lowerCAmelCase__ : int = gcd(__a , __a ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: lowerCAmelCase__ : List[str] = add_three( __a , __a , __a , __a , __a , __a ) unique_s.add(__a ) # n=-1 lowerCAmelCase__ : str = x_num * y_num lowerCAmelCase__ : Optional[Any] = x_den * y_num + x_num * y_den lowerCAmelCase__ : Any = gcd(__a , __a ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: lowerCAmelCase__ : List[str] = add_three( __a , __a , __a , __a , __a , __a ) unique_s.add(__a ) # n=2 lowerCAmelCase__ : Optional[Any] = x_num * x_num * y_num * y_num lowerCAmelCase__ : str = ( x_den * x_den * y_num * y_num + x_num * x_num * y_den * y_den ) if is_sq(__a ) and is_sq(__a ): lowerCAmelCase__ : List[Any] = int(sqrt(__a ) ) lowerCAmelCase__ : Union[str, Any] = int(sqrt(__a ) ) lowerCAmelCase__ : Dict = gcd(__a , __a ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: lowerCAmelCase__ : Union[str, Any] = add_three( __a , __a , __a , __a , __a , __a ) unique_s.add(__a ) for num, den in unique_s: total += Fraction(__a , __a ) return total.denominator + total.numerator if __name__ == "__main__": print(f'''{solution() = }''')
131
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import KarrasVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" __UpperCAmelCase : UNetaDModel __UpperCAmelCase : KarrasVeScheduler def __init__( self : Union[str, Any] ,_a : UNetaDModel ,_a : KarrasVeScheduler ): '''simple docstring''' super().__init__() self.register_modules(unet=_a ,scheduler=_a ) @torch.no_grad() def __call__( self : List[Any] ,_a : int = 1 ,_a : int = 50 ,_a : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,_a : Optional[str] = "pil" ,_a : bool = True ,**_a : List[Any] ,): '''simple docstring''' _a : Any = self.unet.config.sample_size _a : Optional[int] = (batch_size, 3, img_size, img_size) _a : Dict = self.unet # sample x_0 ~ N(0, sigma_0^2 * I) _a : Dict = randn_tensor(_a ,generator=_a ,device=self.device ) * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(_a ) for t in self.progress_bar(self.scheduler.timesteps ): # here sigma_t == t_i from the paper _a : Optional[int] = self.scheduler.schedule[t] _a : List[str] = self.scheduler.schedule[t - 1] if t > 0 else 0 # 1. Select temporarily increased noise level sigma_hat # 2. Add new noise to move from sample_i to sample_hat _a, _a : List[Any] = self.scheduler.add_noise_to_input(_a ,_a ,generator=_a ) # 3. Predict the noise residual given the noise magnitude `sigma_hat` # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_hat / 2) * model((sample_hat + 1) / 2 ,sigma_hat / 2 ).sample # 4. Evaluate dx/dt at sigma_hat # 5. Take Euler step from sigma to sigma_prev _a : Tuple = self.scheduler.step(_a ,_a ,_a ,_a ) if sigma_prev != 0: # 6. Apply 2nd order correction # The model inputs and output are adjusted by following eq. (213) in [1]. _a : Optional[int] = (sigma_prev / 2) * model((step_output.prev_sample + 1) / 2 ,sigma_prev / 2 ).sample _a : Optional[Any] = self.scheduler.step_correct( _a ,_a ,_a ,_a ,step_output.prev_sample ,step_output['derivative'] ,) _a : Dict = step_output.prev_sample _a : Tuple = (sample / 2 + 0.5).clamp(0 ,1 ) _a : Optional[Any] = sample.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": _a : List[str] = self.numpy_to_pil(_a ) if not return_dict: return (image,) return ImagePipelineOutput(images=_a )
271
0
"""simple docstring""" import argparse import os import torch from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) lowerCAmelCase__ = { '''sample_size''': 32, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 2, '''num_class_embeds''': 1_000, '''block_out_channels''': [32, 64], '''attention_head_dim''': 8, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''scale_shift''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } lowerCAmelCase__ = { '''sample_size''': 64, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 3, '''num_class_embeds''': 1_000, '''block_out_channels''': [192, 192 * 2, 192 * 3, 192 * 4], '''attention_head_dim''': 64, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''scale_shift''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } lowerCAmelCase__ = { '''sample_size''': 256, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 2, '''num_class_embeds''': None, '''block_out_channels''': [256, 256, 256 * 2, 256 * 2, 256 * 4, 256 * 4], '''attention_head_dim''': 64, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''ResnetDownsampleBlock2D''', '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', '''ResnetUpsampleBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''default''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } lowerCAmelCase__ = { '''num_train_timesteps''': 40, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } lowerCAmelCase__ = { '''num_train_timesteps''': 201, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } lowerCAmelCase__ = { '''num_train_timesteps''': 151, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } def a__ ( SCREAMING_SNAKE_CASE : int ): '''simple docstring''' if isinstance(__a , __a ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise argparse.ArgumentTypeError("boolean value expected" ) def a__ ( SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : Any , SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : Union[str, Any]=False ): '''simple docstring''' lowerCAmelCase : Optional[Any] = checkpoint[f"""{old_prefix}.in_layers.0.weight"""] lowerCAmelCase : Optional[int] = checkpoint[f"""{old_prefix}.in_layers.0.bias"""] lowerCAmelCase : List[str] = checkpoint[f"""{old_prefix}.in_layers.2.weight"""] lowerCAmelCase : str = checkpoint[f"""{old_prefix}.in_layers.2.bias"""] lowerCAmelCase : Any = checkpoint[f"""{old_prefix}.emb_layers.1.weight"""] lowerCAmelCase : Any = checkpoint[f"""{old_prefix}.emb_layers.1.bias"""] lowerCAmelCase : int = checkpoint[f"""{old_prefix}.out_layers.0.weight"""] lowerCAmelCase : Union[str, Any] = checkpoint[f"""{old_prefix}.out_layers.0.bias"""] lowerCAmelCase : Union[str, Any] = checkpoint[f"""{old_prefix}.out_layers.3.weight"""] lowerCAmelCase : str = checkpoint[f"""{old_prefix}.out_layers.3.bias"""] if has_skip: lowerCAmelCase : Optional[Any] = checkpoint[f"""{old_prefix}.skip_connection.weight"""] lowerCAmelCase : List[str] = checkpoint[f"""{old_prefix}.skip_connection.bias"""] return new_checkpoint def a__ ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : Optional[Any] , SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : Union[str, Any]=None ): '''simple docstring''' lowerCAmelCase : List[str] = checkpoint[f"""{old_prefix}.qkv.weight"""].chunk(3 , dim=0 ) lowerCAmelCase : List[str] = checkpoint[f"""{old_prefix}.qkv.bias"""].chunk(3 , dim=0 ) lowerCAmelCase : Dict = checkpoint[f"""{old_prefix}.norm.weight"""] lowerCAmelCase : str = checkpoint[f"""{old_prefix}.norm.bias"""] lowerCAmelCase : List[Any] = weight_q.squeeze(-1 ).squeeze(-1 ) lowerCAmelCase : Union[str, Any] = bias_q.squeeze(-1 ).squeeze(-1 ) lowerCAmelCase : Tuple = weight_k.squeeze(-1 ).squeeze(-1 ) lowerCAmelCase : Optional[int] = bias_k.squeeze(-1 ).squeeze(-1 ) lowerCAmelCase : Tuple = weight_v.squeeze(-1 ).squeeze(-1 ) lowerCAmelCase : List[str] = bias_v.squeeze(-1 ).squeeze(-1 ) lowerCAmelCase : Optional[int] = ( checkpoint[f"""{old_prefix}.proj_out.weight"""].squeeze(-1 ).squeeze(-1 ) ) lowerCAmelCase : Optional[Any] = checkpoint[f"""{old_prefix}.proj_out.bias"""].squeeze(-1 ).squeeze(-1 ) return new_checkpoint def a__ ( SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : Tuple ): '''simple docstring''' lowerCAmelCase : Optional[Any] = torch.load(__a , map_location="cpu" ) lowerCAmelCase : List[str] = {} lowerCAmelCase : Any = checkpoint['time_embed.0.weight'] lowerCAmelCase : Any = checkpoint['time_embed.0.bias'] lowerCAmelCase : List[Any] = checkpoint['time_embed.2.weight'] lowerCAmelCase : Any = checkpoint['time_embed.2.bias'] if unet_config["num_class_embeds"] is not None: lowerCAmelCase : List[Any] = checkpoint['label_emb.weight'] lowerCAmelCase : Any = checkpoint['input_blocks.0.0.weight'] lowerCAmelCase : Optional[Any] = checkpoint['input_blocks.0.0.bias'] lowerCAmelCase : List[Any] = unet_config['down_block_types'] lowerCAmelCase : int = unet_config['layers_per_block'] lowerCAmelCase : Union[str, Any] = unet_config['attention_head_dim'] lowerCAmelCase : Union[str, Any] = unet_config['block_out_channels'] lowerCAmelCase : Union[str, Any] = 1 lowerCAmelCase : List[str] = channels_list[0] for i, layer_type in enumerate(__a ): lowerCAmelCase : Optional[Any] = channels_list[i] lowerCAmelCase : List[Any] = current_channels != prev_channels if layer_type == "ResnetDownsampleBlock2D": for j in range(__a ): lowerCAmelCase : List[Any] = f"""down_blocks.{i}.resnets.{j}""" lowerCAmelCase : Tuple = f"""input_blocks.{current_layer}.0""" lowerCAmelCase : List[Any] = True if j == 0 and downsample_block_has_skip else False lowerCAmelCase : List[Any] = convert_resnet(__a , __a , __a , __a , has_skip=__a ) current_layer += 1 elif layer_type == "AttnDownBlock2D": for j in range(__a ): lowerCAmelCase : Optional[Any] = f"""down_blocks.{i}.resnets.{j}""" lowerCAmelCase : List[Any] = f"""input_blocks.{current_layer}.0""" lowerCAmelCase : Union[str, Any] = True if j == 0 and downsample_block_has_skip else False lowerCAmelCase : str = convert_resnet(__a , __a , __a , __a , has_skip=__a ) lowerCAmelCase : Tuple = f"""down_blocks.{i}.attentions.{j}""" lowerCAmelCase : Optional[int] = f"""input_blocks.{current_layer}.1""" lowerCAmelCase : Dict = convert_attention( __a , __a , __a , __a , __a ) current_layer += 1 if i != len(__a ) - 1: lowerCAmelCase : Tuple = f"""down_blocks.{i}.downsamplers.0""" lowerCAmelCase : str = f"""input_blocks.{current_layer}.0""" lowerCAmelCase : List[Any] = convert_resnet(__a , __a , __a , __a ) current_layer += 1 lowerCAmelCase : List[Any] = current_channels # hardcoded the mid-block for now lowerCAmelCase : Any = 'mid_block.resnets.0' lowerCAmelCase : Optional[Any] = 'middle_block.0' lowerCAmelCase : Optional[Any] = convert_resnet(__a , __a , __a , __a ) lowerCAmelCase : Union[str, Any] = 'mid_block.attentions.0' lowerCAmelCase : Any = 'middle_block.1' lowerCAmelCase : Tuple = convert_attention(__a , __a , __a , __a , __a ) lowerCAmelCase : Optional[int] = 'mid_block.resnets.1' lowerCAmelCase : Dict = 'middle_block.2' lowerCAmelCase : Optional[Any] = convert_resnet(__a , __a , __a , __a ) lowerCAmelCase : str = 0 lowerCAmelCase : str = unet_config['up_block_types'] for i, layer_type in enumerate(__a ): if layer_type == "ResnetUpsampleBlock2D": for j in range(layers_per_block + 1 ): lowerCAmelCase : List[str] = f"""up_blocks.{i}.resnets.{j}""" lowerCAmelCase : str = f"""output_blocks.{current_layer}.0""" lowerCAmelCase : Dict = convert_resnet(__a , __a , __a , __a , has_skip=__a ) current_layer += 1 if i != len(__a ) - 1: lowerCAmelCase : Dict = f"""up_blocks.{i}.upsamplers.0""" lowerCAmelCase : Union[str, Any] = f"""output_blocks.{current_layer-1}.1""" lowerCAmelCase : Dict = convert_resnet(__a , __a , __a , __a ) elif layer_type == "AttnUpBlock2D": for j in range(layers_per_block + 1 ): lowerCAmelCase : Union[str, Any] = f"""up_blocks.{i}.resnets.{j}""" lowerCAmelCase : int = f"""output_blocks.{current_layer}.0""" lowerCAmelCase : List[str] = convert_resnet(__a , __a , __a , __a , has_skip=__a ) lowerCAmelCase : Union[str, Any] = f"""up_blocks.{i}.attentions.{j}""" lowerCAmelCase : int = f"""output_blocks.{current_layer}.1""" lowerCAmelCase : List[Any] = convert_attention( __a , __a , __a , __a , __a ) current_layer += 1 if i != len(__a ) - 1: lowerCAmelCase : str = f"""up_blocks.{i}.upsamplers.0""" lowerCAmelCase : int = f"""output_blocks.{current_layer-1}.2""" lowerCAmelCase : List[str] = convert_resnet(__a , __a , __a , __a ) lowerCAmelCase : List[str] = checkpoint['out.0.weight'] lowerCAmelCase : int = checkpoint['out.0.bias'] lowerCAmelCase : Optional[int] = checkpoint['out.2.weight'] lowerCAmelCase : Optional[int] = checkpoint['out.2.bias'] return new_checkpoint if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument('''--unet_path''', default=None, type=str, required=True, help='''Path to the unet.pt to convert.''') parser.add_argument( '''--dump_path''', default=None, type=str, required=True, help='''Path to output the converted UNet model.''' ) parser.add_argument('''--class_cond''', default=True, type=str, help='''Whether the model is class-conditional.''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = strabool(args.class_cond) lowerCAmelCase__ = os.path.basename(args.unet_path) print(F"Checkpoint: {ckpt_name}") # Get U-Net config if "imagenet64" in ckpt_name: lowerCAmelCase__ = IMAGENET_64_UNET_CONFIG elif "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): lowerCAmelCase__ = LSUN_256_UNET_CONFIG elif "test" in ckpt_name: lowerCAmelCase__ = TEST_UNET_CONFIG else: raise ValueError(F"Checkpoint type {ckpt_name} is not currently supported.") if not args.class_cond: lowerCAmelCase__ = None lowerCAmelCase__ = con_pt_to_diffuser(args.unet_path, unet_config) lowerCAmelCase__ = UNetaDModel(**unet_config) image_unet.load_state_dict(converted_unet_ckpt) # Get scheduler config if "cd" in ckpt_name or "test" in ckpt_name: lowerCAmelCase__ = CD_SCHEDULER_CONFIG elif "ct" in ckpt_name and "imagenet64" in ckpt_name: lowerCAmelCase__ = CT_IMAGENET_64_SCHEDULER_CONFIG elif "ct" in ckpt_name and "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): lowerCAmelCase__ = CT_LSUN_256_SCHEDULER_CONFIG else: raise ValueError(F"Checkpoint type {ckpt_name} is not currently supported.") lowerCAmelCase__ = CMStochasticIterativeScheduler(**scheduler_config) lowerCAmelCase__ = ConsistencyModelPipeline(unet=image_unet, scheduler=cm_scheduler) consistency_model.save_pretrained(args.dump_path)
108
'''simple docstring''' import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging __lowerCAmelCase = ( """https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py""" ) __lowerCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def UpperCAmelCase_ (): """simple docstring""" _a : Optional[int] = 'https://pypi.org/pypi/diffusers/json' _a : int = json.loads(request.urlopen(__a ).read() )['releases'].keys() return sorted(__a , key=lambda __a : version.Version(__a ) ) def UpperCAmelCase_ (): """simple docstring""" if HF_MODULES_CACHE in sys.path: return sys.path.append(__a ) os.makedirs(__a , exist_ok=__a ) _a : str = Path(__a ) / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : Union[str, os.PathLike] ): """simple docstring""" init_hf_modules() _a : Dict = Path(__a ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(__a , exist_ok=__a ) _a : Optional[int] = dynamic_module_path / '__init__.py' if not init_path.exists(): init_path.touch() def UpperCAmelCase_ (__a : str ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : int = f.read() # Imports of the form `import .xxx` _a : Tuple = re.findall('^\s*import\s+\.(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall('^\s*from\s+\.(\S+)\s+import' , __a , flags=re.MULTILINE ) # Unique-ify return list(set(__a ) ) def UpperCAmelCase_ (__a : Any ): """simple docstring""" _a : Optional[int] = False _a : Optional[int] = [module_file] _a : List[str] = [] # Let's recurse through all relative imports while not no_change: _a : str = [] for f in files_to_check: new_imports.extend(get_relative_imports(__a ) ) _a : Union[str, Any] = Path(__a ).parent _a : str = [str(module_path / m ) for m in new_imports] _a : Tuple = [f for f in new_import_files if f not in all_relative_imports] _a : Dict = [f"""{f}.py""" for f in new_import_files] _a : List[str] = len(__a ) == 0 all_relative_imports.extend(__a ) return all_relative_imports def UpperCAmelCase_ (__a : Tuple ): """simple docstring""" with open(__a , 'r' , encoding='utf-8' ) as f: _a : Dict = f.read() # Imports of the form `import xxx` _a : Optional[int] = re.findall('^\s*import\s+(\S+)\s*$' , __a , flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall('^\s*from\s+(\S+)\s+import' , __a , flags=re.MULTILINE ) # Only keep the top-level module _a : List[str] = [imp.split('.' )[0] for imp in imports if not imp.startswith('.' )] # Unique-ify and test we got them all _a : Optional[int] = list(set(__a ) ) _a : List[str] = [] for imp in imports: try: importlib.import_module(__a ) except ImportError: missing_packages.append(__a ) if len(__a ) > 0: raise ImportError( 'This modeling file requires the following packages that were not found in your environment: ' f"""{', '.join(__a )}. Run `pip install {' '.join(__a )}`""" ) return get_relative_imports(__a ) def UpperCAmelCase_ (__a : Any , __a : str ): """simple docstring""" _a : Any = module_path.replace(os.path.sep , '.' ) _a : Union[str, Any] = importlib.import_module(__a ) if class_name is None: return find_pipeline_class(__a ) return getattr(__a , __a ) def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" from ..pipelines import DiffusionPipeline _a : List[str] = dict(inspect.getmembers(__a , inspect.isclass ) ) _a : str = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls , __a ) and cls.__module__.split('.' )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( f"""Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:""" f""" {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in""" f""" {loaded_module}.""" ) _a : Any = cls return pipeline_class def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , ): """simple docstring""" _a : str = str(__a ) _a : Optional[Any] = os.path.join(__a , __a ) if os.path.isfile(__a ): _a : Tuple = module_file_or_url _a : Optional[Any] = 'local' elif pretrained_model_name_or_path.count('/' ) == 0: _a : int = get_diffusers_versions() # cut ".dev0" _a : Any = 'v' + '.'.join(__version__.split('.' )[:3] ) # retrieve github version that matches if revision is None: _a : Any = latest_version if latest_version[1:] in available_versions else 'main' logger.info(f"""Defaulting to latest_version: {revision}.""" ) elif revision in available_versions: _a : Any = f"""v{revision}""" elif revision == "main": _a : Optional[int] = revision else: raise ValueError( f"""`custom_revision`: {revision} does not exist. Please make sure to choose one of""" f""" {', '.join(available_versions + ['main'] )}.""" ) # community pipeline on GitHub _a : Tuple = COMMUNITY_PIPELINES_URL.format(revision=__a , pipeline=__a ) try: _a : Any = cached_download( __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = 'git' _a : Any = pretrained_model_name_or_path + '.py' except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise else: try: # Load from URL or cache if already cached _a : Optional[Any] = hf_hub_download( __a , __a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , ) _a : List[Any] = os.path.join('local' , '--'.join(pretrained_model_name_or_path.split('/' ) ) ) except EnvironmentError: logger.error(f"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise # Check we have all the requirements in our environment _a : Optional[int] = check_imports(__a ) # Now we move the module inside our cached dynamic modules. _a : Optional[Any] = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(__a ) _a : Any = Path(__a ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(__a , submodule_path / module_file ) for module_needed in modules_needed: _a : Dict = f"""{module_needed}.py""" shutil.copy(os.path.join(__a , __a ) , submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(__a , __a ): _a : Optional[Any] = use_auth_token elif use_auth_token is True: _a : List[Any] = HfFolder.get_token() else: _a : Dict = None _a : int = model_info(__a , revision=__a , token=__a ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. _a : Optional[int] = submodule_path / commit_hash _a : str = full_submodule + os.path.sep + commit_hash create_dynamic_module(__a ) if not (submodule_path / module_file).exists(): shutil.copy(__a , submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( __a , f"""{module_needed}.py""" , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return os.path.join(__a , __a ) def UpperCAmelCase_ (__a : Union[str, os.PathLike] , __a : str , __a : Optional[str] = None , __a : Optional[Union[str, os.PathLike]] = None , __a : bool = False , __a : bool = False , __a : Optional[Dict[str, str]] = None , __a : Optional[Union[bool, str]] = None , __a : Optional[str] = None , __a : bool = False , **__a : str , ): """simple docstring""" _a : Dict = get_cached_module_file( __a , __a , cache_dir=__a , force_download=__a , resume_download=__a , proxies=__a , use_auth_token=__a , revision=__a , local_files_only=__a , ) return get_class_in_module(__a , final_module.replace('.py' , '' ) )
271
0
"""simple docstring""" import inspect import os import torch from transformers import AutoModel from transformers.testing_utils import mockenv_context from transformers.trainer_utils import set_seed import accelerate from accelerate.accelerator import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils.testing import ( AccelerateTestCase, TempDirTestCase, execute_subprocess_async, require_cuda, require_fsdp, require_multi_gpu, slow, ) from accelerate.utils.constants import ( FSDP_AUTO_WRAP_POLICY, FSDP_BACKWARD_PREFETCH, FSDP_SHARDING_STRATEGY, FSDP_STATE_DICT_TYPE, ) from accelerate.utils.dataclasses import FullyShardedDataParallelPlugin from accelerate.utils.other import patch_environment set_seed(42) SCREAMING_SNAKE_CASE : Optional[Any] = """bert-base-cased""" SCREAMING_SNAKE_CASE : List[Any] = """fp16""" SCREAMING_SNAKE_CASE : Union[str, Any] = """bf16""" SCREAMING_SNAKE_CASE : Optional[int] = [FPaa, BFaa] @require_fsdp @require_cuda class _UpperCAmelCase ( lowercase__ ): '''simple docstring''' def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' super().setUp() __snake_case : int = dict( ACCELERATE_USE_FSDP='''true''' , MASTER_ADDR='''localhost''' , MASTER_PORT='''10999''' , RANK='''0''' , LOCAL_RANK='''0''' , WORLD_SIZE='''1''' , ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import ShardingStrategy for i, strategy in enumerate(_a ): __snake_case : Union[str, Any] = self.dist_env.copy() __snake_case : Optional[Any] = f"""{i + 1}""" __snake_case : str = strategy with mockenv_context(**_a ): __snake_case : Optional[int] = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.sharding_strategy , ShardingStrategy(i + 1 ) ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import BackwardPrefetch for i, prefetch_policy in enumerate(_a ): __snake_case : Tuple = self.dist_env.copy() __snake_case : List[str] = prefetch_policy with mockenv_context(**_a ): __snake_case : Optional[int] = FullyShardedDataParallelPlugin() if prefetch_policy == "NO_PREFETCH": self.assertIsNone(fsdp_plugin.backward_prefetch ) else: self.assertEqual(fsdp_plugin.backward_prefetch , BackwardPrefetch(i + 1 ) ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType for i, state_dict_type in enumerate(_a ): __snake_case : List[Any] = self.dist_env.copy() __snake_case : str = state_dict_type with mockenv_context(**_a ): __snake_case : Optional[Any] = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.state_dict_type , StateDictType(i + 1 ) ) if state_dict_type == "FULL_STATE_DICT": self.assertTrue(fsdp_plugin.state_dict_config.offload_to_cpu ) self.assertTrue(fsdp_plugin.state_dict_config.ranka_only ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Dict = AutoModel.from_pretrained(_a ) for policy in FSDP_AUTO_WRAP_POLICY: __snake_case : Optional[int] = self.dist_env.copy() __snake_case : Union[str, Any] = policy if policy == "TRANSFORMER_BASED_WRAP": __snake_case : int = 'BertLayer' elif policy == "SIZE_BASED_WRAP": __snake_case : Optional[int] = '2000' with mockenv_context(**_a ): __snake_case : Union[str, Any] = FullyShardedDataParallelPlugin() fsdp_plugin.set_auto_wrap_policy(_a ) if policy == "NO_WRAP": self.assertIsNone(fsdp_plugin.auto_wrap_policy ) else: self.assertIsNotNone(fsdp_plugin.auto_wrap_policy ) __snake_case : Tuple = self.dist_env.copy() __snake_case : List[str] = 'TRANSFORMER_BASED_WRAP' __snake_case : Optional[Any] = 'T5Layer' with mockenv_context(**_a ): __snake_case : int = FullyShardedDataParallelPlugin() with self.assertRaises(_a ) as cm: fsdp_plugin.set_auto_wrap_policy(_a ) self.assertTrue('''Could not find the transformer layer class to wrap in the model.''' in str(cm.exception ) ) __snake_case : Tuple = self.dist_env.copy() __snake_case : Optional[int] = 'SIZE_BASED_WRAP' __snake_case : str = '0' with mockenv_context(**_a ): __snake_case : Optional[int] = FullyShardedDataParallelPlugin() fsdp_plugin.set_auto_wrap_policy(_a ) self.assertIsNone(fsdp_plugin.auto_wrap_policy ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import MixedPrecision from torch.distributed.fsdp.sharded_grad_scaler import ShardedGradScaler for mp_dtype in dtypes: __snake_case : Any = self.dist_env.copy() __snake_case : List[str] = mp_dtype with mockenv_context(**_a ): __snake_case : Dict = Accelerator() if mp_dtype == "fp16": __snake_case : List[str] = torch.floataa elif mp_dtype == "bf16": __snake_case : Any = torch.bfloataa __snake_case : Any = MixedPrecision(param_dtype=_a , reduce_dtype=_a , buffer_dtype=_a ) self.assertEqual(accelerator.state.fsdp_plugin.mixed_precision_policy , _a ) if mp_dtype == FPaa: self.assertTrue(isinstance(accelerator.scaler , _a ) ) elif mp_dtype == BFaa: self.assertIsNone(accelerator.scaler ) AcceleratorState._reset_state(_a ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' from torch.distributed.fsdp.fully_sharded_data_parallel import CPUOffload for flag in [True, False]: __snake_case : List[str] = self.dist_env.copy() __snake_case : Tuple = str(_a ).lower() with mockenv_context(**_a ): __snake_case : List[str] = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.cpu_offload , CPUOffload(offload_params=_a ) ) @require_fsdp @require_multi_gpu @slow class _UpperCAmelCase ( lowercase__ ): '''simple docstring''' def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' super().setUp() __snake_case : int = 0.82 __snake_case : Union[str, Any] = [ 'fsdp_shard_grad_op_transformer_based_wrap', 'fsdp_full_shard_transformer_based_wrap', ] __snake_case : Dict = { 'multi_gpu_fp16': 32_00, 'fsdp_shard_grad_op_transformer_based_wrap_fp16': 20_00, 'fsdp_full_shard_transformer_based_wrap_fp16': 19_00, # Disabling below test as it overwhelms the RAM memory usage # on CI self-hosted runner leading to tests getting killed. # "fsdp_full_shard_cpu_offload_transformer_based_wrap_fp32": 1500, # fp16 was leading to indefinite hang } __snake_case : Tuple = 1_60 __snake_case : Dict = 1_60 __snake_case : List[str] = inspect.getfile(accelerate.test_utils ) __snake_case : Optional[Any] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''external_deps'''] ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[str] = os.path.join(self.test_scripts_folder , '''test_performance.py''' ) __snake_case : Any = ['accelerate', 'launch', '--num_processes=2', '--num_machines=1', '--machine_rank=0', '--use_fsdp'] for config in self.performance_configs: __snake_case : int = cmd.copy() for i, strategy in enumerate(_a ): if strategy.lower() in config: cmd_config.append(f"""--fsdp_sharding_strategy={i+1}""" ) break if "fp32" in config: cmd_config.append('''--mixed_precision=no''' ) else: cmd_config.append('''--mixed_precision=fp16''' ) if "cpu_offload" in config: cmd_config.append('''--fsdp_offload_params=True''' ) for policy in FSDP_AUTO_WRAP_POLICY: if policy.lower() in config: cmd_config.append(f"""--fsdp_auto_wrap_policy={policy}""" ) break if policy == "TRANSFORMER_BASED_WRAP": cmd_config.append('''--fsdp_transformer_layer_cls_to_wrap=BertLayer''' ) elif policy == "SIZE_BASED_WRAP": cmd_config.append('''--fsdp_min_num_params=2000''' ) cmd_config.extend( [ self.test_file_path, f"""--output_dir={self.tmpdir}""", f"""--performance_lower_bound={self.performance_lower_bound}""", ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_a , env=os.environ.copy() ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[Any] = os.path.join(self.test_scripts_folder , '''test_checkpointing.py''' ) __snake_case : Union[str, Any] = [ 'accelerate', 'launch', '--num_processes=2', '--num_machines=1', '--machine_rank=0', '--use_fsdp', '--mixed_precision=fp16', '--fsdp_transformer_layer_cls_to_wrap=BertLayer', ] for i, strategy in enumerate(_a ): __snake_case : int = cmd.copy() cmd_config.append(f"""--fsdp_sharding_strategy={i+1}""" ) if strategy != "FULL_SHARD": continue __snake_case : int = len(_a ) for state_dict_type in FSDP_STATE_DICT_TYPE: __snake_case : str = cmd_config[:state_dict_config_index] cmd_config.append(f"""--fsdp_state_dict_type={state_dict_type}""" ) cmd_config.extend( [ self.test_file_path, f"""--output_dir={self.tmpdir}""", '''--partial_train_epoch=1''', ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_a , env=os.environ.copy() ) __snake_case : Optional[int] = cmd_config[:-1] __snake_case : Tuple = os.path.join(self.tmpdir , '''epoch_0''' ) cmd_config.extend( [ f"""--resume_from_checkpoint={resume_from_checkpoint}""", ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_a , env=os.environ.copy() ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : int = os.path.join(self.test_scripts_folder , '''test_peak_memory_usage.py''' ) __snake_case : Optional[int] = [ 'accelerate', 'launch', '--num_processes=2', '--num_machines=1', '--machine_rank=0', ] for spec, peak_mem_upper_bound in self.peak_memory_usage_upper_bound.items(): __snake_case : Any = cmd.copy() if "fp16" in spec: cmd_config.extend(['''--mixed_precision=fp16'''] ) else: cmd_config.extend(['''--mixed_precision=no'''] ) if "multi_gpu" in spec: continue else: cmd_config.extend(['''--use_fsdp'''] ) for i, strategy in enumerate(_a ): if strategy.lower() in spec: cmd_config.append(f"""--fsdp_sharding_strategy={i+1}""" ) break if "cpu_offload" in spec: cmd_config.append('''--fsdp_offload_params=True''' ) for policy in FSDP_AUTO_WRAP_POLICY: if policy.lower() in spec: cmd_config.append(f"""--fsdp_auto_wrap_policy={policy}""" ) break if policy == "TRANSFORMER_BASED_WRAP": cmd_config.append('''--fsdp_transformer_layer_cls_to_wrap=BertLayer''' ) elif policy == "SIZE_BASED_WRAP": cmd_config.append('''--fsdp_min_num_params=2000''' ) cmd_config.extend( [ self.test_file_path, f"""--output_dir={self.tmpdir}""", f"""--peak_memory_upper_bound={peak_mem_upper_bound}""", f"""--n_train={self.n_train}""", f"""--n_val={self.n_val}""", ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_a , env=os.environ.copy() )
102
'''simple docstring''' def UpperCAmelCase_ (__a : list , __a : list , __a : int ): """simple docstring""" _a : Optional[Any] = len(__a ) _a : int = [[0] * n for i in range(__a )] for i in range(__a ): _a : Tuple = y_points[i] for i in range(2 , __a ): for j in range(__a , __a ): _a : Tuple = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
271
0
"""simple docstring""" import unittest import numpy as np def __magic_name__ ( __snake_case : np.ndarray , __snake_case : np.ndarray , __snake_case : np.ndarray , __snake_case : np.ndarray | None = None , ) -> Dict: lowercase : Optional[int] = np.shape(__a ) lowercase : Dict = np.shape(__a ) lowercase : List[Any] = np.shape(__a ) if shape_a[0] != shape_b[0]: lowercase : int = ( 'Expected the same number of rows for A and B. ' f"""Instead found A of size {shape_a} and B of size {shape_b}""" ) raise ValueError(__a ) if shape_b[1] != shape_c[1]: lowercase : Tuple = ( 'Expected the same number of columns for B and C. ' f"""Instead found B of size {shape_b} and C of size {shape_c}""" ) raise ValueError(__a ) lowercase : Optional[Any] = pseudo_inv if a_inv is None: try: lowercase : Union[str, Any] = np.linalg.inv(__a ) except np.linalg.LinAlgError: raise ValueError( "Input matrix A is not invertible. Cannot compute Schur complement." ) return mat_c - mat_b.T @ a_inv @ mat_b class a__ ( unittest.TestCase ): def __magic_name__ ( self ): lowercase : int = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) lowercase : Dict = np.array([[0, 3], [3, 0], [2, 3]] ) lowercase : List[str] = np.array([[2, 1], [6, 3]] ) lowercase : Any = schur_complement(_a , _a , _a ) lowercase : Any = np.block([[a, b], [b.T, c]] ) lowercase : str = np.linalg.det(_a ) lowercase : Dict = np.linalg.det(_a ) lowercase : Union[str, Any] = np.linalg.det(_a ) self.assertAlmostEqual(_a , det_a * det_s ) def __magic_name__ ( self ): lowercase : Optional[Any] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) lowercase : Optional[Any] = np.array([[0, 3], [3, 0], [2, 3]] ) lowercase : str = np.array([[2, 1], [6, 3]] ) with self.assertRaises(_a ): schur_complement(_a , _a , _a ) def __magic_name__ ( self ): lowercase : Optional[Any] = np.array([[1, 2, 1], [2, 1, 2], [3, 2, 4]] ) lowercase : Optional[Any] = np.array([[0, 3], [3, 0], [2, 3]] ) lowercase : Optional[Any] = np.array([[2, 1, 3], [6, 3, 5]] ) with self.assertRaises(_a ): schur_complement(_a , _a , _a ) if __name__ == "__main__": import doctest doctest.testmod() unittest.main()
202
'''simple docstring''' import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = inspect.getfile(accelerate.test_utils ) __UpperCAmelCase : List[str] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_cli.py'''] ) __UpperCAmelCase : Dict = ['''accelerate''', '''launch'''] __UpperCAmelCase : Dict = Path.home() / '''.cache/huggingface/accelerate''' __UpperCAmelCase : Dict = '''default_config.yaml''' __UpperCAmelCase : Optional[Any] = config_folder / config_file __UpperCAmelCase : Dict = config_folder / '''_default_config.yaml''' __UpperCAmelCase : Any = Path('''tests/test_configs''' ) @classmethod def __lowercase ( cls : int ): '''simple docstring''' if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def __lowercase ( cls : List[Any] ): '''simple docstring''' if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Dict = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' for config in sorted(self.test_config_path.glob('**/*.yaml' ) ): with self.subTest(config_file=_a ): execute_subprocess_async( self.base_cmd + ['--config_file', str(_a ), self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[int] ): '''simple docstring''' execute_subprocess_async(['accelerate', 'test'] ,env=os.environ.copy() ) class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = '''test-tpu''' __UpperCAmelCase : Any = '''us-central1-a''' __UpperCAmelCase : List[Any] = '''ls''' __UpperCAmelCase : Any = ['''accelerate''', '''tpu-config'''] __UpperCAmelCase : Dict = '''cd /usr/share''' __UpperCAmelCase : Any = '''tests/test_samples/test_command_file.sh''' __UpperCAmelCase : List[Any] = '''Running gcloud compute tpus tpu-vm ssh''' def __lowercase ( self : Dict ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Any = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--debug'] ,return_stdout=_a ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : int ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : str ): '''simple docstring''' _a : List[str] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--command', 'echo "Hello World"', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo \"Hello World\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Any = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command_file', self.command_file, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Union[str, Any] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command_file', self.command_file, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Any ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--accelerate_version', '12.0.0', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,)
271
0
"""simple docstring""" import unittest from transformers import PegasusTokenizer, PegasusTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowercase__ : int = get_tests_dir("""fixtures/test_sentencepiece_no_bos.model""") @require_sentencepiece @require_tokenizers class UpperCamelCase__ ( lowercase__, unittest.TestCase ): """simple docstring""" _SCREAMING_SNAKE_CASE = PegasusTokenizer _SCREAMING_SNAKE_CASE = PegasusTokenizerFast _SCREAMING_SNAKE_CASE = True _SCREAMING_SNAKE_CASE = True def SCREAMING_SNAKE_CASE__ ( self : Any ): super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase_ : Optional[int] = PegasusTokenizer(_a ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ): return PegasusTokenizer.from_pretrained('google/pegasus-large' ) def SCREAMING_SNAKE_CASE__ ( self : str , **SCREAMING_SNAKE_CASE_ : str ): return PegasusTokenizer.from_pretrained(self.tmpdirname , **_a ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Dict ): return ("This is a test", "This is a test") def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): lowerCAmelCase_ : Any = '</s>' lowerCAmelCase_ : Optional[int] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_a ) , _a ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_a ) , _a ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ): lowerCAmelCase_ : List[str] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<pad>' ) self.assertEqual(vocab_keys[1] , '</s>' ) self.assertEqual(vocab_keys[-1] , 'v' ) self.assertEqual(len(_a ) , 1_1_0_3 ) def SCREAMING_SNAKE_CASE__ ( self : Dict ): self.assertEqual(self.get_tokenizer().vocab_size , 1_1_0_3 ) def SCREAMING_SNAKE_CASE__ ( self : Any ): lowerCAmelCase_ : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) lowerCAmelCase_ : Tuple = self.tokenizer_class.from_pretrained(self.tmpdirname ) lowerCAmelCase_ : Union[str, Any] = ( 'Let\'s see which <unk> is the better <unk_token_11> one <mask_1> It seems like this <mask_2> was important' ' </s> <pad> <pad> <pad>' ) lowerCAmelCase_ : Any = rust_tokenizer([raw_input_str] , return_tensors=_a , add_special_tokens=_a ).input_ids[0] lowerCAmelCase_ : List[str] = py_tokenizer([raw_input_str] , return_tensors=_a , add_special_tokens=_a ).input_ids[0] self.assertListEqual(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self : str ): lowerCAmelCase_ : Union[str, Any] = self._large_tokenizer # <mask_1> masks whole sentence while <mask_2> masks single word lowerCAmelCase_ : List[Any] = '<mask_1> To ensure a <mask_2> flow of bank resolutions.' lowerCAmelCase_ : List[Any] = [2, 4_1_3, 6_1_5, 1_1_4, 3, 1_9_7_1, 1_1_3, 1_6_7_9, 1_0_7_1_0, 1_0_7, 1] lowerCAmelCase_ : Union[str, Any] = tokenizer([raw_input_str] , return_tensors=_a ).input_ids[0] self.assertListEqual(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self : Any ): lowerCAmelCase_ : str = self._large_tokenizer # The tracebacks for the following asserts are **better** without messages or self.assertEqual assert tokenizer.vocab_size == 9_6_1_0_3 assert tokenizer.pad_token_id == 0 assert tokenizer.eos_token_id == 1 assert tokenizer.offset == 1_0_3 assert tokenizer.unk_token_id == tokenizer.offset + 2 == 1_0_5 assert tokenizer.unk_token == "<unk>" assert tokenizer.model_max_length == 1_0_2_4 lowerCAmelCase_ : Any = 'To ensure a smooth flow of bank resolutions.' lowerCAmelCase_ : int = [4_1_3, 6_1_5, 1_1_4, 2_2_9_1, 1_9_7_1, 1_1_3, 1_6_7_9, 1_0_7_1_0, 1_0_7, 1] lowerCAmelCase_ : Union[str, Any] = tokenizer([raw_input_str] , return_tensors=_a ).input_ids[0] self.assertListEqual(_a , _a ) assert tokenizer.convert_ids_to_tokens([0, 1, 2, 3] ) == ["<pad>", "</s>", "<mask_1>", "<mask_2>"] @require_torch def SCREAMING_SNAKE_CASE__ ( self : str ): lowerCAmelCase_ : Tuple = ['This is going to be way too long.' * 1_5_0, 'short example'] lowerCAmelCase_ : List[str] = ['not super long but more than 5 tokens', 'tiny'] lowerCAmelCase_ : str = self._large_tokenizer(_a , padding=_a , truncation=_a , return_tensors='pt' ) lowerCAmelCase_ : int = self._large_tokenizer( text_target=_a , max_length=5 , padding=_a , truncation=_a , return_tensors='pt' ) assert batch.input_ids.shape == (2, 1_0_2_4) assert batch.attention_mask.shape == (2, 1_0_2_4) assert targets["input_ids"].shape == (2, 5) assert len(_a ) == 2 # input_ids, attention_mask. @slow def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): lowerCAmelCase_ : Tuple = {'input_ids': [[3_8_9_7_9, 1_4_3, 1_8_4_8_5, 6_0_6, 1_3_0, 2_6_6_6_9, 8_7_6_8_6, 1_2_1, 5_4_1_8_9, 1_1_2_9, 1_1_1, 2_6_6_6_9, 8_7_6_8_6, 1_2_1, 9_1_1_4, 1_4_7_8_7, 1_2_1, 1_3_2_4_9, 1_5_8, 5_9_2, 9_5_6, 1_2_1, 1_4_6_2_1, 3_1_5_7_6, 1_4_3, 6_2_6_1_3, 1_0_8, 9_6_8_8, 9_3_0, 4_3_4_3_0, 1_1_5_6_2, 6_2_6_1_3, 3_0_4, 1_0_8, 1_1_4_4_3, 8_9_7, 1_0_8, 9_3_1_4, 1_7_4_1_5, 6_3_3_9_9, 1_0_8, 1_1_4_4_3, 7_6_1_4, 1_8_3_1_6, 1_1_8, 4_2_8_4, 7_1_4_8, 1_2_4_3_0, 1_4_3, 1_4_0_0, 2_5_7_0_3, 1_5_8, 1_1_1, 4_2_8_4, 7_1_4_8, 1_1_7_7_2, 1_4_3, 2_1_2_9_7, 1_0_6_4, 1_5_8, 1_2_2, 2_0_4, 3_5_0_6, 1_7_5_4, 1_1_3_3, 1_4_7_8_7, 1_5_8_1, 1_1_5, 3_3_2_2_4, 4_4_8_2, 1_1_1, 1_3_5_5, 1_1_0, 2_9_1_7_3, 3_1_7, 5_0_8_3_3, 1_0_8, 2_0_1_4_7, 9_4_6_6_5, 1_1_1, 7_7_1_9_8, 1_0_7, 1], [1_1_0, 6_2_6_1_3, 1_1_7, 6_3_8, 1_1_2, 1_1_3_3, 1_2_1, 2_0_0_9_8, 1_3_5_5, 7_9_0_5_0, 1_3_8_7_2, 1_3_5, 1_5_9_6, 5_3_5_4_1, 1_3_5_2, 1_4_1, 1_3_0_3_9, 5_5_4_2, 1_2_4, 3_0_2, 5_1_8, 1_1_1, 2_6_8, 2_9_5_6, 1_1_5, 1_4_9, 4_4_2_7, 1_0_7, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_3_9, 1_2_3_5, 2_7_9_9, 1_8_2_8_9, 1_7_7_8_0, 2_0_4, 1_0_9, 9_4_7_4, 1_2_9_6, 1_0_7, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_a , model_name='google/bigbird-pegasus-large-arxiv' , revision='ba85d0851d708441f91440d509690f1ab6353415' , ) @require_sentencepiece @require_tokenizers class UpperCamelCase__ ( lowercase__, unittest.TestCase ): """simple docstring""" _SCREAMING_SNAKE_CASE = PegasusTokenizer _SCREAMING_SNAKE_CASE = PegasusTokenizerFast _SCREAMING_SNAKE_CASE = True _SCREAMING_SNAKE_CASE = True def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase_ : Tuple = PegasusTokenizer(_a , offset=0 , mask_token_sent=_a , mask_token='[MASK]' ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): return PegasusTokenizer.from_pretrained('google/bigbird-pegasus-large-arxiv' ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] , **SCREAMING_SNAKE_CASE_ : Optional[Any] ): return PegasusTokenizer.from_pretrained(self.tmpdirname , **_a ) def SCREAMING_SNAKE_CASE__ ( self : Dict , SCREAMING_SNAKE_CASE_ : Union[str, Any] ): return ("This is a test", "This is a test") def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): lowerCAmelCase_ : str = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) lowerCAmelCase_ : List[str] = self.tokenizer_class.from_pretrained(self.tmpdirname ) lowerCAmelCase_ : Optional[int] = ( 'Let\'s see which <unk> is the better <unk_token> one [MASK] It seems like this [MASK] was important </s>' ' <pad> <pad> <pad>' ) lowerCAmelCase_ : int = rust_tokenizer([raw_input_str] , return_tensors=_a , add_special_tokens=_a ).input_ids[0] lowerCAmelCase_ : List[Any] = py_tokenizer([raw_input_str] , return_tensors=_a , add_special_tokens=_a ).input_ids[0] self.assertListEqual(_a , _a ) @require_torch def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): lowerCAmelCase_ : Tuple = ['This is going to be way too long.' * 1_0_0_0, 'short example'] lowerCAmelCase_ : Optional[Any] = ['not super long but more than 5 tokens', 'tiny'] lowerCAmelCase_ : int = self._large_tokenizer(_a , padding=_a , truncation=_a , return_tensors='pt' ) lowerCAmelCase_ : str = self._large_tokenizer( text_target=_a , max_length=5 , padding=_a , truncation=_a , return_tensors='pt' ) assert batch.input_ids.shape == (2, 4_0_9_6) assert batch.attention_mask.shape == (2, 4_0_9_6) assert targets["input_ids"].shape == (2, 5) assert len(_a ) == 2 # input_ids, attention_mask. def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): lowerCAmelCase_ : Optional[int] = ( 'This is an example string that is used to test the original TF implementation against the HF' ' implementation' ) lowerCAmelCase_ : Optional[Any] = self._large_tokenizer(_a ).input_ids self.assertListEqual( _a , [1_8_2, 1_1_7, 1_4_2, 5_8_7, 4_2_1_1, 1_2_0, 1_1_7, 2_6_3, 1_1_2, 8_0_4, 1_0_9, 8_5_6, 2_5_0_1_6, 3_1_3_7, 4_6_4, 1_0_9, 2_6_9_5_5, 3_1_3_7, 1] , )
224
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __lowerCAmelCase = TypeVar("""T""") class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Tuple ,_a : T ): '''simple docstring''' _a : List[str] = data _a : Node[T] | None = None def __str__( self : Dict ): '''simple docstring''' return F"""{self.data}""" class UpperCAmelCase__ ( Generic[T] ): """simple docstring""" def __init__( self : Optional[int] ): '''simple docstring''' _a : Node[T] | None = None def __iter__( self : str ): '''simple docstring''' _a : Tuple = self.top while node: yield node.data _a : int = node.next def __str__( self : str ): '''simple docstring''' return "->".join([str(_a ) for item in self] ) def __len__( self : Optional[Any] ): '''simple docstring''' return len(tuple(iter(self ) ) ) def __lowercase ( self : str ): '''simple docstring''' return self.top is None def __lowercase ( self : List[Any] ,_a : T ): '''simple docstring''' _a : int = Node(_a ) if not self.is_empty(): _a : Optional[Any] = self.top _a : List[str] = node def __lowercase ( self : Tuple ): '''simple docstring''' if self.is_empty(): raise IndexError('pop from empty stack' ) assert isinstance(self.top ,_a ) _a : List[Any] = self.top _a : int = self.top.next return pop_node.data def __lowercase ( self : List[str] ): '''simple docstring''' if self.is_empty(): raise IndexError('peek from empty stack' ) assert self.top is not None return self.top.data def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = None if __name__ == "__main__": from doctest import testmod testmod()
271
0
from ...processing_utils import ProcessorMixin class UpperCAmelCase_ ( lowercase__ ): '''simple docstring''' __A : int = ['''image_processor''', '''feature_extractor'''] __A : Optional[int] = '''TvltImageProcessor''' __A : Optional[int] = '''TvltFeatureExtractor''' def __init__( self , __A , __A ): """simple docstring""" super().__init__(image_processor=_a , feature_extractor=_a ) lowerCamelCase : List[str] = image_processor lowerCamelCase : Optional[int] = feature_extractor def __call__( self , __A=None , __A=None , __A=None , __A=None , __A=False , __A=False , *__A , **__A , ): """simple docstring""" if images is None and audio is None: raise ValueError("You need to specify either an `images` or `audio` input to process." ) lowerCamelCase : Optional[Any] = None if images is not None: lowerCamelCase : Optional[Any] = self.image_processor(_a , mask_pixel=_a , *_a , **_a ) if images_mixed is not None: lowerCamelCase : str = self.image_processor(_a , is_mixed=_a , *_a , **_a ) if audio is not None: lowerCamelCase : str = self.feature_extractor( _a , *_a , sampling_rate=_a , mask_audio=_a , **_a ) lowerCamelCase : List[str] = {} if audio is not None: output_dict.update(_a ) if images is not None: output_dict.update(_a ) if images_mixed_dict is not None: output_dict.update(_a ) return output_dict @property def _snake_case ( self ): """simple docstring""" lowerCamelCase : List[str] = self.image_processor.model_input_names lowerCamelCase : List[str] = self.feature_extractor.model_input_names return list(dict.fromkeys(image_processor_input_names + feature_extractor_input_names ) )
283
'''simple docstring''' import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : List[str] ): '''simple docstring''' torch.manual_seed(0 ) _a : int = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : str = self.dummy_uncond_unet _a : int = PNDMScheduler() _a : str = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Any = pndm(generator=_a ,num_inference_steps=20 ,output_type='numpy' ,return_dict=_a )[0] _a : List[Any] = image[0, -3:, -3:, -1] _a : Any = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : List[Any] = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[str] = 'google/ddpm-cifar10-32' _a : str = UNetaDModel.from_pretrained(_a ) _a : Union[str, Any] = PNDMScheduler() _a : Tuple = PNDMPipeline(unet=_a ,scheduler=_a ) pndm.to(_a ) pndm.set_progress_bar_config(disable=_a ) _a : str = torch.manual_seed(0 ) _a : Optional[Any] = pndm(generator=_a ,output_type='numpy' ).images _a : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) _a : Tuple = np.array([0.1564, 0.1_4645, 0.1406, 0.1_4715, 0.1_2425, 0.1_4045, 0.1_3115, 0.1_2175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
271
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _A = { '''configuration_rag''': ['''RagConfig'''], '''retrieval_rag''': ['''RagRetriever'''], '''tokenization_rag''': ['''RagTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ '''RagModel''', '''RagPreTrainedModel''', '''RagSequenceForGeneration''', '''RagTokenForGeneration''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ '''TFRagModel''', '''TFRagPreTrainedModel''', '''TFRagSequenceForGeneration''', '''TFRagTokenForGeneration''', ] if TYPE_CHECKING: from .configuration_rag import RagConfig from .retrieval_rag import RagRetriever from .tokenization_rag import RagTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rag import RagModel, RagPreTrainedModel, RagSequenceForGeneration, RagTokenForGeneration try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rag import ( TFRagModel, TFRagPreTrainedModel, TFRagSequenceForGeneration, TFRagTokenForGeneration, ) else: import sys _A = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
122
'''simple docstring''' import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __lowerCAmelCase = logging.getLogger() @unittest.skip('''Temporarily disable the doc tests.''' ) @require_torch @require_tf @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : str ,_a : Path ,_a : Union[str, None] = None ,_a : Union[List[str], None] = None ,_a : Union[str, List[str], None] = None ,_a : bool = True ,): '''simple docstring''' _a : Optional[int] = [file for file in os.listdir(_a ) if os.path.isfile(os.path.join(_a ,_a ) )] if identifier is not None: _a : List[str] = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(_a ,_a ): for n_ in n_identifier: _a : Tuple = [file for file in files if n_ not in file] else: _a : Optional[Any] = [file for file in files if n_identifier not in file] _a : List[str] = ignore_files or [] ignore_files.append('__init__.py' ) _a : Tuple = [file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' ,_a ) if only_modules: _a : Any = file.split('.' )[0] try: _a : List[str] = getattr(_a ,_a ) _a : int = doctest.DocTestSuite(_a ) _a : Any = unittest.TextTestRunner().run(_a ) self.assertIs(len(result.failures ) ,0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: _a : Union[str, Any] = doctest.testfile(str('..' / directory / file ) ,optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed ,0 ) def __lowercase ( self : Any ): '''simple docstring''' _a : int = Path('src/transformers' ) _a : List[Any] = 'modeling' _a : Optional[Any] = [ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(_a ,identifier=_a ,ignore_files=_a ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Optional[Any] = Path('src/transformers' ) _a : Optional[Any] = 'tokenization' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Dict = Path('src/transformers' ) _a : str = 'configuration' self.analyze_directory(_a ,identifier=_a ) def __lowercase ( self : Any ): '''simple docstring''' _a : Tuple = Path('src/transformers' ) _a : List[Any] = ['configuration', 'modeling', 'tokenization'] self.analyze_directory(_a ,n_identifier=_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : List[Any] = Path('docs/source' ) _a : List[str] = ['favicon.ico'] self.analyze_directory(_a ,ignore_files=_a ,only_modules=_a )
271
0
'''simple docstring''' import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _a : def __init__( self : Optional[Any] , lowercase : Tuple , lowercase : Union[str, Any]=13 , lowercase : int=32 , lowercase : Optional[Any]=2 , lowercase : Union[str, Any]=3 , lowercase : Union[str, Any]=16 , lowercase : List[str]=[1, 2, 1] , lowercase : Dict=[2, 2, 4] , lowercase : Optional[Any]=2 , lowercase : Dict=2.0 , lowercase : Optional[int]=True , lowercase : List[str]=0.0 , lowercase : Tuple=0.0 , lowercase : Union[str, Any]=0.1 , lowercase : Union[str, Any]="gelu" , lowercase : Any=False , lowercase : Optional[int]=True , lowercase : Optional[Any]=0.02 , lowercase : Optional[int]=1E-5 , lowercase : Optional[int]=True , lowercase : Optional[int]=None , lowercase : List[Any]=True , lowercase : int=10 , lowercase : Optional[int]=8 , ): '''simple docstring''' UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = embed_dim UpperCAmelCase = depths UpperCAmelCase = num_heads UpperCAmelCase = window_size UpperCAmelCase = mlp_ratio UpperCAmelCase = qkv_bias UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = drop_path_rate UpperCAmelCase = hidden_act UpperCAmelCase = use_absolute_embeddings UpperCAmelCase = patch_norm UpperCAmelCase = layer_norm_eps UpperCAmelCase = initializer_range UpperCAmelCase = is_training UpperCAmelCase = scope UpperCAmelCase = use_labels UpperCAmelCase = type_sequence_label_size UpperCAmelCase = encoder_stride def A ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def A ( self : List[str] ): '''simple docstring''' return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def A ( self : Tuple , lowercase : str , lowercase : Tuple , lowercase : Dict ): '''simple docstring''' UpperCAmelCase = SwinvaModel(config=_a ) model.to(_a ) model.eval() UpperCAmelCase = model(_a ) UpperCAmelCase = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) UpperCAmelCase = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def A ( self : Tuple , lowercase : Union[str, Any] , lowercase : Dict , lowercase : int ): '''simple docstring''' UpperCAmelCase = SwinvaForMaskedImageModeling(config=_a ) model.to(_a ) model.eval() UpperCAmelCase = model(_a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase = 1 UpperCAmelCase = SwinvaForMaskedImageModeling(_a ) model.to(_a ) model.eval() UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase = model(_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def A ( self : Tuple , lowercase : Optional[Any] , lowercase : Optional[int] , lowercase : str ): '''simple docstring''' UpperCAmelCase = self.type_sequence_label_size UpperCAmelCase = SwinvaForImageClassification(_a ) model.to(_a ) model.eval() UpperCAmelCase = model(_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def A ( self : str ): '''simple docstring''' UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase = config_and_inputs UpperCAmelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class _a ( lowercase__ , lowercase__ , unittest.TestCase ): __a : Optional[int] = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) __a : Tuple = ( {'''feature-extraction''': SwinvaModel, '''image-classification''': SwinvaForImageClassification} if is_torch_available() else {} ) __a : Optional[Any] = False __a : List[Any] = False __a : List[str] = False __a : Tuple = False def A ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase = SwinvaModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=_a , embed_dim=37 ) def A ( self : List[str] ): '''simple docstring''' self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A ( self : Tuple ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) @unittest.skip(reason='''Got `CUDA error: misaligned address` with PyTorch 2.0.0.''' ) def A ( self : Optional[int] ): '''simple docstring''' pass @unittest.skip(reason='''Swinv2 does not use inputs_embeds''' ) def A ( self : Dict ): '''simple docstring''' pass def A ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(_a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_a , nn.Linear ) ) def A ( self : Any ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(_a ) UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , _a ) def A ( self : List[Any] ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase = True for model_class in self.all_model_classes: UpperCAmelCase = True UpperCAmelCase = False UpperCAmelCase = True UpperCAmelCase = model_class(_a ) model.to(_a ) model.eval() with torch.no_grad(): UpperCAmelCase = model(**self._prepare_for_class(_a , _a ) ) UpperCAmelCase = outputs.attentions UpperCAmelCase = len(self.model_tester.depths ) self.assertEqual(len(_a ) , _a ) # check that output_attentions also work using config del inputs_dict["output_attentions"] UpperCAmelCase = True UpperCAmelCase = config.window_size**2 UpperCAmelCase = model_class(_a ) model.to(_a ) model.eval() with torch.no_grad(): UpperCAmelCase = model(**self._prepare_for_class(_a , _a ) ) UpperCAmelCase = outputs.attentions self.assertEqual(len(_a ) , _a ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) UpperCAmelCase = len(_a ) # Check attention is always last and order is fine UpperCAmelCase = True UpperCAmelCase = True UpperCAmelCase = model_class(_a ) model.to(_a ) model.eval() with torch.no_grad(): UpperCAmelCase = model(**self._prepare_for_class(_a , _a ) ) if hasattr(self.model_tester , '''num_hidden_states_types''' ): UpperCAmelCase = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states UpperCAmelCase = 2 self.assertEqual(out_len + added_hidden_states , len(_a ) ) UpperCAmelCase = outputs.attentions self.assertEqual(len(_a ) , _a ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def A ( self : Dict , lowercase : List[str] , lowercase : Any , lowercase : List[str] , lowercase : Optional[int] ): '''simple docstring''' UpperCAmelCase = model_class(_a ) model.to(_a ) model.eval() with torch.no_grad(): UpperCAmelCase = model(**self._prepare_for_class(_a , _a ) ) UpperCAmelCase = outputs.hidden_states UpperCAmelCase = getattr( self.model_tester , '''expected_num_hidden_layers''' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(_a ) , _a ) # Swinv2 has a different seq_length UpperCAmelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) UpperCAmelCase = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) UpperCAmelCase = outputs.reshaped_hidden_states self.assertEqual(len(_a ) , _a ) UpperCAmelCase = reshaped_hidden_states[0].shape UpperCAmelCase = ( reshaped_hidden_states[0].view(_a , _a , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def A ( self : Tuple ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: UpperCAmelCase = True self.check_hidden_states_output(_a , _a , _a , _a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase = True self.check_hidden_states_output(_a , _a , _a , _a ) def A ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase = 3 UpperCAmelCase = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) UpperCAmelCase = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) UpperCAmelCase = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) UpperCAmelCase = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: UpperCAmelCase = True self.check_hidden_states_output(_a , _a , _a , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase = True self.check_hidden_states_output(_a , _a , _a , (padded_height, padded_width) ) def A ( self : Dict ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_a ) def A ( self : List[str] ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_a ) @slow def A ( self : Optional[Any] ): '''simple docstring''' for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = SwinvaModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def A ( self : Any ): '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase = _config_zero_init(_a ) for model_class in self.all_model_classes: UpperCAmelCase = model_class(config=_a ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f"Parameter {name} of model {model_class} seems not properly initialized" , ) @require_vision @require_torch class _a ( unittest.TestCase ): @cached_property def A ( self : Dict ): '''simple docstring''' return ( AutoImageProcessor.from_pretrained('''microsoft/swinv2-tiny-patch4-window8-256''' ) if is_vision_available() else None ) @slow def A ( self : List[Any] ): '''simple docstring''' UpperCAmelCase = SwinvaForImageClassification.from_pretrained('''microsoft/swinv2-tiny-patch4-window8-256''' ).to( _a ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) UpperCAmelCase = image_processor(images=_a , return_tensors='''pt''' ).to(_a ) # forward pass with torch.no_grad(): UpperCAmelCase = model(**_a ) # verify the logits UpperCAmelCase = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , _a ) UpperCAmelCase = torch.tensor([-0.3947, -0.4306, 0.0026] ).to(_a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _a , atol=1E-4 ) )
34
'''simple docstring''' import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def UpperCAmelCase_ (__a : Optional[Any] , __a : str , __a : Optional[Any]=None ): """simple docstring""" assert torch_layer.weight.shape == weight.shape, f"""{torch_layer} layer.weight does not match""" _a : str = nn.Parameter(__a ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f"""{torch_layer} layer.bias does not match""" _a : Any = nn.Parameter(__a ) def UpperCAmelCase_ (__a : int , __a : Optional[Any] , __a : int ): """simple docstring""" _a : Tuple = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : Dict = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Optional[Any] , __a : Optional[int] , __a : List[str] ): """simple docstring""" _a : Dict = np.asarray(weights[0] ) _a : Union[str, Any] = np.asarray(weights[1] ) _a : str = np.asarray(weights[2] ) _a : int = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.key , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__a ).transpose(1 , 2 ).contiguous().view(-1 , __a ) , ) set_param( torch_layer.output.dense , torch.tensor(__a ).view(-1 , __a ).contiguous().transpose(0 , 1 ) , ) def UpperCAmelCase_ (__a : Any , __a : Any , __a : Optional[Any] ): """simple docstring""" _a : List[str] = weights[0][0][0] _a : List[Any] = np.asarray(layer_norm_a[0] ) _a : List[str] = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # lsh weights + output _a : List[str] = weights[0][1] if len(__a ) < 4: set_layer_weights_in_torch_lsh(__a , torch_block.attention , __a ) else: set_layer_weights_in_torch_local(__a , torch_block.attention , __a ) # intermediate weighs _a : Optional[Any] = weights[2][0][1][2] # Chunked Feed Forward if len(__a ) == 4: _a : Union[str, Any] = intermediate_weights[2] # layernorm 2 _a : Any = np.asarray(intermediate_weights[0][0] ) _a : List[Any] = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # intermediate dense _a : Any = np.asarray(intermediate_weights[1][0] ) _a : Any = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) # intermediate out _a : Optional[int] = np.asarray(intermediate_weights[4][0] ) _a : int = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Dict , __a : Dict , __a : List[Any] ): """simple docstring""" _a : Optional[int] = torch_model.reformer # word embeds _a : Tuple = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(__a ) , ) if isinstance(weights[3] , __a ): _a : Any = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _a : List[Any] = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f"""{position_embeddings[emb_idx]} emb does not match""" _a : Any = nn.Parameter(torch.tensor(__a ) ) _a : List[str] = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( __a ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _a : Tuple = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(__a , __a , __a ) # output layer norm _a : Optional[Any] = np.asarray(weights[7][0] ) _a : int = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(__a ) , torch.tensor(__a ) , ) # output embeddings _a : List[str] = np.asarray(weights[9][0] ) _a : int = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(__a ).transpose(0 , 1 ).contiguous() , torch.tensor(__a ) , ) def UpperCAmelCase_ (__a : Tuple , __a : Optional[Any] , __a : Dict ): """simple docstring""" _a : List[Any] = ReformerConfig.from_json_file(__a ) print(f"""Building PyTorch model from configuration: {config}""" ) _a : int = ReformerModelWithLMHead(__a ) with open(__a , 'rb' ) as f: _a : Optional[Any] = pickle.load(__a )['weights'] set_model_weights_in_torch(__a , __a , config.hidden_size ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , __a ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--trax_model_pkl_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained Reformer model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __lowerCAmelCase = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
271
0
import logging import os from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union from filelock import FileLock from transformers import PreTrainedTokenizer, is_tf_available, is_torch_available __UpperCamelCase : Tuple = logging.getLogger(__name__) @dataclass class __SCREAMING_SNAKE_CASE: _UpperCAmelCase = 42 _UpperCAmelCase = 42 _UpperCAmelCase = 42 @dataclass class __SCREAMING_SNAKE_CASE: _UpperCAmelCase = 42 _UpperCAmelCase = 42 _UpperCAmelCase = None _UpperCAmelCase = None class __SCREAMING_SNAKE_CASE( lowercase__ ): _UpperCAmelCase = '''train''' _UpperCAmelCase = '''dev''' _UpperCAmelCase = '''test''' class __SCREAMING_SNAKE_CASE: @staticmethod def lowerCAmelCase_ ( UpperCamelCase: int , UpperCamelCase: Union[Split, str] ) -> str: raise NotImplementedError @staticmethod def lowerCAmelCase_ ( UpperCamelCase: str ) -> Any: raise NotImplementedError @staticmethod def lowerCAmelCase_ ( UpperCamelCase: List[InputExample] , UpperCamelCase: List[str] , UpperCamelCase: int , UpperCamelCase: PreTrainedTokenizer , UpperCamelCase: List[str]=False , UpperCamelCase: Tuple="[CLS]" , UpperCamelCase: Optional[int]=1 , UpperCamelCase: List[Any]="[SEP]" , UpperCamelCase: List[str]=False , UpperCamelCase: Optional[int]=False , UpperCamelCase: Optional[Any]=0 , UpperCamelCase: List[str]=0 , UpperCamelCase: Tuple=-1_00 , UpperCamelCase: Optional[int]=0 , UpperCamelCase: int=True , ) -> Any: snake_case__ = {label: i for i, label in enumerate(_a )} snake_case__ = [] for ex_index, example in enumerate(_a ): if ex_index % 1_00_00 == 0: logger.info('Writing example %d of %d' , _a , len(_a ) ) snake_case__ = [] snake_case__ = [] for word, label in zip(example.words , example.labels ): snake_case__ = tokenizer.tokenize(_a ) # bert-base-multilingual-cased sometimes output "nothing ([]) when calling tokenize with just a space. if len(_a ) > 0: tokens.extend(_a ) # Use the real label id for the first token of the word, and padding ids for the remaining tokens label_ids.extend([label_map[label]] + [pad_token_label_id] * (len(_a ) - 1) ) # Account for [CLS] and [SEP] with "- 2" and with "- 3" for RoBERTa. snake_case__ = tokenizer.num_special_tokens_to_add() if len(_a ) > max_seq_length - special_tokens_count: snake_case__ = tokens[: (max_seq_length - special_tokens_count)] snake_case__ = label_ids[: (max_seq_length - special_tokens_count)] # The convention in BERT is: # (a) For sequence pairs: # tokens: [CLS] is this jack ##son ##ville ? [SEP] no it is not . [SEP] # type_ids: 0 0 0 0 0 0 0 0 1 1 1 1 1 1 # (b) For single sequences: # tokens: [CLS] the dog is hairy . [SEP] # type_ids: 0 0 0 0 0 0 0 # # Where "type_ids" are used to indicate whether this is the first # sequence or the second sequence. The embedding vectors for `type=0` and # `type=1` were learned during pre-training and are added to the wordpiece # embedding vector (and position vector). This is not *strictly* necessary # since the [SEP] token unambiguously separates the sequences, but it makes # it easier for the model to learn the concept of sequences. # # For classification tasks, the first vector (corresponding to [CLS]) is # used as the "sentence vector". Note that this only makes sense because # the entire model is fine-tuned. tokens += [sep_token] label_ids += [pad_token_label_id] if sep_token_extra: # roberta uses an extra separator b/w pairs of sentences tokens += [sep_token] label_ids += [pad_token_label_id] snake_case__ = [sequence_a_segment_id] * len(_a ) if cls_token_at_end: tokens += [cls_token] label_ids += [pad_token_label_id] segment_ids += [cls_token_segment_id] else: snake_case__ = [cls_token] + tokens snake_case__ = [pad_token_label_id] + label_ids snake_case__ = [cls_token_segment_id] + segment_ids snake_case__ = tokenizer.convert_tokens_to_ids(_a ) # The mask has 1 for real tokens and 0 for padding tokens. Only real # tokens are attended to. snake_case__ = [1 if mask_padding_with_zero else 0] * len(_a ) # Zero-pad up to the sequence length. snake_case__ = max_seq_length - len(_a ) if pad_on_left: snake_case__ = ([pad_token] * padding_length) + input_ids snake_case__ = ([0 if mask_padding_with_zero else 1] * padding_length) + input_mask snake_case__ = ([pad_token_segment_id] * padding_length) + segment_ids snake_case__ = ([pad_token_label_id] * padding_length) + label_ids else: input_ids += [pad_token] * padding_length input_mask += [0 if mask_padding_with_zero else 1] * padding_length segment_ids += [pad_token_segment_id] * padding_length label_ids += [pad_token_label_id] * padding_length assert len(_a ) == max_seq_length assert len(_a ) == max_seq_length assert len(_a ) == max_seq_length assert len(_a ) == max_seq_length if ex_index < 5: logger.info('*** Example ***' ) logger.info('guid: %s' , example.guid ) logger.info('tokens: %s' , ' '.join([str(_a ) for x in tokens] ) ) logger.info('input_ids: %s' , ' '.join([str(_a ) for x in input_ids] ) ) logger.info('input_mask: %s' , ' '.join([str(_a ) for x in input_mask] ) ) logger.info('segment_ids: %s' , ' '.join([str(_a ) for x in segment_ids] ) ) logger.info('label_ids: %s' , ' '.join([str(_a ) for x in label_ids] ) ) if "token_type_ids" not in tokenizer.model_input_names: snake_case__ = None features.append( InputFeatures( input_ids=_a , attention_mask=_a , token_type_ids=_a , label_ids=_a ) ) return features if is_torch_available(): import torch from torch import nn from torch.utils.data import Dataset class __SCREAMING_SNAKE_CASE( lowercase__ ): _UpperCAmelCase = 42 _UpperCAmelCase = nn.CrossEntropyLoss().ignore_index def __init__( self: int , UpperCamelCase: TokenClassificationTask , UpperCamelCase: str , UpperCamelCase: PreTrainedTokenizer , UpperCamelCase: List[str] , UpperCamelCase: str , UpperCamelCase: Optional[int] = None , UpperCamelCase: Any=False , UpperCamelCase: Split = Split.train , ) -> Optional[Any]: snake_case__ = os.path.join( _a , 'cached_{}_{}_{}'.format(mode.value , tokenizer.__class__.__name__ , str(_a ) ) , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. snake_case__ = cached_features_file + '.lock' with FileLock(_a ): if os.path.exists(_a ) and not overwrite_cache: logger.info(F'''Loading features from cached file {cached_features_file}''' ) snake_case__ = torch.load(_a ) else: logger.info(F'''Creating features from dataset file at {data_dir}''' ) snake_case__ = token_classification_task.read_examples_from_file(_a , _a ) # TODO clean up all this to leverage built-in features of tokenizers snake_case__ = token_classification_task.convert_examples_to_features( _a , _a , _a , _a , cls_token_at_end=bool(model_type in ['xlnet'] ) , cls_token=tokenizer.cls_token , cls_token_segment_id=2 if model_type in ['xlnet'] else 0 , sep_token=tokenizer.sep_token , sep_token_extra=_a , pad_on_left=bool(tokenizer.padding_side == 'left' ) , pad_token=tokenizer.pad_token_id , pad_token_segment_id=tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info(F'''Saving features into cached file {cached_features_file}''' ) torch.save(self.features , _a ) def __len__( self: List[Any] ) -> str: return len(self.features ) def __getitem__( self: Tuple , UpperCamelCase: Any ) -> Any: return self.features[i] if is_tf_available(): import tensorflow as tf class __SCREAMING_SNAKE_CASE: _UpperCAmelCase = 42 _UpperCAmelCase = -1_0_0 def __init__( self: Tuple , UpperCamelCase: TokenClassificationTask , UpperCamelCase: str , UpperCamelCase: PreTrainedTokenizer , UpperCamelCase: List[str] , UpperCamelCase: str , UpperCamelCase: Optional[int] = None , UpperCamelCase: Union[str, Any]=False , UpperCamelCase: Split = Split.train , ) -> int: snake_case__ = token_classification_task.read_examples_from_file(_a , _a ) # TODO clean up all this to leverage built-in features of tokenizers snake_case__ = token_classification_task.convert_examples_to_features( _a , _a , _a , _a , cls_token_at_end=bool(model_type in ['xlnet'] ) , cls_token=tokenizer.cls_token , cls_token_segment_id=2 if model_type in ['xlnet'] else 0 , sep_token=tokenizer.sep_token , sep_token_extra=_a , pad_on_left=bool(tokenizer.padding_side == 'left' ) , pad_token=tokenizer.pad_token_id , pad_token_segment_id=tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) def gen(): for ex in self.features: if ex.token_type_ids is None: yield ( {"input_ids": ex.input_ids, "attention_mask": ex.attention_mask}, ex.label_ids, ) else: yield ( { "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label_ids, ) if "token_type_ids" not in tokenizer.model_input_names: snake_case__ = tf.data.Dataset.from_generator( _a , ({'input_ids': tf.intaa, 'attention_mask': tf.intaa}, tf.intaa) , ( {'input_ids': tf.TensorShape([None] ), 'attention_mask': tf.TensorShape([None] )}, tf.TensorShape([None] ), ) , ) else: snake_case__ = tf.data.Dataset.from_generator( _a , ({'input_ids': tf.intaa, 'attention_mask': tf.intaa, 'token_type_ids': tf.intaa}, tf.intaa) , ( { 'input_ids': tf.TensorShape([None] ), 'attention_mask': tf.TensorShape([None] ), 'token_type_ids': tf.TensorShape([None] ), }, tf.TensorShape([None] ), ) , ) def lowerCAmelCase_ ( self: int ) -> Optional[int]: snake_case__ = self.dataset.apply(tf.data.experimental.assert_cardinality(len(self.features ) ) ) return self.dataset def __len__( self: Optional[int] ) -> Any: return len(self.features ) def __getitem__( self: Any , UpperCamelCase: Any ) -> Optional[Any]: return self.features[i]
307
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @property def __lowercase ( self : Union[str, Any] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Union[str, Any] = VQModel( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=3 ,) return model @property def __lowercase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) _a : Any = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1000 ,) return CLIPTextModel(_a ) def __lowercase ( self : Tuple ): '''simple docstring''' _a : Dict = self.dummy_uncond_unet _a : List[Any] = DDIMScheduler() _a : List[Any] = self.dummy_vq_model _a : str = LDMPipeline(unet=_a ,vqvae=_a ,scheduler=_a ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : List[str] = torch.manual_seed(0 ) _a : List[str] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ).images _a : List[str] = torch.manual_seed(0 ) _a : Union[str, Any] = ldm(generator=_a ,num_inference_steps=2 ,output_type='numpy' ,return_dict=_a )[0] _a : Tuple = image[0, -3:, -3:, -1] _a : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _a : int = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) _a : Any = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self : Optional[Any] ): '''simple docstring''' _a : List[str] = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(_a ) ldm.set_progress_bar_config(disable=_a ) _a : Optional[int] = torch.manual_seed(0 ) _a : Dict = ldm(generator=_a ,num_inference_steps=5 ,output_type='numpy' ).images _a : str = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _a : Optional[int] = np.array([0.4399, 0.4_4975, 0.4_6825, 0.474, 0.4359, 0.4581, 0.4_5095, 0.4341, 0.4447] ) _a : int = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
271
0
import math import tensorflow as tf from packaging import version def lowerCAmelCase_ (lowerCAmelCase__: Tuple ): """simple docstring""" UpperCAmelCase_: Optional[int] = tf.convert_to_tensor(__a ) UpperCAmelCase_: Any = 0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ) , x.dtype ) )) return x * cdf def lowerCAmelCase_ (lowerCAmelCase__: Union[str, Any] ): """simple docstring""" UpperCAmelCase_: Any = tf.convert_to_tensor(__a ) UpperCAmelCase_: Optional[Any] = tf.cast(math.pi , x.dtype ) UpperCAmelCase_: Any = tf.cast(0.044715 , x.dtype ) UpperCAmelCase_: str = 0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(__a , 3 )) )) return x * cdf def lowerCAmelCase_ (lowerCAmelCase__: List[str] ): """simple docstring""" UpperCAmelCase_: List[Any] = tf.convert_to_tensor(__a ) return x * tf.tanh(tf.math.softplus(__a ) ) def lowerCAmelCase_ (lowerCAmelCase__: Optional[int] ): """simple docstring""" UpperCAmelCase_: Any = tf.convert_to_tensor(__a ) UpperCAmelCase_: int = tf.cast(0.044715 , x.dtype ) UpperCAmelCase_: str = tf.cast(0.7978845608 , x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def lowerCAmelCase_ (lowerCAmelCase__: str ): """simple docstring""" UpperCAmelCase_: Union[str, Any] = tf.convert_to_tensor(__a ) UpperCAmelCase_: List[str] = tf.cast(1.702 , x.dtype ) return x * tf.math.sigmoid(coeff * x ) def lowerCAmelCase_ (lowerCAmelCase__: List[str] ): """simple docstring""" return tf.clip_by_value(_gelu(__a ) , -1_0 , 1_0 ) def lowerCAmelCase_ (lowerCAmelCase__: Union[str, Any] , lowerCAmelCase__: Optional[Any]=-1 ): """simple docstring""" UpperCAmelCase_: int = tf.split(__a , 2 , axis=__a ) return a * tf.math.sigmoid(__a ) if version.parse(tf.version.VERSION) >= version.parse('2.4'): def lowerCAmelCase_ (lowerCAmelCase__: List[str] ): """simple docstring""" return tf.keras.activations.gelu(__a , approximate=__a ) a : List[Any] = tf.keras.activations.gelu a : Optional[int] = approximate_gelu_wrap else: a : Optional[int] = _gelu a : Tuple = _gelu_new a : Tuple = { 'gelu': gelu, 'gelu_10': gelu_aa, 'gelu_fast': gelu_fast, 'gelu_new': gelu_new, 'glu': glu, 'mish': mish, 'quick_gelu': quick_gelu, 'relu': tf.keras.activations.relu, 'sigmoid': tf.keras.activations.sigmoid, 'silu': tf.keras.activations.swish, 'swish': tf.keras.activations.swish, 'tanh': tf.keras.activations.tanh, } def lowerCAmelCase_ (lowerCAmelCase__: Any ): """simple docstring""" if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(F'function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}' )
147
'''simple docstring''' import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __lowerCAmelCase = logging.get_logger(__name__) class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : int ,*_a : Optional[int] ,**_a : str ): '''simple docstring''' warnings.warn( 'The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use BeitImageProcessor instead.' ,_a ,) super().__init__(*_a ,**_a )
271
0
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class _UpperCAmelCase ( lowercase__ , unittest.TestCase ): """simple docstring""" a_ = BarthezTokenizer a_ = BarthezTokenizerFast a_ = True a_ = True def lowercase ( self : Tuple ) -> List[Any]: super().setUp() __lowerCAmelCase = BarthezTokenizerFast.from_pretrained('moussaKam/mbarthez' ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=_a ) __lowerCAmelCase = tokenizer def lowercase ( self : List[Any] ) -> Optional[int]: __lowerCAmelCase = '<pad>' __lowerCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_a ) , _a ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_a ) , _a ) def lowercase ( self : List[str] ) -> Optional[Any]: __lowerCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(vocab_keys[-1] , '<mask>' ) self.assertEqual(len(_a ) , 1_0_1_1_2_2 ) def lowercase ( self : List[Any] ) -> Dict: self.assertEqual(self.get_tokenizer().vocab_size , 1_0_1_1_2_2 ) @require_torch def lowercase ( self : Optional[int] ) -> Optional[Any]: __lowerCAmelCase = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] __lowerCAmelCase = [0, 5_7, 3_0_1_8, 7_0_3_0_7, 9_1, 2] __lowerCAmelCase = self.tokenizer( _a , max_length=len(_a ) , padding=_a , truncation=_a , return_tensors='pt' ) self.assertIsInstance(_a , _a ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) __lowerCAmelCase = batch.input_ids.tolist()[0] self.assertListEqual(_a , _a ) def lowercase ( self : Optional[Any] ) -> str: if not self.test_rust_tokenizer: return __lowerCAmelCase = self.get_tokenizer() __lowerCAmelCase = self.get_rust_tokenizer() __lowerCAmelCase = 'I was born in 92000, and this is falsé.' __lowerCAmelCase = tokenizer.tokenize(_a ) __lowerCAmelCase = rust_tokenizer.tokenize(_a ) self.assertListEqual(_a , _a ) __lowerCAmelCase = tokenizer.encode(_a , add_special_tokens=_a ) __lowerCAmelCase = rust_tokenizer.encode(_a , add_special_tokens=_a ) self.assertListEqual(_a , _a ) __lowerCAmelCase = self.get_rust_tokenizer() __lowerCAmelCase = tokenizer.encode(_a ) __lowerCAmelCase = rust_tokenizer.encode(_a ) self.assertListEqual(_a , _a ) @slow def lowercase ( self : List[Any] ) -> Optional[Any]: __lowerCAmelCase = {'input_ids': [[0, 4_9_0, 1_4_3_2_8, 4_5_0_7, 3_5_4, 4_7, 4_3_6_6_9, 9_5, 2_5, 7_8_1_1_7, 2_0_2_1_5, 1_9_7_7_9, 1_9_0, 2_2, 4_0_0, 4, 3_5_3_4_3, 8_0_3_1_0, 6_0_3, 8_6, 2_4_9_3_7, 1_0_5, 3_3_4_3_8, 9_4_7_6_2, 1_9_6, 3_9_6_4_2, 7, 1_5, 1_5_9_3_3, 1_7_3, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_0_5_3_4, 8_7, 2_5, 6_6, 3_3_5_8, 1_9_6, 5_5_2_8_9, 8, 8_2_9_6_1, 8_1, 2_2_0_4, 7_5_2_0_3, 7, 1_5, 7_6_3, 1_2_9_5_6, 2_1_6, 1_7_8, 1_4_3_2_8, 9_5_9_5, 1_3_7_7, 6_9_6_9_3, 7, 4_4_8, 7_1_0_2_1, 1_9_6, 1_8_1_0_6, 1_4_3_7, 1_3_9_7_4, 1_0_8, 9_0_8_3, 4, 4_9_3_1_5, 7, 3_9, 8_6, 1_3_2_6, 2_7_9_3, 4_6_3_3_3, 4, 4_4_8, 1_9_6, 7_4_5_8_8, 7, 4_9_3_1_5, 7, 3_9, 2_1, 8_2_2, 3_8_4_7_0, 7_4, 2_1, 6_6_7_2_3, 6_2_4_8_0, 8, 2_2_0_5_0, 5, 2]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. __lowerCAmelCase = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=_a , model_name='moussaKam/mbarthez' , revision='c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6' , sequences=_a , )
284
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { """linear""": get_linear_schedule_with_warmup, """cosine""": get_cosine_schedule_with_warmup, """cosine_w_restarts""": get_cosine_with_hard_restarts_schedule_with_warmup, """polynomial""": get_polynomial_decay_schedule_with_warmup, """constant""": get_constant_schedule, """constant_w_warmup""": get_constant_schedule_with_warmup, } class UpperCAmelCase__ ( lowercase__ ): """simple docstring""" def __init__( self : Optional[int] ,_a : Optional[Any]=None ,_a : Dict=None ,*_a : int ,**_a : str ): '''simple docstring''' super().__init__(*_a ,**_a ) if config is None: assert isinstance(self.model ,_a ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F""" {self.model.__class__}""" ) _a : List[Any] = self.model.config else: _a : Optional[int] = config _a : List[str] = data_args _a : List[Any] = self.config.tgt_vocab_size if isinstance(self.config ,_a ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ' padding..' ) if self.args.label_smoothing == 0: _a : List[str] = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss _a : Tuple = label_smoothed_nll_loss def __lowercase ( self : List[str] ,_a : int ): '''simple docstring''' if self.optimizer is None: _a : Union[str, Any] = ['bias', 'LayerNorm.weight'] _a : Tuple = [ { 'params': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], 'weight_decay': self.args.weight_decay, }, { 'params': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], 'weight_decay': 0.0, }, ] _a : Optional[int] = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: _a : Any = Adafactor _a : Dict = {'scale_parameter': False, 'relative_step': False} else: _a : Union[str, Any] = AdamW _a : str = { 'betas': (self.args.adam_betaa, self.args.adam_betaa), 'eps': self.args.adam_epsilon, } _a : Union[str, Any] = self.args.learning_rate if self.sharded_ddp: _a : str = OSS( params=_a ,optim=_a ,**_a ,) else: _a : Tuple = optimizer_cls(_a ,**_a ) if self.lr_scheduler is None: _a : List[Any] = self._get_lr_scheduler(_a ) else: # ignoring --lr_scheduler logger.warning('scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.' ) def __lowercase ( self : List[Any] ,_a : List[Any] ): '''simple docstring''' _a : str = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": _a : int = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": _a : List[str] = schedule_func(self.optimizer ,num_warmup_steps=self.args.warmup_steps ) else: _a : Optional[int] = schedule_func( self.optimizer ,num_warmup_steps=self.args.warmup_steps ,num_training_steps=_a ) return scheduler def __lowercase ( self : Tuple ): '''simple docstring''' if isinstance(self.train_dataset ,torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size ,distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) ,) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def __lowercase ( self : Dict ,_a : Dict ,_a : Any ,_a : Dict ): '''simple docstring''' if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Union[str, Any] = self.loss_fn(logits.view(-1 ,logits.shape[-1] ) ,labels.view(-1 ) ) else: # compute usual loss via models _a, _a : Union[str, Any] = model(**_a ,labels=_a ,use_cache=_a )[:2] else: # compute label smoothed loss _a : List[Any] = model(**_a ,use_cache=_a )[0] _a : Any = torch.nn.functional.log_softmax(_a ,dim=-1 ) _a, _a : List[str] = self.loss_fn(_a ,_a ,self.args.label_smoothing ,ignore_index=self.config.pad_token_id ) return loss, logits def __lowercase ( self : Optional[int] ,_a : Union[str, Any] ,_a : List[Any] ): '''simple docstring''' _a : Optional[int] = inputs.pop('labels' ) _a, _a : int = self._compute_loss(_a ,_a ,_a ) return loss def __lowercase ( self : Optional[Any] ,_a : nn.Module ,_a : Dict[str, Union[torch.Tensor, Any]] ,_a : bool ,_a : Optional[List[str]] = None ,): '''simple docstring''' _a : int = self._prepare_inputs(_a ) _a : Any = { 'max_length': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, 'num_beams': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: _a : int = self.model.generate( inputs['input_ids'] ,attention_mask=inputs['attention_mask'] ,**_a ,) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: _a : int = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) _a : Union[str, Any] = inputs.pop('labels' ) with torch.no_grad(): # compute loss on predict data _a, _a : Optional[int] = self._compute_loss(_a ,_a ,_a ) _a : Optional[Any] = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) _a : Optional[Any] = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: _a : Dict = self._pad_tensors_to_max_len(_a ,gen_kwargs['max_length'] ) return (loss, logits, labels) def __lowercase ( self : str ,_a : Tuple ,_a : Tuple ): '''simple docstring''' _a : List[Any] = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( 'Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be' F""" padded to `max_length`={max_length}""" ) _a : int = pad_token_id * torch.ones( (tensor.shape[0], max_length) ,dtype=tensor.dtype ,device=tensor.device ) _a : Union[str, Any] = tensor return padded_tensor
271
0
import argparse from typing import List import evaluate import numpy as np import torch from datasets import DatasetDict, load_dataset # New Code # # We'll be using StratifiedKFold for this example from sklearn.model_selection import StratifiedKFold from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to perform Cross Validation, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## lowerCamelCase = 16 lowerCamelCase = 32 def lowerCamelCase_ ( _a , _a , _a , _a , _a = 16 ): """simple docstring""" lowerCAmelCase__ : Union[str, Any] = AutoTokenizer.from_pretrained('''bert-base-cased''' ) lowerCAmelCase__ : str = DatasetDict( { '''train''': dataset['''train'''].select(__a ), '''validation''': dataset['''train'''].select(__a ), '''test''': dataset['''validation'''], } ) def tokenize_function(_a ): # max_length=None => use the model max length (it's actually the default) lowerCAmelCase__ : Optional[int] = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=__a , max_length=__a ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowerCAmelCase__ : List[str] = datasets.map( __a , batched=__a , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowerCAmelCase__ : List[Any] = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(_a ): # On TPU it's best to pad everything to the same length or training will be very slow. lowerCAmelCase__ : Dict = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowerCAmelCase__ : Tuple = 16 elif accelerator.mixed_precision != "no": lowerCAmelCase__ : List[Any] = 8 else: lowerCAmelCase__ : List[Any] = None return tokenizer.pad( __a , padding='''longest''' , max_length=__a , pad_to_multiple_of=__a , return_tensors='''pt''' , ) # Instantiate dataloaders. lowerCAmelCase__ : Any = DataLoader( tokenized_datasets['''train'''] , shuffle=__a , collate_fn=__a , batch_size=__a ) lowerCAmelCase__ : Optional[int] = DataLoader( tokenized_datasets['''validation'''] , shuffle=__a , collate_fn=__a , batch_size=__a ) lowerCAmelCase__ : Optional[Any] = DataLoader( tokenized_datasets['''test'''] , shuffle=__a , collate_fn=__a , batch_size=__a ) return train_dataloader, eval_dataloader, test_dataloader def lowerCamelCase_ ( _a , _a ): """simple docstring""" lowerCAmelCase__ : Dict = [] # Download the dataset lowerCAmelCase__ : Tuple = load_dataset('''glue''' , '''mrpc''' ) # Create our splits lowerCAmelCase__ : Union[str, Any] = StratifiedKFold(n_splits=int(args.num_folds ) ) # Initialize accelerator lowerCAmelCase__ : Any = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowerCAmelCase__ : Optional[Any] = config['lr'] lowerCAmelCase__ : Optional[int] = int(config['''num_epochs'''] ) lowerCAmelCase__ : Dict = int(config['''seed'''] ) lowerCAmelCase__ : Dict = int(config['''batch_size'''] ) lowerCAmelCase__ : Optional[int] = evaluate.load('''glue''' , '''mrpc''' ) # If the batch size is too big we use gradient accumulation lowerCAmelCase__ : List[Any] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: lowerCAmelCase__ : Any = batch_size // MAX_GPU_BATCH_SIZE lowerCAmelCase__ : List[str] = MAX_GPU_BATCH_SIZE set_seed(__a ) # New Code # # Create our folds: lowerCAmelCase__ : int = kfold.split(np.zeros(datasets['''train'''].num_rows ) , datasets['''train''']['''label'''] ) lowerCAmelCase__ : Any = [] # Iterate over them for i, (train_idxs, valid_idxs) in enumerate(__a ): lowerCAmelCase__ : Optional[Any] = get_fold_dataloaders( __a , __a , __a , __a , ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowerCAmelCase__ : Dict = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=__a ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowerCAmelCase__ : List[Any] = model.to(accelerator.device ) # Instantiate optimizer lowerCAmelCase__ : List[str] = AdamW(params=model.parameters() , lr=__a ) # Instantiate scheduler lowerCAmelCase__ : List[Any] = get_linear_schedule_with_warmup( optimizer=__a , num_warmup_steps=100 , num_training_steps=(len(__a ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowerCAmelCase__ : Union[str, Any] = accelerator.prepare( __a , __a , __a , __a , __a ) # Now we train the model for epoch in range(__a ): model.train() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) lowerCAmelCase__ : Dict = model(**__a ) lowerCAmelCase__ : int = outputs.loss lowerCAmelCase__ : Any = loss / gradient_accumulation_steps accelerator.backward(__a ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowerCAmelCase__ : Union[str, Any] = model(**__a ) lowerCAmelCase__ : Tuple = outputs.logits.argmax(dim=-1 ) lowerCAmelCase__ : Any = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=__a , references=__a , ) lowerCAmelCase__ : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'epoch {epoch}:' , __a ) # New Code # # We also run predictions on the test set at the very end lowerCAmelCase__ : Any = [] for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowerCAmelCase__ : Tuple = model(**__a ) lowerCAmelCase__ : Dict = outputs.logits lowerCAmelCase__ : Optional[int] = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) fold_predictions.append(predictions.cpu() ) if i == 0: # We need all of the test predictions test_references.append(references.cpu() ) # Use accelerator.print to print only on the main process. test_predictions.append(torch.cat(__a , dim=0 ) ) # We now need to release all our memory and get rid of the current model, optimizer, etc accelerator.free_memory() # New Code # # Finally we check the accuracy of our folded results: lowerCAmelCase__ : Dict = torch.cat(__a , dim=0 ) lowerCAmelCase__ : Any = torch.stack(__a , dim=0 ).sum(dim=0 ).div(int(args.num_folds ) ).argmax(dim=-1 ) lowerCAmelCase__ : str = metric.compute(predictions=__a , references=__a ) accelerator.print('''Average test metrics from all folds:''' , __a ) def lowerCamelCase_ ( ): """simple docstring""" lowerCAmelCase__ : Any = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=__a , default=__a , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) # New Code # parser.add_argument('''--num_folds''' , type=__a , default=3 , help='''The number of splits to perform across the dataset''' ) lowerCAmelCase__ : Any = parser.parse_args() lowerCAmelCase__ : int = {'lr': 2e-5, 'num_epochs': 3, 'seed': 42, 'batch_size': 16} training_function(__a , __a ) if __name__ == "__main__": main()
131
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from transformers import AutoTokenizer, HfArgumentParser __lowerCAmelCase = re.compile(r"""\s+""") def UpperCAmelCase_ (__a : Any ): """simple docstring""" return {"hash": hashlib.mda(re.sub(__a , '' , example['content'] ).encode('utf-8' ) ).hexdigest()} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : List[str] = [len(__a ) for line in example['content'].splitlines()] return {"line_mean": np.mean(__a ), "line_max": max(__a )} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Union[str, Any] = np.mean([c.isalnum() for c in example['content']] ) return {"alpha_frac": alpha_frac} def UpperCAmelCase_ (__a : Optional[int] , __a : Any ): """simple docstring""" if example["hash"] in uniques: uniques.remove(example['hash'] ) return True else: return False def UpperCAmelCase_ (__a : int , __a : Union[str, Any]=5 ): """simple docstring""" _a : Optional[int] = ['auto-generated', 'autogenerated', 'automatically generated'] _a : List[str] = example['content'].splitlines() for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"autogenerated": True} else: return {"autogenerated": False} def UpperCAmelCase_ (__a : List[str] , __a : Dict=5 , __a : Tuple=0.05 ): """simple docstring""" _a : Optional[int] = ['unit tests', 'test file', 'configuration file'] _a : int = example['content'].splitlines() _a : int = 0 _a : Dict = 0 # first test for _, line in zip(range(__a ) , __a ): for keyword in keywords: if keyword in line.lower(): return {"config_or_test": True} # second test _a : int = example['content'].count('\n' ) _a : int = int(coeff * nlines ) for line in lines: count_config += line.lower().count('config' ) count_test += line.lower().count('test' ) if count_config > threshold or count_test > threshold: return {"config_or_test": True} return {"config_or_test": False} def UpperCAmelCase_ (__a : Optional[int] ): """simple docstring""" _a : List[str] = ['def ', 'class ', 'for ', 'while '] _a : str = example['content'].splitlines() for line in lines: for keyword in keywords: if keyword in line.lower(): return {"has_no_keywords": False} return {"has_no_keywords": True} def UpperCAmelCase_ (__a : int , __a : Any=4 ): """simple docstring""" _a : List[str] = example['content'].splitlines() _a : Dict = 0 for line in lines: counter += line.lower().count('=' ) if counter > minimum: return {"has_few_assignments": False} return {"has_few_assignments": True} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Optional[Any] = tokenizer(example['content'] , truncation=__a )['input_ids'] _a : Optional[int] = len(example['content'] ) / len(__a ) return {"ratio": ratio} def UpperCAmelCase_ (__a : str ): """simple docstring""" _a : Dict = {} results.update(get_hash(__a ) ) results.update(line_stats(__a ) ) results.update(alpha_stats(__a ) ) results.update(char_token_ratio(__a ) ) results.update(is_autogenerated(__a ) ) results.update(is_config_or_test(__a ) ) results.update(has_no_keywords(__a ) ) results.update(has_few_assignments(__a ) ) return results def UpperCAmelCase_ (__a : Any , __a : Any , __a : str ): """simple docstring""" if not check_uniques(__a , __a ): return False elif example["autogenerated"]: return False elif example["line_max"] > args.line_max: return False elif example["line_mean"] > args.line_mean: return False elif example["alpha_frac"] < args.alpha_frac: return False elif example["ratio"] < args.min_token_ratio: return False elif example["config_or_test"] and np.random.rand() <= args.filter_proba: return False elif example["has_no_keywords"] and np.random.rand() <= args.filter_proba: return False elif example["has_few_assignments"]: return False else: return True def UpperCAmelCase_ (__a : Union[str, Any] ): """simple docstring""" with open(__a , 'rb' ) as f_in: with gzip.open(str(__a ) + '.gz' , 'wb' , compresslevel=6 ) as f_out: shutil.copyfileobj(__a , __a ) os.unlink(__a ) # Settings __lowerCAmelCase = HfArgumentParser(PreprocessingArguments) __lowerCAmelCase = parser.parse_args() if args.num_workers is None: __lowerCAmelCase = multiprocessing.cpu_count() __lowerCAmelCase = AutoTokenizer.from_pretrained(args.tokenizer_dir) # Load dataset __lowerCAmelCase = time.time() __lowerCAmelCase = load_dataset(args.dataset_name, split="""train""") print(f'''Time to load dataset: {time.time()-t_start:.2f}''') # Run preprocessing __lowerCAmelCase = time.time() __lowerCAmelCase = ds.map(preprocess, num_proc=args.num_workers) print(f'''Time to preprocess dataset: {time.time()-t_start:.2f}''') # Deduplicate hashes __lowerCAmelCase = set(ds.unique("""hash""")) __lowerCAmelCase = len(uniques) / len(ds) print(f'''Fraction of duplicates: {1-frac:.2%}''') # Deduplicate data and apply heuristics __lowerCAmelCase = time.time() __lowerCAmelCase = ds.filter(filter, fn_kwargs={"""uniques""": uniques, """args""": args}) print(f'''Time to filter dataset: {time.time()-t_start:.2f}''') print(f'''Size of filtered dataset: {len(ds_filter)}''') # Deduplicate with minhash and jaccard similarity if args.near_deduplication: __lowerCAmelCase = time.time() __lowerCAmelCase , __lowerCAmelCase = deduplicate_dataset(ds_filter, args.jaccard_threshold) print(f'''Time to deduplicate dataset: {time.time()-t_start:.2f}''') print(f'''Size of deduplicate dataset: {len(ds_filter)}''') # Save data in batches of samples_per_file __lowerCAmelCase = Path(args.output_dir) output_dir.mkdir(exist_ok=True) # save duplicate_clusters in the output_dir as artifacts # not sure it is the right place the save it if args.near_deduplication: with open(output_dir / """duplicate_clusters.json""", """w""") as f: json.dump(duplicate_clusters, f) __lowerCAmelCase = output_dir / """data""" data_dir.mkdir(exist_ok=True) __lowerCAmelCase = time.time() for file_number, index in enumerate(range(0, len(ds_filter), args.samples_per_file)): __lowerCAmelCase = str(data_dir / f'''file-{file_number+1:012}.json''') __lowerCAmelCase = min(len(ds_filter), index + args.samples_per_file) ds_filter.select(list(range(index, end_index))).to_json(file_path) compress_file(file_path) print(f'''Time to save dataset: {time.time()-t_start:.2f}''')
271
0
"""simple docstring""" def a__ ( SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' stooge(__a , 0 , len(__a ) - 1 ) return arr def a__ ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: lowerCAmelCase : Any = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: lowerCAmelCase : Dict = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(__a , __a , (h - t) ) # Recursively sort last 2/3 elements stooge(__a , i + t , (__a) ) # Recursively sort first 2/3 elements stooge(__a , __a , (h - t) ) if __name__ == "__main__": lowerCAmelCase__ = input('''Enter numbers separated by a comma:\n''').strip() lowerCAmelCase__ = [int(item) for item in user_input.split(''',''')] print(stooge_sort(unsorted))
108
'''simple docstring''' import argparse from typing import List import evaluate import numpy as np import torch from datasets import DatasetDict, load_dataset # New Code # # We'll be using StratifiedKFold for this example from sklearn.model_selection import StratifiedKFold from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to perform Cross Validation, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## __lowerCAmelCase = 1_6 __lowerCAmelCase = 3_2 def UpperCAmelCase_ (__a : Accelerator , __a : DatasetDict , __a : List[int] , __a : List[int] , __a : int = 1_6 ): """simple docstring""" _a : Union[str, Any] = AutoTokenizer.from_pretrained('bert-base-cased' ) _a : str = DatasetDict( { 'train': dataset['train'].select(__a ), 'validation': dataset['train'].select(__a ), 'test': dataset['validation'], } ) def tokenize_function(__a : List[Any] ): # max_length=None => use the model max length (it's actually the default) _a : Optional[int] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=__a , max_length=__a ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _a : List[str] = datasets.map( __a , batched=__a , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _a : List[Any] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(__a : int ): # On TPU it's best to pad everything to the same length or training will be very slow. _a : Dict = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _a : Tuple = 1_6 elif accelerator.mixed_precision != "no": _a : List[Any] = 8 else: _a : List[Any] = None return tokenizer.pad( __a , padding='longest' , max_length=__a , pad_to_multiple_of=__a , return_tensors='pt' , ) # Instantiate dataloaders. _a : Any = DataLoader( tokenized_datasets['train'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[int] = DataLoader( tokenized_datasets['validation'] , shuffle=__a , collate_fn=__a , batch_size=__a ) _a : Optional[Any] = DataLoader( tokenized_datasets['test'] , shuffle=__a , collate_fn=__a , batch_size=__a ) return train_dataloader, eval_dataloader, test_dataloader def UpperCAmelCase_ (__a : Any , __a : Union[str, Any] ): """simple docstring""" _a : Dict = [] # Download the dataset _a : Tuple = load_dataset('glue' , 'mrpc' ) # Create our splits _a : Union[str, Any] = StratifiedKFold(n_splits=int(args.num_folds ) ) # Initialize accelerator _a : Any = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _a : Optional[Any] = config['lr'] _a : Optional[int] = int(config['num_epochs'] ) _a : Dict = int(config['seed'] ) _a : Dict = int(config['batch_size'] ) _a : Optional[int] = evaluate.load('glue' , 'mrpc' ) # If the batch size is too big we use gradient accumulation _a : List[Any] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: _a : Any = batch_size // MAX_GPU_BATCH_SIZE _a : List[str] = MAX_GPU_BATCH_SIZE set_seed(__a ) # New Code # # Create our folds: _a : int = kfold.split(np.zeros(datasets['train'].num_rows ) , datasets['train']['label'] ) _a : Any = [] # Iterate over them for i, (train_idxs, valid_idxs) in enumerate(__a ): _a, _a, _a : Optional[Any] = get_fold_dataloaders( __a , __a , __a , __a , ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _a : Dict = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=__a ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _a : List[Any] = model.to(accelerator.device ) # Instantiate optimizer _a : List[str] = AdamW(params=model.parameters() , lr=__a ) # Instantiate scheduler _a : List[Any] = get_linear_schedule_with_warmup( optimizer=__a , num_warmup_steps=1_0_0 , num_training_steps=(len(__a ) * num_epochs) // gradient_accumulation_steps , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _a, _a, _a, _a, _a : Union[str, Any] = accelerator.prepare( __a , __a , __a , __a , __a ) # Now we train the model for epoch in range(__a ): model.train() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) _a : Dict = model(**__a ) _a : int = outputs.loss _a : Any = loss / gradient_accumulation_steps accelerator.backward(__a ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Union[str, Any] = model(**__a ) _a : Tuple = outputs.logits.argmax(dim=-1 ) _a, _a : Any = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=__a , references=__a , ) _a : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , __a ) # New Code # # We also run predictions on the test set at the very end _a : Any = [] for step, batch in enumerate(__a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _a : Tuple = model(**__a ) _a : Dict = outputs.logits _a, _a : Optional[int] = accelerator.gather_for_metrics((predictions, batch['labels']) ) fold_predictions.append(predictions.cpu() ) if i == 0: # We need all of the test predictions test_references.append(references.cpu() ) # Use accelerator.print to print only on the main process. test_predictions.append(torch.cat(__a , dim=0 ) ) # We now need to release all our memory and get rid of the current model, optimizer, etc accelerator.free_memory() # New Code # # Finally we check the accuracy of our folded results: _a : Dict = torch.cat(__a , dim=0 ) _a : Any = torch.stack(__a , dim=0 ).sum(dim=0 ).div(int(args.num_folds ) ).argmax(dim=-1 ) _a : str = metric.compute(predictions=__a , references=__a ) accelerator.print('Average test metrics from all folds:' , __a ) def UpperCAmelCase_ (): """simple docstring""" _a : Any = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=__a , default=__a , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) # New Code # parser.add_argument('--num_folds' , type=__a , default=3 , help='The number of splits to perform across the dataset' ) _a : Any = parser.parse_args() _a : int = {'lr': 2e-5, 'num_epochs': 3, 'seed': 4_2, 'batch_size': 1_6} training_function(__a , __a ) if __name__ == "__main__": main()
271
0
"""simple docstring""" from __future__ import annotations def lowercase ( _snake_case : list[int] ) ->Optional[Any]: """simple docstring""" __snake_case : Optional[Any] = len(__a ) // 2 # choose the middle 3 elements __snake_case : Optional[int] = lst[m - 1 : m + 2] # if middle element is peak if three[1] > three[0] and three[1] > three[2]: return three[1] # if increasing, recurse on right elif three[0] < three[2]: if len(lst[:m] ) == 2: m -= 1 return peak(lst[m:] ) # decreasing else: if len(lst[:m] ) == 2: m += 1 return peak(lst[:m] ) if __name__ == "__main__": import doctest doctest.testmod()
102
'''simple docstring''' from __future__ import annotations __lowerCAmelCase = [-1_0, -5, 0, 5, 5.1, 1_1, 1_3, 2_1, 3, 4, -2_1, -1_0, -5, -1, 0] __lowerCAmelCase = [-5, 0, 5, 5.1, 1_1, 1_3, 2_1, -1, 4, -1, -1_0, -5, -1, 0, -1] def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Optional[int] = [] _a : int = len(__a ) for i in range(__a ): _a : float = -1 for j in range(i + 1 , __a ): if arr[i] < arr[j]: _a : Any = arr[j] break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : Tuple = [] for i, outer in enumerate(__a ): _a : float = -1 for inner in arr[i + 1 :]: if outer < inner: _a : Dict = inner break result.append(__a ) return result def UpperCAmelCase_ (__a : list[float] ): """simple docstring""" _a : int = len(__a ) _a : list[float] = [] _a : list[float] = [-1] * arr_size for index in reversed(range(__a ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: _a : Dict = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) __lowerCAmelCase = ( """from __main__ import arr, next_greatest_element_slow, """ """next_greatest_element_fast, next_greatest_element""" ) print( """next_greatest_element_slow():""", timeit("""next_greatest_element_slow(arr)""", setup=setup), ) print( """next_greatest_element_fast():""", timeit("""next_greatest_element_fast(arr)""", setup=setup), ) print( """ next_greatest_element():""", timeit("""next_greatest_element(arr)""", setup=setup), )
271
0