code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import platform import numpy as np import psutil import torch from accelerate import __version__ as version from accelerate.commands.config import default_config_file, load_config_from_file from ..utils import is_npu_available, is_xpu_available def A__ ( __lowerCamelCase=None ): if subparsers is not None: SCREAMING_SNAKE_CASE_ = subparsers.add_parser('''env''' ) else: SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser('''Accelerate env command''' ) parser.add_argument( '''--config_file''', default=__lowerCamelCase, help='''The config file to use for the default values in the launching script.''' ) if subparsers is not None: parser.set_defaults(func=__lowerCamelCase ) return parser def A__ ( __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = torch.__version__ SCREAMING_SNAKE_CASE_ = torch.cuda.is_available() SCREAMING_SNAKE_CASE_ = is_xpu_available() SCREAMING_SNAKE_CASE_ = is_npu_available() SCREAMING_SNAKE_CASE_ = '''Not found''' # Get the default from the config file. if args.config_file is not None or os.path.isfile(__lowerCamelCase ): SCREAMING_SNAKE_CASE_ = load_config_from_file(args.config_file ).to_dict() SCREAMING_SNAKE_CASE_ = { '''`Accelerate` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''Numpy version''': np.__version__, '''PyTorch version (GPU?)''': F'''{pt_version} ({pt_cuda_available})''', '''PyTorch XPU available''': str(__lowerCamelCase ), '''PyTorch NPU available''': str(__lowerCamelCase ), '''System RAM''': F'''{psutil.virtual_memory().total / 10_24 ** 3:.2f} GB''', } if pt_cuda_available: SCREAMING_SNAKE_CASE_ = torch.cuda.get_device_name() print('''\nCopy-and-paste the text below in your GitHub issue\n''' ) print('''\n'''.join([F'''- {prop}: {val}''' for prop, val in info.items()] ) ) print('''- `Accelerate` default config:''' if args.config_file is None else '''- `Accelerate` config passed:''' ) SCREAMING_SNAKE_CASE_ = ( '''\n'''.join([F'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(__lowerCamelCase, __lowerCamelCase ) else F'''\t{accelerate_config}''' ) print(__lowerCamelCase ) SCREAMING_SNAKE_CASE_ = accelerate_config return info def A__ ( ): SCREAMING_SNAKE_CASE_ = env_command_parser() SCREAMING_SNAKE_CASE_ = parser.parse_args() env_command(__lowerCamelCase ) return 0 if __name__ == "__main__": raise SystemExit(main())
299
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): __UpperCAmelCase = "pt" elif is_tf_available(): __UpperCAmelCase = "tf" else: __UpperCAmelCase = "jax" class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" UpperCAmelCase_ =ByTaTokenizer UpperCAmelCase_ =False def _UpperCamelCase ( self ) -> Tuple: super().setUp() SCREAMING_SNAKE_CASE_ = ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def _UpperCamelCase ( self ) -> List[str]: return ByTaTokenizer.from_pretrained('''google/byt5-small''' ) def _UpperCamelCase ( self , **_A ) -> ByTaTokenizer: return self.tokenizer_class.from_pretrained(self.tmpdirname , **_A ) def _UpperCamelCase ( self , _A , _A=False , _A=20 , _A=5 ) -> Tuple[str, list]: # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for ByT5 because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. SCREAMING_SNAKE_CASE_ = [] for i in range(len(_A ) ): try: SCREAMING_SNAKE_CASE_ = tokenizer.decode([i] , clean_up_tokenization_spaces=_A ) except UnicodeDecodeError: pass toks.append((i, tok) ) SCREAMING_SNAKE_CASE_ = list(filter(lambda _A : re.match(R'''^[ a-zA-Z]+$''' , t[1] ) , _A ) ) SCREAMING_SNAKE_CASE_ = list(filter(lambda _A : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=_A ) , _A ) ) if max_length is not None and len(_A ) > max_length: SCREAMING_SNAKE_CASE_ = toks[:max_length] if min_length is not None and len(_A ) < min_length and len(_A ) > 0: while len(_A ) < min_length: SCREAMING_SNAKE_CASE_ = toks + toks # toks_str = [t[1] for t in toks] SCREAMING_SNAKE_CASE_ = [t[0] for t in toks] # Ensure consistency SCREAMING_SNAKE_CASE_ = tokenizer.decode(_A , clean_up_tokenization_spaces=_A ) if " " not in output_txt and len(_A ) > 1: SCREAMING_SNAKE_CASE_ = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=_A ) + ''' ''' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=_A ) ) if with_prefix_space: SCREAMING_SNAKE_CASE_ = ''' ''' + output_txt SCREAMING_SNAKE_CASE_ = tokenizer.encode(_A , add_special_tokens=_A ) return output_txt, output_ids def _UpperCamelCase ( self ) -> str: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = tokenizer(['''hi</s>''', '''I went to the gym</s>''', '''</s>'''] ) SCREAMING_SNAKE_CASE_ = tokenizer(['''hi''', '''I went to the gym''', ''''''] ) self.assertListEqual(batch_with_eos_added['''input_ids'''] , batch_without_eos_added['''input_ids'''] ) def _UpperCamelCase ( self ) -> Any: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = '''Unicode €.''' SCREAMING_SNAKE_CASE_ = tokenizer(_A ) SCREAMING_SNAKE_CASE_ = [88, 113, 108, 102, 114, 103, 104, 35, 229, 133, 175, 49, 1] self.assertEqual(encoded['''input_ids'''] , _A ) # decoding SCREAMING_SNAKE_CASE_ = tokenizer.decode(_A ) self.assertEqual(_A , '''Unicode €.</s>''' ) SCREAMING_SNAKE_CASE_ = tokenizer('''e è é ê ë''' ) SCREAMING_SNAKE_CASE_ = [104, 35, 198, 171, 35, 198, 172, 35, 198, 173, 35, 198, 174, 1] self.assertEqual(encoded['''input_ids'''] , _A ) # decoding SCREAMING_SNAKE_CASE_ = tokenizer.decode(_A ) self.assertEqual(_A , '''e è é ê ë</s>''' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('''e è é ê ë''' ) ) , '''e è é ê ë</s>''' ) def _UpperCamelCase ( self ) -> List[str]: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] # fmt: off SCREAMING_SNAKE_CASE_ = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 1, 0] # fmt: on SCREAMING_SNAKE_CASE_ = tokenizer(_A , padding=_A , return_tensors=_A ) self.assertIsInstance(_A , _A ) if FRAMEWORK != "jax": SCREAMING_SNAKE_CASE_ = list(batch.input_ids.numpy()[0] ) else: SCREAMING_SNAKE_CASE_ = list(batch.input_ids.tolist()[0] ) self.assertListEqual(_A , _A ) self.assertEqual((2, 37) , batch.input_ids.shape ) self.assertEqual((2, 37) , batch.attention_mask.shape ) def _UpperCamelCase ( self ) -> str: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] SCREAMING_SNAKE_CASE_ = tokenizer(_A , padding=_A , return_tensors=_A ) # check if input_ids are returned and no decoder_input_ids self.assertIn('''input_ids''' , _A ) self.assertIn('''attention_mask''' , _A ) self.assertNotIn('''decoder_input_ids''' , _A ) self.assertNotIn('''decoder_attention_mask''' , _A ) def _UpperCamelCase ( self ) -> Tuple: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = [ '''Summary of the text.''', '''Another summary.''', ] SCREAMING_SNAKE_CASE_ = tokenizer( text_target=_A , max_length=32 , padding='''max_length''' , truncation=_A , return_tensors=_A ) self.assertEqual(32 , targets['''input_ids'''].shape[1] ) def _UpperCamelCase ( self ) -> List[Any]: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = ['''A long paragraph for summarization. </s>'''] SCREAMING_SNAKE_CASE_ = ['''Summary of the text. </s>'''] # fmt: off SCREAMING_SNAKE_CASE_ = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 35, 1] SCREAMING_SNAKE_CASE_ = [86, 120, 112, 112, 100, 117, 124, 35, 114, 105, 35, 119, 107, 104, 35, 119, 104, 123, 119, 49, 35, 1] # fmt: on SCREAMING_SNAKE_CASE_ = tokenizer(_A , text_target=_A ) self.assertEqual(_A , batch['''input_ids'''][0] ) self.assertEqual(_A , batch['''labels'''][0] ) def _UpperCamelCase ( self ) -> Dict: # safety check on max_len default value so we are sure the test works SCREAMING_SNAKE_CASE_ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test SCREAMING_SNAKE_CASE_ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc SCREAMING_SNAKE_CASE_ = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ = ''' He is very happy, UNwant\u00E9d,running''' SCREAMING_SNAKE_CASE_ = tokenizer.encode(_A , add_special_tokens=_A ) tokenizer.save_pretrained(_A ) SCREAMING_SNAKE_CASE_ = tokenizer.__class__.from_pretrained(_A ) SCREAMING_SNAKE_CASE_ = after_tokenizer.encode(_A , add_special_tokens=_A ) self.assertListEqual(_A , _A ) shutil.rmtree(_A ) SCREAMING_SNAKE_CASE_ = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc SCREAMING_SNAKE_CASE_ = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ = ''' He is very happy, UNwant\u00E9d,running''' tokenizer.add_tokens(['''bim''', '''bambam'''] ) SCREAMING_SNAKE_CASE_ = tokenizer.additional_special_tokens additional_special_tokens.append('''new_additional_special_token''' ) tokenizer.add_special_tokens({'''additional_special_tokens''': additional_special_tokens} ) SCREAMING_SNAKE_CASE_ = tokenizer.encode(_A , add_special_tokens=_A ) tokenizer.save_pretrained(_A ) SCREAMING_SNAKE_CASE_ = tokenizer.__class__.from_pretrained(_A ) SCREAMING_SNAKE_CASE_ = after_tokenizer.encode(_A , add_special_tokens=_A ) self.assertListEqual(_A , _A ) self.assertIn('''new_additional_special_token''' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) SCREAMING_SNAKE_CASE_ = tokenizer.__class__.from_pretrained(_A , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(_A ) def _UpperCamelCase ( self ) -> int: SCREAMING_SNAKE_CASE_ = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_A ) with open(os.path.join(_A , '''special_tokens_map.json''' ) , encoding='''utf-8''' ) as json_file: SCREAMING_SNAKE_CASE_ = json.load(_A ) with open(os.path.join(_A , '''tokenizer_config.json''' ) , encoding='''utf-8''' ) as json_file: SCREAMING_SNAKE_CASE_ = json.load(_A ) SCREAMING_SNAKE_CASE_ = [F'''<extra_id_{i}>''' for i in range(125 )] SCREAMING_SNAKE_CASE_ = added_tokens_extra_ids + [ '''an_additional_special_token''' ] SCREAMING_SNAKE_CASE_ = added_tokens_extra_ids + [ '''an_additional_special_token''' ] with open(os.path.join(_A , '''special_tokens_map.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(_A , _A ) with open(os.path.join(_A , '''tokenizer_config.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(_A , _A ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files SCREAMING_SNAKE_CASE_ = tokenizer_class.from_pretrained( _A , ) self.assertIn( '''an_additional_special_token''' , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ['''an_additional_special_token'''] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['''an_additional_special_token'''] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained SCREAMING_SNAKE_CASE_ = added_tokens_extra_ids + [AddedToken('''a_new_additional_special_token''' , lstrip=_A )] SCREAMING_SNAKE_CASE_ = tokenizer_class.from_pretrained( _A , additional_special_tokens=_A , ) self.assertIn('''a_new_additional_special_token''' , tokenizer.additional_special_tokens ) self.assertEqual( ['''a_new_additional_special_token'''] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['''a_new_additional_special_token'''] ) ) , ) def _UpperCamelCase ( self ) -> str: SCREAMING_SNAKE_CASE_ = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_A ) SCREAMING_SNAKE_CASE_ = tokenizer_class.from_pretrained(_A ) self.assertTrue(tokenizer.decode([255] ) == '''''' ) def _UpperCamelCase ( self ) -> int: pass def _UpperCamelCase ( self ) -> Any: pass def _UpperCamelCase ( self ) -> Any: pass def _UpperCamelCase ( self ) -> Optional[int]: pass def _UpperCamelCase ( self ) -> Union[str, Any]: # The default common tokenizer tests uses invalid tokens for ByT5 that can only accept one-character strings # and special added tokens as tokens SCREAMING_SNAKE_CASE_ = self.get_tokenizers(fast=_A , do_lower_case=_A ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE_ = ['''t''', '''h''', '''i''', '''s''', ''' ''', '''i''', '''s''', ''' ''', '''a''', ''' ''', '''t''', '''e''', '''x''', '''t''', '''</s>'''] SCREAMING_SNAKE_CASE_ = tokenizer.convert_tokens_to_string(_A ) self.assertIsInstance(_A , _A ) def _UpperCamelCase ( self ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE_ = [ '''bos_token''', '''eos_token''', '''unk_token''', '''sep_token''', '''pad_token''', '''cls_token''', '''mask_token''', ] SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = tokenizer.convert_ids_to_tokens( _A , skip_special_tokens=_A ) for attr in attributes_list: setattr(_A , attr + '''_id''' , _A ) self.assertEqual(getattr(_A , _A ) , _A ) self.assertEqual(getattr(_A , attr + '''_id''' ) , _A ) setattr(_A , attr + '''_id''' , _A ) self.assertEqual(getattr(_A , _A ) , _A ) self.assertEqual(getattr(_A , attr + '''_id''' ) , _A ) setattr(_A , '''additional_special_tokens_ids''' , [] ) self.assertListEqual(getattr(_A , '''additional_special_tokens''' ) , [] ) self.assertListEqual(getattr(_A , '''additional_special_tokens_ids''' ) , [] ) setattr(_A , '''additional_special_tokens_ids''' , [token_id_to_test_setters] ) self.assertListEqual(getattr(_A , '''additional_special_tokens''' ) , [token_to_test_setters] ) self.assertListEqual(getattr(_A , '''additional_special_tokens_ids''' ) , [token_id_to_test_setters] )
299
1
'''simple docstring''' import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () __snake_case = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). __snake_case = [0, 25, 50] __snake_case = [25, 50, 75] __snake_case = fuzz.membership.trimf(X, abca) __snake_case = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. __snake_case = np.ones(75) __snake_case = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) __snake_case = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) __snake_case = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) __snake_case = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) __snake_case = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] __snake_case = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) __snake_case = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] __snake_case = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] __snake_case = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('''Young''') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('''Middle aged''') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('''union''') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('''intersection''') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('''complement_a''') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('''difference a/b''') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('''alg_sum''') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('''alg_product''') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('''bdd_sum''') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('''bdd_difference''') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
361
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from .config import config_command_parser from .config_args import default_config_file, load_config_from_file # noqa: F401 from .default import default_command_parser from .update import update_command_parser def a ( __a=None ) -> List[str]: '''simple docstring''' UpperCamelCase__ :Optional[Any] = argparse.ArgumentParser(add_help=__a , allow_abbrev=__a ) # The main config parser UpperCamelCase__ :str = config_command_parser(__a ) # The subparser to add commands to UpperCamelCase__ :Union[str, Any] = config_parser.add_subparsers(title='''subcommands''' , dest='''subcommand''' ) # Then add other parsers with the parent parser default_command_parser(__a , parents=[parent_parser] ) update_command_parser(__a , parents=[parent_parser] ) return config_parser def a ( ) -> Any: '''simple docstring''' UpperCamelCase__ :int = get_config_parser() UpperCamelCase__ :List[Any] = config_parser.parse_args() if not hasattr(__a , '''func''' ): config_parser.print_help() exit(1 ) # Run args.func(__a ) if __name__ == "__main__": main()
219
0
import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class __snake_case ( __UpperCamelCase , unittest.TestCase ): _a : Any= BioGptTokenizer _a : Union[str, Any]= False def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowercase : Any = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''w</w>''', '''r</w>''', '''t</w>''', '''lo''', '''low''', '''er</w>''', '''low</w>''', '''lowest</w>''', '''newer</w>''', '''wider</w>''', '''<unk>''', ] lowercase : Tuple = dict(zip(a_ ,range(len(a_ ) ) ) ) lowercase : Any = ['''l o 123''', '''lo w 1456''', '''e r</w> 1789''', ''''''] lowercase : str = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) lowercase : List[str] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file ,"""w""" ) as fp: fp.write(json.dumps(a_ ) ) with open(self.merges_file ,"""w""" ) as fp: fp.write("""\n""".join(a_ ) ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ): '''simple docstring''' lowercase : Union[str, Any] = '''lower newer''' lowercase : str = '''lower newer''' return input_text, output_text def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowercase : List[Any] = BioGptTokenizer(self.vocab_file ,self.merges_file ) lowercase : Optional[Any] = '''lower''' lowercase : int = ['''low''', '''er</w>'''] lowercase : int = tokenizer.tokenize(a_ ) self.assertListEqual(a_ ,a_ ) lowercase : int = tokens + ['''<unk>'''] lowercase : int = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) ,a_ ) @slow def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowercase : Tuple = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) lowercase : Union[str, Any] = tokenizer.encode("""sequence builders""" ,add_special_tokens=a_ ) lowercase : Any = tokenizer.encode("""multi-sequence build""" ,add_special_tokens=a_ ) lowercase : Dict = tokenizer.build_inputs_with_special_tokens(a_ ) lowercase : List[str] = tokenizer.build_inputs_with_special_tokens(a_ ,a_ ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
20
import glob import os import random from string import ascii_lowercase, digits import cva __A ="" __A ="" __A ="" __A =1 # (0 is vertical, 1 is horizontal) def a ( ): '''simple docstring''' __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = get_dataset(_UpperCAmelCase , _UpperCAmelCase ) print('''Processing...''' ) __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : List[Any] = update_image_and_anno(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) for index, image in enumerate(_UpperCAmelCase ): # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __UpperCAmelCase : Any = random_chars(32 ) __UpperCAmelCase : List[str] = paths[index].split(os.sep )[-1].rsplit('''.''' , 1 )[0] __UpperCAmelCase : Optional[Any] = f'{OUTPUT_DIR}/{file_name}_FLIP_{letter_code}' cva.imwrite(f'/{file_root}.jpg' , _UpperCAmelCase , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(f'Success {index+1}/{len(_UpperCAmelCase )} with {file_name}' ) __UpperCAmelCase : Optional[Any] = [] for anno in new_annos[index]: __UpperCAmelCase : Union[str, Any] = f'{anno[0]} {anno[1]} {anno[2]} {anno[3]} {anno[4]}' annos_list.append(_UpperCAmelCase ) with open(f'/{file_root}.txt' , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def a ( _UpperCAmelCase : str , _UpperCAmelCase : str ): '''simple docstring''' __UpperCAmelCase : Any = [] __UpperCAmelCase : Any = [] for label_file in glob.glob(os.path.join(_UpperCAmelCase , '''*.txt''' ) ): __UpperCAmelCase : Optional[Any] = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(_UpperCAmelCase ) as in_file: __UpperCAmelCase : List[str] = in_file.readlines() __UpperCAmelCase : Optional[Any] = os.path.join(_UpperCAmelCase , f'{label_name}.jpg' ) __UpperCAmelCase : str = [] for obj_list in obj_lists: __UpperCAmelCase : str = obj_list.rstrip('''\n''' ).split(''' ''' ) boxes.append( [ int(obj[0] ), float(obj[1] ), float(obj[2] ), float(obj[3] ), float(obj[4] ), ] ) if not boxes: continue img_paths.append(_UpperCAmelCase ) labels.append(_UpperCAmelCase ) return img_paths, labels def a ( _UpperCAmelCase : list , _UpperCAmelCase : list , _UpperCAmelCase : int = 1 ): '''simple docstring''' __UpperCAmelCase : Dict = [] __UpperCAmelCase : Optional[Any] = [] __UpperCAmelCase : Any = [] for idx in range(len(_UpperCAmelCase ) ): __UpperCAmelCase : Tuple = [] __UpperCAmelCase : List[Any] = img_list[idx] path_list.append(_UpperCAmelCase ) __UpperCAmelCase : str = anno_list[idx] __UpperCAmelCase : str = cva.imread(_UpperCAmelCase ) if flip_type == 1: __UpperCAmelCase : Any = cva.flip(_UpperCAmelCase , _UpperCAmelCase ) for bbox in img_annos: __UpperCAmelCase : List[str] = 1 - bbox[1] new_annos.append([bbox[0], x_center_new, bbox[2], bbox[3], bbox[4]] ) elif flip_type == 0: __UpperCAmelCase : Any = cva.flip(_UpperCAmelCase , _UpperCAmelCase ) for bbox in img_annos: __UpperCAmelCase : Union[str, Any] = 1 - bbox[2] new_annos.append([bbox[0], bbox[1], y_center_new, bbox[3], bbox[4]] ) new_annos_lists.append(_UpperCAmelCase ) new_imgs_list.append(_UpperCAmelCase ) return new_imgs_list, new_annos_lists, path_list def a ( _UpperCAmelCase : int = 32 ): '''simple docstring''' assert number_char > 1, "The number of character should greater than 1" __UpperCAmelCase : Union[str, Any] = ascii_lowercase + digits return "".join(random.choice(_UpperCAmelCase ) for _ in range(_UpperCAmelCase ) ) if __name__ == "__main__": main() print("DONE ✅")
226
0
'''simple docstring''' from __future__ import annotations from random import random class lowercase : def __init__( self , UpperCamelCase_ = None ): '''simple docstring''' UpperCamelCase__ :Tuple = value UpperCamelCase__ :Any = random() UpperCamelCase__ :Node | None = None UpperCamelCase__ :Node | None = None def __repr__( self ): '''simple docstring''' from pprint import pformat if self.left is None and self.right is None: return F'''\'{self.value}: {self.prior:.5}\'''' else: return pformat( {F'''{self.value}: {self.prior:.5}''': (self.left, self.right)} , indent=1 ) def __str__( self ): '''simple docstring''' UpperCamelCase__ :Tuple = str(self.value ) + " " UpperCamelCase__ :Optional[int] = str(self.left or '''''' ) UpperCamelCase__ :Union[str, Any] = str(self.right or '''''' ) return value + left + right def a ( __a , __a ) -> tuple[Node | None, Node | None]: '''simple docstring''' if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: UpperCamelCase__ :List[Any] = split(root.left , _lowerCAmelCase ) return left, root else: UpperCamelCase__ :Union[str, Any] = split(root.right , _lowerCAmelCase ) return root, right def a ( __a , __a ) -> Node | None: '''simple docstring''' if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: UpperCamelCase__ :Optional[Any] = merge(left.right , _lowerCAmelCase ) return left else: UpperCamelCase__ :int = merge(_lowerCAmelCase , right.left ) return right def a ( __a , __a ) -> Node | None: '''simple docstring''' UpperCamelCase__ :Union[str, Any] = Node(_lowerCAmelCase ) UpperCamelCase__ :str = split(_lowerCAmelCase , _lowerCAmelCase ) return merge(merge(_lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase ) def a ( __a , __a ) -> Node | None: '''simple docstring''' UpperCamelCase__ :Tuple = split(_lowerCAmelCase , value - 1 ) UpperCamelCase__ :Optional[int] = split(_lowerCAmelCase , _lowerCAmelCase ) return merge(_lowerCAmelCase , _lowerCAmelCase ) def a ( __a ) -> None: '''simple docstring''' if not root: # None return else: inorder(root.left ) print(root.value , end=''',''' ) inorder(root.right ) def a ( __a , __a ) -> Node | None: '''simple docstring''' for arg in args.split(): if arg[0] == "+": UpperCamelCase__ :str = insert(_lowerCAmelCase , int(arg[1:] ) ) elif arg[0] == "-": UpperCamelCase__ :Tuple = erase(_lowerCAmelCase , int(arg[1:] ) ) else: print('''Unknown command''' ) return root def a ( ) -> None: '''simple docstring''' UpperCamelCase__ :str = None print( '''enter numbers to create a tree, + value to add value into treap, ''' '''- value to erase all nodes with value. \'q\' to quit. ''' ) UpperCamelCase__ :Any = input() while args != "q": UpperCamelCase__ :Tuple = interact_treap(_lowerCAmelCase , _lowerCAmelCase ) print(_lowerCAmelCase ) UpperCamelCase__ :Optional[Any] = input() print('''good by!''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
354
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __snake_case = logging.get_logger(__name__) __snake_case = { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/config.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/config.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json''' ), '''distilbert-base-uncased-finetuned-sst-2-english''': ( '''https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json''' ), } class lowercase ( A__ ): """simple docstring""" _a = 'distilbert' _a = { 'hidden_size': 'dim', 'num_attention_heads': 'n_heads', 'num_hidden_layers': 'n_layers', } def __init__( self , UpperCamelCase_=30522 , UpperCamelCase_=512 , UpperCamelCase_=False , UpperCamelCase_=6 , UpperCamelCase_=12 , UpperCamelCase_=768 , UpperCamelCase_=4 * 768 , UpperCamelCase_=0.1 , UpperCamelCase_=0.1 , UpperCamelCase_="gelu" , UpperCamelCase_=0.02 , UpperCamelCase_=0.1 , UpperCamelCase_=0.2 , UpperCamelCase_=0 , **UpperCamelCase_ , ): '''simple docstring''' UpperCamelCase__ :List[Any] = vocab_size UpperCamelCase__ :Dict = max_position_embeddings UpperCamelCase__ :str = sinusoidal_pos_embds UpperCamelCase__ :Any = n_layers UpperCamelCase__ :str = n_heads UpperCamelCase__ :Tuple = dim UpperCamelCase__ :str = hidden_dim UpperCamelCase__ :Dict = dropout UpperCamelCase__ :int = attention_dropout UpperCamelCase__ :Optional[Any] = activation UpperCamelCase__ :Optional[int] = initializer_range UpperCamelCase__ :Union[str, Any] = qa_dropout UpperCamelCase__ :Dict = seq_classif_dropout super().__init__(**UpperCamelCase_ , pad_token_id=UpperCamelCase_ ) class lowercase ( A__ ): """simple docstring""" @property def lowerCAmelCase__ ( self ): '''simple docstring''' if self.task == "multiple-choice": UpperCamelCase__ :str = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: UpperCamelCase__ :str = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
219
0
from __future__ import annotations import numpy as np def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' return np.maximum(0 , lowerCAmelCase__ ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
101
import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowercase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): lowercase_ : Any =KandinskyVaaPriorPipeline lowercase_ : str =['''prompt'''] lowercase_ : Union[str, Any] =['''prompt''', '''negative_prompt'''] lowercase_ : Dict =[ '''num_images_per_prompt''', '''generator''', '''num_inference_steps''', '''latents''', '''negative_prompt''', '''guidance_scale''', '''output_type''', '''return_dict''', ] lowercase_ : List[str] =False @property def A__ ( self): return 3_2 @property def A__ ( self): return 3_2 @property def A__ ( self): return self.time_input_dim @property def A__ ( self): return self.time_input_dim * 4 @property def A__ ( self): return 1_0_0 @property def A__ ( self): lowercase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') return tokenizer @property def A__ ( self): torch.manual_seed(0) lowercase = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=self.text_embedder_hidden_size ,projection_dim=self.text_embedder_hidden_size ,intermediate_size=3_7 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1_0_0_0 ,) return CLIPTextModelWithProjection(A__) @property def A__ ( self): torch.manual_seed(0) lowercase = { '''num_attention_heads''': 2, '''attention_head_dim''': 1_2, '''embedding_dim''': self.text_embedder_hidden_size, '''num_layers''': 1, } lowercase = PriorTransformer(**A__) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 lowercase = nn.Parameter(torch.ones(model.clip_std.shape)) return model @property def A__ ( self): torch.manual_seed(0) lowercase = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size ,image_size=2_2_4 ,projection_dim=self.text_embedder_hidden_size ,intermediate_size=3_7 ,num_attention_heads=4 ,num_channels=3 ,num_hidden_layers=5 ,patch_size=1_4 ,) lowercase = CLIPVisionModelWithProjection(A__) return model @property def A__ ( self): lowercase = CLIPImageProcessor( crop_size=2_2_4 ,do_center_crop=A__ ,do_normalize=A__ ,do_resize=A__ ,image_mean=[0.48145466, 0.4578275, 0.40821073] ,image_std=[0.26862954, 0.26130258, 0.27577711] ,resample=3 ,size=2_2_4 ,) return image_processor def A__ ( self): lowercase = self.dummy_prior lowercase = self.dummy_image_encoder lowercase = self.dummy_text_encoder lowercase = self.dummy_tokenizer lowercase = self.dummy_image_processor lowercase = UnCLIPScheduler( variance_type='''fixed_small_log''' ,prediction_type='''sample''' ,num_train_timesteps=1_0_0_0 ,clip_sample=A__ ,clip_sample_range=10.0 ,) lowercase = { '''prior''': prior, '''image_encoder''': image_encoder, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''scheduler''': scheduler, '''image_processor''': image_processor, } return components def A__ ( self ,A__ ,A__=0): if str(A__).startswith('''mps'''): lowercase = torch.manual_seed(A__) else: lowercase = torch.Generator(device=A__).manual_seed(A__) lowercase = { '''prompt''': '''horse''', '''generator''': generator, '''guidance_scale''': 4.0, '''num_inference_steps''': 2, '''output_type''': '''np''', } return inputs def A__ ( self): lowercase = '''cpu''' lowercase = self.get_dummy_components() lowercase = self.pipeline_class(**A__) lowercase = pipe.to(A__) pipe.set_progress_bar_config(disable=A__) lowercase = pipe(**self.get_dummy_inputs(A__)) lowercase = output.image_embeds lowercase = pipe( **self.get_dummy_inputs(A__) ,return_dict=A__ ,)[0] lowercase = image[0, -1_0:] lowercase = image_from_tuple[0, -1_0:] assert image.shape == (1, 3_2) lowercase = np.array( [-0.0532, 1.7120, 0.3656, -1.0852, -0.8946, -1.1756, 0.4348, 0.2482, 0.5146, -0.1156]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1E-2 @skip_mps def A__ ( self): lowercase = torch_device == '''cpu''' lowercase = True lowercase = False self._test_inference_batch_single_identical( test_max_difference=A__ ,relax_max_difference=A__ ,test_mean_pixel_difference=A__ ,) @skip_mps def A__ ( self): lowercase = torch_device == '''cpu''' lowercase = False self._test_attention_slicing_forward_pass( test_max_difference=A__ ,test_mean_pixel_difference=A__ ,)
101
1
'''simple docstring''' from __future__ import annotations def __lowerCamelCase ( _lowercase ) -> bool: return len(set(_lowercase ) ) == len(_lowercase ) if __name__ == "__main__": import doctest doctest.testmod()
354
'''simple docstring''' import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatchSampler, SkipDataLoader, skip_first_batches, ) class UpperCamelCase_ ( __magic_name__ ): def __init__( self , A=0.0_1 , A=1000 ) -> List[str]: UpperCAmelCase : List[Any] = p_stop UpperCAmelCase : Optional[int] = max_length def __iter__( self ) -> Union[str, Any]: UpperCAmelCase : Dict = 0 UpperCAmelCase : Union[str, Any] = False while not stop and count < self.max_length: yield count count += 1 UpperCAmelCase : Any = random.random() < self.p_stop class UpperCamelCase_ ( unittest.TestCase ): def _lowercase( self , A , A , A=False , A=True ) -> Union[str, Any]: UpperCAmelCase : List[str] = [ BatchSamplerShard(A , 2 , A , split_batches=A , even_batches=A ) for i in range(2 ) ] UpperCAmelCase : List[str] = [list(A ) for batch_sampler_shard in batch_sampler_shards] if not split_batches: self.assertListEqual([len(A ) for shard in batch_sampler_shards] , [len(A ) for e in expected] ) self.assertListEqual(A , A ) def _lowercase( self ) -> Union[str, Any]: # Check the shards when the dataset is a round multiple of total batch size. UpperCAmelCase : int = BatchSampler(range(24 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : Any = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(A , A ) UpperCAmelCase : Union[str, Any] = BatchSampler(range(24 ) , batch_size=3 , drop_last=A ) # Expected shouldn't change self.check_batch_sampler_shards(A , A ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. UpperCAmelCase : Tuple = BatchSampler(range(21 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : Optional[int] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [0, 1, 2]], ] self.check_batch_sampler_shards(A , A ) UpperCAmelCase : Optional[int] = BatchSampler(range(21 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : Optional[int] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(A , A ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. UpperCAmelCase : Tuple = BatchSampler(range(22 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : Tuple = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 0, 1]], ] self.check_batch_sampler_shards(A , A ) UpperCAmelCase : int = BatchSampler(range(22 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : List[Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(A , A ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. UpperCAmelCase : Union[str, Any] = BatchSampler(range(20 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : Optional[int] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 0]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [1, 2, 3]], ] self.check_batch_sampler_shards(A , A ) UpperCAmelCase : Optional[Any] = BatchSampler(range(20 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : int = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(A , A ) # Check the shards when the dataset is very small. UpperCAmelCase : Any = BatchSampler(range(2 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : Union[str, Any] = [[[0, 1, 0]], [[1, 0, 1]]] self.check_batch_sampler_shards(A , A ) UpperCAmelCase : Dict = BatchSampler(range(2 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : List[Any] = [[], []] self.check_batch_sampler_shards(A , A ) def _lowercase( self ) -> Tuple: # Check the shards when the dataset is a round multiple of batch size. UpperCAmelCase : Any = BatchSampler(range(24 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : List[str] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(A , A , split_batches=A ) UpperCAmelCase : List[Any] = BatchSampler(range(24 ) , batch_size=4 , drop_last=A ) # Expected shouldn't change self.check_batch_sampler_shards(A , A , split_batches=A ) # Check the shards when the dataset is not a round multiple of batch size. UpperCAmelCase : Optional[Any] = BatchSampler(range(22 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : List[str] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [0, 1]], ] self.check_batch_sampler_shards(A , A , split_batches=A ) UpperCAmelCase : Union[str, Any] = BatchSampler(range(22 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : Union[str, Any] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(A , A , split_batches=A ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. UpperCAmelCase : Any = BatchSampler(range(21 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : Any = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 0]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [1, 2]], ] self.check_batch_sampler_shards(A , A , split_batches=A ) UpperCAmelCase : int = BatchSampler(range(21 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : Optional[int] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(A , A , split_batches=A ) # Check the shards when the dataset is very small. UpperCAmelCase : Optional[int] = BatchSampler(range(2 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : Optional[Any] = [[[0, 1]], [[0, 1]]] self.check_batch_sampler_shards(A , A , split_batches=A ) UpperCAmelCase : Dict = BatchSampler(range(2 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : Any = [[], []] self.check_batch_sampler_shards(A , A , split_batches=A ) def _lowercase( self ) -> Any: # Check the shards when the dataset is a round multiple of total batch size. UpperCAmelCase : str = BatchSampler(range(24 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : Optional[Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(A , A , even_batches=A ) UpperCAmelCase : Union[str, Any] = BatchSampler(range(24 ) , batch_size=3 , drop_last=A ) # Expected shouldn't change self.check_batch_sampler_shards(A , A , even_batches=A ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. UpperCAmelCase : Optional[Any] = BatchSampler(range(21 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : Optional[int] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(A , A , even_batches=A ) UpperCAmelCase : str = BatchSampler(range(21 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : List[Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(A , A , even_batches=A ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. UpperCAmelCase : List[Any] = BatchSampler(range(22 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : Dict = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21]], ] self.check_batch_sampler_shards(A , A , even_batches=A ) UpperCAmelCase : Dict = BatchSampler(range(22 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : Optional[Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(A , A , even_batches=A ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. UpperCAmelCase : List[str] = BatchSampler(range(20 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : Union[str, Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(A , A , even_batches=A ) UpperCAmelCase : Optional[int] = BatchSampler(range(20 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : Optional[Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(A , A , even_batches=A ) # Check the shards when the dataset is very small. UpperCAmelCase : Dict = BatchSampler(range(2 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : str = [[[0, 1]], []] self.check_batch_sampler_shards(A , A , even_batches=A ) UpperCAmelCase : List[str] = BatchSampler(range(2 ) , batch_size=3 , drop_last=A ) UpperCAmelCase : Tuple = [[], []] self.check_batch_sampler_shards(A , A , even_batches=A ) def _lowercase( self ) -> List[Any]: # Check the shards when the dataset is a round multiple of batch size. UpperCAmelCase : Dict = BatchSampler(range(24 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : List[Any] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(A , A , split_batches=A , even_batches=A ) UpperCAmelCase : int = BatchSampler(range(24 ) , batch_size=4 , drop_last=A ) # Expected shouldn't change self.check_batch_sampler_shards(A , A , split_batches=A , even_batches=A ) # Check the shards when the dataset is not a round multiple of batch size. UpperCAmelCase : List[str] = BatchSampler(range(22 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : Optional[Any] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(A , A , split_batches=A , even_batches=A ) UpperCAmelCase : Dict = BatchSampler(range(22 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : Dict = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(A , A , split_batches=A , even_batches=A ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. UpperCAmelCase : Dict = BatchSampler(range(21 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : Union[str, Any] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(A , A , split_batches=A , even_batches=A ) UpperCAmelCase : Any = BatchSampler(range(21 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : Dict = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(A , A , split_batches=A , even_batches=A ) # Check the shards when the dataset is very small. UpperCAmelCase : str = BatchSampler(range(2 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : Dict = [[[0, 1]], []] self.check_batch_sampler_shards(A , A , split_batches=A , even_batches=A ) UpperCAmelCase : Any = BatchSampler(range(2 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : Dict = [[], []] self.check_batch_sampler_shards(A , A , split_batches=A , even_batches=A ) def _lowercase( self ) -> Optional[int]: UpperCAmelCase : Optional[int] = [[0, 1, 2], [3, 4], [5, 6, 7, 8], [9, 10, 11], [12, 13]] UpperCAmelCase : List[str] = [BatchSamplerShard(A , 2 , A , even_batches=A ) for i in range(2 )] self.assertEqual(len(batch_sampler_shards[0] ) , 3 ) self.assertEqual(len(batch_sampler_shards[1] ) , 2 ) self.assertListEqual(list(batch_sampler_shards[0] ) , [[0, 1, 2], [5, 6, 7, 8], [12, 13]] ) self.assertListEqual(list(batch_sampler_shards[1] ) , [[3, 4], [9, 10, 11]] ) def _lowercase( self , A , A , A , A=False , A=2 , A=False ) -> Tuple: random.seed(A ) UpperCAmelCase : Dict = list(A ) UpperCAmelCase : Any = [ IterableDatasetShard( A , batch_size=A , drop_last=A , num_processes=A , process_index=A , split_batches=A , ) for i in range(A ) ] UpperCAmelCase : Dict = [] for iterable_dataset_shard in iterable_dataset_shards: # Since our random iterable dataset will be... random... we need to use a seed to get reproducible results. random.seed(A ) iterable_dataset_lists.append(list(A ) ) UpperCAmelCase : Optional[Any] = batch_size // num_processes if split_batches else batch_size # All iterable dataset shard should have the same length, a round multiple of shard_batch_size UpperCAmelCase : List[Any] = iterable_dataset_lists[0] for l in iterable_dataset_lists[1:]: self.assertEqual(len(A ) , len(A ) ) self.assertTrue(len(A ) % shard_batch_size == 0 ) UpperCAmelCase : List[Any] = [] for idx in range(0 , len(A ) , A ): for l in iterable_dataset_lists: observed += l[idx : idx + shard_batch_size] if not drop_last: while len(A ) < len(A ): reference += reference self.assertListEqual(A , reference[: len(A )] ) def _lowercase( self ) -> str: UpperCAmelCase : Tuple = 42 UpperCAmelCase : List[Any] = RandomIterableDataset() self.check_iterable_dataset_shards(A , A , batch_size=4 , drop_last=A , split_batches=A ) self.check_iterable_dataset_shards(A , A , batch_size=4 , drop_last=A , split_batches=A ) self.check_iterable_dataset_shards(A , A , batch_size=4 , drop_last=A , split_batches=A ) self.check_iterable_dataset_shards(A , A , batch_size=4 , drop_last=A , split_batches=A ) # Edge case with a very small dataset UpperCAmelCase : List[Any] = RandomIterableDataset(max_length=2 ) self.check_iterable_dataset_shards(A , A , batch_size=4 , drop_last=A , split_batches=A ) self.check_iterable_dataset_shards(A , A , batch_size=4 , drop_last=A , split_batches=A ) self.check_iterable_dataset_shards(A , A , batch_size=4 , drop_last=A , split_batches=A ) self.check_iterable_dataset_shards(A , A , batch_size=4 , drop_last=A , split_batches=A ) def _lowercase( self ) -> Tuple: UpperCAmelCase : Dict = BatchSampler(range(16 ) , batch_size=4 , drop_last=A ) UpperCAmelCase : Any = SkipBatchSampler(A , 2 ) self.assertListEqual(list(A ) , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def _lowercase( self ) -> int: UpperCAmelCase : Any = SkipDataLoader(list(range(16 ) ) , batch_size=4 , skip_batches=2 ) self.assertListEqual([t.tolist() for t in dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def _lowercase( self ) -> Union[str, Any]: UpperCAmelCase : List[Any] = DataLoader(list(range(16 ) ) , batch_size=4 ) UpperCAmelCase : Optional[Any] = skip_first_batches(A , num_batches=2 ) self.assertListEqual([t.tolist() for t in new_dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def _lowercase( self ) -> Optional[Any]: UpperCAmelCase : Optional[int] = DataLoaderShard(list(range(16 ) ) , batch_size=4 ) for idx, _ in enumerate(A ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(A ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) def _lowercase( self ) -> Dict: Accelerator() UpperCAmelCase : Union[str, Any] = DataLoaderDispatcher(range(16 ) , batch_size=4 ) for idx, _ in enumerate(A ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(A ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 )
338
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available lowerCAmelCase__ = {'configuration_speech_encoder_decoder': ['SpeechEncoderDecoderConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['SpeechEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['FlaxSpeechEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_speech_encoder_decoder import SpeechEncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_encoder_decoder import SpeechEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_speech_encoder_decoder import FlaxSpeechEncoderDecoderModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
11
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''facebook/dpr-ctx_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-reader-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-ctx_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-reader-multiset-base''': ( '''https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json''' ), } class SCREAMING_SNAKE_CASE__ ( lowercase ): """simple docstring""" a : int ="dpr" def __init__( self , snake_case__=30_522 , snake_case__=768 , snake_case__=12 , snake_case__=12 , snake_case__=3_072 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=512 , snake_case__=2 , snake_case__=0.02 , snake_case__=1e-12 , snake_case__=0 , snake_case__="absolute" , snake_case__ = 0 , **snake_case__ , ): """simple docstring""" super().__init__(pad_token_id=snake_case__ , **snake_case__ ) lowerCAmelCase : Union[str, Any] = vocab_size lowerCAmelCase : str = hidden_size lowerCAmelCase : Any = num_hidden_layers lowerCAmelCase : Optional[int] = num_attention_heads lowerCAmelCase : Union[str, Any] = hidden_act lowerCAmelCase : Dict = intermediate_size lowerCAmelCase : Union[str, Any] = hidden_dropout_prob lowerCAmelCase : Dict = attention_probs_dropout_prob lowerCAmelCase : Dict = max_position_embeddings lowerCAmelCase : Tuple = type_vocab_size lowerCAmelCase : Any = initializer_range lowerCAmelCase : Any = layer_norm_eps lowerCAmelCase : Dict = projection_dim lowerCAmelCase : Dict = position_embedding_type
108
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_funnel import FunnelTokenizer _UpperCamelCase = logging.get_logger(__name__) _UpperCamelCase = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} _UpperCamelCase = [ '''small''', '''small-base''', '''medium''', '''medium-base''', '''intermediate''', '''intermediate-base''', '''large''', '''large-base''', '''xlarge''', '''xlarge-base''', ] _UpperCamelCase = { '''vocab_file''': { '''funnel-transformer/small''': '''https://huggingface.co/funnel-transformer/small/resolve/main/vocab.txt''', '''funnel-transformer/small-base''': '''https://huggingface.co/funnel-transformer/small-base/resolve/main/vocab.txt''', '''funnel-transformer/medium''': '''https://huggingface.co/funnel-transformer/medium/resolve/main/vocab.txt''', '''funnel-transformer/medium-base''': ( '''https://huggingface.co/funnel-transformer/medium-base/resolve/main/vocab.txt''' ), '''funnel-transformer/intermediate''': ( '''https://huggingface.co/funnel-transformer/intermediate/resolve/main/vocab.txt''' ), '''funnel-transformer/intermediate-base''': ( '''https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/vocab.txt''' ), '''funnel-transformer/large''': '''https://huggingface.co/funnel-transformer/large/resolve/main/vocab.txt''', '''funnel-transformer/large-base''': '''https://huggingface.co/funnel-transformer/large-base/resolve/main/vocab.txt''', '''funnel-transformer/xlarge''': '''https://huggingface.co/funnel-transformer/xlarge/resolve/main/vocab.txt''', '''funnel-transformer/xlarge-base''': ( '''https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''funnel-transformer/small''': '''https://huggingface.co/funnel-transformer/small/resolve/main/tokenizer.json''', '''funnel-transformer/small-base''': ( '''https://huggingface.co/funnel-transformer/small-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/medium''': '''https://huggingface.co/funnel-transformer/medium/resolve/main/tokenizer.json''', '''funnel-transformer/medium-base''': ( '''https://huggingface.co/funnel-transformer/medium-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/intermediate''': ( '''https://huggingface.co/funnel-transformer/intermediate/resolve/main/tokenizer.json''' ), '''funnel-transformer/intermediate-base''': ( '''https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/large''': '''https://huggingface.co/funnel-transformer/large/resolve/main/tokenizer.json''', '''funnel-transformer/large-base''': ( '''https://huggingface.co/funnel-transformer/large-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/xlarge''': '''https://huggingface.co/funnel-transformer/xlarge/resolve/main/tokenizer.json''', '''funnel-transformer/xlarge-base''': ( '''https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/tokenizer.json''' ), }, } _UpperCamelCase = {F'funnel-transformer/{name}': 512 for name in _model_names} _UpperCamelCase = {F'funnel-transformer/{name}': {'''do_lower_case''': True} for name in _model_names} class _A ( __SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : int = VOCAB_FILES_NAMES _SCREAMING_SNAKE_CASE : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP _SCREAMING_SNAKE_CASE : Dict = PRETRAINED_INIT_CONFIGURATION _SCREAMING_SNAKE_CASE : Optional[int] = FunnelTokenizer _SCREAMING_SNAKE_CASE : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _SCREAMING_SNAKE_CASE : int = 2 def __init__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=True , __UpperCAmelCase="<unk>" , __UpperCAmelCase="<sep>" , __UpperCAmelCase="<pad>" , __UpperCAmelCase="<cls>" , __UpperCAmelCase="<mask>" , __UpperCAmelCase="<s>" , __UpperCAmelCase="</s>" , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=None , __UpperCAmelCase="##" , **__UpperCAmelCase , ) -> List[Any]: '''simple docstring''' super().__init__( __UpperCAmelCase , tokenizer_file=__UpperCAmelCase , do_lower_case=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , clean_text=__UpperCAmelCase , tokenize_chinese_chars=__UpperCAmelCase , strip_accents=__UpperCAmelCase , wordpieces_prefix=__UpperCAmelCase , **__UpperCAmelCase , ) __UpperCAmelCase : Optional[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , __UpperCAmelCase ) != do_lower_case or normalizer_state.get("""strip_accents""" , __UpperCAmelCase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , __UpperCAmelCase ) != tokenize_chinese_chars ): __UpperCAmelCase : Optional[int] = getattr(__UpperCAmelCase , normalizer_state.pop("""type""" ) ) __UpperCAmelCase : List[Any] = do_lower_case __UpperCAmelCase : List[str] = strip_accents __UpperCAmelCase : Optional[Any] = tokenize_chinese_chars __UpperCAmelCase : List[Any] = normalizer_class(**__UpperCAmelCase ) __UpperCAmelCase : Union[str, Any] = do_lower_case def __A ( self , __UpperCAmelCase , __UpperCAmelCase=None ) -> Dict: '''simple docstring''' __UpperCAmelCase : Tuple = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __A ( self , __UpperCAmelCase , __UpperCAmelCase = None ) -> List[int]: '''simple docstring''' __UpperCAmelCase : Optional[Any] = [self.sep_token_id] __UpperCAmelCase : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __A ( self , __UpperCAmelCase , __UpperCAmelCase = None ) -> Tuple[str]: '''simple docstring''' __UpperCAmelCase : List[str] = self._tokenizer.model.save(__UpperCAmelCase , name=__UpperCAmelCase ) return tuple(__UpperCAmelCase )
16
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LDMTextToImagePipeline, UNetaDConditionModel from diffusers.utils.testing_utils import ( enable_full_determinism, load_numpy, nightly, require_torch_gpu, slow, torch_device, ) from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class _A ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): _SCREAMING_SNAKE_CASE : Dict = LDMTextToImagePipeline _SCREAMING_SNAKE_CASE : Tuple = TEXT_TO_IMAGE_PARAMS - { "negative_prompt", "negative_prompt_embeds", "cross_attention_kwargs", "prompt_embeds", } _SCREAMING_SNAKE_CASE : List[Any] = PipelineTesterMixin.required_optional_params - { "num_images_per_prompt", "callback", "callback_steps", } _SCREAMING_SNAKE_CASE : Dict = TEXT_TO_IMAGE_BATCH_PARAMS _SCREAMING_SNAKE_CASE : List[str] = False def __A ( self ) -> Optional[int]: '''simple docstring''' torch.manual_seed(0 ) __UpperCAmelCase : Dict = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) __UpperCAmelCase : List[Any] = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=__UpperCAmelCase , set_alpha_to_one=__UpperCAmelCase , ) torch.manual_seed(0 ) __UpperCAmelCase : Any = AutoencoderKL( block_out_channels=(32, 64) , in_channels=3 , out_channels=3 , down_block_types=("""DownEncoderBlock2D""", """DownEncoderBlock2D""") , up_block_types=("""UpDecoderBlock2D""", """UpDecoderBlock2D""") , latent_channels=4 , ) torch.manual_seed(0 ) __UpperCAmelCase : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) __UpperCAmelCase : Tuple = CLIPTextModel(__UpperCAmelCase ) __UpperCAmelCase : Tuple = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) __UpperCAmelCase : Dict = { """unet""": unet, """scheduler""": scheduler, """vqvae""": vae, """bert""": text_encoder, """tokenizer""": tokenizer, } return components def __A ( self , __UpperCAmelCase , __UpperCAmelCase=0 ) -> Any: '''simple docstring''' if str(__UpperCAmelCase ).startswith("""mps""" ): __UpperCAmelCase : int = torch.manual_seed(__UpperCAmelCase ) else: __UpperCAmelCase : List[str] = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) __UpperCAmelCase : Dict = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def __A ( self ) -> Optional[Any]: '''simple docstring''' __UpperCAmelCase : Optional[int] = """cpu""" # ensure determinism for the device-dependent torch.Generator __UpperCAmelCase : Dict = self.get_dummy_components() __UpperCAmelCase : Tuple = LDMTextToImagePipeline(**__UpperCAmelCase ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __UpperCAmelCase : Optional[Any] = self.get_dummy_inputs(__UpperCAmelCase ) __UpperCAmelCase : Union[str, Any] = pipe(**__UpperCAmelCase ).images __UpperCAmelCase : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 16, 16, 3) __UpperCAmelCase : Dict = np.array([0.6101, 0.6156, 0.5622, 0.4895, 0.6661, 0.3804, 0.5748, 0.6136, 0.5014] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class _A ( unittest.TestCase ): def __A ( self ) -> List[str]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __A ( self , __UpperCAmelCase , __UpperCAmelCase=torch.floataa , __UpperCAmelCase=0 ) -> int: '''simple docstring''' __UpperCAmelCase : Tuple = torch.manual_seed(__UpperCAmelCase ) __UpperCAmelCase : int = np.random.RandomState(__UpperCAmelCase ).standard_normal((1, 4, 32, 32) ) __UpperCAmelCase : int = torch.from_numpy(__UpperCAmelCase ).to(device=__UpperCAmelCase , dtype=__UpperCAmelCase ) __UpperCAmelCase : Tuple = { """prompt""": """A painting of a squirrel eating a burger""", """latents""": latents, """generator""": generator, """num_inference_steps""": 3, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def __A ( self ) -> str: '''simple docstring''' __UpperCAmelCase : Any = LDMTextToImagePipeline.from_pretrained("""CompVis/ldm-text2im-large-256""" ).to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __UpperCAmelCase : Optional[Any] = self.get_inputs(__UpperCAmelCase ) __UpperCAmelCase : int = pipe(**__UpperCAmelCase ).images __UpperCAmelCase : Tuple = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 256, 256, 3) __UpperCAmelCase : Tuple = np.array([0.5_1825, 0.5_2850, 0.5_2543, 0.5_4258, 0.5_2304, 0.5_2569, 0.5_4363, 0.5_5276, 0.5_6878] ) __UpperCAmelCase : Union[str, Any] = np.abs(expected_slice - image_slice ).max() assert max_diff < 1E-3 @nightly @require_torch_gpu class _A ( unittest.TestCase ): def __A ( self ) -> Optional[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __A ( self , __UpperCAmelCase , __UpperCAmelCase=torch.floataa , __UpperCAmelCase=0 ) -> List[str]: '''simple docstring''' __UpperCAmelCase : Optional[Any] = torch.manual_seed(__UpperCAmelCase ) __UpperCAmelCase : List[Any] = np.random.RandomState(__UpperCAmelCase ).standard_normal((1, 4, 32, 32) ) __UpperCAmelCase : int = torch.from_numpy(__UpperCAmelCase ).to(device=__UpperCAmelCase , dtype=__UpperCAmelCase ) __UpperCAmelCase : Optional[Any] = { """prompt""": """A painting of a squirrel eating a burger""", """latents""": latents, """generator""": generator, """num_inference_steps""": 50, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def __A ( self ) -> Optional[Any]: '''simple docstring''' __UpperCAmelCase : Union[str, Any] = LDMTextToImagePipeline.from_pretrained("""CompVis/ldm-text2im-large-256""" ).to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) __UpperCAmelCase : Union[str, Any] = self.get_inputs(__UpperCAmelCase ) __UpperCAmelCase : Optional[int] = pipe(**__UpperCAmelCase ).images[0] __UpperCAmelCase : Tuple = load_numpy( """https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/ldm_text2img/ldm_large_256_ddim.npy""" ) __UpperCAmelCase : Dict = np.abs(expected_image - image ).max() assert max_diff < 1E-3
16
1
"""simple docstring""" from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class lowerCamelCase__ : """simple docstring""" __a = 42 # [batch_size x 3] __a = 42 # [batch_size x 3] __a = 42 # [batch_size x 3] __a = 42 # [batch_size x 3] __a = 42 __a = 42 __a = 42 __a = 42 __a = 42 def lowerCamelCase__ ( self : Any ): '''simple docstring''' assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def lowerCamelCase__ ( self : Optional[int] ): '''simple docstring''' return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def lowerCamelCase__ ( self : Dict ): '''simple docstring''' return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def lowerCamelCase__ ( self : str ): '''simple docstring''' __UpperCAmelCase : int = torch.arange(self.height * self.width ) __UpperCAmelCase : Union[str, Any] = torch.stack( [ pixel_indices % self.width, torch.div(_lowercase , self.width , rounding_mode="""trunc""" ), ] , axis=1 , ) return coords @property def lowerCamelCase__ ( self : str ): '''simple docstring''' __UpperCAmelCase ,*__UpperCAmelCase : Optional[int] = self.shape __UpperCAmelCase : Optional[int] = int(np.prod(_lowercase ) ) __UpperCAmelCase : Any = self.get_image_coords() __UpperCAmelCase : Optional[Any] = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) __UpperCAmelCase : Union[str, Any] = self.get_camera_rays(_lowercase ) __UpperCAmelCase : Dict = rays.view(_lowercase , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def lowerCamelCase__ ( self : str , UpperCamelCase : torch.Tensor ): '''simple docstring''' __UpperCAmelCase ,*__UpperCAmelCase ,__UpperCAmelCase : List[str] = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] __UpperCAmelCase : List[str] = coords.view(_lowercase , -1 , 2 ) __UpperCAmelCase : Dict = self.resolution() __UpperCAmelCase : Tuple = self.fov() __UpperCAmelCase : Union[str, Any] = (flat.float() / (res - 1)) * 2 - 1 __UpperCAmelCase : int = fracs * torch.tan(fov / 2 ) __UpperCAmelCase : Any = fracs.view(_lowercase , -1 , 2 ) __UpperCAmelCase : str = ( self.z.view(_lowercase , 1 , 3 ) + self.x.view(_lowercase , 1 , 3 ) * fracs[:, :, :1] + self.y.view(_lowercase , 1 , 3 ) * fracs[:, :, 1:] ) __UpperCAmelCase : str = directions / directions.norm(dim=-1 , keepdim=_lowercase ) __UpperCAmelCase : str = torch.stack( [ torch.broadcast_to(self.origin.view(_lowercase , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(_lowercase , *_lowercase , 2 , 3 ) def lowerCamelCase__ ( self : Dict , UpperCamelCase : int , UpperCamelCase : int ): '''simple docstring''' assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=_lowercase , height=_lowercase , x_fov=self.x_fov , y_fov=self.y_fov , ) def lowerCamelCase ( _UpperCamelCase : int ) -> DifferentiableProjectiveCamera: '''simple docstring''' __UpperCAmelCase : List[str] = [] __UpperCAmelCase : List[str] = [] __UpperCAmelCase : Optional[int] = [] __UpperCAmelCase : Dict = [] for theta in np.linspace(0 , 2 * np.pi , num=2_0 ): __UpperCAmelCase : List[Any] = np.array([np.sin(__UpperCamelCase ), np.cos(__UpperCamelCase ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) __UpperCAmelCase : str = -z * 4 __UpperCAmelCase : int = np.array([np.cos(__UpperCamelCase ), -np.sin(__UpperCamelCase ), 0.0] ) __UpperCAmelCase : List[str] = np.cross(__UpperCamelCase , __UpperCamelCase ) origins.append(__UpperCamelCase ) xs.append(__UpperCamelCase ) ys.append(__UpperCamelCase ) zs.append(__UpperCamelCase ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(__UpperCamelCase , axis=0 ) ).float() , x=torch.from_numpy(np.stack(__UpperCamelCase , axis=0 ) ).float() , y=torch.from_numpy(np.stack(__UpperCamelCase , axis=0 ) ).float() , z=torch.from_numpy(np.stack(__UpperCamelCase , axis=0 ) ).float() , width=__UpperCamelCase , height=__UpperCamelCase , x_fov=0.7 , y_fov=0.7 , shape=(1, len(__UpperCamelCase )) , )
115
import socket def __SCREAMING_SNAKE_CASE ( ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = socket.socket(socket.AF_INET , socket.SOCK_STREAM ) SCREAMING_SNAKE_CASE__ = socket.gethostname() SCREAMING_SNAKE_CASE__ = 1_23_12 sock.connect((host, port) ) sock.send(B"""Hello server!""" ) with open("""Received_file""" , """wb""" ) as out_file: print("""File opened""" ) print("""Receiving data...""" ) while True: SCREAMING_SNAKE_CASE__ = sock.recv(10_24 ) if not data: break out_file.write(__UpperCamelCase ) print("""Successfully received the file""" ) sock.close() print("""Connection closed""" ) if __name__ == "__main__": main()
219
0
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase : Tuple = logging.get_logger(__name__) def snake_case_ ( lowerCAmelCase_ : str , lowerCAmelCase_ : str=False ): __lowercase : Dict = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"blocks.{i}.norm1.weight", F"vit.encoder.layer.{i}.layernorm_before.weight") ) rename_keys.append((F"blocks.{i}.norm1.bias", F"vit.encoder.layer.{i}.layernorm_before.bias") ) rename_keys.append((F"blocks.{i}.attn.proj.weight", F"vit.encoder.layer.{i}.attention.output.dense.weight") ) rename_keys.append((F"blocks.{i}.attn.proj.bias", F"vit.encoder.layer.{i}.attention.output.dense.bias") ) rename_keys.append((F"blocks.{i}.norm2.weight", F"vit.encoder.layer.{i}.layernorm_after.weight") ) rename_keys.append((F"blocks.{i}.norm2.bias", F"vit.encoder.layer.{i}.layernorm_after.bias") ) rename_keys.append((F"blocks.{i}.mlp.fc1.weight", F"vit.encoder.layer.{i}.intermediate.dense.weight") ) rename_keys.append((F"blocks.{i}.mlp.fc1.bias", F"vit.encoder.layer.{i}.intermediate.dense.bias") ) rename_keys.append((F"blocks.{i}.mlp.fc2.weight", F"vit.encoder.layer.{i}.output.dense.weight") ) rename_keys.append((F"blocks.{i}.mlp.fc2.bias", F"vit.encoder.layer.{i}.output.dense.bias") ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """vit.embeddings.cls_token"""), ("""patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """vit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" __lowercase : int = [(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def snake_case_ ( lowerCAmelCase_ : int , lowerCAmelCase_ : Any , lowerCAmelCase_ : Dict=False ): for i in range(config.num_hidden_layers ): if base_model: __lowercase : str = '''''' else: __lowercase : Optional[Any] = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) __lowercase : Optional[int] = state_dict.pop(F"blocks.{i}.attn.qkv.weight" ) __lowercase : Any = state_dict.pop(F"blocks.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict __lowercase : Dict = in_proj_weight[ : config.hidden_size, : ] __lowercase : List[Any] = in_proj_bias[: config.hidden_size] __lowercase : List[str] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __lowercase : Optional[int] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] __lowercase : str = in_proj_weight[ -config.hidden_size :, : ] __lowercase : Union[str, Any] = in_proj_bias[-config.hidden_size :] def snake_case_ ( lowerCAmelCase_ : Optional[int] ): __lowercase : str = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(UpperCamelCase__ , UpperCamelCase__ ) def snake_case_ ( lowerCAmelCase_ : Tuple , lowerCAmelCase_ : Any , lowerCAmelCase_ : str ): __lowercase : int = dct.pop(UpperCamelCase__ ) __lowercase : Optional[int] = val def snake_case_ ( ): __lowercase : List[str] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __lowercase : Any = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return im @torch.no_grad() def snake_case_ ( lowerCAmelCase_ : Dict , lowerCAmelCase_ : Optional[Any] ): __lowercase : Tuple = ViTConfig() __lowercase : Optional[int] = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": __lowercase : Optional[int] = True __lowercase : List[str] = int(vit_name[-12:-10] ) __lowercase : Dict = int(vit_name[-9:-6] ) else: __lowercase : Optional[int] = 1000 __lowercase : Dict = '''huggingface/label-files''' __lowercase : Optional[Any] = '''imagenet-1k-id2label.json''' __lowercase : Tuple = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) __lowercase : Optional[int] = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} __lowercase : Optional[int] = idalabel __lowercase : Union[str, Any] = {v: k for k, v in idalabel.items()} __lowercase : List[Any] = int(vit_name[-6:-4] ) __lowercase : Any = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith("""tiny""" ): __lowercase : Any = 192 __lowercase : Tuple = 768 __lowercase : str = 12 __lowercase : int = 3 elif vit_name[9:].startswith("""small""" ): __lowercase : Optional[int] = 384 __lowercase : List[Any] = 1536 __lowercase : Optional[Any] = 12 __lowercase : List[Any] = 6 else: pass else: if vit_name[4:].startswith("""small""" ): __lowercase : Dict = 768 __lowercase : List[Any] = 2304 __lowercase : Optional[int] = 8 __lowercase : Any = 8 elif vit_name[4:].startswith("""base""" ): pass elif vit_name[4:].startswith("""large""" ): __lowercase : Union[str, Any] = 1024 __lowercase : Optional[Any] = 4096 __lowercase : str = 24 __lowercase : List[Any] = 16 elif vit_name[4:].startswith("""huge""" ): __lowercase : Any = 1280 __lowercase : Optional[int] = 5120 __lowercase : Dict = 32 __lowercase : str = 16 # load original model from timm __lowercase : Any = timm.create_model(UpperCamelCase__ , pretrained=UpperCamelCase__ ) timm_model.eval() # load state_dict of original model, remove and rename some keys __lowercase : Union[str, Any] = timm_model.state_dict() if base_model: remove_classification_head_(UpperCamelCase__ ) __lowercase : Tuple = create_rename_keys(UpperCamelCase__ , UpperCamelCase__ ) for src, dest in rename_keys: rename_key(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) read_in_q_k_v(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # load HuggingFace model if vit_name[-5:] == "in21k": __lowercase : Any = ViTModel(UpperCamelCase__ ).eval() else: __lowercase : Dict = ViTForImageClassification(UpperCamelCase__ ).eval() model.load_state_dict(UpperCamelCase__ ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: __lowercase : Union[str, Any] = DeiTImageProcessor(size=config.image_size ) else: __lowercase : Optional[Any] = ViTImageProcessor(size=config.image_size ) __lowercase : Union[str, Any] = image_processor(images=prepare_img() , return_tensors="""pt""" ) __lowercase : Optional[int] = encoding['''pixel_values'''] __lowercase : List[Any] = model(UpperCamelCase__ ) if base_model: __lowercase : int = timm_model.forward_features(UpperCamelCase__ ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(UpperCamelCase__ , outputs.pooler_output , atol=1e-3 ) else: __lowercase : Tuple = timm_model(UpperCamelCase__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(UpperCamelCase__ , outputs.logits , atol=1e-3 ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) print(F"Saving model {vit_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(UpperCamelCase__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": lowerCamelCase : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--vit_name''', default='''vit_base_patch16_224''', type=str, help='''Name of the ViT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) lowerCamelCase : Dict = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
352
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def lowerCAmelCase ( self : Union[str, Any] ) -> Optional[Any]: """simple docstring""" __lowercase : Dict = TFCamembertModel.from_pretrained("""jplu/tf-camembert-base""" ) __lowercase : List[str] = tf.convert_to_tensor( [[5, 121, 11, 660, 16, 730, 25543, 110, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" __lowercase : Optional[Any] = model(__a )["""last_hidden_state"""] __lowercase : Any = tf.TensorShape((1, 10, 768) ) self.assertEqual(output.shape , __a ) # compare the actual values for a slice. __lowercase : Dict = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
306
0
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class A : def __init__(self : Any , __UpperCAmelCase : Any , __UpperCAmelCase : Tuple=1_3 , __UpperCAmelCase : int=7 , __UpperCAmelCase : List[Any]=True , __UpperCAmelCase : Any=True , __UpperCAmelCase : Dict=True , __UpperCAmelCase : str=True , __UpperCAmelCase : Union[str, Any]=9_9 , __UpperCAmelCase : Tuple=3_2 , __UpperCAmelCase : List[Any]=2 , __UpperCAmelCase : List[Any]=4 , __UpperCAmelCase : str=3_7 , __UpperCAmelCase : Union[str, Any]="gelu" , __UpperCAmelCase : Tuple=0.1 , __UpperCAmelCase : List[Any]=0.1 , __UpperCAmelCase : Optional[Any]=5_1_2 , __UpperCAmelCase : Tuple=1_6 , __UpperCAmelCase : Any=2 , __UpperCAmelCase : Union[str, Any]=0.02 , __UpperCAmelCase : Optional[int]=3 , __UpperCAmelCase : str=4 , __UpperCAmelCase : str=None , ) -> Any: """simple docstring""" UpperCAmelCase__ = parent UpperCAmelCase__ = 1_3 UpperCAmelCase__ = 7 UpperCAmelCase__ = True UpperCAmelCase__ = True UpperCAmelCase__ = True UpperCAmelCase__ = True UpperCAmelCase__ = 9_9 UpperCAmelCase__ = 3_8_4 UpperCAmelCase__ = 2 UpperCAmelCase__ = 4 UpperCAmelCase__ = 3_7 UpperCAmelCase__ = "gelu" UpperCAmelCase__ = 0.1 UpperCAmelCase__ = 0.1 UpperCAmelCase__ = 5_1_2 UpperCAmelCase__ = 1_6 UpperCAmelCase__ = 2 UpperCAmelCase__ = 0.02 UpperCAmelCase__ = 3 UpperCAmelCase__ = 4 UpperCAmelCase__ = 1_2_8 UpperCAmelCase__ = 2 UpperCAmelCase__ = 9 UpperCAmelCase__ = 1 UpperCAmelCase__ = None def lowercase_ (self : Any ) -> str: """simple docstring""" UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase__ = None if self.use_input_mask: UpperCAmelCase__ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ = None if self.use_token_type_ids: UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None if self.use_labels: UpperCAmelCase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase__ = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase__ = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=__UpperCAmelCase , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowercase_ (self : Optional[int] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Dict , __UpperCAmelCase : List[Any] , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : Tuple ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = TFConvBertModel(config=__UpperCAmelCase ) UpperCAmelCase__ = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} UpperCAmelCase__ = [input_ids, input_mask] UpperCAmelCase__ = model(__UpperCAmelCase ) UpperCAmelCase__ = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Tuple , __UpperCAmelCase : List[str] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Dict , __UpperCAmelCase : Dict , __UpperCAmelCase : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = TFConvBertForMaskedLM(config=__UpperCAmelCase ) UpperCAmelCase__ = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCAmelCase__ = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowercase_ (self : List[Any] , __UpperCAmelCase : int , __UpperCAmelCase : str , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : List[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase__ = self.num_labels UpperCAmelCase__ = TFConvBertForSequenceClassification(config=__UpperCAmelCase ) UpperCAmelCase__ = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCAmelCase__ = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : List[str] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : List[str] ) -> str: """simple docstring""" UpperCAmelCase__ = self.num_choices UpperCAmelCase__ = TFConvBertForMultipleChoice(config=__UpperCAmelCase ) UpperCAmelCase__ = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) UpperCAmelCase__ = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) UpperCAmelCase__ = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) UpperCAmelCase__ = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } UpperCAmelCase__ = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowercase_ (self : Any , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : str , __UpperCAmelCase : int ) -> int: """simple docstring""" UpperCAmelCase__ = self.num_labels UpperCAmelCase__ = TFConvBertForTokenClassification(config=__UpperCAmelCase ) UpperCAmelCase__ = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCAmelCase__ = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowercase_ (self : str , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Any , __UpperCAmelCase : int , __UpperCAmelCase : List[str] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Any ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = TFConvBertForQuestionAnswering(config=__UpperCAmelCase ) UpperCAmelCase__ = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCAmelCase__ = model(__UpperCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowercase_ (self : Tuple ) -> Dict: """simple docstring""" UpperCAmelCase__ = self.prepare_config_and_inputs() ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) = config_and_inputs UpperCAmelCase__ = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class A ( UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): __UpperCAmelCase : Tuple = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) __UpperCAmelCase : Optional[Any] = ( { 'feature-extraction': TFConvBertModel, 'fill-mask': TFConvBertForMaskedLM, 'question-answering': TFConvBertForQuestionAnswering, 'text-classification': TFConvBertForSequenceClassification, 'token-classification': TFConvBertForTokenClassification, 'zero-shot': TFConvBertForSequenceClassification, } if is_tf_available() else {} ) __UpperCAmelCase : List[Any] = False __UpperCAmelCase : Optional[Any] = False __UpperCAmelCase : str = False def lowercase_ (self : Tuple ) -> str: """simple docstring""" UpperCAmelCase__ = TFConvBertModelTester(self ) UpperCAmelCase__ = ConfigTester(self , config_class=__UpperCAmelCase , hidden_size=3_7 ) def lowercase_ (self : str ) -> Dict: """simple docstring""" self.config_tester.run_common_tests() def lowercase_ (self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def lowercase_ (self : str ) -> Tuple: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCAmelCase ) def lowercase_ (self : int ) -> Dict: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCAmelCase ) def lowercase_ (self : Tuple ) -> str: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCAmelCase ) def lowercase_ (self : Optional[Any] ) -> int: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCAmelCase ) def lowercase_ (self : Optional[int] ) -> str: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCAmelCase ) @slow def lowercase_ (self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ = True UpperCAmelCase__ = True if hasattr(__UpperCAmelCase , "use_cache" ): UpperCAmelCase__ = True UpperCAmelCase__ = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) UpperCAmelCase__ = getattr(self.model_tester , "key_length" , __UpperCAmelCase ) for model_class in self.all_model_classes: UpperCAmelCase__ = self._prepare_for_class(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = model_class(__UpperCAmelCase ) UpperCAmelCase__ = len(model(__UpperCAmelCase ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__UpperCAmelCase , saved_model=__UpperCAmelCase ) UpperCAmelCase__ = os.path.join(__UpperCAmelCase , "saved_model" , "1" ) UpperCAmelCase__ = tf.keras.models.load_model(__UpperCAmelCase ) UpperCAmelCase__ = model(__UpperCAmelCase ) if self.is_encoder_decoder: UpperCAmelCase__ = outputs["encoder_hidden_states"] UpperCAmelCase__ = outputs["encoder_attentions"] else: UpperCAmelCase__ = outputs["hidden_states"] UpperCAmelCase__ = outputs["attentions"] self.assertEqual(len(__UpperCAmelCase ) , __UpperCAmelCase ) UpperCAmelCase__ = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(__UpperCAmelCase ) , __UpperCAmelCase ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(__UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def lowercase_ (self : Any ) -> Any: """simple docstring""" UpperCAmelCase__ = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) self.assertIsNotNone(__UpperCAmelCase ) def lowercase_ (self : int ) -> List[str]: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ = True UpperCAmelCase__ = getattr(self.model_tester , "decoder_seq_length" , self.model_tester.seq_length ) UpperCAmelCase__ = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) UpperCAmelCase__ = getattr(self.model_tester , "key_length" , __UpperCAmelCase ) UpperCAmelCase__ = getattr(self.model_tester , "key_length" , __UpperCAmelCase ) def check_decoder_attentions_output(__UpperCAmelCase : List[str] ): UpperCAmelCase__ = len(__UpperCAmelCase ) self.assertEqual(out_len % 2 , 0 ) UpperCAmelCase__ = outputs.decoder_attentions self.assertEqual(len(__UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(__UpperCAmelCase : List[str] ): UpperCAmelCase__ = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(__UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: UpperCAmelCase__ = True UpperCAmelCase__ = False UpperCAmelCase__ = model_class(__UpperCAmelCase ) UpperCAmelCase__ = model(self._prepare_for_class(__UpperCAmelCase , __UpperCAmelCase ) ) UpperCAmelCase__ = len(__UpperCAmelCase ) self.assertEqual(config.output_hidden_states , __UpperCAmelCase ) check_encoder_attentions_output(__UpperCAmelCase ) if self.is_encoder_decoder: UpperCAmelCase__ = model_class(__UpperCAmelCase ) UpperCAmelCase__ = model(self._prepare_for_class(__UpperCAmelCase , __UpperCAmelCase ) ) self.assertEqual(config.output_hidden_states , __UpperCAmelCase ) check_decoder_attentions_output(__UpperCAmelCase ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] UpperCAmelCase__ = True UpperCAmelCase__ = model_class(__UpperCAmelCase ) UpperCAmelCase__ = model(self._prepare_for_class(__UpperCAmelCase , __UpperCAmelCase ) ) self.assertEqual(config.output_hidden_states , __UpperCAmelCase ) check_encoder_attentions_output(__UpperCAmelCase ) # Check attention is always last and order is fine UpperCAmelCase__ = True UpperCAmelCase__ = True UpperCAmelCase__ = model_class(__UpperCAmelCase ) UpperCAmelCase__ = model(self._prepare_for_class(__UpperCAmelCase , __UpperCAmelCase ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(__UpperCAmelCase ) ) self.assertEqual(model.config.output_hidden_states , __UpperCAmelCase ) check_encoder_attentions_output(__UpperCAmelCase ) @require_tf class A ( unittest.TestCase ): @slow def lowercase_ (self : Dict ) -> Any: """simple docstring""" UpperCAmelCase__ = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) UpperCAmelCase__ = tf.constant([[0, 1, 2, 3, 4, 5]] ) UpperCAmelCase__ = model(__UpperCAmelCase )[0] UpperCAmelCase__ = [1, 6, 7_6_8] self.assertEqual(output.shape , __UpperCAmelCase ) UpperCAmelCase__ = tf.constant( [ [ [-0.03475493, -0.4686034, -0.30638832], [0.22637248, -0.26988646, -0.7423424], [0.10324868, -0.45013508, -0.58280784], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , __UpperCAmelCase , atol=1E-4 )
65
from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": __lowerCamelCase : str = input('''Enter image url: ''').strip() print(F"""Downloading image from {url} ...""") __lowerCamelCase : Any = BeautifulSoup(requests.get(url).content, '''html.parser''') # The image URL is in the content field of the first meta tag with property og:image __lowerCamelCase : List[Any] = soup.find('''meta''', {'''property''': '''og:image'''})['''content'''] __lowerCamelCase : Tuple = requests.get(image_url).content __lowerCamelCase : Union[str, Any] = F"""{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg""" with open(file_name, '''wb''') as fp: fp.write(image_data) print(F"""Done. Image saved to disk as {file_name}.""")
219
0
import argparse import torch from torch import nn from transformers import SpeechaTextConfig, SpeechaTextForConditionalGeneration def __lowerCamelCase (UpperCAmelCase__ : str ): SCREAMING_SNAKE_CASE = [ "encoder.version", "decoder.version", "model.encoder.version", "model.decoder.version", "decoder.output_projection.weight", "_float_tensor", "encoder.embed_positions._float_tensor", "decoder.embed_positions._float_tensor", ] for k in ignore_keys: state_dict.pop(UpperCAmelCase__ , UpperCAmelCase__ ) def __lowerCamelCase (UpperCAmelCase__ : Dict ): SCREAMING_SNAKE_CASE = list(s_dict.keys() ) for key in keys: if "transformer_layers" in key: SCREAMING_SNAKE_CASE = s_dict.pop(UpperCAmelCase__ ) elif "subsample" in key: SCREAMING_SNAKE_CASE = s_dict.pop(UpperCAmelCase__ ) def __lowerCamelCase (UpperCAmelCase__ : str ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = emb.weight.shape SCREAMING_SNAKE_CASE = nn.Linear(UpperCAmelCase__ , UpperCAmelCase__ , bias=UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = emb.weight.data return lin_layer def __lowerCamelCase (UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[Any] ): SCREAMING_SNAKE_CASE = torch.load(UpperCAmelCase__ , map_location="cpu" ) SCREAMING_SNAKE_CASE = mam_aaa["args"] SCREAMING_SNAKE_CASE = mam_aaa["model"] SCREAMING_SNAKE_CASE = state_dict["decoder.output_projection.weight"] remove_ignore_keys_(UpperCAmelCase__ ) rename_keys(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = state_dict["decoder.embed_tokens.weight"].shape[0] SCREAMING_SNAKE_CASE = args.share_decoder_input_output_embed SCREAMING_SNAKE_CASE = [int(UpperCAmelCase__ ) for i in args.conv_kernel_sizes.split("," )] SCREAMING_SNAKE_CASE = SpeechaTextConfig( vocab_size=UpperCAmelCase__ , max_source_positions=args.max_source_positions , max_target_positions=args.max_target_positions , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function="relu" , num_conv_layers=len(UpperCAmelCase__ ) , conv_channels=args.conv_channels , conv_kernel_sizes=UpperCAmelCase__ , input_feat_per_channel=args.input_feat_per_channel , input_channels=args.input_channels , tie_word_embeddings=UpperCAmelCase__ , num_beams=5 , max_length=2_0_0 , use_cache=UpperCAmelCase__ , decoder_start_token_id=2 , early_stopping=UpperCAmelCase__ , ) SCREAMING_SNAKE_CASE = SpeechaTextForConditionalGeneration(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = model.model.load_state_dict(UpperCAmelCase__ , strict=UpperCAmelCase__ ) if len(UpperCAmelCase__ ) > 0 and not set(UpperCAmelCase__ ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( "Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing," F" but all the following weights are missing {missing}" ) if tie_embeds: SCREAMING_SNAKE_CASE = make_linear_from_emb(model.model.decoder.embed_tokens ) else: SCREAMING_SNAKE_CASE = lm_head_weights model.save_pretrained(UpperCAmelCase__ ) if __name__ == "__main__": _lowerCamelCase : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument('''--fairseq_path''', type=str, help='''Path to the fairseq model (.pt) file.''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') _lowerCamelCase : Any = parser.parse_args() convert_fairseq_sat_checkpoint_to_tfms(args.fairseq_path, args.pytorch_dump_folder_path)
206
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_funnel import FunnelTokenizer _lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) _lowerCamelCase : Dict = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} _lowerCamelCase : List[Any] = [ '''small''', '''small-base''', '''medium''', '''medium-base''', '''intermediate''', '''intermediate-base''', '''large''', '''large-base''', '''xlarge''', '''xlarge-base''', ] _lowerCamelCase : Optional[Any] = { '''vocab_file''': { '''funnel-transformer/small''': '''https://huggingface.co/funnel-transformer/small/resolve/main/vocab.txt''', '''funnel-transformer/small-base''': '''https://huggingface.co/funnel-transformer/small-base/resolve/main/vocab.txt''', '''funnel-transformer/medium''': '''https://huggingface.co/funnel-transformer/medium/resolve/main/vocab.txt''', '''funnel-transformer/medium-base''': ( '''https://huggingface.co/funnel-transformer/medium-base/resolve/main/vocab.txt''' ), '''funnel-transformer/intermediate''': ( '''https://huggingface.co/funnel-transformer/intermediate/resolve/main/vocab.txt''' ), '''funnel-transformer/intermediate-base''': ( '''https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/vocab.txt''' ), '''funnel-transformer/large''': '''https://huggingface.co/funnel-transformer/large/resolve/main/vocab.txt''', '''funnel-transformer/large-base''': '''https://huggingface.co/funnel-transformer/large-base/resolve/main/vocab.txt''', '''funnel-transformer/xlarge''': '''https://huggingface.co/funnel-transformer/xlarge/resolve/main/vocab.txt''', '''funnel-transformer/xlarge-base''': ( '''https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''funnel-transformer/small''': '''https://huggingface.co/funnel-transformer/small/resolve/main/tokenizer.json''', '''funnel-transformer/small-base''': ( '''https://huggingface.co/funnel-transformer/small-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/medium''': '''https://huggingface.co/funnel-transformer/medium/resolve/main/tokenizer.json''', '''funnel-transformer/medium-base''': ( '''https://huggingface.co/funnel-transformer/medium-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/intermediate''': ( '''https://huggingface.co/funnel-transformer/intermediate/resolve/main/tokenizer.json''' ), '''funnel-transformer/intermediate-base''': ( '''https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/large''': '''https://huggingface.co/funnel-transformer/large/resolve/main/tokenizer.json''', '''funnel-transformer/large-base''': ( '''https://huggingface.co/funnel-transformer/large-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/xlarge''': '''https://huggingface.co/funnel-transformer/xlarge/resolve/main/tokenizer.json''', '''funnel-transformer/xlarge-base''': ( '''https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/tokenizer.json''' ), }, } _lowerCamelCase : Any = {f"""funnel-transformer/{name}""": 5_12 for name in _model_names} _lowerCamelCase : Optional[Any] = {f"""funnel-transformer/{name}""": {'''do_lower_case''': True} for name in _model_names} class lowercase ( a ): lowercase__ : Optional[int] = VOCAB_FILES_NAMES lowercase__ : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP lowercase__ : Optional[Any] = PRETRAINED_INIT_CONFIGURATION lowercase__ : Union[str, Any] = FunnelTokenizer lowercase__ : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : int = 2 def __init__( self : str , _UpperCamelCase : str=None , _UpperCamelCase : str=None , _UpperCamelCase : Union[str, Any]=True , _UpperCamelCase : str="<unk>" , _UpperCamelCase : Optional[Any]="<sep>" , _UpperCamelCase : Optional[int]="<pad>" , _UpperCamelCase : int="<cls>" , _UpperCamelCase : Dict="<mask>" , _UpperCamelCase : Union[str, Any]="<s>" , _UpperCamelCase : Optional[int]="</s>" , _UpperCamelCase : Dict=True , _UpperCamelCase : List[Any]=True , _UpperCamelCase : Any=None , _UpperCamelCase : Dict="##" , **_UpperCamelCase : Dict , ) -> Optional[int]: '''simple docstring''' super().__init__( _UpperCamelCase , tokenizer_file=_UpperCamelCase , do_lower_case=_UpperCamelCase , unk_token=_UpperCamelCase , sep_token=_UpperCamelCase , pad_token=_UpperCamelCase , cls_token=_UpperCamelCase , mask_token=_UpperCamelCase , bos_token=_UpperCamelCase , eos_token=_UpperCamelCase , clean_text=_UpperCamelCase , tokenize_chinese_chars=_UpperCamelCase , strip_accents=_UpperCamelCase , wordpieces_prefix=_UpperCamelCase , **_UpperCamelCase , ) SCREAMING_SNAKE_CASE = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , _UpperCamelCase ) != do_lower_case or normalizer_state.get("strip_accents" , _UpperCamelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , _UpperCamelCase ) != tokenize_chinese_chars ): SCREAMING_SNAKE_CASE = getattr(_UpperCamelCase , normalizer_state.pop("type" ) ) SCREAMING_SNAKE_CASE = do_lower_case SCREAMING_SNAKE_CASE = strip_accents SCREAMING_SNAKE_CASE = tokenize_chinese_chars SCREAMING_SNAKE_CASE = normalizer_class(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = do_lower_case def __snake_case( self : Union[str, Any] , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Dict=None ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __snake_case( self : int , _UpperCamelCase : List[int] , _UpperCamelCase : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = [self.sep_token_id] SCREAMING_SNAKE_CASE = [self.cls_token_id] if token_ids_a is None: return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __snake_case( self : Optional[Any] , _UpperCamelCase : str , _UpperCamelCase : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = self._tokenizer.model.save(_UpperCamelCase , name=_UpperCamelCase ) return tuple(_UpperCamelCase )
206
1
'''simple docstring''' import os from argparse import ArgumentParser from typing import List import torch.utils.data from datasets import Dataset, IterableDataset from datasets.distributed import split_dataset_by_node __lowercase : List[Any] = 4 __lowercase : Dict = 3 class __lowercase ( UpperCamelCase_ ): pass def lowercase_ ( _lowercase ) -> Optional[Any]: '''simple docstring''' for shard in shards: for i in range(snake_case__ ): yield {"i": i, "shard": shard} def lowercase_ ( ) -> Tuple: '''simple docstring''' lowerCamelCase_ : Union[str, Any] = int(os.environ['''RANK'''] ) lowerCamelCase_ : List[Any] = int(os.environ['''WORLD_SIZE'''] ) lowerCamelCase_ : Optional[int] = ArgumentParser() parser.add_argument('''--streaming''' , type=snake_case__ ) parser.add_argument('''--local_rank''' , type=snake_case__ ) parser.add_argument('''--num_workers''' , type=snake_case__ , default=0 ) lowerCamelCase_ : Tuple = parser.parse_args() lowerCamelCase_ : Union[str, Any] = args.streaming lowerCamelCase_ : int = args.num_workers lowerCamelCase_ : Optional[int] = {'''shards''': [F"""shard_{shard_idx}""" for shard_idx in range(snake_case__ )]} lowerCamelCase_ : Any = IterableDataset.from_generator(snake_case__ , gen_kwargs=snake_case__ ) if not streaming: lowerCamelCase_ : Any = Dataset.from_list(list(snake_case__ ) ) lowerCamelCase_ : List[str] = split_dataset_by_node(snake_case__ , rank=snake_case__ , world_size=snake_case__ ) lowerCamelCase_ : int = torch.utils.data.DataLoader(snake_case__ , num_workers=snake_case__ ) lowerCamelCase_ : str = NUM_SHARDS * NUM_ITEMS_PER_SHARD lowerCamelCase_ : str = full_size // world_size expected_local_size += int(rank < (full_size % world_size) ) lowerCamelCase_ : Optional[int] = sum(1 for _ in dataloader ) if local_size != expected_local_size: raise FailedTestError(F"""local_size {local_size} != expected_local_size {expected_local_size}""" ) if __name__ == "__main__": main()
318
from ...processing_utils import ProcessorMixin class lowercase_ ( UpperCamelCase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = ["""image_processor""", """feature_extractor"""] UpperCAmelCase_ : Optional[int] = """TvltImageProcessor""" UpperCAmelCase_ : Optional[int] = """TvltFeatureExtractor""" def __init__( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ->Optional[int]: super().__init__(image_processor=__SCREAMING_SNAKE_CASE , feature_extractor=__SCREAMING_SNAKE_CASE ) lowerCAmelCase = image_processor lowerCAmelCase = feature_extractor def __call__( self , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=False , __SCREAMING_SNAKE_CASE=False , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , ) ->List[Any]: if images is None and audio is None: raise ValueError('''You need to specify either an `images` or `audio` input to process.''' ) lowerCAmelCase = None if images is not None: lowerCAmelCase = self.image_processor(__SCREAMING_SNAKE_CASE , mask_pixel=__SCREAMING_SNAKE_CASE , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) if images_mixed is not None: lowerCAmelCase = self.image_processor(__SCREAMING_SNAKE_CASE , is_mixed=__SCREAMING_SNAKE_CASE , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) if audio is not None: lowerCAmelCase = self.feature_extractor( __SCREAMING_SNAKE_CASE , *__SCREAMING_SNAKE_CASE , sampling_rate=__SCREAMING_SNAKE_CASE , mask_audio=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) lowerCAmelCase = {} if audio is not None: output_dict.update(__SCREAMING_SNAKE_CASE ) if images is not None: output_dict.update(__SCREAMING_SNAKE_CASE ) if images_mixed_dict is not None: output_dict.update(__SCREAMING_SNAKE_CASE ) return output_dict @property def SCREAMING_SNAKE_CASE_ ( self ) ->Any: lowerCAmelCase = self.image_processor.model_input_names lowerCAmelCase = self.feature_extractor.model_input_names return list(dict.fromkeys(image_processor_input_names + feature_extractor_input_names ) )
338
0
"""simple docstring""" from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : Optional[int] = logging.get_logger(__name__) lowerCamelCase : Dict = { '''huggingface/informer-tourism-monthly''': ( '''https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json''' ), # See all Informer models at https://huggingface.co/models?filter=informer } class A( a_ ): '''simple docstring''' UpperCamelCase = '''informer''' UpperCamelCase = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', '''num_hidden_layers''': '''encoder_layers''', } def __init__( self : str , A_ : Optional[int] = None , A_ : Optional[int] = None , A_ : str = "student_t" , A_ : str = "nll" , A_ : int = 1 , A_ : List[int] = None , A_ : Optional[Union[str, bool]] = "mean" , A_ : int = 0 , A_ : int = 0 , A_ : int = 0 , A_ : int = 0 , A_ : Optional[List[int]] = None , A_ : Optional[List[int]] = None , A_ : int = 64 , A_ : int = 32 , A_ : int = 32 , A_ : int = 2 , A_ : int = 2 , A_ : int = 2 , A_ : int = 2 , A_ : bool = True , A_ : str = "gelu" , A_ : float = 0.05 , A_ : float = 0.1 , A_ : float = 0.1 , A_ : float = 0.1 , A_ : float = 0.1 , A_ : int = 100 , A_ : float = 0.02 , A_ : Union[str, Any]=True , A_ : str = "prob" , A_ : int = 5 , A_ : bool = True , **A_ : List[str] , ) -> Any: """simple docstring""" lowerCamelCase_ = prediction_length lowerCamelCase_ = context_length or prediction_length lowerCamelCase_ = distribution_output lowerCamelCase_ = loss lowerCamelCase_ = input_size lowerCamelCase_ = num_time_features lowerCamelCase_ = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] lowerCamelCase_ = scaling lowerCamelCase_ = num_dynamic_real_features lowerCamelCase_ = num_static_real_features lowerCamelCase_ = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(lowercase_ ) != num_static_categorical_features: raise ValueError( 'The cardinality should be a list of the same length as `num_static_categorical_features`' ) lowerCamelCase_ = cardinality else: lowerCamelCase_ = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(lowercase_ ) != num_static_categorical_features: raise ValueError( 'The embedding dimension should be a list of the same length as `num_static_categorical_features`' ) lowerCamelCase_ = embedding_dimension else: lowerCamelCase_ = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] lowerCamelCase_ = num_parallel_samples # Transformer architecture configuration lowerCamelCase_ = input_size * len(self.lags_sequence ) + self._number_of_features lowerCamelCase_ = d_model lowerCamelCase_ = encoder_attention_heads lowerCamelCase_ = decoder_attention_heads lowerCamelCase_ = encoder_ffn_dim lowerCamelCase_ = decoder_ffn_dim lowerCamelCase_ = encoder_layers lowerCamelCase_ = decoder_layers lowerCamelCase_ = dropout lowerCamelCase_ = attention_dropout lowerCamelCase_ = activation_dropout lowerCamelCase_ = encoder_layerdrop lowerCamelCase_ = decoder_layerdrop lowerCamelCase_ = activation_function lowerCamelCase_ = init_std lowerCamelCase_ = use_cache # Informer lowerCamelCase_ = attention_type lowerCamelCase_ = sampling_factor lowerCamelCase_ = distil super().__init__(is_encoder_decoder=lowercase_ , **lowercase_ ) @property def a__ ( self : Optional[int] ) -> Tuple: """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
358
# HF Trainer benchmarking tool # # This tool can be used to run and compare multiple dimensions of the HF Trainers args. # # It then prints a report once in github format with all the information that needs to be shared # with others and second time in a console-friendly format, so it's easier to use for tuning things up. # # The main idea is: # # ./trainer-benchmark.py --base-cmd '<cmd args that don't change>' \ # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' \ # --target-metric-key train_samples_per_second # # The variations can be any command line argument that you want to compare and not just dtype as in # the example. # # --variations allows you to compare variations in multiple dimensions. # # as the first dimention has 2 options and the second 3 in our example, this will run the trainer 6 # times adding one of: # # 1. --tf32 0 --fp16 0 # 2. --tf32 0 --fp16 1 # 3. --tf32 0 --bf16 1 # 4. --tf32 1 --fp16 0 # 5. --tf32 1 --fp16 1 # 6. --tf32 1 --bf16 1 # # and print the results. This is just a cartesian product - and more than 2 dimensions can be used. # # If you want to rely on defaults, this: # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' # is identical to this: # --variations '--tf32 0|--tf32 1' '|--fp16|--bf16' # # the leading empty variation in the 2nd dimension is a valid variation. # # So here we get the following 6 variations: # # 1. --tf32 0 # 2. --tf32 0 --fp16 # 3. --tf32 0 --bf16 # 4. --tf32 1 # 5. --tf32 1 --fp16 # 6. --tf32 1 --bf16 # # In this particular case we don't know what the default tf32 setting is as it's normally # pytorch-version dependent). That's why it's best to do an explicit setting of each variation: # `--tf32 0|--tf32 1` # # Here is a full example of a train: # # CUDA_VISIBLE_DEVICES=0 python ./scripts/benchmark/trainer-benchmark.py \ # --base-cmd \ # ' examples/pytorch/translation/run_translation.py --model_name_or_path t5-small \ # --output_dir output_dir --do_train --label_smoothing 0.1 --logging_strategy no \ # --save_strategy no --per_device_train_batch_size 32 --max_source_length 512 \ # --max_target_length 512 --num_train_epochs 1 --overwrite_output_dir \ # --source_lang en --target_lang ro --dataset_name wmt16 --dataset_config "ro-en" \ # --source_prefix "translate English to Romanian: " --warmup_steps 50 \ # --max_train_samples 20000 --dataloader_num_workers 2 ' \ # --target-metric-key train_samples_per_second --repeat-times 1 --variations \ # '|--fp16|--bf16' '--tf32 0|--tf32 1' --report-metric-keys train_loss \ # --repeat-times 1 --base-variation '--tf32 0' # # and here is a possible output: # # # | Variation | Train | Diff | Train | # | | samples | % | loss | # | | per | | | # | | second | | | # |:----------------|----------:|-------:|--------:| # | --tf32 0 | 285.11 | 0 | 2.51 | # | --tf32 1 | 342.09 | 20 | 2.51 | # | --fp16 --tf32 0 | 423.49 | 49 | 2.51 | # | --fp16 --tf32 1 | 423.13 | 48 | 2.51 | # | --bf16 --tf32 0 | 416.80 | 46 | 2.52 | # | --bf16 --tf32 1 | 415.87 | 46 | 2.52 | # # # So you can quickly compare the different outcomes. # # Typically running each experiment once is enough, but if the environment is unstable you can # re-run each multiple times, e.g., 3 using --repeat-times 3 and it will report the averaged results. # # By default it'll use the lowest result as the base line to use as 100% and then compare the rest to # it as can be seen from the table above, but you can also specify which combination is the one to use as # the baseline, e.g., to change to another entry use: --base-variation '--tf32 1 --fp16 0' # # --target-metric-key is there to tell the program which metrics to compare - the different metric keys are # inside output_dir/all_results.json. e.g., to measure eval performance instead of train use: # --target-metric-key eval_samples_per_second # but of course you will need to adjust the --base-cmd value in the example to perform evaluation as # well (as currently it doesn't) # import argparse import datetime import io import itertools import json import math import os import platform import re import shlex import subprocess import sys from pathlib import Path from statistics import fmean import pandas as pd import torch from tqdm import tqdm import transformers lowerCamelCase : Optional[int] = float("nan") class A: '''simple docstring''' def __init__( self : Optional[Any] , A_ : int ) -> Dict: """simple docstring""" lowerCamelCase_ = sys.stdout lowerCamelCase_ = open(A_ , 'a' ) def __getattr__( self : List[Any] , A_ : Optional[int] ) -> str: """simple docstring""" return getattr(self.stdout , A_ ) def a__ ( self : int , A_ : int ) -> List[str]: """simple docstring""" self.stdout.write(A_ ) # strip tqdm codes self.file.write(re.sub(r'^.*\r' , '' , A_ , 0 , re.M ) ) def _SCREAMING_SNAKE_CASE ( lowercase : str=80 , lowercase : Tuple=False ): '''simple docstring''' lowerCamelCase_ = [] # deal with critical env vars lowerCamelCase_ = ['CUDA_VISIBLE_DEVICES'] for key in env_keys: lowerCamelCase_ = os.environ.get(lowercase , lowercase ) if val is not None: cmd.append(f"""{key}={val}""" ) # python executable (not always needed if the script is executable) lowerCamelCase_ = sys.executable if full_python_path else sys.executable.split('/' )[-1] cmd.append(lowercase ) # now the normal args cmd += list(map(shlex.quote , sys.argv ) ) # split up into up to MAX_WIDTH lines with shell multi-line escapes lowerCamelCase_ = [] lowerCamelCase_ = '' while len(lowercase ) > 0: current_line += f"""{cmd.pop(0 )} """ if len(lowercase ) == 0 or len(lowercase ) + len(cmd[0] ) + 1 > max_width - 1: lines.append(lowercase ) lowerCamelCase_ = '' return "\\\n".join(lowercase ) def _SCREAMING_SNAKE_CASE ( lowercase : Optional[int] , lowercase : Tuple ): '''simple docstring''' lowerCamelCase_ = re.sub(r'[\\\n]+' , ' ' , args.base_cmd ) # remove --output_dir if any and set our own lowerCamelCase_ = re.sub('--output_dir\s+[^\s]+' , '' , args.base_cmd ) args.base_cmd += f""" --output_dir {output_dir}""" # ensure we have --overwrite_output_dir lowerCamelCase_ = re.sub('--overwrite_output_dir\s+' , '' , args.base_cmd ) args.base_cmd += " --overwrite_output_dir" return [sys.executable] + shlex.split(args.base_cmd ) def _SCREAMING_SNAKE_CASE ( lowercase : Dict , lowercase : int , lowercase : Dict , lowercase : List[str] , lowercase : List[str] , lowercase : List[str] , lowercase : Dict ): '''simple docstring''' if 0: import random from time import sleep sleep(0 ) return dict( {k: random.uniform(0 , 1_00 ) for k in metric_keys} , **{target_metric_key: random.choice([nan, 10.31, 100.2, 55.6666, 222.2222_2222] )} , ) lowerCamelCase_ = subprocess.run(lowercase , capture_output=lowercase , text=lowercase ) if verbose: print('STDOUT' , result.stdout ) print('STDERR' , result.stderr ) # save the streams lowerCamelCase_ = variation.replace(' ' , '-' ) with open(Path(lowercase ) / f"""log.{prefix}.stdout.txt""" , 'w' ) as f: f.write(result.stdout ) with open(Path(lowercase ) / f"""log.{prefix}.stderr.txt""" , 'w' ) as f: f.write(result.stderr ) if result.returncode != 0: if verbose: print('failed' ) return {target_metric_key: nan} with io.open(f"""{output_dir}/all_results.json""" , 'r' , encoding='utf-8' ) as f: lowerCamelCase_ = json.load(lowercase ) # filter out just the keys we want return {k: v for k, v in metrics.items() if k in metric_keys} def _SCREAMING_SNAKE_CASE ( lowercase : Dict , lowercase : Dict , lowercase : Optional[Any] , lowercase : List[Any] , lowercase : Optional[Any] , lowercase : Optional[int] , lowercase : List[str] , lowercase : Dict , lowercase : Any , lowercase : int , ): '''simple docstring''' lowerCamelCase_ = [] lowerCamelCase_ = [] lowerCamelCase_ = f"""{id}: {variation:<{longest_variation_len}}""" lowerCamelCase_ = f"""{preamble}: """ lowerCamelCase_ = set(report_metric_keys + [target_metric_key] ) for i in tqdm(range(lowercase ) , desc=lowercase , leave=lowercase ): lowerCamelCase_ = process_run_single( lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ) lowerCamelCase_ = single_run_metrics[target_metric_key] if not math.isnan(lowercase ): metrics.append(lowercase ) results.append(lowercase ) outcome += "✓" else: outcome += "✘" lowerCamelCase_ = f"""\33[2K\r{outcome}""" if len(lowercase ) > 0: lowerCamelCase_ = {k: fmean([x[k] for x in metrics] ) for k in metrics[0].keys()} lowerCamelCase_ = round(mean_metrics[target_metric_key] , 2 ) lowerCamelCase_ = f"""{outcome} {mean_target}""" if len(lowercase ) > 1: results_str += f""" {tuple(round(lowercase , 2 ) for x in results )}""" print(lowercase ) lowerCamelCase_ = variation return mean_metrics else: print(lowercase ) return {variation_key: variation, target_metric_key: nan} def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = torch.cuda.get_device_properties(torch.device('cuda' ) ) return f""" Datetime : {datetime.datetime.now().strftime("%Y-%m-%d %H:%M:%S" )} Software: transformers: {transformers.__version__} torch : {torch.__version__} cuda : {torch.version.cuda} python : {platform.python_version()} Hardware: {torch.cuda.device_count()} GPUs : {properties.name}, {properties.total_memory/2**30:0.2f}GB """ def _SCREAMING_SNAKE_CASE ( lowercase : Any , lowercase : Union[str, Any] , lowercase : Optional[Any] , lowercase : Union[str, Any] , lowercase : Union[str, Any] ): '''simple docstring''' lowerCamelCase_ = pd.DataFrame(lowercase ) lowerCamelCase_ = 'variation' lowerCamelCase_ = 'diff_%' lowerCamelCase_ = nan if base_variation is not None and len(df[df[variation_key] == base_variation] ): # this may still return nan lowerCamelCase_ = df.loc[df[variation_key] == base_variation][target_metric_key].item() if math.isnan(lowercase ): # as a fallback, use the minimal value as the sentinel lowerCamelCase_ = df.loc[df[target_metric_key] != nan][target_metric_key].min() # create diff column if possible if not math.isnan(lowercase ): lowerCamelCase_ = df.apply( lambda lowercase : round(1_00 * (r[target_metric_key] - sentinel_value) / sentinel_value ) if not math.isnan(r[target_metric_key] ) else 0 , axis='columns' , ) # re-order columns lowerCamelCase_ = [variation_key, target_metric_key, diff_key, *report_metric_keys] lowerCamelCase_ = df.reindex(lowercase , axis='columns' ) # reorder cols # capitalize lowerCamelCase_ = df.rename(str.capitalize , axis='columns' ) # make the cols as narrow as possible lowerCamelCase_ = df.rename(lambda lowercase : c.replace('_' , '<br>' ) , axis='columns' ) lowerCamelCase_ = df.rename(lambda lowercase : c.replace('_' , '\n' ) , axis='columns' ) lowerCamelCase_ = ['', 'Copy between the cut-here-lines and paste as is to github or a forum'] report += ["----------8<-----------------8<--------"] report += ["*** Results:", df_github.to_markdown(index=lowercase , floatfmt='.2f' )] report += ["```"] report += ["*** Setup:", get_versions()] report += ["*** The benchmark command line was:", get_original_command()] report += ["```"] report += ["----------8<-----------------8<--------"] report += ["*** Results (console):", df_console.to_markdown(index=lowercase , floatfmt='.2f' )] print('\n\n'.join(lowercase ) ) def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = argparse.ArgumentParser() parser.add_argument( '--base-cmd' , default=lowercase , type=lowercase , required=lowercase , help='Base cmd' , ) parser.add_argument( '--variations' , default=lowercase , type=lowercase , nargs='+' , required=lowercase , help='Multi-dimensional variations, example: \'|--fp16|--bf16\' \'|--tf32\'' , ) parser.add_argument( '--base-variation' , default=lowercase , type=lowercase , help='Baseline variation to compare to. if None the minimal target value will be used to compare against' , ) parser.add_argument( '--target-metric-key' , default=lowercase , type=lowercase , required=lowercase , help='Target metric key in output_dir/all_results.json, e.g., train_samples_per_second' , ) parser.add_argument( '--report-metric-keys' , default='' , type=lowercase , help='Report metric keys - other metric keys from output_dir/all_results.json to report, e.g., train_loss. Use a single argument e.g., \'train_loss train_samples' , ) parser.add_argument( '--repeat-times' , default=1 , type=lowercase , help='How many times to re-run each variation - an average will be reported' , ) parser.add_argument( '--output_dir' , default='output_benchmark' , type=lowercase , help='The output directory where all the benchmark reports will go to and additionally this directory will be used to override --output_dir in the script that is being benchmarked' , ) parser.add_argument( '--verbose' , default=lowercase , action='store_true' , help='Whether to show the outputs of each run or just the benchmark progress' , ) lowerCamelCase_ = parser.parse_args() lowerCamelCase_ = args.output_dir Path(lowercase ).mkdir(exist_ok=lowercase ) lowerCamelCase_ = get_base_command(lowercase , lowercase ) # split each dimension into its --foo variations lowerCamelCase_ = [list(map(str.strip , re.split(r'\|' , lowercase ) ) ) for x in args.variations] # build a cartesian product of dimensions and convert those back into cmd-line arg strings, # while stripping white space for inputs that were empty lowerCamelCase_ = list(map(str.strip , map(' '.join , itertools.product(*lowercase ) ) ) ) lowerCamelCase_ = max(len(lowercase ) for x in variations ) # split wanted keys lowerCamelCase_ = args.report_metric_keys.split() # capture prints into a log file for convenience lowerCamelCase_ = f"""benchmark-report-{datetime.datetime.now().strftime("%Y-%m-%d-%H-%M-%S" )}.txt""" print(f"""\nNote: each run's output is also logged under {output_dir}/log.*.std*.txt""" ) print(f"""and this script's output is also piped into {report_fn}""" ) lowerCamelCase_ = Tee(lowercase ) print(f"""\n*** Running {len(lowercase )} benchmarks:""" ) print(f"""Base command: {" ".join(lowercase )}""" ) lowerCamelCase_ = 'variation' lowerCamelCase_ = [] for id, variation in enumerate(tqdm(lowercase , desc='Total completion: ' , leave=lowercase ) ): lowerCamelCase_ = base_cmd + variation.split() results.append( process_run( id + 1 , lowercase , lowercase , lowercase , lowercase , args.target_metric_key , lowercase , args.repeat_times , lowercase , args.verbose , ) ) process_results(lowercase , args.target_metric_key , lowercase , args.base_variation , lowercase ) if __name__ == "__main__": main()
208
0
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_funnel import FunnelTokenizer lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} lowerCAmelCase_ = [ 'small', 'small-base', 'medium', 'medium-base', 'intermediate', 'intermediate-base', 'large', 'large-base', 'xlarge', 'xlarge-base', ] lowerCAmelCase_ = { 'vocab_file': { 'funnel-transformer/small': 'https://huggingface.co/funnel-transformer/small/resolve/main/vocab.txt', 'funnel-transformer/small-base': 'https://huggingface.co/funnel-transformer/small-base/resolve/main/vocab.txt', 'funnel-transformer/medium': 'https://huggingface.co/funnel-transformer/medium/resolve/main/vocab.txt', 'funnel-transformer/medium-base': ( 'https://huggingface.co/funnel-transformer/medium-base/resolve/main/vocab.txt' ), 'funnel-transformer/intermediate': ( 'https://huggingface.co/funnel-transformer/intermediate/resolve/main/vocab.txt' ), 'funnel-transformer/intermediate-base': ( 'https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/vocab.txt' ), 'funnel-transformer/large': 'https://huggingface.co/funnel-transformer/large/resolve/main/vocab.txt', 'funnel-transformer/large-base': 'https://huggingface.co/funnel-transformer/large-base/resolve/main/vocab.txt', 'funnel-transformer/xlarge': 'https://huggingface.co/funnel-transformer/xlarge/resolve/main/vocab.txt', 'funnel-transformer/xlarge-base': ( 'https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'funnel-transformer/small': 'https://huggingface.co/funnel-transformer/small/resolve/main/tokenizer.json', 'funnel-transformer/small-base': ( 'https://huggingface.co/funnel-transformer/small-base/resolve/main/tokenizer.json' ), 'funnel-transformer/medium': 'https://huggingface.co/funnel-transformer/medium/resolve/main/tokenizer.json', 'funnel-transformer/medium-base': ( 'https://huggingface.co/funnel-transformer/medium-base/resolve/main/tokenizer.json' ), 'funnel-transformer/intermediate': ( 'https://huggingface.co/funnel-transformer/intermediate/resolve/main/tokenizer.json' ), 'funnel-transformer/intermediate-base': ( 'https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/tokenizer.json' ), 'funnel-transformer/large': 'https://huggingface.co/funnel-transformer/large/resolve/main/tokenizer.json', 'funnel-transformer/large-base': ( 'https://huggingface.co/funnel-transformer/large-base/resolve/main/tokenizer.json' ), 'funnel-transformer/xlarge': 'https://huggingface.co/funnel-transformer/xlarge/resolve/main/tokenizer.json', 'funnel-transformer/xlarge-base': ( 'https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/tokenizer.json' ), }, } lowerCAmelCase_ = {F'''funnel-transformer/{name}''': 512 for name in _model_names} lowerCAmelCase_ = {F'''funnel-transformer/{name}''': {'do_lower_case': True} for name in _model_names} class __A ( A_ ): '''simple docstring''' lowerCAmelCase : str = VOCAB_FILES_NAMES lowerCAmelCase : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase : int = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase : Dict = FunnelTokenizer lowerCAmelCase : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase : int = 2 def __init__( self : Optional[int] ,_snake_case : Dict=None ,_snake_case : List[str]=None ,_snake_case : Optional[Any]=True ,_snake_case : Optional[int]="<unk>" ,_snake_case : Dict="<sep>" ,_snake_case : Any="<pad>" ,_snake_case : str="<cls>" ,_snake_case : Optional[Any]="<mask>" ,_snake_case : int="<s>" ,_snake_case : Dict="</s>" ,_snake_case : Optional[int]=True ,_snake_case : List[str]=True ,_snake_case : Dict=None ,_snake_case : str="##" ,**_snake_case : Optional[Any] ,) -> Any: """simple docstring""" super().__init__( _snake_case ,tokenizer_file=_snake_case ,do_lower_case=_snake_case ,unk_token=_snake_case ,sep_token=_snake_case ,pad_token=_snake_case ,cls_token=_snake_case ,mask_token=_snake_case ,bos_token=_snake_case ,eos_token=_snake_case ,clean_text=_snake_case ,tokenize_chinese_chars=_snake_case ,strip_accents=_snake_case ,wordpieces_prefix=_snake_case ,**_snake_case ,) lowercase__ : Union[str, Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' ,_snake_case ) != do_lower_case or normalizer_state.get('''strip_accents''' ,_snake_case ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' ,_snake_case ) != tokenize_chinese_chars ): lowercase__ : List[str] = getattr(_snake_case ,normalizer_state.pop('''type''' ) ) lowercase__ : List[str] = do_lower_case lowercase__ : Any = strip_accents lowercase__ : Union[str, Any] = tokenize_chinese_chars lowercase__ : Union[str, Any] = normalizer_class(**_snake_case ) lowercase__ : List[str] = do_lower_case def UpperCAmelCase ( self : int ,_snake_case : Tuple ,_snake_case : Tuple=None ) -> Optional[Any]: """simple docstring""" lowercase__ : Any = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCAmelCase ( self : Tuple ,_snake_case : List[int] ,_snake_case : Optional[List[int]] = None ) -> List[int]: """simple docstring""" lowercase__ : Optional[int] = [self.sep_token_id] lowercase__ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCAmelCase ( self : Tuple ,_snake_case : str ,_snake_case : Optional[str] = None ) -> Tuple[str]: """simple docstring""" lowercase__ : str = self._tokenizer.model.save(_snake_case ,name=_snake_case ) return tuple(_snake_case )
16
"""simple docstring""" import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## lowerCAmelCase_ = 16 lowerCAmelCase_ = 32 def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase = 16 ) -> Optional[Any]: lowercase__ : Optional[Any] = AutoTokenizer.from_pretrained('''bert-base-cased''' ) lowercase__ : int = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(__lowerCamelCase ): # max_length=None => use the model max length (it's actually the default) lowercase__ : str = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=__lowerCamelCase , max_length=__lowerCamelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowercase__ : str = datasets.map( __lowerCamelCase , batched=__lowerCamelCase , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowercase__ : Union[str, Any] = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(__lowerCamelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. lowercase__ : List[str] = 1_28 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowercase__ : Optional[int] = 16 elif accelerator.mixed_precision != "no": lowercase__ : List[Any] = 8 else: lowercase__ : int = None return tokenizer.pad( __lowerCamelCase , padding='''longest''' , max_length=__lowerCamelCase , pad_to_multiple_of=__lowerCamelCase , return_tensors='''pt''' , ) # Instantiate dataloaders. lowercase__ : List[Any] = DataLoader( tokenized_datasets['''train'''] , shuffle=__lowerCamelCase , collate_fn=__lowerCamelCase , batch_size=__lowerCamelCase ) lowercase__ : str = DataLoader( tokenized_datasets['''validation'''] , shuffle=__lowerCamelCase , collate_fn=__lowerCamelCase , batch_size=__lowerCamelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders lowerCAmelCase_ = mocked_dataloaders # noqa: F811 def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase ) -> str: # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , __lowerCamelCase ) == "1": lowercase__ : List[Any] = 2 # Initialize accelerator lowercase__ : Optional[int] = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowercase__ : str = config['''lr'''] lowercase__ : str = int(config['''num_epochs'''] ) lowercase__ : Optional[int] = int(config['''seed'''] ) lowercase__ : Tuple = int(config['''batch_size'''] ) lowercase__ : List[Any] = evaluate.load('''glue''' , '''mrpc''' ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=__lowerCamelCase ) def inner_training_loop(__lowerCamelCase ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(__lowerCamelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowercase__ : List[str] = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=__lowerCamelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowercase__ : Tuple = model.to(accelerator.device ) # Instantiate optimizer lowercase__ : List[str] = AdamW(params=model.parameters() , lr=__lowerCamelCase ) lowercase__ , lowercase__ : List[Any] = get_dataloaders(__lowerCamelCase , __lowerCamelCase ) # Instantiate scheduler lowercase__ : Optional[int] = get_linear_schedule_with_warmup( optimizer=__lowerCamelCase , num_warmup_steps=1_00 , num_training_steps=(len(__lowerCamelCase ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ : Optional[int] = accelerator.prepare( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Now we train the model for epoch in range(__lowerCamelCase ): model.train() for step, batch in enumerate(__lowerCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) lowercase__ : Dict = model(**__lowerCamelCase ) lowercase__ : List[Any] = outputs.loss accelerator.backward(__lowerCamelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__lowerCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowercase__ : Tuple = model(**__lowerCamelCase ) lowercase__ : Any = outputs.logits.argmax(dim=-1 ) lowercase__ , lowercase__ : int = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=__lowerCamelCase , references=__lowerCamelCase , ) lowercase__ : List[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , __lowerCamelCase ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def __UpperCAmelCase ( ) -> Dict: lowercase__ : Optional[int] = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=__lowerCamelCase , default=__lowerCamelCase , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) lowercase__ : int = parser.parse_args() lowercase__ : Union[str, Any] = {'''lr''': 2E-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": main()
16
1
"""simple docstring""" import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class lowercase_ ( __lowerCAmelCase ): '''simple docstring''' UpperCAmelCase : Union[str, Any] = (PNDMScheduler,) UpperCAmelCase : List[Any] = (('''num_inference_steps''', 50),) def lowerCAmelCase_ ( self : Tuple , **_UpperCAmelCase : str ): _A = { 'num_train_timesteps': 1_000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', } config.update(**_UpperCAmelCase ) return config def lowerCAmelCase_ ( self : List[str] , _UpperCAmelCase : int=0 , **_UpperCAmelCase : str ): _A = dict(self.forward_default_kwargs ) _A = kwargs.pop('num_inference_steps' , _UpperCAmelCase ) _A = self.dummy_sample _A = 0.1 * sample _A = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: _A = self.get_scheduler_config(**_UpperCAmelCase ) _A = scheduler_class(**_UpperCAmelCase ) scheduler.set_timesteps(_UpperCAmelCase ) # copy over dummy past residuals _A = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_UpperCAmelCase ) _A = scheduler_class.from_pretrained(_UpperCAmelCase ) new_scheduler.set_timesteps(_UpperCAmelCase ) # copy over dummy past residuals _A = dummy_past_residuals[:] _A = scheduler.step_prk(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = new_scheduler.step_prk(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _A = scheduler.step_plms(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = new_scheduler.step_plms(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def lowerCAmelCase_ ( self : Dict ): pass def lowerCAmelCase_ ( self : int , _UpperCAmelCase : Any=0 , **_UpperCAmelCase : Optional[int] ): _A = dict(self.forward_default_kwargs ) _A = kwargs.pop('num_inference_steps' , _UpperCAmelCase ) _A = self.dummy_sample _A = 0.1 * sample _A = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: _A = self.get_scheduler_config() _A = scheduler_class(**_UpperCAmelCase ) scheduler.set_timesteps(_UpperCAmelCase ) # copy over dummy past residuals (must be after setting timesteps) _A = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_UpperCAmelCase ) _A = scheduler_class.from_pretrained(_UpperCAmelCase ) # copy over dummy past residuals new_scheduler.set_timesteps(_UpperCAmelCase ) # copy over dummy past residual (must be after setting timesteps) _A = dummy_past_residuals[:] _A = scheduler.step_prk(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = new_scheduler.step_prk(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _A = scheduler.step_plms(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = new_scheduler.step_plms(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def lowerCAmelCase_ ( self : Any , **_UpperCAmelCase : str ): _A = self.scheduler_classes[0] _A = self.get_scheduler_config(**_UpperCAmelCase ) _A = scheduler_class(**_UpperCAmelCase ) _A = 10 _A = self.dummy_model() _A = self.dummy_sample_deter scheduler.set_timesteps(_UpperCAmelCase ) for i, t in enumerate(scheduler.prk_timesteps ): _A = model(_UpperCAmelCase , _UpperCAmelCase ) _A = scheduler.step_prk(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): _A = model(_UpperCAmelCase , _UpperCAmelCase ) _A = scheduler.step_plms(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ).prev_sample return sample def lowerCAmelCase_ ( self : List[str] ): _A = dict(self.forward_default_kwargs ) _A = kwargs.pop('num_inference_steps' , _UpperCAmelCase ) for scheduler_class in self.scheduler_classes: _A = self.get_scheduler_config() _A = scheduler_class(**_UpperCAmelCase ) _A = self.dummy_sample _A = 0.1 * sample if num_inference_steps is not None and hasattr(_UpperCAmelCase , 'set_timesteps' ): scheduler.set_timesteps(_UpperCAmelCase ) elif num_inference_steps is not None and not hasattr(_UpperCAmelCase , 'set_timesteps' ): _A = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) _A = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] _A = dummy_past_residuals[:] _A = scheduler.step_prk(_UpperCAmelCase , 0 , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = scheduler.step_prk(_UpperCAmelCase , 1 , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) _A = scheduler.step_plms(_UpperCAmelCase , 0 , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = scheduler.step_plms(_UpperCAmelCase , 1 , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def lowerCAmelCase_ ( self : Optional[int] ): for timesteps in [100, 1_000]: self.check_over_configs(num_train_timesteps=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Tuple ): for steps_offset in [0, 1]: self.check_over_configs(steps_offset=_UpperCAmelCase ) _A = self.scheduler_classes[0] _A = self.get_scheduler_config(steps_offset=1 ) _A = scheduler_class(**_UpperCAmelCase ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [901, 851, 851, 801, 801, 751, 751, 701, 701, 651, 651, 601, 601, 501, 401, 301, 201, 101, 1] ) , ) def lowerCAmelCase_ ( self : str ): for beta_start, beta_end in zip([0.0001, 0.001] , [0.002, 0.02] ): self.check_over_configs(beta_start=_UpperCAmelCase , beta_end=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Optional[Any] ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=_UpperCAmelCase ) def lowerCAmelCase_ ( self : str ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_UpperCAmelCase ) def lowerCAmelCase_ ( self : int ): for t in [1, 5, 10]: self.check_over_forward(time_step=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Dict ): for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Tuple ): # earlier version of set_timesteps() caused an error indexing alpha's with inference steps as power of 3 _A = 27 for scheduler_class in self.scheduler_classes: _A = self.dummy_sample _A = 0.1 * sample _A = self.get_scheduler_config() _A = scheduler_class(**_UpperCAmelCase ) scheduler.set_timesteps(_UpperCAmelCase ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): _A = scheduler.step_prk(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ).prev_sample def lowerCAmelCase_ ( self : Optional[Any] ): with self.assertRaises(_UpperCAmelCase ): _A = self.scheduler_classes[0] _A = self.get_scheduler_config() _A = scheduler_class(**_UpperCAmelCase ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def lowerCAmelCase_ ( self : Optional[int] ): _A = self.full_loop() _A = torch.sum(torch.abs(_UpperCAmelCase ) ) _A = torch.mean(torch.abs(_UpperCAmelCase ) ) assert abs(result_sum.item() - 198.1318 ) < 1E-2 assert abs(result_mean.item() - 0.2580 ) < 1E-3 def lowerCAmelCase_ ( self : int ): _A = self.full_loop(prediction_type='v_prediction' ) _A = torch.sum(torch.abs(_UpperCAmelCase ) ) _A = torch.mean(torch.abs(_UpperCAmelCase ) ) assert abs(result_sum.item() - 67.3986 ) < 1E-2 assert abs(result_mean.item() - 0.0878 ) < 1E-3 def lowerCAmelCase_ ( self : str ): # We specify different beta, so that the first alpha is 0.99 _A = self.full_loop(set_alpha_to_one=_UpperCAmelCase , beta_start=0.01 ) _A = torch.sum(torch.abs(_UpperCAmelCase ) ) _A = torch.mean(torch.abs(_UpperCAmelCase ) ) assert abs(result_sum.item() - 230.0399 ) < 1E-2 assert abs(result_mean.item() - 0.2995 ) < 1E-3 def lowerCAmelCase_ ( self : int ): # We specify different beta, so that the first alpha is 0.99 _A = self.full_loop(set_alpha_to_one=_UpperCAmelCase , beta_start=0.01 ) _A = torch.sum(torch.abs(_UpperCAmelCase ) ) _A = torch.mean(torch.abs(_UpperCAmelCase ) ) assert abs(result_sum.item() - 186.9482 ) < 1E-2 assert abs(result_mean.item() - 0.2434 ) < 1E-3
271
"""simple docstring""" import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging a = { '''cola''': 2, '''mnli''': 3, '''mrpc''': 2, '''sst-2''': 2, '''sts-b''': 1, '''qqp''': 2, '''qnli''': 2, '''rte''': 2, '''wnli''': 2, } logging.set_verbosity_info() def _snake_case ( _snake_case : str , _snake_case : Optional[int] , _snake_case : List[Any] , _snake_case : Tuple=None ) -> List[str]: '''simple docstring''' _A = XLNetConfig.from_json_file(_snake_case ) _A = finetuning_task.lower() if finetuning_task is not None else '' if finetuning_task in GLUE_TASKS_NUM_LABELS: print(F'''Building PyTorch XLNetForSequenceClassification model from configuration: {config}''' ) _A = finetuning_task _A = GLUE_TASKS_NUM_LABELS[finetuning_task] _A = XLNetForSequenceClassification(_snake_case ) elif "squad" in finetuning_task: _A = finetuning_task _A = XLNetForQuestionAnswering(_snake_case ) else: _A = XLNetLMHeadModel(_snake_case ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(_snake_case , _snake_case , _snake_case ) # Save pytorch-model _A = os.path.join(_snake_case , _snake_case ) _A = os.path.join(_snake_case , _snake_case ) print(F'''Save PyTorch model to {os.path.abspath(_snake_case )}''' ) torch.save(model.state_dict() , _snake_case ) print(F'''Save configuration file to {os.path.abspath(_snake_case )}''' ) with open(_snake_case , 'w' , encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": a = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--xlnet_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained XLNet model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the folder to store the PyTorch model or dataset/vocab.''', ) parser.add_argument( '''--finetuning_task''', default=None, type=str, help='''Name of a task on which the XLNet TensorFlow model was fine-tuned''', ) a = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
271
1
'''simple docstring''' import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class __lowercase ( tf.keras.layers.Layer ): def __init__(self , A , A , A = None , A = None ): super().__init__() lowerCamelCase_ : List[Any] = pad_token_id lowerCamelCase_ : Union[str, Any] = max_length lowerCamelCase_ : List[Any] = vocab lowerCamelCase_ : Optional[int] = merges lowerCamelCase_ : List[str] = BytePairTokenizer(A , A , sequence_length=A ) @classmethod def UpperCAmelCase__ (cls , A , *A , **A ): lowerCamelCase_ : int = [''' '''.join(A ) for m in tokenizer.bpe_ranks.keys()] lowerCamelCase_ : Dict = tokenizer.get_vocab() return cls(A , A , *A , **A ) @classmethod def UpperCAmelCase__ (cls , A , *A , **A ): lowerCamelCase_ : Optional[int] = GPTaTokenizer.from_pretrained(A , *A , **A ) return cls.from_tokenizer(A , *A , **A ) @classmethod def UpperCAmelCase__ (cls , A ): return cls(**A ) def UpperCAmelCase__ (self ): return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def UpperCAmelCase__ (self , A , A = None ): lowerCamelCase_ : str = self.tf_tokenizer(A ) lowerCamelCase_ : Any = tf.ones_like(A ) if self.pad_token_id is not None: # pad the tokens up to max length lowerCamelCase_ : Tuple = max_length if max_length is not None else self.max_length if max_length is not None: lowerCamelCase_, lowerCamelCase_ : Tuple = pad_model_inputs( A , max_seq_length=A , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
318
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('''4.31.0''') __lowercase : Dict = logging.getLogger(__name__) @dataclass class __lowercase : lowerCamelCase : str = field( metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) lowerCamelCase : Optional[str] = field( default=_lowercase , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) lowerCamelCase : Optional[str] = field( default=_lowercase , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) lowerCamelCase : Optional[str] = field( default=_lowercase , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) lowerCamelCase : bool = field( default=_lowercase , metadata={"help": "Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."} , ) lowerCamelCase : str = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) lowerCamelCase : bool = field( default=_lowercase , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) @dataclass class __lowercase : lowerCamelCase : Optional[str] = field(default=_lowercase , metadata={"help": "The input training data file (a text file)."} ) lowerCamelCase : Optional[str] = field( default=_lowercase , metadata={"help": "An optional input evaluation data file to evaluate the perplexity on (a text file)."} , ) lowerCamelCase : bool = field( default=_lowercase , metadata={"help": "Overwrite the cached training and evaluation sets"} ) lowerCamelCase : Optional[int] = field( default=_lowercase , metadata={"help": "The number of processes to use for the preprocessing."} , ) lowerCamelCase : Optional[int] = field( default=_lowercase , metadata={ "help": ( "The maximum total input sequence length after tokenization. If passed, sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) lowerCamelCase : bool = field( default=_lowercase , metadata={ "help": ( "Whether to pad all samples to the maximum sentence length. " "If False, will pad the samples dynamically when batching to the maximum length in the batch. More " "efficient on GPU but very bad for TPU." ) } , ) lowerCamelCase : Optional[int] = field( default=_lowercase , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) lowerCamelCase : Optional[int] = field( default=_lowercase , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) def UpperCAmelCase__ (self ): if self.train_file is not None: lowerCamelCase_ : Optional[Any] = self.train_file.split('''.''' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: lowerCamelCase_ : Optional[Any] = self.validation_file.split('''.''' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class __lowercase : lowerCamelCase : PreTrainedTokenizerBase lowerCamelCase : Union[bool, str, PaddingStrategy] = True lowerCamelCase : Optional[int] = None lowerCamelCase : Optional[int] = None def __call__(self , A ): lowerCamelCase_ : List[str] = '''label''' if '''label''' in features[0].keys() else '''labels''' lowerCamelCase_ : str = [feature.pop(A ) for feature in features] lowerCamelCase_ : Any = len(A ) lowerCamelCase_ : List[Any] = len(features[0]['''input_ids'''] ) lowerCamelCase_ : Union[str, Any] = [ [{k: v[i] for k, v in feature.items()} for i in range(A )] for feature in features ] lowerCamelCase_ : str = list(chain(*A ) ) lowerCamelCase_ : Any = self.tokenizer.pad( A , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' , ) # Un-flatten lowerCamelCase_ : int = {k: v.view(A , A , -1 ) for k, v in batch.items()} # Add back labels lowerCamelCase_ : Tuple = torch.tensor(A , dtype=torch.intaa ) return batch def lowercase_ ( ) -> Union[str, Any]: '''simple docstring''' lowerCamelCase_ : int = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. lowerCamelCase_, lowerCamelCase_, lowerCamelCase_ : List[str] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: lowerCamelCase_, lowerCamelCase_, lowerCamelCase_ : str = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_swag''' , _lowercase , _lowercase ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() lowerCamelCase_ : Optional[int] = training_args.get_process_log_level() logger.setLevel(_lowercase ) datasets.utils.logging.set_verbosity(_lowercase ) transformers.utils.logging.set_verbosity(_lowercase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(F"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. lowerCamelCase_ : Any = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowerCamelCase_ : str = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: lowerCamelCase_ : Optional[Any] = {} if data_args.train_file is not None: lowerCamelCase_ : Union[str, Any] = data_args.train_file if data_args.validation_file is not None: lowerCamelCase_ : Tuple = data_args.validation_file lowerCamelCase_ : Optional[Any] = data_args.train_file.split('''.''' )[-1] lowerCamelCase_ : Dict = load_dataset( _lowercase , data_files=_lowercase , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. lowerCamelCase_ : Optional[Any] = load_dataset( '''swag''' , '''regular''' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCamelCase_ : str = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) lowerCamelCase_ : Any = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) lowerCamelCase_ : List[Any] = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=_lowercase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. lowerCamelCase_ : int = [F"""ending{i}""" for i in range(4 )] lowerCamelCase_ : List[Any] = '''sent1''' lowerCamelCase_ : Dict = '''sent2''' if data_args.max_seq_length is None: lowerCamelCase_ : str = tokenizer.model_max_length if max_seq_length > 1_024: logger.warning( '''The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value''' ''' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can''' ''' override this default with `--block_size xxx`.''' ) lowerCamelCase_ : Optional[int] = 1_024 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F"""The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the""" F"""model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.""" ) lowerCamelCase_ : Optional[int] = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(_lowercase ): lowerCamelCase_ : Tuple = [[context] * 4 for context in examples[context_name]] lowerCamelCase_ : List[Any] = examples[question_header_name] lowerCamelCase_ : Optional[Any] = [ [F"""{header} {examples[end][i]}""" for end in ending_names] for i, header in enumerate(_lowercase ) ] # Flatten out lowerCamelCase_ : Optional[Any] = list(chain(*_lowercase ) ) lowerCamelCase_ : List[Any] = list(chain(*_lowercase ) ) # Tokenize lowerCamelCase_ : List[str] = tokenizer( _lowercase , _lowercase , truncation=_lowercase , max_length=_lowercase , padding='''max_length''' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(_lowercase ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('''--do_train requires a train dataset''' ) lowerCamelCase_ : Union[str, Any] = raw_datasets['''train'''] if data_args.max_train_samples is not None: lowerCamelCase_ : List[str] = min(len(_lowercase ) , data_args.max_train_samples ) lowerCamelCase_ : List[str] = train_dataset.select(range(_lowercase ) ) with training_args.main_process_first(desc='''train dataset map pre-processing''' ): lowerCamelCase_ : Dict = train_dataset.map( _lowercase , batched=_lowercase , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('''--do_eval requires a validation dataset''' ) lowerCamelCase_ : Optional[int] = raw_datasets['''validation'''] if data_args.max_eval_samples is not None: lowerCamelCase_ : Optional[int] = min(len(_lowercase ) , data_args.max_eval_samples ) lowerCamelCase_ : Any = eval_dataset.select(range(_lowercase ) ) with training_args.main_process_first(desc='''validation dataset map pre-processing''' ): lowerCamelCase_ : Tuple = eval_dataset.map( _lowercase , batched=_lowercase , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator lowerCamelCase_ : int = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=_lowercase , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(_lowercase ): lowerCamelCase_, lowerCamelCase_ : Optional[Any] = eval_predictions lowerCamelCase_ : Any = np.argmax(_lowercase , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer lowerCamelCase_ : Any = Trainer( model=_lowercase , args=_lowercase , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=_lowercase , data_collator=_lowercase , compute_metrics=_lowercase , ) # Training if training_args.do_train: lowerCamelCase_ : int = None if training_args.resume_from_checkpoint is not None: lowerCamelCase_ : Dict = training_args.resume_from_checkpoint elif last_checkpoint is not None: lowerCamelCase_ : List[Any] = last_checkpoint lowerCamelCase_ : Dict = trainer.train(resume_from_checkpoint=_lowercase ) trainer.save_model() # Saves the tokenizer too for easy upload lowerCamelCase_ : Any = train_result.metrics lowerCamelCase_ : Union[str, Any] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(_lowercase ) ) lowerCamelCase_ : List[Any] = min(_lowercase , len(_lowercase ) ) trainer.log_metrics('''train''' , _lowercase ) trainer.save_metrics('''train''' , _lowercase ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('''*** Evaluate ***''' ) lowerCamelCase_ : str = trainer.evaluate() lowerCamelCase_ : Dict = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(_lowercase ) lowerCamelCase_ : Union[str, Any] = min(_lowercase , len(_lowercase ) ) trainer.log_metrics('''eval''' , _lowercase ) trainer.save_metrics('''eval''' , _lowercase ) lowerCamelCase_ : List[str] = { '''finetuned_from''': model_args.model_name_or_path, '''tasks''': '''multiple-choice''', '''dataset_tags''': '''swag''', '''dataset_args''': '''regular''', '''dataset''': '''SWAG''', '''language''': '''en''', } if training_args.push_to_hub: trainer.push_to_hub(**_lowercase ) else: trainer.create_model_card(**_lowercase ) def lowercase_ ( _lowercase ) -> Dict: '''simple docstring''' main() if __name__ == "__main__": main()
318
1
"""simple docstring""" import argparse import json import math import os import time import traceback import zipfile from collections import Counter import requests def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase=None ) -> Optional[int]: '''simple docstring''' lowercase_ = None if token is not None: lowercase_ = {"""Accept""": """application/vnd.github+json""", """Authorization""": F'''Bearer {token}'''} lowercase_ = F'''https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100''' lowercase_ = requests.get(snake_case_ , headers=snake_case_ ).json() lowercase_ = {} try: job_links.update({job["""name"""]: job["""html_url"""] for job in result["""jobs"""]} ) lowercase_ = math.ceil((result["""total_count"""] - 1_00) / 1_00 ) for i in range(snake_case_ ): lowercase_ = requests.get(url + F'''&page={i + 2}''' , headers=snake_case_ ).json() job_links.update({job["""name"""]: job["""html_url"""] for job in result["""jobs"""]} ) return job_links except Exception: print(F'''Unknown error, could not fetch links:\n{traceback.format_exc()}''' ) return {} def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase=None ) -> Optional[Any]: '''simple docstring''' lowercase_ = None if token is not None: lowercase_ = {"""Accept""": """application/vnd.github+json""", """Authorization""": F'''Bearer {token}'''} lowercase_ = F'''https://api.github.com/repos/huggingface/transformers/actions/runs/{worflow_run_id}/artifacts?per_page=100''' lowercase_ = requests.get(snake_case_ , headers=snake_case_ ).json() lowercase_ = {} try: artifacts.update({artifact["""name"""]: artifact["""archive_download_url"""] for artifact in result["""artifacts"""]} ) lowercase_ = math.ceil((result["""total_count"""] - 1_00) / 1_00 ) for i in range(snake_case_ ): lowercase_ = requests.get(url + F'''&page={i + 2}''' , headers=snake_case_ ).json() artifacts.update({artifact["""name"""]: artifact["""archive_download_url"""] for artifact in result["""artifacts"""]} ) return artifacts except Exception: print(F'''Unknown error, could not fetch links:\n{traceback.format_exc()}''' ) return {} def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Union[str, Any]: '''simple docstring''' lowercase_ = None if token is not None: lowercase_ = {"""Accept""": """application/vnd.github+json""", """Authorization""": F'''Bearer {token}'''} lowercase_ = requests.get(snake_case_ , headers=snake_case_ , allow_redirects=snake_case_ ) lowercase_ = result.headers["""Location"""] lowercase_ = requests.get(snake_case_ , allow_redirects=snake_case_ ) lowercase_ = os.path.join(snake_case_ , F'''{artifact_name}.zip''' ) with open(snake_case_ , """wb""" ) as fp: fp.write(response.content ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase=None ) -> Dict: '''simple docstring''' lowercase_ = [] lowercase_ = [] lowercase_ = None with zipfile.ZipFile(snake_case_ ) as z: for filename in z.namelist(): if not os.path.isdir(snake_case_ ): # read the file if filename in ["failures_line.txt", "summary_short.txt", "job_name.txt"]: with z.open(snake_case_ ) as f: for line in f: lowercase_ = line.decode("""UTF-8""" ).strip() if filename == "failures_line.txt": try: # `error_line` is the place where `error` occurs lowercase_ = line[: line.index(""": """ )] lowercase_ = line[line.index(""": """ ) + len(""": """ ) :] errors.append([error_line, error] ) except Exception: # skip un-related lines pass elif filename == "summary_short.txt" and line.startswith("""FAILED """ ): # `test` is the test method that failed lowercase_ = line[len("""FAILED """ ) :] failed_tests.append(snake_case_ ) elif filename == "job_name.txt": lowercase_ = line if len(snake_case_ ) != len(snake_case_ ): raise ValueError( F'''`errors` and `failed_tests` should have the same number of elements. Got {len(snake_case_ )} for `errors` ''' F'''and {len(snake_case_ )} for `failed_tests` instead. The test reports in {artifact_zip_path} have some''' """ problem.""" ) lowercase_ = None if job_name and job_links: lowercase_ = job_links.get(snake_case_ , snake_case_ ) # A list with elements of the form (line of error, error, failed test) lowercase_ = [x + [y] + [job_link] for x, y in zip(snake_case_ , snake_case_ )] return result def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase=None ) -> str: '''simple docstring''' lowercase_ = [] lowercase_ = [os.path.join(snake_case_ , snake_case_ ) for p in os.listdir(snake_case_ ) if p.endswith(""".zip""" )] for p in paths: errors.extend(get_errors_from_single_artifact(snake_case_ , job_links=snake_case_ ) ) return errors def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase=None ) -> Optional[int]: '''simple docstring''' lowercase_ = Counter() counter.update([x[1] for x in logs] ) lowercase_ = counter.most_common() lowercase_ = {} for error, count in counts: if error_filter is None or error not in error_filter: lowercase_ = {"""count""": count, """failed_tests""": [(x[2], x[0]) for x in logs if x[1] == error]} lowercase_ = dict(sorted(r.items() , key=lambda __lowerCAmelCase : item[1]["count"] , reverse=snake_case_ ) ) return r def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> Optional[Any]: '''simple docstring''' lowercase_ = test.split("""::""" )[0] if test.startswith("""tests/models/""" ): lowercase_ = test.split("""/""" )[2] else: lowercase_ = None return test def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase=None ) -> Optional[int]: '''simple docstring''' lowercase_ = [(x[0], x[1], get_model(x[2] )) for x in logs] lowercase_ = [x for x in logs if x[2] is not None] lowercase_ = {x[2] for x in logs} lowercase_ = {} for test in tests: lowercase_ = Counter() # count by errors in `test` counter.update([x[1] for x in logs if x[2] == test] ) lowercase_ = counter.most_common() lowercase_ = {error: count for error, count in counts if (error_filter is None or error not in error_filter)} lowercase_ = sum(error_counts.values() ) if n_errors > 0: lowercase_ = {"""count""": n_errors, """errors""": error_counts} lowercase_ = dict(sorted(r.items() , key=lambda __lowerCAmelCase : item[1]["count"] , reverse=snake_case_ ) ) return r def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> Optional[Any]: '''simple docstring''' lowercase_ = """| no. | error | status |""" lowercase_ = """|-:|:-|:-|""" lowercase_ = [header, sep] for error in reduced_by_error: lowercase_ = reduced_by_error[error]["""count"""] lowercase_ = F'''| {count} | {error[:1_00]} | |''' lines.append(snake_case_ ) return "\n".join(snake_case_ ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> str: '''simple docstring''' lowercase_ = """| model | no. of errors | major error | count |""" lowercase_ = """|-:|-:|-:|-:|""" lowercase_ = [header, sep] for model in reduced_by_model: lowercase_ = reduced_by_model[model]["""count"""] lowercase_ = list(reduced_by_model[model]["""errors"""].items() )[0] lowercase_ = F'''| {model} | {count} | {error[:60]} | {_count} |''' lines.append(snake_case_ ) return "\n".join(snake_case_ ) if __name__ == "__main__": UpperCAmelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument("--workflow_run_id", type=str, required=True, help="A GitHub Actions workflow run id.") parser.add_argument( "--output_dir", type=str, required=True, help="Where to store the downloaded artifacts and other result files.", ) parser.add_argument("--token", default=None, type=str, help="A token that has actions:read permission.") UpperCAmelCase : Tuple = parser.parse_args() os.makedirs(args.output_dir, exist_ok=True) UpperCAmelCase : Any = get_job_links(args.workflow_run_id, token=args.token) UpperCAmelCase : int = {} # To deal with `workflow_call` event, where a job name is the combination of the job names in the caller and callee. # For example, `PyTorch 1.11 / Model tests (models/albert, single-gpu)`. if _job_links: for k, v in _job_links.items(): # This is how GitHub actions combine job names. if " / " in k: UpperCAmelCase : List[Any] = k.find(" / ") UpperCAmelCase : Tuple = k[index + len(" / ") :] UpperCAmelCase : Any = v with open(os.path.join(args.output_dir, "job_links.json"), "w", encoding="UTF-8") as fp: json.dump(job_links, fp, ensure_ascii=False, indent=4) UpperCAmelCase : Any = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, "artifacts.json"), "w", encoding="UTF-8") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) for idx, (name, url) in enumerate(artifacts.items()): download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) UpperCAmelCase : Optional[int] = get_all_errors(args.output_dir, job_links=job_links) # `e[1]` is the error UpperCAmelCase : Any = Counter() counter.update([e[1] for e in errors]) # print the top 30 most common test errors UpperCAmelCase : Dict = counter.most_common(30) for item in most_common: print(item) with open(os.path.join(args.output_dir, "errors.json"), "w", encoding="UTF-8") as fp: json.dump(errors, fp, ensure_ascii=False, indent=4) UpperCAmelCase : str = reduce_by_error(errors) UpperCAmelCase : str = reduce_by_model(errors) UpperCAmelCase : List[Any] = make_github_table(reduced_by_error) UpperCAmelCase : Tuple = make_github_table_per_model(reduced_by_model) with open(os.path.join(args.output_dir, "reduced_by_error.txt"), "w", encoding="UTF-8") as fp: fp.write(sa) with open(os.path.join(args.output_dir, "reduced_by_model.txt"), "w", encoding="UTF-8") as fp: fp.write(sa)
355
"""simple docstring""" from __future__ import annotations import numpy as np def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> tuple[np.ndarray, np.ndarray]: '''simple docstring''' lowercase_ , lowercase_ = np.shape(__lowerCAmelCase ) if rows != columns: lowercase_ = ( """'table' has to be of square shaped array but got a """ F'''{rows}x{columns} array:\n{table}''' ) raise ValueError(__lowerCAmelCase ) lowercase_ = np.zeros((rows, columns) ) lowercase_ = np.zeros((rows, columns) ) for i in range(__lowerCAmelCase ): for j in range(__lowerCAmelCase ): lowercase_ = sum(lower[i][k] * upper[k][j] for k in range(__lowerCAmelCase ) ) if upper[j][j] == 0: raise ArithmeticError("""No LU decomposition exists""" ) lowercase_ = (table[i][j] - total) / upper[j][j] lowercase_ = 1 for j in range(__lowerCAmelCase , __lowerCAmelCase ): lowercase_ = sum(lower[i][k] * upper[k][j] for k in range(__lowerCAmelCase ) ) lowercase_ = table[i][j] - total return lower, upper if __name__ == "__main__": import doctest doctest.testmod()
313
0
'''simple docstring''' from __future__ import annotations import time lowerCamelCase :Tuple = list[tuple[int, int]] lowerCamelCase :Optional[Any] = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] lowerCamelCase :int = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right class _lowerCAmelCase : def __init__(self , lowercase , lowercase , lowercase , lowercase , lowercase ): A_ : Dict = pos_x A_ : Tuple = pos_y A_ : List[str] = (pos_y, pos_x) A_ : Tuple = goal_x A_ : Optional[int] = goal_y A_ : Optional[int] = parent class _lowerCAmelCase : def __init__(self , lowercase , lowercase ): A_ : Tuple = Node(start[1] , start[0] , goal[1] , goal[0] , lowercase ) A_ : List[str] = Node(goal[1] , goal[0] , goal[1] , goal[0] , lowercase ) A_ : Optional[int] = [self.start] A_ : List[Any] = False def _a (self ): while self.node_queue: A_ : List[str] = self.node_queue.pop(0 ) if current_node.pos == self.target.pos: A_ : List[str] = True return self.retrace_path(lowercase ) A_ : Optional[Any] = self.get_successors(lowercase ) for node in successors: self.node_queue.append(lowercase ) if not self.reached: return [self.start.pos] return None def _a (self , lowercase ): A_ : Tuple = [] for action in delta: A_ : Dict = parent.pos_x + action[1] A_ : Tuple = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowercase ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node(lowercase , lowercase , self.target.pos_y , self.target.pos_x , lowercase ) ) return successors def _a (self , lowercase ): A_ : Optional[Any] = node A_ : List[str] = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) A_ : Optional[int] = current_node.parent path.reverse() return path class _lowerCAmelCase : def __init__(self , lowercase , lowercase ): A_ : int = BreadthFirstSearch(lowercase , lowercase ) A_ : str = BreadthFirstSearch(lowercase , lowercase ) A_ : List[str] = False def _a (self ): while self.fwd_bfs.node_queue or self.bwd_bfs.node_queue: A_ : Optional[int] = self.fwd_bfs.node_queue.pop(0 ) A_ : Dict = self.bwd_bfs.node_queue.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: A_ : List[str] = True return self.retrace_bidirectional_path( lowercase , lowercase ) A_ : Optional[int] = current_bwd_node A_ : Optional[int] = current_fwd_node A_ : List[Any] = { self.fwd_bfs: self.fwd_bfs.get_successors(lowercase ), self.bwd_bfs: self.bwd_bfs.get_successors(lowercase ), } for bfs in [self.fwd_bfs, self.bwd_bfs]: for node in successors[bfs]: bfs.node_queue.append(lowercase ) if not self.reached: return [self.fwd_bfs.start.pos] return None def _a (self , lowercase , lowercase ): A_ : List[str] = self.fwd_bfs.retrace_path(lowercase ) A_ : List[str] = self.bwd_bfs.retrace_path(lowercase ) bwd_path.pop() bwd_path.reverse() A_ : Union[str, Any] = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] import doctest doctest.testmod() lowerCamelCase :List[str] = (0, 0) lowerCamelCase :str = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) lowerCamelCase :int = time.time() lowerCamelCase :Dict = BreadthFirstSearch(init, goal) lowerCamelCase :Tuple = bfs.search() lowerCamelCase :Optional[int] = time.time() - start_bfs_time print('''Unidirectional BFS computation time : ''', bfs_time) lowerCamelCase :Union[str, Any] = time.time() lowerCamelCase :str = BidirectionalBreadthFirstSearch(init, goal) lowerCamelCase :Optional[Any] = bd_bfs.search() lowerCamelCase :Optional[int] = time.time() - start_bd_bfs_time print('''Bidirectional BFS computation time : ''', bd_bfs_time)
206
'''simple docstring''' import argparse import struct import unittest class _lowerCAmelCase : def __init__(self , lowercase ): A_ : List[str] = data # Initialize hash values A_ : Tuple = [ 0X6A09_E667, 0XBB67_AE85, 0X3C6E_F372, 0XA54F_F53A, 0X510E_527F, 0X9B05_688C, 0X1F83_D9AB, 0X5BE0_CD19, ] # Initialize round constants A_ : List[Any] = [ 0X428A_2F98, 0X7137_4491, 0XB5C0_FBCF, 0XE9B5_DBA5, 0X3956_C25B, 0X59F1_11F1, 0X923F_82A4, 0XAB1C_5ED5, 0XD807_AA98, 0X1283_5B01, 0X2431_85BE, 0X550C_7DC3, 0X72BE_5D74, 0X80DE_B1FE, 0X9BDC_06A7, 0XC19B_F174, 0XE49B_69C1, 0XEFBE_4786, 0X0FC1_9DC6, 0X240C_A1CC, 0X2DE9_2C6F, 0X4A74_84AA, 0X5CB0_A9DC, 0X76F9_88DA, 0X983E_5152, 0XA831_C66D, 0XB003_27C8, 0XBF59_7FC7, 0XC6E0_0BF3, 0XD5A7_9147, 0X06CA_6351, 0X1429_2967, 0X27B7_0A85, 0X2E1B_2138, 0X4D2C_6DFC, 0X5338_0D13, 0X650A_7354, 0X766A_0ABB, 0X81C2_C92E, 0X9272_2C85, 0XA2BF_E8A1, 0XA81A_664B, 0XC24B_8B70, 0XC76C_51A3, 0XD192_E819, 0XD699_0624, 0XF40E_3585, 0X106A_A070, 0X19A4_C116, 0X1E37_6C08, 0X2748_774C, 0X34B0_BCB5, 0X391C_0CB3, 0X4ED8_AA4A, 0X5B9C_CA4F, 0X682E_6FF3, 0X748F_82EE, 0X78A5_636F, 0X84C8_7814, 0X8CC7_0208, 0X90BE_FFFA, 0XA450_6CEB, 0XBEF9_A3F7, 0XC671_78F2, ] A_ : Tuple = self.preprocessing(self.data ) self.final_hash() @staticmethod def _a (lowercase ): A_ : Any = b"""\x80""" + (b"""\x00""" * (63 - (len(lowercase ) + 8) % 64)) A_ : Optional[Any] = struct.pack(""">Q""" , (len(lowercase ) * 8) ) return data + padding + big_endian_integer def _a (self ): # Convert into blocks of 64 bytes A_ : str = [ self.preprocessed_data[x : x + 64] for x in range(0 , len(self.preprocessed_data ) , 64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers A_ : Tuple = list(struct.unpack(""">16L""" , lowercase ) ) # add 48 0-ed integers words += [0] * 48 A_, A_, A_, A_, A_, A_, A_, A_ : Dict = self.hashes for index in range(0 , 64 ): if index > 15: # modify the zero-ed indexes at the end of the array A_ : Optional[Any] = ( self.ror(words[index - 15] , 7 ) ^ self.ror(words[index - 15] , 18 ) ^ (words[index - 15] >> 3) ) A_ : Tuple = ( self.ror(words[index - 2] , 17 ) ^ self.ror(words[index - 2] , 19 ) ^ (words[index - 2] >> 10) ) A_ : Any = ( words[index - 16] + sa + words[index - 7] + sa ) % 0X1_0000_0000 # Compression A_ : Union[str, Any] = self.ror(lowercase , 6 ) ^ self.ror(lowercase , 11 ) ^ self.ror(lowercase , 25 ) A_ : List[Any] = (e & f) ^ ((~e & 0XFFFF_FFFF) & g) A_ : Dict = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0X1_0000_0000 A_ : List[str] = self.ror(lowercase , 2 ) ^ self.ror(lowercase , 13 ) ^ self.ror(lowercase , 22 ) A_ : str = (a & b) ^ (a & c) ^ (b & c) A_ : int = (sa + maj) % 0X1_0000_0000 A_, A_, A_, A_, A_, A_, A_, A_ : List[Any] = ( g, f, e, ((d + tempa) % 0X1_0000_0000), c, b, a, ((tempa + tempa) % 0X1_0000_0000), ) A_ : int = [a, b, c, d, e, f, g, h] # Modify final values A_ : Dict = [ ((element + mutated_hash_values[index]) % 0X1_0000_0000) for index, element in enumerate(self.hashes ) ] A_ : Any = """""".join([hex(lowercase )[2:].zfill(8 ) for value in self.hashes] ) def _a (self , lowercase , lowercase ): return 0XFFFF_FFFF & (value << (32 - rotations)) | (value >> rotations) class _lowerCAmelCase ( unittest.TestCase ): def _a (self ): import hashlib A_ : Optional[Any] = bytes("""Test String""" , """utf-8""" ) self.assertEqual(SHAaaa(lowercase ).hash , hashlib.shaaaa(lowercase ).hexdigest() ) def a ( ): '''simple docstring''' import doctest doctest.testmod() A_ : Any = argparse.ArgumentParser() parser.add_argument( """-s""" , """--string""" , dest="""input_string""" , default="""Hello World!! Welcome to Cryptography""" , help="""Hash the string""" , ) parser.add_argument( """-f""" , """--file""" , dest="""input_file""" , help="""Hash contents of a file""" ) A_ : List[str] = parser.parse_args() A_ : Tuple = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , """rb""" ) as f: A_ : Union[str, Any] = f.read() else: A_ : Optional[Any] = bytes(lowerCamelCase__ , """utf-8""" ) print(SHAaaa(lowerCamelCase__ ).hash ) if __name__ == "__main__": main()
206
1
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING lowercase__ : List[Any] = logging.get_logger(__name__) lowercase__ : str = { '''salesforce/blip2-opt-2.7b''': '''https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json''', } class SCREAMING_SNAKE_CASE (a__ ): lowerCAmelCase = '''blip_2_vision_model''' def __init__( self , _UpperCAmelCase=1408 , _UpperCAmelCase=6144 , _UpperCAmelCase=39 , _UpperCAmelCase=16 , _UpperCAmelCase=224 , _UpperCAmelCase=14 , _UpperCAmelCase="gelu" , _UpperCAmelCase=0.00001 , _UpperCAmelCase=0.0 , _UpperCAmelCase=1e-1_0 , _UpperCAmelCase=True , **_UpperCAmelCase , ): '''simple docstring''' super().__init__(**_UpperCAmelCase) __A : str = hidden_size __A : Optional[int] = intermediate_size __A : Tuple = num_hidden_layers __A : Union[str, Any] = num_attention_heads __A : List[str] = patch_size __A : List[Any] = image_size __A : Union[str, Any] = initializer_range __A : Optional[Any] = attention_dropout __A : Dict = layer_norm_eps __A : List[str] = hidden_act __A : Optional[int] = qkv_bias @classmethod def SCREAMING_SNAKE_CASE ( cls , _UpperCAmelCase , **_UpperCAmelCase): '''simple docstring''' cls._set_token_in_kwargs(_UpperCAmelCase) __A ,__A : Optional[Any] = cls.get_config_dict(_UpperCAmelCase , **_UpperCAmelCase) # get the vision config dict if we are loading from Blip2Config if config_dict.get('model_type') == "blip-2": __A : Optional[Any] = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type') and config_dict["model_type"] != cls.model_type: logger.warning( F'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' F'{cls.model_type}. This is not supported for all configurations of models and can yield errors.') return cls.from_dict(_UpperCAmelCase , **_UpperCAmelCase) class SCREAMING_SNAKE_CASE (a__ ): lowerCAmelCase = '''blip_2_qformer''' def __init__( self , _UpperCAmelCase=3_0522 , _UpperCAmelCase=768 , _UpperCAmelCase=12 , _UpperCAmelCase=12 , _UpperCAmelCase=3072 , _UpperCAmelCase="gelu" , _UpperCAmelCase=0.1 , _UpperCAmelCase=0.1 , _UpperCAmelCase=512 , _UpperCAmelCase=0.02 , _UpperCAmelCase=1e-1_2 , _UpperCAmelCase=0 , _UpperCAmelCase="absolute" , _UpperCAmelCase=2 , _UpperCAmelCase=1408 , **_UpperCAmelCase , ): '''simple docstring''' super().__init__(pad_token_id=_UpperCAmelCase , **_UpperCAmelCase) __A : Union[str, Any] = vocab_size __A : str = hidden_size __A : List[Any] = num_hidden_layers __A : Any = num_attention_heads __A : Union[str, Any] = hidden_act __A : Any = intermediate_size __A : Tuple = hidden_dropout_prob __A : Dict = attention_probs_dropout_prob __A : Tuple = max_position_embeddings __A : str = initializer_range __A : Union[str, Any] = layer_norm_eps __A : Union[str, Any] = position_embedding_type __A : Optional[Any] = cross_attention_frequency __A : List[str] = encoder_hidden_size @classmethod def SCREAMING_SNAKE_CASE ( cls , _UpperCAmelCase , **_UpperCAmelCase): '''simple docstring''' cls._set_token_in_kwargs(_UpperCAmelCase) __A ,__A : int = cls.get_config_dict(_UpperCAmelCase , **_UpperCAmelCase) # get the qformer config dict if we are loading from Blip2Config if config_dict.get('model_type') == "blip-2": __A : List[str] = config_dict['qformer_config'] if "model_type" in config_dict and hasattr(cls , 'model_type') and config_dict["model_type"] != cls.model_type: logger.warning( F'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' F'{cls.model_type}. This is not supported for all configurations of models and can yield errors.') return cls.from_dict(_UpperCAmelCase , **_UpperCAmelCase) class SCREAMING_SNAKE_CASE (a__ ): lowerCAmelCase = '''blip-2''' lowerCAmelCase = True def __init__( self , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=32 , **_UpperCAmelCase): '''simple docstring''' super().__init__(**_UpperCAmelCase) if vision_config is None: __A : str = {} logger.info('vision_config is None. initializing the Blip2VisionConfig with default values.') if qformer_config is None: __A : Optional[Any] = {} logger.info('qformer_config is None. Initializing the Blip2QFormerConfig with default values.') if text_config is None: __A : str = {} logger.info('text_config is None. Initializing the text config with default values (`OPTConfig`).') __A : Optional[Any] = BlipaVisionConfig(**_UpperCAmelCase) __A : Optional[Any] = BlipaQFormerConfig(**_UpperCAmelCase) __A : int = text_config['model_type'] if 'model_type' in text_config else 'opt' __A : Any = CONFIG_MAPPING[text_model_type](**_UpperCAmelCase) __A : int = self.text_config.tie_word_embeddings __A : Union[str, Any] = self.text_config.is_encoder_decoder __A : int = num_query_tokens __A : str = self.vision_config.hidden_size __A : Tuple = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES __A : Any = 1.0 __A : Optional[int] = 0.02 @classmethod def SCREAMING_SNAKE_CASE ( cls , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase , ): '''simple docstring''' return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **_UpperCAmelCase , ) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Dict = copy.deepcopy(self.__dict__) __A : int = self.vision_config.to_dict() __A : Optional[int] = self.qformer_config.to_dict() __A : Dict = self.text_config.to_dict() __A : Union[str, Any] = self.__class__.model_type return output
190
'''simple docstring''' import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse('''9.1.0'''): lowercase__ : Dict = { '''linear''': PIL.Image.Resampling.BILINEAR, '''bilinear''': PIL.Image.Resampling.BILINEAR, '''bicubic''': PIL.Image.Resampling.BICUBIC, '''lanczos''': PIL.Image.Resampling.LANCZOS, '''nearest''': PIL.Image.Resampling.NEAREST, } else: lowercase__ : Any = { '''linear''': PIL.Image.LINEAR, '''bilinear''': PIL.Image.BILINEAR, '''bicubic''': PIL.Image.BICUBIC, '''lanczos''': PIL.Image.LANCZOS, '''nearest''': PIL.Image.NEAREST, } def _lowerCAmelCase ( __snake_case : Any ) -> Optional[Any]: __A : Dict = (images / 2 + 0.5).clamp(0 , 1 ) __A : str = images.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() __A : Dict = numpy_to_pil(__snake_case ) return images def _lowerCAmelCase ( __snake_case : List[Any] ) -> Optional[Any]: if images.ndim == 3: __A : List[Any] = images[None, ...] __A : List[str] = (images * 2_55).round().astype('uint8' ) if images.shape[-1] == 1: # special case for grayscale (single channel) images __A : str = [Image.fromarray(image.squeeze() , mode='L' ) for image in images] else: __A : str = [Image.fromarray(__snake_case ) for image in images] return pil_images
190
1
"""simple docstring""" import torch from transformers import AutoModel class UpperCamelCase ( torch.nn.Module ): def __init__( self, lowerCAmelCase__="sayef/fsner-bert-base-uncased") -> List[Any]: super(lowerCAmelCase__, self).__init__() snake_case_ = AutoModel.from_pretrained(lowerCAmelCase__, return_dict=lowerCAmelCase__) snake_case_ = torch.nn.CosineSimilarity(3, 1e-08) snake_case_ = torch.nn.Softmax(dim=1) def a_ ( self, **lowerCAmelCase__) -> Union[str, Any]: return self.bert(**lowerCAmelCase__).last_hidden_state def a_ ( self, lowerCAmelCase__) -> Union[str, Any]: return token_embeddings.sum(2, keepdim=lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__=1) -> List[Any]: return self.softmax(T * self.cos(lowerCAmelCase__, lowerCAmelCase__)) def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Dict: snake_case_ = W_supports['sizes'].tolist() snake_case_ = W_supports['start_token_id'].item() snake_case_ = W_supports['end_token_id'].item() del W_supports["sizes"] del W_supports["start_token_id"] del W_supports["end_token_id"] snake_case_ = self.BERT(**lowerCAmelCase__) snake_case_ = self.BERT(**lowerCAmelCase__) snake_case_ = None snake_case_ = None snake_case_ = W_supports['input_ids'] == start_token_id snake_case_ = W_supports['input_ids'] == end_token_id for i, size in enumerate(lowerCAmelCase__): if i == 0: snake_case_ = 0 else: snake_case_ = support_sizes[i - 1] snake_case_ = S[s : s + size][start_token_masks[s : s + size]] snake_case_ = S[s : s + size][end_token_masks[s : s + size]] snake_case_ = torch.matmul(q[i], s_start.T).sum(1).softmax(0) snake_case_ = torch.matmul(q[i], s_end.T).sum(1).softmax(0) if p_starts is not None: snake_case_ = torch.vstack((p_starts, p_start)) snake_case_ = torch.vstack((p_ends, p_end)) else: snake_case_ = p_start snake_case_ = p_end return p_starts, p_ends
69
'''simple docstring''' from collections.abc import Sequence def a_ ( _lowerCAmelCase ,_lowerCAmelCase ) -> float: return sum(c * (x**i) for i, c in enumerate(_lowerCAmelCase ) ) def a_ ( _lowerCAmelCase ,_lowerCAmelCase ) -> float: __lowerCamelCase : Any = 0.0 for coeff in reversed(_lowerCAmelCase ): __lowerCamelCase : Tuple = result * x + coeff return result if __name__ == "__main__": _UpperCamelCase = (0.0, 0.0, 5.0, 9.3, 7.0) _UpperCamelCase = 1_0.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
208
0
from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class UpperCamelCase ( __a ): lowerCAmelCase : int = ["""image_processor""", """tokenizer"""] lowerCAmelCase : Union[str, Any] = """BlipImageProcessor""" lowerCAmelCase : List[str] = """AutoTokenizer""" def __init__( self , UpperCAmelCase__ , UpperCAmelCase__ ): A__ = False super().__init__(a__ , a__ ) A__ = self.image_processor def __call__( self , UpperCAmelCase__ = None , UpperCAmelCase__ = None , UpperCAmelCase__ = True , UpperCAmelCase__ = False , UpperCAmelCase__ = None , UpperCAmelCase__ = None , UpperCAmelCase__ = 0 , UpperCAmelCase__ = None , UpperCAmelCase__ = None , UpperCAmelCase__ = False , UpperCAmelCase__ = False , UpperCAmelCase__ = False , UpperCAmelCase__ = False , UpperCAmelCase__ = False , UpperCAmelCase__ = True , UpperCAmelCase__ = None , **UpperCAmelCase__ , ): if images is None and text is None: raise ValueError("You have to specify either images or text." ) # Get only text if images is None: A__ = self.tokenizer A__ = self.tokenizer( text=a__ , add_special_tokens=a__ , padding=a__ , truncation=a__ , max_length=a__ , stride=a__ , pad_to_multiple_of=a__ , return_attention_mask=a__ , return_overflowing_tokens=a__ , return_special_tokens_mask=a__ , return_offsets_mapping=a__ , return_token_type_ids=a__ , return_length=a__ , verbose=a__ , return_tensors=a__ , **a__ , ) return text_encoding # add pixel_values A__ = self.image_processor(a__ , return_tensors=a__ ) if text is not None: A__ = self.tokenizer( text=a__ , add_special_tokens=a__ , padding=a__ , truncation=a__ , max_length=a__ , stride=a__ , pad_to_multiple_of=a__ , return_attention_mask=a__ , return_overflowing_tokens=a__ , return_special_tokens_mask=a__ , return_offsets_mapping=a__ , return_token_type_ids=a__ , return_length=a__ , verbose=a__ , return_tensors=a__ , **a__ , ) else: A__ = None if text_encoding is not None: encoding_image_processor.update(a__ ) return encoding_image_processor def __A ( self , *UpperCAmelCase__ , **UpperCAmelCase__ ): return self.tokenizer.batch_decode(*a__ , **a__ ) def __A ( self , *UpperCAmelCase__ , **UpperCAmelCase__ ): return self.tokenizer.decode(*a__ , **a__ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def __A ( self ): A__ = self.tokenizer.model_input_names A__ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
360
import math def UpperCamelCase ( _A : float , _A : float )-> float: """simple docstring""" if ( not isinstance(_A , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError("power_factor must be a valid float value between -1 and 1." ) return apparent_power * power_factor def UpperCamelCase ( _A : float , _A : float )-> float: """simple docstring""" if ( not isinstance(_A , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError("power_factor must be a valid float value between -1 and 1." ) return apparent_power * math.sqrt(1 - power_factor**2 ) if __name__ == "__main__": import doctest doctest.testmod()
198
0
'''simple docstring''' from __future__ import annotations from collections.abc import Callable def UpperCAmelCase_ (__a : Callable[[int | float], int | float] , __a : int | float , __a : int | float , __a : int = 1_0_0 , ): """simple docstring""" _a : Dict = x_start _a : Dict = fnc(__a ) _a : Union[str, Any] = 0.0 for _ in range(__a ): # Approximates small segments of curve as linear and solve # for trapezoidal area _a : List[Any] = (x_end - x_start) / steps + xa _a : Dict = fnc(__a ) area += abs(fxa + fxa ) * (xa - xa) / 2 # Increment step _a : Any = xa _a : Union[str, Any] = fxa return area if __name__ == "__main__": def UpperCAmelCase_ (__a : int ): """simple docstring""" return x**3 + x**2 print("""f(x) = x^3 + x^2""") print("""The area between the curve, x = -5, x = 5 and the x axis is:""") __lowerCAmelCase = 1_0 while i <= 1_0_0_0_0_0: print(f'''with {i} steps: {trapezoidal_area(f, -5, 5, i)}''') i *= 1_0
271
'''simple docstring''' import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = inspect.getfile(accelerate.test_utils ) __UpperCAmelCase : List[str] = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''test_cli.py'''] ) __UpperCAmelCase : Dict = ['''accelerate''', '''launch'''] __UpperCAmelCase : Dict = Path.home() / '''.cache/huggingface/accelerate''' __UpperCAmelCase : Dict = '''default_config.yaml''' __UpperCAmelCase : Optional[Any] = config_folder / config_file __UpperCAmelCase : Dict = config_folder / '''_default_config.yaml''' __UpperCAmelCase : Any = Path('''tests/test_configs''' ) @classmethod def __lowercase ( cls : int ): '''simple docstring''' if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def __lowercase ( cls : List[Any] ): '''simple docstring''' if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def __lowercase ( self : Optional[int] ): '''simple docstring''' _a : Dict = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' for config in sorted(self.test_config_path.glob('**/*.yaml' ) ): with self.subTest(config_file=_a ): execute_subprocess_async( self.base_cmd + ['--config_file', str(_a ), self.test_file_path] ,env=os.environ.copy() ) def __lowercase ( self : Optional[int] ): '''simple docstring''' execute_subprocess_async(['accelerate', 'test'] ,env=os.environ.copy() ) class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = '''test-tpu''' __UpperCAmelCase : Any = '''us-central1-a''' __UpperCAmelCase : List[Any] = '''ls''' __UpperCAmelCase : Any = ['''accelerate''', '''tpu-config'''] __UpperCAmelCase : Dict = '''cd /usr/share''' __UpperCAmelCase : Any = '''tests/test_samples/test_command_file.sh''' __UpperCAmelCase : List[Any] = '''Running gcloud compute tpus tpu-vm ssh''' def __lowercase ( self : Dict ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Any = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command', self.command, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--debug'] ,return_stdout=_a ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : int ): '''simple docstring''' _a : Optional[Any] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" ,_a ,) def __lowercase ( self : str ): '''simple docstring''' _a : List[str] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--command', self.command, '--command', 'echo "Hello World"', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo \"Hello World\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Any = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--command_file', self.command_file, '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Union[str, Any] ): '''simple docstring''' _a : Union[str, Any] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/0_12_0.yaml', '--command_file', self.command_file, '--tpu_zone', self.tpu_zone, '--tpu_name', self.tpu_name, '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : Any ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + ['--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--debug'] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,) def __lowercase ( self : List[str] ): '''simple docstring''' _a : Optional[int] = run_command( self.cmd + [ '--config_file', 'tests/test_configs/latest.yaml', '--install_accelerate', '--accelerate_version', '12.0.0', '--debug', ] ,return_stdout=_a ,) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo \"hello world\"; echo \"this is a second command\" --worker all""" ,_a ,)
271
1
"""simple docstring""" from __future__ import annotations def _snake_case ( lowercase__ , lowercase__ , lowercase__ ): if len(lowercase__ ) == 0: raise ValueError('find_max() arg is an empty sequence' ) if ( left >= len(lowercase__ ) or left < -len(lowercase__ ) or right >= len(lowercase__ ) or right < -len(lowercase__ ) ): raise IndexError('list index out of range' ) if left == right: return nums[left] _lowerCamelCase : int = (left + right) >> 1 # the middle _lowerCamelCase : Tuple = find_max(lowercase__ , lowercase__ , lowercase__ ) # find max in range[left, mid] _lowerCamelCase : str = find_max(lowercase__ , mid + 1 , lowercase__ ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
12
"""simple docstring""" import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer lowercase__ = logging.get_logger(__name__) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """AutoTokenizer""" lowerCamelCase__ = ["""tokenizer"""] lowerCamelCase__ = { """semantic_prompt""": 1, """coarse_prompt""": 2, """fine_prompt""": 2, } def __init__( self , lowercase , lowercase=None ): super().__init__(lowercase ) _lowerCamelCase : Optional[int] = speaker_embeddings @classmethod def A_ ( cls , lowercase , lowercase="speaker_embeddings_path.json" , **lowercase ): if speaker_embeddings_dict_path is not None: _lowerCamelCase : Optional[Any] = get_file_from_repo( lowercase , lowercase , subfolder=kwargs.pop('subfolder' , lowercase ) , cache_dir=kwargs.pop('cache_dir' , lowercase ) , force_download=kwargs.pop('force_download' , lowercase ) , proxies=kwargs.pop('proxies' , lowercase ) , resume_download=kwargs.pop('resume_download' , lowercase ) , local_files_only=kwargs.pop('local_files_only' , lowercase ) , use_auth_token=kwargs.pop('use_auth_token' , lowercase ) , revision=kwargs.pop('revision' , lowercase ) , ) if speaker_embeddings_path is None: logger.warning( F'''`{os.path.join(lowercase , lowercase )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.''' ) _lowerCamelCase : List[Any] = None else: with open(lowercase ) as speaker_embeddings_json: _lowerCamelCase : Union[str, Any] = json.load(lowercase ) else: _lowerCamelCase : Tuple = None _lowerCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(lowercase , **lowercase ) return cls(tokenizer=lowercase , speaker_embeddings=lowercase ) def A_ ( self , lowercase , lowercase="speaker_embeddings_path.json" , lowercase="speaker_embeddings" , lowercase = False , **lowercase , ): if self.speaker_embeddings is not None: os.makedirs(os.path.join(lowercase , lowercase , 'v2' ) , exist_ok=lowercase ) _lowerCamelCase : int = {} _lowerCamelCase : List[Any] = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _lowerCamelCase : Optional[Any] = self._load_voice_preset(lowercase ) _lowerCamelCase : Any = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['repo_or_path'] , lowercase , F'''{prompt_key}_{key}''' ) , voice_preset[key] , allow_pickle=lowercase , ) _lowerCamelCase : List[str] = os.path.join(lowercase , F'''{prompt_key}_{key}.npy''' ) _lowerCamelCase : Optional[Any] = tmp_dict with open(os.path.join(lowercase , lowercase ) , 'w' ) as fp: json.dump(lowercase , lowercase ) super().save_pretrained(lowercase , lowercase , **lowercase ) def A_ ( self , lowercase = None , **lowercase ): _lowerCamelCase : Tuple = self.speaker_embeddings[voice_preset] _lowerCamelCase : Any = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F'''Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].''' ) _lowerCamelCase : Union[str, Any] = get_file_from_repo( self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] , subfolder=kwargs.pop('subfolder' , lowercase ) , cache_dir=kwargs.pop('cache_dir' , lowercase ) , force_download=kwargs.pop('force_download' , lowercase ) , proxies=kwargs.pop('proxies' , lowercase ) , resume_download=kwargs.pop('resume_download' , lowercase ) , local_files_only=kwargs.pop('local_files_only' , lowercase ) , use_auth_token=kwargs.pop('use_auth_token' , lowercase ) , revision=kwargs.pop('revision' , lowercase ) , ) if path is None: raise ValueError( F'''`{os.path.join(self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.''' ) _lowerCamelCase : List[str] = np.load(lowercase ) return voice_preset_dict def A_ ( self , lowercase = None ): for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F'''Voice preset unrecognized, missing {key} as a key.''' ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) def __call__( self , lowercase=None , lowercase=None , lowercase="pt" , lowercase=256 , lowercase=False , lowercase=True , lowercase=False , **lowercase , ): if voice_preset is not None and not isinstance(lowercase , lowercase ): if ( isinstance(lowercase , lowercase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _lowerCamelCase : Any = self._load_voice_preset(lowercase ) else: if isinstance(lowercase , lowercase ) and not voice_preset.endswith('.npz' ): _lowerCamelCase : Optional[Any] = voice_preset + '.npz' _lowerCamelCase : Union[str, Any] = np.load(lowercase ) if voice_preset is not None: self._validate_voice_preset_dict(lowercase , **lowercase ) _lowerCamelCase : Tuple = BatchFeature(data=lowercase , tensor_type=lowercase ) _lowerCamelCase : Any = self.tokenizer( lowercase , return_tensors=lowercase , padding='max_length' , max_length=lowercase , return_attention_mask=lowercase , return_token_type_ids=lowercase , add_special_tokens=lowercase , **lowercase , ) if voice_preset is not None: _lowerCamelCase : Optional[int] = voice_preset return encoded_text
12
1
'''simple docstring''' import json import sys def lowerCamelCase (_SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : int ): with open(_SCREAMING_SNAKE_CASE , encoding='utf-8' ) as f: __a : Optional[Any] = json.load(_SCREAMING_SNAKE_CASE ) __a : str = ['<details>', '<summary>Show updated benchmarks!</summary>', ' '] for benchmark_name in sorted(_SCREAMING_SNAKE_CASE ): __a : Optional[Any] = results[benchmark_name] __a : List[str] = benchmark_name.split('/' )[-1] output_md.append(F"""### Benchmark: {benchmark_file_name}""" ) __a : List[Any] = '| metric |' __a : List[Any] = '|--------|' __a : Union[str, Any] = '| new / old (diff) |' for metric_name in sorted(_SCREAMING_SNAKE_CASE ): __a : str = benchmark_res[metric_name] __a : str = metric_vals['new'] __a : Dict = metric_vals.get('old' , _SCREAMING_SNAKE_CASE ) __a : Dict = metric_vals.get('diff' , _SCREAMING_SNAKE_CASE ) __a : Union[str, Any] = F""" {new_val:f}""" if isinstance(_SCREAMING_SNAKE_CASE , (int, float) ) else 'None' if old_val is not None: val_str += F""" / {old_val:f}""" if isinstance(_SCREAMING_SNAKE_CASE , (int, float) ) else "None" if dif_val is not None: val_str += F""" ({dif_val:f})""" if isinstance(_SCREAMING_SNAKE_CASE , (int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append('</details>' ) with open(_SCREAMING_SNAKE_CASE , 'w' , encoding='utf-8' ) as f: f.writelines('\n'.join(_SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": __lowercase : List[str] = sys.argv[1] __lowercase : str = sys.argv[2] format_json_to_md(input_json_file, output_md_file)
27
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModel, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableUnCLIPImgaImgPipeline, UNetaDConditionModel from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, load_image, load_numpy, require_torch_gpu, skip_mps, slow, torch_device, ) from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class a_ ( a__ , a__ , a__ , unittest.TestCase ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = StableUnCLIPImgaImgPipeline __SCREAMING_SNAKE_CASE : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS __SCREAMING_SNAKE_CASE : List[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS __SCREAMING_SNAKE_CASE : Any = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __SCREAMING_SNAKE_CASE : Tuple = frozenset([] ) def __lowerCAmelCase ( self ) ->Union[str, Any]: SCREAMING_SNAKE_CASE : Optional[Any] = 32 SCREAMING_SNAKE_CASE : Tuple = embedder_hidden_size # image encoding components SCREAMING_SNAKE_CASE : int = CLIPImageProcessor(crop_size=32 , size=32 ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE : Optional[int] = CLIPVisionModelWithProjection( CLIPVisionConfig( hidden_size=_lowerCamelCase , projection_dim=_lowerCamelCase , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , ) ) # regular denoising components torch.manual_seed(0 ) SCREAMING_SNAKE_CASE : int = StableUnCLIPImageNormalizer(embedding_dim=_lowerCamelCase ) SCREAMING_SNAKE_CASE : Tuple = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''' ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE : List[str] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE : Any = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=_lowerCamelCase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE : Any = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=_lowerCamelCase , layers_per_block=1 , upcast_attention=_lowerCamelCase , use_linear_projection=_lowerCamelCase , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE : Optional[int] = DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , prediction_type='''v_prediction''' , set_alpha_to_one=_lowerCamelCase , steps_offset=1 , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE : Dict = AutoencoderKL() SCREAMING_SNAKE_CASE : Optional[Any] = { # image encoding components '''feature_extractor''': feature_extractor, '''image_encoder''': image_encoder.eval(), # image noising components '''image_normalizer''': image_normalizer.eval(), '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder.eval(), '''unet''': unet.eval(), '''scheduler''': scheduler, '''vae''': vae.eval(), } return components def __lowerCAmelCase ( self , _lowerCamelCase , _lowerCamelCase=0 , _lowerCamelCase=True ) ->Optional[int]: if str(_lowerCamelCase ).startswith('''mps''' ): SCREAMING_SNAKE_CASE : Optional[int] = torch.manual_seed(_lowerCamelCase ) else: SCREAMING_SNAKE_CASE : Union[str, Any] = torch.Generator(device=_lowerCamelCase ).manual_seed(_lowerCamelCase ) SCREAMING_SNAKE_CASE : List[str] = floats_tensor((1, 3, 32, 32) , rng=random.Random(_lowerCamelCase ) ).to(_lowerCamelCase ) if pil_image: SCREAMING_SNAKE_CASE : Any = input_image * 0.5 + 0.5 SCREAMING_SNAKE_CASE : int = input_image.clamp(0 , 1 ) SCREAMING_SNAKE_CASE : Union[str, Any] = input_image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() SCREAMING_SNAKE_CASE : List[str] = DiffusionPipeline.numpy_to_pil(_lowerCamelCase )[0] return { "prompt": "An anime racoon running a marathon", "image": input_image, "generator": generator, "num_inference_steps": 2, "output_type": "np", } @skip_mps def __lowerCAmelCase ( self ) ->Tuple: SCREAMING_SNAKE_CASE : Any = '''cpu''' # ensure determinism for the device-dependent torch.Generator SCREAMING_SNAKE_CASE : Union[str, Any] = self.get_dummy_components() SCREAMING_SNAKE_CASE : Tuple = StableUnCLIPImgaImgPipeline(**_lowerCamelCase ) SCREAMING_SNAKE_CASE : str = sd_pipe.to(_lowerCamelCase ) sd_pipe.set_progress_bar_config(disable=_lowerCamelCase ) SCREAMING_SNAKE_CASE : List[Any] = self.get_dummy_inputs(_lowerCamelCase ) inputs.update({'''image_embeds''': None} ) SCREAMING_SNAKE_CASE : Optional[int] = sd_pipe(**_lowerCamelCase ).images SCREAMING_SNAKE_CASE : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) SCREAMING_SNAKE_CASE : Tuple = np.array([0.3_8_7_2, 0.7_2_2_4, 0.5_6_0_1, 0.4_7_4_1, 0.6_8_7_2, 0.5_8_1_4, 0.4_6_3_6, 0.3_8_6_7, 0.5_0_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def __lowerCAmelCase ( self ) ->Tuple: SCREAMING_SNAKE_CASE : str = torch_device in ['''cpu''', '''mps'''] self._test_attention_slicing_forward_pass(test_max_difference=_lowerCamelCase ) def __lowerCAmelCase ( self ) ->List[Any]: SCREAMING_SNAKE_CASE : Tuple = torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=_lowerCamelCase ) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def __lowerCAmelCase ( self ) ->Optional[int]: self._test_xformers_attention_forwardGenerator_pass(test_max_difference=_lowerCamelCase ) @slow @require_torch_gpu class a_ ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ) ->int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self ) ->Tuple: SCREAMING_SNAKE_CASE : List[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''' ) SCREAMING_SNAKE_CASE : Optional[Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_img2img_anime_turtle_fp16.npy''' ) SCREAMING_SNAKE_CASE : List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-l-img2img''' , torch_dtype=torch.floataa ) pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() SCREAMING_SNAKE_CASE : Union[str, Any] = torch.Generator(device='''cpu''' ).manual_seed(0 ) SCREAMING_SNAKE_CASE : Tuple = pipe(_lowerCamelCase , '''anime turle''' , generator=_lowerCamelCase , output_type='''np''' ) SCREAMING_SNAKE_CASE : List[str] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(_lowerCamelCase , _lowerCamelCase ) def __lowerCAmelCase ( self ) ->Optional[Any]: SCREAMING_SNAKE_CASE : List[str] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''' ) SCREAMING_SNAKE_CASE : Optional[int] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_h_img2img_anime_turtle_fp16.npy''' ) SCREAMING_SNAKE_CASE : List[str] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa ) pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() SCREAMING_SNAKE_CASE : Any = torch.Generator(device='''cpu''' ).manual_seed(0 ) SCREAMING_SNAKE_CASE : Tuple = pipe(_lowerCamelCase , '''anime turle''' , generator=_lowerCamelCase , output_type='''np''' ) SCREAMING_SNAKE_CASE : List[str] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(_lowerCamelCase , _lowerCamelCase ) def __lowerCAmelCase ( self ) ->Any: SCREAMING_SNAKE_CASE : Union[str, Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''' ) torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() SCREAMING_SNAKE_CASE : str = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa ) SCREAMING_SNAKE_CASE : Dict = pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() SCREAMING_SNAKE_CASE : Dict = pipe( _lowerCamelCase , '''anime turtle''' , num_inference_steps=2 , output_type='''np''' , ) SCREAMING_SNAKE_CASE : Any = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
313
0
"""simple docstring""" import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed lowerCAmelCase : str = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"""{bindir}/../../examples/pytorch/translation"""): from run_translation import main # noqa set_seed(42) lowerCAmelCase : List[Any] = """sshleifer/student_marian_en_ro_6_1""" lowerCAmelCase : int = """sshleifer/tiny-mbart""" @require_torch class __magic_name__ ( UpperCAmelCase__ ): '''simple docstring''' def _lowerCAmelCase ( self , _a=False , _a=None , _a=True , _a=True , _a=True , _a=True , ): """simple docstring""" lowerCamelCase = self.run_trainer( eval_steps=1 , max_len=12 , model_name=_a , num_train_epochs=1 , distributed=_a , extra_args_str=_a , predict_with_generate=_a , do_train=_a , do_eval=_a , do_predict=_a , ) lowerCamelCase = TrainerState.load_from_json(os.path.join(_a , """trainer_state.json""" ) ).log_history if not do_eval: return lowerCamelCase = [log for log in logs if """eval_loss""" in log.keys()] lowerCamelCase = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats lowerCamelCase = eval_metrics[-1] assert isinstance(last_step_stats["""eval_bleu"""] , _a ) assert not math.isnan(float(last_step_stats["""eval_loss"""] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def _lowerCAmelCase ( self ): """simple docstring""" self.run_seqaseq_quick() @require_torch_multi_gpu def _lowerCAmelCase ( self ): """simple docstring""" self.run_seqaseq_quick(distributed=_a ) @require_torch_multi_gpu def _lowerCAmelCase ( self ): """simple docstring""" self.run_seqaseq_quick(distributed=_a ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowerCAmelCase ( self ): """simple docstring""" self.run_seqaseq_quick(distributed=_a , extra_args_str="""--sharded_ddp simple""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowerCAmelCase ( self ): """simple docstring""" self.run_seqaseq_quick(distributed=_a , extra_args_str="""--sharded_ddp simple --fp16""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowerCAmelCase ( self ): """simple docstring""" self.run_seqaseq_quick(distributed=_a , extra_args_str="""--sharded_ddp zero_dp_2""" , predict_with_generate=_a ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def _lowerCAmelCase ( self ): """simple docstring""" self.run_seqaseq_quick( distributed=_a , extra_args_str="""--sharded_ddp zero_dp_2 --fp16""" , predict_with_generate=_a ) @require_apex @require_torch_gpu def _lowerCAmelCase ( self ): """simple docstring""" # XXX: apex breaks the trainer if it's run twice e.g. run_seq2seq.main() from the same # program and it breaks other tests that run from the same pytest worker, therefore until this is # sorted out it must be run only in an external program, that is distributed=True in this # test and only under one or more gpus - if we want cpu will need to make a special test # # specifically to the problem traced it to self.optimizer.step() - if it's run 2nd time via # 2nd main() call it botches the future eval. # self.run_seqaseq_quick(distributed=_a , extra_args_str="""--fp16 --fp16_backend=apex""" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=_a , extra_args_str="""--fp16 --fp16_backend=apex""" ) @parameterized.expand(["""base""", """low""", """high""", """mixed"""] ) @require_torch_multi_gpu def _lowerCAmelCase ( self , _a ): """simple docstring""" # as each sub-test is slow-ish split into multiple sub-tests to avoid CI timeout lowerCamelCase = { # test with the default log_level - should be info and thus log info once """base""": {"""extra_args_str""": """""", """n_matches""": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes """low""": {"""extra_args_str""": """--log_level debug --log_level_replica debug""", """n_matches""": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica """high""": {"""extra_args_str""": """--log_level error --log_level_replica debug""", """n_matches""": 1}, # test with high log_level and log_level_replica - should be quiet on all processes """mixed""": {"""extra_args_str""": """--log_level error --log_level_replica error""", """n_matches""": 0}, } lowerCamelCase = experiments[experiment_id] lowerCamelCase = {"""distributed""": True, """predict_with_generate""": False, """do_eval""": False, """do_predict""": False} lowerCamelCase = """Running training""" with CaptureStderr() as cl: self.run_seqaseq_quick(**_a , extra_args_str=data["""extra_args_str"""] ) lowerCamelCase = len(re.findall(_a , cl.err ) ) self.assertEqual(_a , data["""n_matches"""] ) @slow def _lowerCAmelCase ( self ): """simple docstring""" lowerCamelCase = self.run_trainer( eval_steps=2 , max_len=128 , model_name=_a , learning_rate=3e-4 , num_train_epochs=10 , distributed=_a , ) # Check metrics lowerCamelCase = TrainerState.load_from_json(os.path.join(_a , """trainer_state.json""" ) ).log_history lowerCamelCase = [log for log in logs if """eval_loss""" in log.keys()] lowerCamelCase = eval_metrics[0] lowerCamelCase = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["""eval_bleu"""] , _a ) # test if do_predict saves generations and metrics lowerCamelCase = os.listdir(_a ) lowerCamelCase = {os.path.basename(_a ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def _lowerCAmelCase ( self ): """simple docstring""" from transformers.training_args import OptimizerNames def train_and_return_metrics(_a ) -> Tuple[int, float]: lowerCamelCase = """--skip_memory_metrics 0""" lowerCamelCase = self.run_trainer( max_len=128 , model_name=_a , learning_rate=3e-4 , num_train_epochs=1 , optim=_a , distributed=_a , extra_args_str=_a , do_eval=_a , do_predict=_a , n_gpus_to_use=1 , ) # Check metrics lowerCamelCase = TrainerState.load_from_json(Path(_a , """trainer_state.json""" ) ).log_history lowerCamelCase = int(logs[0]["""train_mem_gpu_peaked_delta"""] / 2**20 ) lowerCamelCase = int(logs[0]["""train_mem_gpu_alloc_delta"""] / 2**20 ) lowerCamelCase = logs[0]["""train_loss"""] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss lowerCamelCase , lowerCamelCase , lowerCamelCase = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) lowerCamelCase , lowerCamelCase , lowerCamelCase = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) lowerCamelCase = gpu_alloc_mem_orig - gpu_alloc_mem_bnb lowerCamelCase = gpu_peak_mem_orig + gpu_alloc_mem_orig lowerCamelCase = gpu_peak_mem_bnb + gpu_alloc_mem_bnb lowerCamelCase = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings lowerCamelCase = 120 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( _a , _a , """should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got""" f' a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and' f' gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB' , ) self.assertGreater( _a , _a , """should use ~150MB less total gpu memory with BNB, compared to without it for this model but got""" f' a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and' f' gpu_total_mem_bnb={gpu_total_mem_bnb}MB' , ) self.assertEqual( _a , _a , f'loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}' ) def _lowerCAmelCase ( self , _a , _a , _a , _a = 3e-3 , _a = "adafactor" , _a = False , _a = None , _a = 0 , _a = True , _a = True , _a = True , _a = True , _a = None , ): """simple docstring""" lowerCamelCase = self.test_file_dir / """../fixtures/tests_samples/wmt_en_ro""" lowerCamelCase = self.get_auto_remove_tmp_dir() lowerCamelCase = f'\n --model_name_or_path {model_name}\n --train_file {data_dir}/train.json\n --validation_file {data_dir}/val.json\n --test_file {data_dir}/test.json\n --output_dir {output_dir}\n --overwrite_output_dir\n --max_train_samples 8\n --max_source_length {max_len}\n --max_target_length {max_len}\n --do_train\n --num_train_epochs {str(_a )}\n --per_device_train_batch_size 4\n --learning_rate {learning_rate}\n --warmup_steps 8\n --logging_steps 0\n --logging_strategy no\n --save_steps {str(_a )}\n --group_by_length\n --label_smoothing_factor 0.1\n --target_lang ro_RO\n --source_lang en_XX\n '.split() lowerCamelCase = f'\n --do_eval\n --per_device_eval_batch_size 4\n --max_eval_samples 8\n --val_max_target_length {max_len}\n --evaluation_strategy steps\n --eval_steps {str(_a )}\n '.split() lowerCamelCase = """ --do_predict """.split() lowerCamelCase = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += f'--optim {optim}'.split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: lowerCamelCase = get_gpu_count() lowerCamelCase = get_torch_dist_unique_port() lowerCamelCase = f'\n -m torch.distributed.run\n --nproc_per_node={n_gpus_to_use}\n --master_port={master_port}\n {self.examples_dir_str}/pytorch/translation/run_translation.py\n '.split() lowerCamelCase = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(_a , env=self.get_env() ) else: lowerCamelCase = ["""run_translation.py"""] + args with patch.object(_a , """argv""" , _a ): main() return output_dir
168
"""simple docstring""" import math import time from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class __magic_name__ ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self , *_a , _a=None , _a=None , **_a ): """simple docstring""" super().__init__(*_a , **_a ) lowerCamelCase = eval_examples lowerCamelCase = post_process_function def _lowerCAmelCase ( self , _a=None , _a=None , _a=None , _a = "eval" ): """simple docstring""" lowerCamelCase = self.eval_dataset if eval_dataset is None else eval_dataset lowerCamelCase = self.get_eval_dataloader(_a ) lowerCamelCase = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. lowerCamelCase = self.compute_metrics lowerCamelCase = None lowerCamelCase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop lowerCamelCase = time.time() try: lowerCamelCase = eval_loop( _a , description="""Evaluation""" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_a , metric_key_prefix=_a , ) finally: lowerCamelCase = compute_metrics lowerCamelCase = self.args.eval_batch_size * self.args.world_size if f'{metric_key_prefix}_jit_compilation_time' in output.metrics: start_time += output.metrics[f'{metric_key_prefix}_jit_compilation_time'] output.metrics.update( speed_metrics( _a , _a , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default lowerCamelCase = self.post_process_function(_a , _a , output.predictions ) lowerCamelCase = self.compute_metrics(_a ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f'{metric_key_prefix}_' ): lowerCamelCase = metrics.pop(_a ) metrics.update(output.metrics ) else: lowerCamelCase = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(_a ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) lowerCamelCase = self.callback_handler.on_evaluate(self.args , self.state , self.control , _a ) return metrics def _lowerCAmelCase ( self , _a , _a , _a=None , _a = "test" ): """simple docstring""" lowerCamelCase = self.get_test_dataloader(_a ) # Temporarily disable metric computation, we will do it in the loop here. lowerCamelCase = self.compute_metrics lowerCamelCase = None lowerCamelCase = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop lowerCamelCase = time.time() try: lowerCamelCase = eval_loop( _a , description="""Prediction""" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_a , metric_key_prefix=_a , ) finally: lowerCamelCase = compute_metrics lowerCamelCase = self.args.eval_batch_size * self.args.world_size if f'{metric_key_prefix}_jit_compilation_time' in output.metrics: start_time += output.metrics[f'{metric_key_prefix}_jit_compilation_time'] output.metrics.update( speed_metrics( _a , _a , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output lowerCamelCase = self.post_process_function(_a , _a , output.predictions , """predict""" ) lowerCamelCase = self.compute_metrics(_a ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(f'{metric_key_prefix}_' ): lowerCamelCase = metrics.pop(_a ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=_a )
168
1
'''simple docstring''' from math import pi, sqrt, tan def _lowerCAmelCase ( __snake_case : float ) -> float: if side_length < 0: raise ValueError('surface_area_cube() only accepts non-negative values' ) return 6 * side_length**2 def _lowerCAmelCase ( __snake_case : float , __snake_case : float , __snake_case : float ) -> float: if length < 0 or breadth < 0 or height < 0: raise ValueError('surface_area_cuboid() only accepts non-negative values' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def _lowerCAmelCase ( __snake_case : float ) -> float: if radius < 0: raise ValueError('surface_area_sphere() only accepts non-negative values' ) return 4 * pi * radius**2 def _lowerCAmelCase ( __snake_case : float ) -> float: if radius < 0: raise ValueError('surface_area_hemisphere() only accepts non-negative values' ) return 3 * pi * radius**2 def _lowerCAmelCase ( __snake_case : float , __snake_case : float ) -> float: if radius < 0 or height < 0: raise ValueError('surface_area_cone() only accepts non-negative values' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def _lowerCAmelCase ( __snake_case : float , __snake_case : float , __snake_case : float ) -> float: if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( 'surface_area_conical_frustum() only accepts non-negative values' ) __A : Any = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def _lowerCAmelCase ( __snake_case : float , __snake_case : float ) -> float: if radius < 0 or height < 0: raise ValueError('surface_area_cylinder() only accepts non-negative values' ) return 2 * pi * radius * (height + radius) def _lowerCAmelCase ( __snake_case : float , __snake_case : float ) -> float: if torus_radius < 0 or tube_radius < 0: raise ValueError('surface_area_torus() only accepts non-negative values' ) if torus_radius < tube_radius: raise ValueError( 'surface_area_torus() does not support spindle or self intersecting tori' ) return 4 * pow(__snake_case , 2 ) * torus_radius * tube_radius def _lowerCAmelCase ( __snake_case : float , __snake_case : float ) -> float: if length < 0 or width < 0: raise ValueError('area_rectangle() only accepts non-negative values' ) return length * width def _lowerCAmelCase ( __snake_case : float ) -> float: if side_length < 0: raise ValueError('area_square() only accepts non-negative values' ) return side_length**2 def _lowerCAmelCase ( __snake_case : float , __snake_case : float ) -> float: if base < 0 or height < 0: raise ValueError('area_triangle() only accepts non-negative values' ) return (base * height) / 2 def _lowerCAmelCase ( __snake_case : float , __snake_case : float , __snake_case : float ) -> float: if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('area_triangle_three_sides() only accepts non-negative values' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('Given three sides do not form a triangle' ) __A : Tuple = (sidea + sidea + sidea) / 2 __A : Any = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def _lowerCAmelCase ( __snake_case : float , __snake_case : float ) -> float: if base < 0 or height < 0: raise ValueError('area_parallelogram() only accepts non-negative values' ) return base * height def _lowerCAmelCase ( __snake_case : float , __snake_case : float , __snake_case : float ) -> float: if basea < 0 or basea < 0 or height < 0: raise ValueError('area_trapezium() only accepts non-negative values' ) return 1 / 2 * (basea + basea) * height def _lowerCAmelCase ( __snake_case : float ) -> float: if radius < 0: raise ValueError('area_circle() only accepts non-negative values' ) return pi * radius**2 def _lowerCAmelCase ( __snake_case : float , __snake_case : float ) -> float: if radius_x < 0 or radius_y < 0: raise ValueError('area_ellipse() only accepts non-negative values' ) return pi * radius_x * radius_y def _lowerCAmelCase ( __snake_case : float , __snake_case : float ) -> float: if diagonal_a < 0 or diagonal_a < 0: raise ValueError('area_rhombus() only accepts non-negative values' ) return 1 / 2 * diagonal_a * diagonal_a def _lowerCAmelCase ( __snake_case : int , __snake_case : float ) -> float: if not isinstance(__snake_case , __snake_case ) or sides < 3: raise ValueError( 'area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides' ) elif length < 0: raise ValueError( 'area_reg_polygon() only accepts non-negative values as \ length of a side' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('''[DEMO] Areas of various geometric shapes: \n''') print(f"""Rectangle: {area_rectangle(10, 20) = }""") print(f"""Square: {area_square(10) = }""") print(f"""Triangle: {area_triangle(10, 10) = }""") print(f"""Triangle: {area_triangle_three_sides(5, 12, 13) = }""") print(f"""Parallelogram: {area_parallelogram(10, 20) = }""") print(f"""Rhombus: {area_rhombus(10, 20) = }""") print(f"""Trapezium: {area_trapezium(10, 20, 30) = }""") print(f"""Circle: {area_circle(20) = }""") print(f"""Ellipse: {area_ellipse(10, 20) = }""") print('''\nSurface Areas of various geometric shapes: \n''') print(f"""Cube: {surface_area_cube(20) = }""") print(f"""Cuboid: {surface_area_cuboid(10, 20, 30) = }""") print(f"""Sphere: {surface_area_sphere(20) = }""") print(f"""Hemisphere: {surface_area_hemisphere(20) = }""") print(f"""Cone: {surface_area_cone(10, 20) = }""") print(f"""Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }""") print(f"""Cylinder: {surface_area_cylinder(10, 20) = }""") print(f"""Torus: {surface_area_torus(20, 10) = }""") print(f"""Equilateral Triangle: {area_reg_polygon(3, 10) = }""") print(f"""Square: {area_reg_polygon(4, 10) = }""") print(f"""Reqular Pentagon: {area_reg_polygon(5, 10) = }""")
190
'''simple docstring''' import itertools from dataclasses import dataclass from typing import Optional import pandas as pd import pyarrow as pa import datasets from datasets.table import table_cast @dataclass class SCREAMING_SNAKE_CASE (datasets.BuilderConfig ): lowerCAmelCase = None class SCREAMING_SNAKE_CASE (datasets.ArrowBasedBuilder ): lowerCAmelCase = PandasConfig def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' return datasets.DatasetInfo(features=self.config.features) def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase): '''simple docstring''' if not self.config.data_files: raise ValueError(F'At least one data file must be specified, but got data_files={self.config.data_files}') __A : Dict = dl_manager.download_and_extract(self.config.data_files) if isinstance(_UpperCAmelCase , (str, list, tuple)): __A : Union[str, Any] = data_files if isinstance(_UpperCAmelCase , _UpperCAmelCase): __A : Optional[int] = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive __A : Optional[Any] = [dl_manager.iter_files(_UpperCAmelCase) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'files': files})] __A : Tuple = [] for split_name, files in data_files.items(): if isinstance(_UpperCAmelCase , _UpperCAmelCase): __A : Dict = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive __A : Optional[Any] = [dl_manager.iter_files(_UpperCAmelCase) for file in files] splits.append(datasets.SplitGenerator(name=_UpperCAmelCase , gen_kwargs={'files': files})) return splits def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase): '''simple docstring''' if self.config.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example __A : List[str] = table_cast(_UpperCAmelCase , self.config.features.arrow_schema) return pa_table def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase): '''simple docstring''' for i, file in enumerate(itertools.chain.from_iterable(_UpperCAmelCase)): with open(_UpperCAmelCase , 'rb') as f: __A : Optional[int] = pa.Table.from_pandas(pd.read_pickle(_UpperCAmelCase)) yield i, self._cast_table(_UpperCAmelCase)
190
1
"""simple docstring""" # Copyright (c) 2021-, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #################################################################################################### # # Note: If when running this conversion script you're getting an exception: # ModuleNotFoundError: No module named 'megatron.model.enums' # you need to tell python where to find the clone of Megatron-LM, e.g.: # # cd /tmp # git clone https://github.com/NVIDIA/Megatron-LM # PYTHONPATH=/tmp/Megatron-LM python src/transformers/models/megatron_gpt2/convert_megatron_gpt2_checkpoint.py ... # # if you already have it cloned elsewhere, simply adjust the path to the existing path # # If the training was done using a Megatron-LM fork, e.g., # https://github.com/microsoft/Megatron-DeepSpeed/ then chances are that you need to have that one # in your path, i.e., /path/to/Megatron-DeepSpeed/ # import argparse import os import re import zipfile import torch from transformers import AutoTokenizer, GPTaConfig def lowerCAmelCase__ ( lowerCamelCase_ : Any ,lowerCamelCase_ : List[Any] ,lowerCamelCase_ : Dict=0): '''simple docstring''' if name is None: lowerCAmelCase__ : Optional[int] = None else: lowerCAmelCase__ : str = "." * max(0 ,spaces - 2) + "# {:" + str(50 - spaces) + "s}" lowerCAmelCase__ : Optional[int] = fmt.format(lowerCamelCase_) # Print and recurse (if needed). if isinstance(lowerCamelCase_ ,lowerCamelCase_): if msg is not None: print(lowerCamelCase_) for k in val.keys(): recursive_print(lowerCamelCase_ ,val[k] ,spaces + 2) elif isinstance(lowerCamelCase_ ,torch.Tensor): print(lowerCamelCase_ ,''':''' ,val.size()) else: print(lowerCamelCase_ ,''':''' ,lowerCamelCase_) def lowerCAmelCase__ ( lowerCamelCase_ : Optional[Any] ,lowerCamelCase_ : Tuple ,lowerCamelCase_ : int ,lowerCamelCase_ : List[Any] ,lowerCamelCase_ : Dict): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = param.size() if checkpoint_version == 1.0: # version 1.0 stores [num_heads * hidden_size * num_splits, :] lowerCAmelCase__ : List[str] = (num_heads, hidden_size, num_splits) + input_shape[1:] lowerCAmelCase__ : int = param.view(*lowerCamelCase_) lowerCAmelCase__ : str = param.transpose(0 ,2) lowerCAmelCase__ : Tuple = param.transpose(1 ,2).contiguous() elif checkpoint_version >= 2.0: # other versions store [num_heads * num_splits * hidden_size, :] lowerCAmelCase__ : Dict = (num_heads, num_splits, hidden_size) + input_shape[1:] lowerCAmelCase__ : Any = param.view(*lowerCamelCase_) lowerCAmelCase__ : int = param.transpose(0 ,1).contiguous() lowerCAmelCase__ : Optional[int] = param.view(*lowerCamelCase_) return param def lowerCAmelCase__ ( lowerCamelCase_ : int ,lowerCamelCase_ : str ,lowerCamelCase_ : Dict): '''simple docstring''' lowerCAmelCase__ : Tuple = {} # old versions did not store training args lowerCAmelCase__ : Tuple = input_state_dict.get('''args''' ,lowerCamelCase_) if ds_args is not None: # do not make the user write a config file when the exact dimensions/sizes are already in the checkpoint # from pprint import pprint # pprint(vars(ds_args)) lowerCAmelCase__ : List[Any] = ds_args.padded_vocab_size lowerCAmelCase__ : List[Any] = ds_args.max_position_embeddings lowerCAmelCase__ : Optional[int] = ds_args.hidden_size lowerCAmelCase__ : Tuple = ds_args.num_layers lowerCAmelCase__ : Any = ds_args.num_attention_heads lowerCAmelCase__ : Union[str, Any] = ds_args.ffn_hidden_size # pprint(config) # The number of heads. lowerCAmelCase__ : int = config.n_head # The hidden_size per head. lowerCAmelCase__ : int = config.n_embd // config.n_head # Megatron-LM checkpoint version if "checkpoint_version" in input_state_dict.keys(): lowerCAmelCase__ : Optional[Any] = input_state_dict["checkpoint_version"] else: lowerCAmelCase__ : Dict = 0.0 # The model. lowerCAmelCase__ : Any = input_state_dict["model"] # The language model. lowerCAmelCase__ : Union[str, Any] = model["language_model"] # The embeddings. lowerCAmelCase__ : Dict = lm["embedding"] # The word embeddings. lowerCAmelCase__ : List[Any] = embeddings["word_embeddings"]["weight"] # Truncate the embedding table to vocab_size rows. lowerCAmelCase__ : str = word_embeddings[: config.vocab_size, :] lowerCAmelCase__ : Union[str, Any] = word_embeddings # The position embeddings. lowerCAmelCase__ : Optional[int] = embeddings["position_embeddings"]["weight"] # Read the causal mask dimension (seqlen). [max_sequence_length, hidden_size] lowerCAmelCase__ : Optional[int] = pos_embeddings.size(0) if n_positions != config.n_positions: raise ValueError( f"""pos_embeddings.max_sequence_length={n_positions} and config.n_positions={config.n_positions} don't match""") # Store the position embeddings. lowerCAmelCase__ : List[str] = pos_embeddings # The transformer. lowerCAmelCase__ : Tuple = lm["transformer"] if "transformer" in lm.keys() else lm["encoder"] # The regex to extract layer names. lowerCAmelCase__ : Tuple = re.compile(r'''layers\.(\d+)\.([a-z0-9_.]+)\.([a-z]+)''') # The simple map of names for "automated" rules. lowerCAmelCase__ : Dict = { "attention.dense": ".attn.c_proj.", "self_attention.dense": ".attn.c_proj.", "mlp.dense_h_to_4h": ".mlp.c_fc.", "mlp.dense_4h_to_h": ".mlp.c_proj.", } # Extract the layers. for key, val in transformer.items(): # Match the name. lowerCAmelCase__ : str = layer_re.match(lowerCamelCase_) # Stop if that's not a layer if m is None: break # The index of the layer. lowerCAmelCase__ : Any = int(m.group(1)) # The name of the operation. lowerCAmelCase__ : Tuple = m.group(2) # Is it a weight or a bias? lowerCAmelCase__ : List[str] = m.group(3) # The name of the layer. lowerCAmelCase__ : Dict = f"""transformer.h.{layer_idx}""" # For layernorm(s), simply store the layer norm. if op_name.endswith('''layernorm'''): lowerCAmelCase__ : Tuple = "ln_1" if op_name.startswith('''input''') else "ln_2" lowerCAmelCase__ : Dict = val # Transpose the QKV matrix. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "weight": # Insert a tensor of 1x1xDxD bias. lowerCAmelCase__ : Dict = torch.tril(torch.ones((n_positions, n_positions) ,dtype=torch.floataa)).view( 1 ,1 ,lowerCamelCase_ ,lowerCamelCase_) lowerCAmelCase__ : int = causal_mask # Insert a "dummy" tensor for masked_bias. lowerCAmelCase__ : Any = torch.tensor(-1E4 ,dtype=torch.floataa) lowerCAmelCase__ : Any = masked_bias lowerCAmelCase__ : Union[str, Any] = fix_query_key_value_ordering(lowerCamelCase_ ,lowerCamelCase_ ,3 ,lowerCamelCase_ ,lowerCamelCase_) # Megatron stores (3*D) x D but transformers-GPT2 expects D x 3*D. lowerCAmelCase__ : List[str] = out_val.transpose(0 ,1).contiguous() # Store. lowerCAmelCase__ : Optional[int] = out_val # Transpose the bias. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "bias": lowerCAmelCase__ : List[str] = fix_query_key_value_ordering(lowerCamelCase_ ,lowerCamelCase_ ,3 ,lowerCamelCase_ ,lowerCamelCase_) # Store. No change of shape. lowerCAmelCase__ : int = out_val # Transpose the weights. elif weight_or_bias == "weight": lowerCAmelCase__ : str = megatron_to_transformers[op_name] lowerCAmelCase__ : Optional[Any] = val.transpose(0 ,1) # Copy the bias. elif weight_or_bias == "bias": lowerCAmelCase__ : Dict = megatron_to_transformers[op_name] lowerCAmelCase__ : str = val # DEBUG. assert config.n_layer == layer_idx + 1 # The final layernorm. lowerCAmelCase__ : List[Any] = transformer["final_layernorm.weight"] lowerCAmelCase__ : List[str] = transformer["final_layernorm.bias"] # For LM head, transformers' wants the matrix to weight embeddings. lowerCAmelCase__ : int = word_embeddings # It should be done! return output_state_dict def lowerCAmelCase__ ( ): '''simple docstring''' lowerCAmelCase__ : Dict = argparse.ArgumentParser() parser.add_argument('''--print-checkpoint-structure''' ,action='''store_true''') parser.add_argument( '''path_to_checkpoint''' ,type=lowerCamelCase_ ,help='''Path to the checkpoint file (.zip archive or direct .pt file)''' ,) parser.add_argument( '''--config_file''' ,default='''''' ,type=lowerCamelCase_ ,help='''An optional config json file describing the pre-trained model.''' ,) lowerCAmelCase__ : Optional[Any] = parser.parse_args() # Extract the basename. lowerCAmelCase__ : Any = os.path.dirname(args.path_to_checkpoint) # Load the model. # the .zip is very optional, let's keep it for backward compatibility print(f"""Extracting PyTorch state dictionary from {args.path_to_checkpoint}""") if args.path_to_checkpoint.endswith('''.zip'''): with zipfile.ZipFile(args.path_to_checkpoint ,'''r''') as checkpoint: with checkpoint.open('''release/mp_rank_00/model_optim_rng.pt''') as pytorch_dict: lowerCAmelCase__ : Any = torch.load(lowerCamelCase_ ,map_location='''cpu''') else: lowerCAmelCase__ : List[Any] = torch.load(args.path_to_checkpoint ,map_location='''cpu''') lowerCAmelCase__ : Any = input_state_dict.get('''args''' ,lowerCamelCase_) # Read the config, or default to the model released by NVIDIA. if args.config_file == "": if ds_args is not None: if ds_args.bias_gelu_fusion: lowerCAmelCase__ : int = "gelu_fast" elif ds_args.openai_gelu: lowerCAmelCase__ : Optional[int] = "gelu_new" else: lowerCAmelCase__ : List[str] = "gelu" else: # in the very early days this used to be "gelu_new" lowerCAmelCase__ : Optional[Any] = "gelu_new" # Spell out all parameters in case the defaults change. lowerCAmelCase__ : Union[str, Any] = GPTaConfig( vocab_size=50257 ,n_positions=1024 ,n_embd=1024 ,n_layer=24 ,n_head=16 ,n_inner=4096 ,activation_function=lowerCamelCase_ ,resid_pdrop=0.1 ,embd_pdrop=0.1 ,attn_pdrop=0.1 ,layer_norm_epsilon=1E-5 ,initializer_range=0.02 ,summary_type='''cls_index''' ,summary_use_proj=lowerCamelCase_ ,summary_activation=lowerCamelCase_ ,summary_proj_to_labels=lowerCamelCase_ ,summary_first_dropout=0.1 ,scale_attn_weights=lowerCamelCase_ ,use_cache=lowerCamelCase_ ,bos_token_id=50256 ,eos_token_id=50256 ,) else: lowerCAmelCase__ : List[Any] = GPTaConfig.from_json_file(args.config_file) lowerCAmelCase__ : str = ["GPT2LMHeadModel"] # Convert. print('''Converting''') lowerCAmelCase__ : List[str] = convert_megatron_checkpoint(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_) # Print the structure of converted state dict. if args.print_checkpoint_structure: recursive_print(lowerCamelCase_ ,lowerCamelCase_) # Add tokenizer class info to config # see https://github.com/huggingface/transformers/issues/13906) if ds_args is not None: lowerCAmelCase__ : Any = ds_args.tokenizer_type if tokenizer_type == "GPT2BPETokenizer": lowerCAmelCase__ : List[Any] = "gpt2" elif tokenizer_type == "PretrainedFromHF": lowerCAmelCase__ : str = ds_args.tokenizer_name_or_path else: raise ValueError(f"""Unrecognized tokenizer_type {tokenizer_type}""") else: lowerCAmelCase__ : Optional[int] = "gpt2" lowerCAmelCase__ : Optional[int] = AutoTokenizer.from_pretrained(lowerCamelCase_) lowerCAmelCase__ : Dict = type(lowerCamelCase_).__name__ lowerCAmelCase__ : Optional[Any] = tokenizer_class # Store the config to file. print('''Saving config''') config.save_pretrained(lowerCamelCase_) # Save tokenizer based on args print(f"""Adding {tokenizer_class} tokenizer files""") tokenizer.save_pretrained(lowerCamelCase_) # Store the state_dict to file. lowerCAmelCase__ : List[Any] = os.path.join(lowerCamelCase_ ,'''pytorch_model.bin''') print(f"""Saving checkpoint to \"{output_checkpoint_file}\"""") torch.save(lowerCamelCase_ ,lowerCamelCase_) #################################################################################################### if __name__ == "__main__": main() ####################################################################################################
370
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( BertTokenizer, ViltConfig, ViltForImageAndTextRetrieval, ViltForImagesAndTextClassification, ViltForMaskedLM, ViltForQuestionAnswering, ViltImageProcessor, ViltProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __snake_case : Tuple =logging.get_logger(__name__) def lowerCAmelCase__ ( lowerCamelCase_ : Tuple ,lowerCamelCase_ : Dict=False ,lowerCamelCase_ : List[Any]=False ,lowerCamelCase_ : int=False): '''simple docstring''' lowerCAmelCase__ : Tuple = [] for i in range(config.num_hidden_layers): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""transformer.blocks.{i}.norm1.weight""", f"""vilt.encoder.layer.{i}.layernorm_before.weight""")) rename_keys.append((f"""transformer.blocks.{i}.norm1.bias""", f"""vilt.encoder.layer.{i}.layernorm_before.bias""")) rename_keys.append( (f"""transformer.blocks.{i}.attn.proj.weight""", f"""vilt.encoder.layer.{i}.attention.output.dense.weight""")) rename_keys.append( (f"""transformer.blocks.{i}.attn.proj.bias""", f"""vilt.encoder.layer.{i}.attention.output.dense.bias""")) rename_keys.append((f"""transformer.blocks.{i}.norm2.weight""", f"""vilt.encoder.layer.{i}.layernorm_after.weight""")) rename_keys.append((f"""transformer.blocks.{i}.norm2.bias""", f"""vilt.encoder.layer.{i}.layernorm_after.bias""")) rename_keys.append( (f"""transformer.blocks.{i}.mlp.fc1.weight""", f"""vilt.encoder.layer.{i}.intermediate.dense.weight""")) rename_keys.append((f"""transformer.blocks.{i}.mlp.fc1.bias""", f"""vilt.encoder.layer.{i}.intermediate.dense.bias""")) rename_keys.append((f"""transformer.blocks.{i}.mlp.fc2.weight""", f"""vilt.encoder.layer.{i}.output.dense.weight""")) rename_keys.append((f"""transformer.blocks.{i}.mlp.fc2.bias""", f"""vilt.encoder.layer.{i}.output.dense.bias""")) # embeddings rename_keys.extend( [ # text embeddings ('''text_embeddings.word_embeddings.weight''', '''vilt.embeddings.text_embeddings.word_embeddings.weight'''), ( '''text_embeddings.position_embeddings.weight''', '''vilt.embeddings.text_embeddings.position_embeddings.weight''', ), ('''text_embeddings.position_ids''', '''vilt.embeddings.text_embeddings.position_ids'''), ( '''text_embeddings.token_type_embeddings.weight''', '''vilt.embeddings.text_embeddings.token_type_embeddings.weight''', ), ('''text_embeddings.LayerNorm.weight''', '''vilt.embeddings.text_embeddings.LayerNorm.weight'''), ('''text_embeddings.LayerNorm.bias''', '''vilt.embeddings.text_embeddings.LayerNorm.bias'''), # patch embeddings ('''transformer.cls_token''', '''vilt.embeddings.cls_token'''), ('''transformer.patch_embed.proj.weight''', '''vilt.embeddings.patch_embeddings.projection.weight'''), ('''transformer.patch_embed.proj.bias''', '''vilt.embeddings.patch_embeddings.projection.bias'''), ('''transformer.pos_embed''', '''vilt.embeddings.position_embeddings'''), # token type embeddings ('''token_type_embeddings.weight''', '''vilt.embeddings.token_type_embeddings.weight'''), ]) # final layernorm + pooler rename_keys.extend( [ ('''transformer.norm.weight''', '''vilt.layernorm.weight'''), ('''transformer.norm.bias''', '''vilt.layernorm.bias'''), ('''pooler.dense.weight''', '''vilt.pooler.dense.weight'''), ('''pooler.dense.bias''', '''vilt.pooler.dense.bias'''), ]) # classifier head(s) if vqa_model: # classification head rename_keys.extend( [ ('''vqa_classifier.0.weight''', '''classifier.0.weight'''), ('''vqa_classifier.0.bias''', '''classifier.0.bias'''), ('''vqa_classifier.1.weight''', '''classifier.1.weight'''), ('''vqa_classifier.1.bias''', '''classifier.1.bias'''), ('''vqa_classifier.3.weight''', '''classifier.3.weight'''), ('''vqa_classifier.3.bias''', '''classifier.3.bias'''), ]) elif nlvr_model: # classification head rename_keys.extend( [ ('''nlvr2_classifier.0.weight''', '''classifier.0.weight'''), ('''nlvr2_classifier.0.bias''', '''classifier.0.bias'''), ('''nlvr2_classifier.1.weight''', '''classifier.1.weight'''), ('''nlvr2_classifier.1.bias''', '''classifier.1.bias'''), ('''nlvr2_classifier.3.weight''', '''classifier.3.weight'''), ('''nlvr2_classifier.3.bias''', '''classifier.3.bias'''), ]) else: pass return rename_keys def lowerCAmelCase__ ( lowerCamelCase_ : Dict ,lowerCamelCase_ : Optional[Any]): '''simple docstring''' for i in range(config.num_hidden_layers): lowerCAmelCase__ : Optional[Any] = '''vilt.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCAmelCase__ : Dict = state_dict.pop(f"""transformer.blocks.{i}.attn.qkv.weight""") lowerCAmelCase__ : Tuple = state_dict.pop(f"""transformer.blocks.{i}.attn.qkv.bias""") # next, add query, keys and values (in that order) to the state dict lowerCAmelCase__ : Dict = in_proj_weight[ : config.hidden_size, : ] lowerCAmelCase__ : Tuple = in_proj_bias[: config.hidden_size] lowerCAmelCase__ : Union[str, Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCAmelCase__ : Optional[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCAmelCase__ : str = in_proj_weight[ -config.hidden_size :, : ] lowerCAmelCase__ : Union[str, Any] = in_proj_bias[-config.hidden_size :] def lowerCAmelCase__ ( lowerCamelCase_ : List[str]): '''simple docstring''' lowerCAmelCase__ : List[Any] = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(lowerCamelCase_ ,lowerCamelCase_) def lowerCAmelCase__ ( lowerCamelCase_ : Optional[Any] ,lowerCamelCase_ : Optional[int] ,lowerCamelCase_ : int): '''simple docstring''' lowerCAmelCase__ : int = dct.pop(lowerCamelCase_) lowerCAmelCase__ : List[Any] = val @torch.no_grad() def lowerCAmelCase__ ( lowerCamelCase_ : int ,lowerCamelCase_ : Union[str, Any]): '''simple docstring''' lowerCAmelCase__ : Optional[int] = ViltConfig(image_size=384 ,patch_size=32 ,tie_word_embeddings=lowerCamelCase_) lowerCAmelCase__ : Optional[Any] = False lowerCAmelCase__ : Union[str, Any] = False lowerCAmelCase__ : Dict = False lowerCAmelCase__ : Any = False if "vqa" in checkpoint_url: lowerCAmelCase__ : List[Any] = True lowerCAmelCase__ : List[Any] = 3129 lowerCAmelCase__ : List[Any] = '''huggingface/label-files''' lowerCAmelCase__ : Union[str, Any] = '''vqa2-id2label.json''' lowerCAmelCase__ : Optional[int] = json.load(open(hf_hub_download(lowerCamelCase_ ,lowerCamelCase_ ,repo_type='''dataset''') ,'''r''')) lowerCAmelCase__ : Optional[Any] = {int(lowerCamelCase_): v for k, v in idalabel.items()} lowerCAmelCase__ : Dict = idalabel lowerCAmelCase__ : List[Any] = {v: k for k, v in idalabel.items()} lowerCAmelCase__ : Optional[int] = ViltForQuestionAnswering(lowerCamelCase_) elif "nlvr" in checkpoint_url: lowerCAmelCase__ : str = True lowerCAmelCase__ : Optional[Any] = 2 lowerCAmelCase__ : Optional[Any] = {0: '''False''', 1: '''True'''} lowerCAmelCase__ : Union[str, Any] = {v: k for k, v in config.idalabel.items()} lowerCAmelCase__ : int = 3 lowerCAmelCase__ : int = ViltForImagesAndTextClassification(lowerCamelCase_) elif "irtr" in checkpoint_url: lowerCAmelCase__ : str = True lowerCAmelCase__ : List[str] = ViltForImageAndTextRetrieval(lowerCamelCase_) elif "mlm_itm" in checkpoint_url: lowerCAmelCase__ : Any = True lowerCAmelCase__ : int = ViltForMaskedLM(lowerCamelCase_) else: raise ValueError('''Unknown model type''') # load state_dict of original model, remove and rename some keys lowerCAmelCase__ : Tuple = torch.hub.load_state_dict_from_url(lowerCamelCase_ ,map_location='''cpu''')['''state_dict'''] lowerCAmelCase__ : Tuple = create_rename_keys(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_) for src, dest in rename_keys: rename_key(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_) read_in_q_k_v(lowerCamelCase_ ,lowerCamelCase_) if mlm_model or irtr_model: lowerCAmelCase__ : int = ['''itm_score.fc.weight''', '''itm_score.fc.bias'''] for k in ignore_keys: state_dict.pop(lowerCamelCase_ ,lowerCamelCase_) # load state dict into HuggingFace model model.eval() if mlm_model: lowerCAmelCase__ , lowerCAmelCase__ : str = model.load_state_dict(lowerCamelCase_ ,strict=lowerCamelCase_) assert missing_keys == ["mlm_score.decoder.bias"] else: model.load_state_dict(lowerCamelCase_) # Define processor lowerCAmelCase__ : List[str] = ViltImageProcessor(size=384) lowerCAmelCase__ : Tuple = BertTokenizer.from_pretrained('''bert-base-uncased''') lowerCAmelCase__ : Union[str, Any] = ViltProcessor(lowerCamelCase_ ,lowerCamelCase_) # Forward pass on example inputs (image + text) if nlvr_model: lowerCAmelCase__ : int = Image.open(requests.get('''https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg''' ,stream=lowerCamelCase_).raw) lowerCAmelCase__ : Union[str, Any] = Image.open(requests.get('''https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg''' ,stream=lowerCamelCase_).raw) lowerCAmelCase__ : Union[str, Any] = ( '''The left image contains twice the number of dogs as the right image, and at least two dogs in total are''' ''' standing.''' ) lowerCAmelCase__ : Optional[int] = processor(lowerCamelCase_ ,lowerCamelCase_ ,return_tensors='''pt''') lowerCAmelCase__ : Tuple = processor(lowerCamelCase_ ,lowerCamelCase_ ,return_tensors='''pt''') lowerCAmelCase__ : Union[str, Any] = model( input_ids=encoding_a.input_ids ,pixel_values=encoding_a.pixel_values ,pixel_values_a=encoding_a.pixel_values ,) else: lowerCAmelCase__ : Dict = Image.open(requests.get('''http://images.cocodataset.org/val2017/000000039769.jpg''' ,stream=lowerCamelCase_).raw) if mlm_model: lowerCAmelCase__ : int = '''a bunch of [MASK] laying on a [MASK].''' else: lowerCAmelCase__ : Optional[int] = '''How many cats are there?''' lowerCAmelCase__ : Optional[int] = processor(lowerCamelCase_ ,lowerCamelCase_ ,return_tensors='''pt''') lowerCAmelCase__ : Any = model(**lowerCamelCase_) # Verify outputs if mlm_model: lowerCAmelCase__ : Dict = torch.Size([1, 11, 30522]) lowerCAmelCase__ : int = torch.tensor([-12.5061, -12.5123, -12.5174]) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] ,lowerCamelCase_ ,atol=1E-4) # verify masked token prediction equals "cats" lowerCAmelCase__ : Optional[Any] = outputs.logits[0, 4, :].argmax(-1).item() assert tokenizer.decode([predicted_id]) == "cats" elif vqa_model: lowerCAmelCase__ : List[Any] = torch.Size([1, 3129]) lowerCAmelCase__ : str = torch.tensor([-15.9495, -18.1472, -10.3041]) assert torch.allclose(outputs.logits[0, :3] ,lowerCamelCase_ ,atol=1E-4) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] ,lowerCamelCase_ ,atol=1E-4) # verify vqa prediction equals "2" lowerCAmelCase__ : List[Any] = outputs.logits.argmax(-1).item() assert model.config.idalabel[predicted_idx] == "2" elif nlvr_model: lowerCAmelCase__ : Union[str, Any] = torch.Size([1, 2]) lowerCAmelCase__ : Dict = torch.tensor([-2.8721, 2.1291]) assert torch.allclose(outputs.logits[0, :3] ,lowerCamelCase_ ,atol=1E-4) assert outputs.logits.shape == expected_shape Path(lowerCamelCase_).mkdir(exist_ok=lowerCamelCase_) print(f"""Saving model and processor to {pytorch_dump_folder_path}""") model.save_pretrained(lowerCamelCase_) processor.save_pretrained(lowerCamelCase_) if __name__ == "__main__": __snake_case : Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://github.com/dandelin/ViLT/releases/download/200k/vilt_200k_mlm_itm.ckpt', type=str, help='URL of the checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) __snake_case : Union[str, Any] =parser.parse_args() convert_vilt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
94
0
import os import re import shutil import sys import tempfile import unittest import black snake_case : Any = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, '''utils''')) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If BertLMPredictionHead is changed in modeling_bert.py, this code needs to be manually updated. snake_case : int = ''' def __init__(self, config): super().__init__() self.transform = BertPredictionHeadTransform(config) # The output weights are the same as the input embeddings, but there is # an output-only bias for each token. self.decoder = nn.Linear(config.hidden_size, config.vocab_size, bias=False) self.bias = nn.Parameter(torch.zeros(config.vocab_size)) # Need a link between the two variables so that the bias is correctly resized with `resize_token_embeddings` self.decoder.bias = self.bias def forward(self, hidden_states): hidden_states = self.transform(hidden_states) hidden_states = self.decoder(hidden_states) return hidden_states ''' class _snake_case ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self ): a :Optional[Any] = tempfile.mkdtemp() os.makedirs(os.path.join(self.transformer_dir , '''models/bert/''' ) ) a :List[str] = self.transformer_dir shutil.copy( os.path.join(_lowerCamelCase , '''src/transformers/models/bert/modeling_bert.py''' ) , os.path.join(self.transformer_dir , '''models/bert/modeling_bert.py''' ) , ) def SCREAMING_SNAKE_CASE__ ( self ): a :Tuple = '''src/transformers''' shutil.rmtree(self.transformer_dir ) def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=None ): a :Dict = comment + F'''\nclass {class_name}(nn.Module):\n''' + class_code if overwrite_result is not None: a :str = comment + F'''\nclass {class_name}(nn.Module):\n''' + overwrite_result a :Dict = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 ) a :Any = black.format_str(_lowerCamelCase , mode=_lowerCamelCase ) a :List[str] = os.path.join(self.transformer_dir , '''new_code.py''' ) with open(_lowerCamelCase , '''w''' , newline='''\n''' ) as f: f.write(_lowerCamelCase ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(_lowerCamelCase ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=_lowerCamelCase ) with open(_lowerCamelCase , '''r''' ) as f: self.assertTrue(f.read() , _lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): a :Any = check_copies.find_code_in_transformers('''models.bert.modeling_bert.BertLMPredictionHead''' ) self.assertEqual(_lowerCamelCase , _lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): # Base copy consistency self.check_copy_consistency( '''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead''' , '''BertLMPredictionHead''' , REFERENCE_CODE + '''\n''' , ) # With no empty line at the end self.check_copy_consistency( '''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead''' , '''BertLMPredictionHead''' , _lowerCamelCase , ) # Copy consistency with rename self.check_copy_consistency( '''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel''' , '''TestModelLMPredictionHead''' , re.sub('''Bert''' , '''TestModel''' , _lowerCamelCase ) , ) # Copy consistency with a really long name a :str = '''TestModelWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason''' self.check_copy_consistency( F'''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->{long_class_name}''' , F'''{long_class_name}LMPredictionHead''' , re.sub('''Bert''' , _lowerCamelCase , _lowerCamelCase ) , ) # Copy consistency with overwrite self.check_copy_consistency( '''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel''' , '''TestModelLMPredictionHead''' , _lowerCamelCase , overwrite_result=re.sub('''Bert''' , '''TestModel''' , _lowerCamelCase ) , ) def SCREAMING_SNAKE_CASE__ ( self ): a :Any = check_copies.LOCALIZED_READMES['''README_zh-hans.md'''] a :List[str] = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the''' ''' Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for''' ''' Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong''' ''' Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.\n1.''' ''' **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (from HuggingFace),''' ''' released together with the paper [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and''' ''' lighter](https://arxiv.org/abs/1910.01108) by Victor Sanh, Lysandre Debut and Thomas Wolf. The same''' ''' method has been applied to compress GPT2 into''' ''' [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into''' ''' [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),''' ''' Multilingual BERT into''' ''' [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German''' ''' version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)**''' ''' (from Google Research/Stanford University) released with the paper [ELECTRA: Pre-training text encoders''' ''' as discriminators rather than generators](https://arxiv.org/abs/2003.10555) by Kevin Clark, Minh-Thang''' ''' Luong, Quoc V. Le, Christopher D. Manning.''' ) a :Union[str, Any] = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the''' ''' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of''' ''' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian''' ''' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n''' ) a :List[str] = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the''' ''' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of''' ''' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian''' ''' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n1.''' ''' **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (来自 HuggingFace) 伴随论文''' ''' [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and''' ''' lighter](https://arxiv.org/abs/1910.01108) 由 Victor Sanh, Lysandre Debut and Thomas Wolf 发布。 The same''' ''' method has been applied to compress GPT2 into''' ''' [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into''' ''' [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),''' ''' Multilingual BERT into''' ''' [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German''' ''' version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)** (来自''' ''' Google Research/Stanford University) 伴随论文 [ELECTRA: Pre-training text encoders as discriminators rather''' ''' than generators](https://arxiv.org/abs/2003.10555) 由 Kevin Clark, Minh-Thang Luong, Quoc V. Le,''' ''' Christopher D. Manning 发布。\n''' ) a , a :Optional[Any] = check_copies.convert_to_localized_md( _lowerCamelCase , _lowerCamelCase , localized_readme['''format_model_list'''] ) self.assertFalse(_lowerCamelCase ) self.assertEqual(_lowerCamelCase , _lowerCamelCase ) a , a :Optional[int] = check_copies.convert_to_localized_md( _lowerCamelCase , _lowerCamelCase , localized_readme['''format_model_list'''] ) # Check whether the number of models is equal to README.md after conversion. self.assertTrue(_lowerCamelCase ) a :str = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the''' ''' Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for''' ''' Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong''' ''' Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.''' ) a :Dict = ( '''1. **[ALBERT](https://huggingface.co/transformers/main/model_doc/albert.html)** (来自 Google Research and''' ''' the Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of''' ''' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian''' ''' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n''' ) a :str = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the''' ''' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of''' ''' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian''' ''' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n''' ) a , a :int = check_copies.convert_to_localized_md( _lowerCamelCase , _lowerCamelCase , localized_readme['''format_model_list'''] ) # Check if the model link is synchronized. self.assertEqual(_lowerCamelCase , _lowerCamelCase )
94
'''simple docstring''' def __UpperCamelCase ( UpperCAmelCase ): lowercase__ , lowercase__ : Optional[int] = [], [] while len(UpperCAmelCase ) > 1: lowercase__ , lowercase__ : List[str] = min(UpperCAmelCase ), max(UpperCAmelCase ) start.append(UpperCAmelCase ) end.append(UpperCAmelCase ) collection.remove(UpperCAmelCase ) collection.remove(UpperCAmelCase ) end.reverse() return start + collection + end if __name__ == "__main__": __a: Optional[int] = input("""Enter numbers separated by a comma:\n""").strip() __a: Tuple = [int(item) for item in user_input.split(""",""")] print(*merge_sort(unsorted), sep=""",""")
198
0
import unittest import numpy as np import torch from diffusers import DDIMPipeline, DDIMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow, torch_device from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class _a ( UpperCamelCase__ , unittest.TestCase ): _lowercase : Optional[int] = DDIMPipeline _lowercase : int = UNCONDITIONAL_IMAGE_GENERATION_PARAMS _lowercase : Optional[int] = PipelineTesterMixin.required_optional_params - { '''num_images_per_prompt''', '''latents''', '''callback''', '''callback_steps''', } _lowercase : Optional[Any] = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS _lowercase : Optional[int] = False def lowerCamelCase_ ( self: List[Any] ) -> int: """simple docstring""" torch.manual_seed(0 ) lowercase__ = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) lowercase__ = DDIMScheduler() lowercase__ = {'''unet''': unet, '''scheduler''': scheduler} return components def lowerCamelCase_ ( self: Any , UpperCamelCase_: Optional[int] , UpperCamelCase_: Union[str, Any]=0 ) -> Dict: """simple docstring""" if str(_snake_case ).startswith('''mps''' ): lowercase__ = torch.manual_seed(_snake_case ) else: lowercase__ = torch.Generator(device=_snake_case ).manual_seed(_snake_case ) lowercase__ = { '''batch_size''': 1, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def lowerCamelCase_ ( self: Dict ) -> Any: """simple docstring""" lowercase__ = '''cpu''' lowercase__ = self.get_dummy_components() lowercase__ = self.pipeline_class(**_snake_case ) pipe.to(_snake_case ) pipe.set_progress_bar_config(disable=_snake_case ) lowercase__ = self.get_dummy_inputs(_snake_case ) lowercase__ = pipe(**_snake_case ).images lowercase__ = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 32, 32, 3) ) lowercase__ = np.array( [1.0_0_0E0_0, 5.7_1_7E-0_1, 4.7_1_7E-0_1, 1.0_0_0E0_0, 0.0_0_0E0_0, 1.0_0_0E0_0, 3.0_0_0E-0_4, 0.0_0_0E0_0, 9.0_0_0E-0_4] ) lowercase__ = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_snake_case , 1E-3 ) def lowerCamelCase_ ( self: str ) -> Optional[Any]: """simple docstring""" super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def lowerCamelCase_ ( self: int ) -> List[Any]: """simple docstring""" super().test_save_load_local(expected_max_difference=3E-3 ) def lowerCamelCase_ ( self: Optional[int] ) -> str: """simple docstring""" super().test_save_load_optional_components(expected_max_difference=3E-3 ) def lowerCamelCase_ ( self: Optional[Any] ) -> Optional[int]: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class _a ( unittest.TestCase ): def lowerCamelCase_ ( self: int ) -> Dict: """simple docstring""" lowercase__ = '''google/ddpm-cifar10-32''' lowercase__ = UNetaDModel.from_pretrained(_snake_case ) lowercase__ = DDIMScheduler() lowercase__ = DDIMPipeline(unet=_snake_case , scheduler=_snake_case ) ddim.to(_snake_case ) ddim.set_progress_bar_config(disable=_snake_case ) lowercase__ = torch.manual_seed(0 ) lowercase__ = ddim(generator=_snake_case , eta=0.0 , output_type='''numpy''' ).images lowercase__ = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowercase__ = np.array([0.1723, 0.1617, 0.1600, 0.1626, 0.1497, 0.1513, 0.1505, 0.1442, 0.1453] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase_ ( self: Optional[int] ) -> Tuple: """simple docstring""" lowercase__ = '''google/ddpm-ema-bedroom-256''' lowercase__ = UNetaDModel.from_pretrained(_snake_case ) lowercase__ = DDIMScheduler.from_pretrained(_snake_case ) lowercase__ = DDIMPipeline(unet=_snake_case , scheduler=_snake_case ) ddpm.to(_snake_case ) ddpm.set_progress_bar_config(disable=_snake_case ) lowercase__ = torch.manual_seed(0 ) lowercase__ = ddpm(generator=_snake_case , output_type='''numpy''' ).images lowercase__ = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) lowercase__ = np.array([0.0060, 0.0201, 0.0344, 0.0024, 0.0018, 0.0002, 0.0022, 0.0000, 0.0069] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
361
def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" return base * power(SCREAMING_SNAKE_CASE , (exponent - 1) ) if exponent else 1 if __name__ == "__main__": print('Raise base to the power of exponent using recursion...') lowerCAmelCase = int(input('Enter the base: ').strip()) lowerCAmelCase = int(input('Enter the exponent: ').strip()) lowerCAmelCase = power(base, abs(exponent)) if exponent < 0: # power() does not properly deal w/ negative exponents lowerCAmelCase = 1 / result print(f"""{base} to the power of {exponent} is {result}""")
93
0
from __future__ import annotations def lowerCamelCase__ ( A__ : list[int | float] , A__ : int , A__ : int ): '''simple docstring''' if len(A__ ) == 0: raise ValueError("""find_max() arg is an empty sequence""" ) if ( left >= len(A__ ) or left < -len(A__ ) or right >= len(A__ ) or right < -len(A__ ) ): raise IndexError("""list index out of range""" ) if left == right: return nums[left] __lowerCamelCase = (left + right) >> 1 # the middle __lowerCamelCase = find_max(A__ , A__ , A__ ) # find max in range[left, mid] __lowerCamelCase = find_max(A__ , mid + 1 , A__ ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
12
# Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version UpperCAmelCase_ = get_logger(__name__) class lowerCamelCase__: UpperCAmelCase__ : List[Any] = 'dummy_data' UpperCAmelCase__ : str = 'datasets' UpperCAmelCase__ : Tuple = False def __init__( self: Optional[Any] , UpperCamelCase_: str , UpperCamelCase_: str , UpperCamelCase_: Union[Version, str] , UpperCamelCase_: Optional[str] = None , UpperCamelCase_: bool = False , UpperCamelCase_: bool = True , UpperCamelCase_: Optional[List[Callable]] = None , ): __lowerCamelCase = 0 __lowerCamelCase = dataset_name __lowerCamelCase = cache_dir __lowerCamelCase = use_local_dummy_data __lowerCamelCase = config # download_callbacks take a single url as input __lowerCamelCase = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root __lowerCamelCase = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general __lowerCamelCase = str(UpperCamelCase_ ) # to be downloaded __lowerCamelCase = None __lowerCamelCase = None @property def lowerCAmelCase__ ( self: List[Any] ): if self._dummy_file is None: __lowerCamelCase = self.download_dummy_data() return self._dummy_file @property def lowerCAmelCase__ ( self: str ): if self.config is not None: # structure is dummy / config_name / version_name return os.path.join("""dummy""" , self.config.name , self.version_name ) # structure is dummy / version_name return os.path.join("""dummy""" , self.version_name ) @property def lowerCAmelCase__ ( self: Optional[Any] ): return os.path.join(self.dummy_data_folder , """dummy_data.zip""" ) def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) __lowerCamelCase = cached_path( UpperCamelCase_ , cache_dir=self.cache_dir , extract_compressed_file=UpperCamelCase_ , force_extract=UpperCamelCase_ ) return os.path.join(UpperCamelCase_ , self.dummy_file_name ) @property def lowerCAmelCase__ ( self: Optional[Any] ): return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file ) @property def lowerCAmelCase__ ( self: Tuple ): if self._bucket_url is None: __lowerCamelCase = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , """/""" ) ) return self._bucket_url @property def lowerCAmelCase__ ( self: str ): # return full path if its a dir if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , """/""" ).split("""/""" )[:-1] ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: Dict , *UpperCamelCase_: str ): if self.load_existing_dummy_data: # dummy data is downloaded and tested __lowerCamelCase = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned __lowerCamelCase = self.dummy_file_name # special case when data_url is a dict if isinstance(UpperCamelCase_ , UpperCamelCase_ ): return self.create_dummy_data_dict(UpperCamelCase_ , UpperCamelCase_ ) elif isinstance(UpperCamelCase_ , (list, tuple) ): return self.create_dummy_data_list(UpperCamelCase_ , UpperCamelCase_ ) else: return self.create_dummy_data_single(UpperCamelCase_ , UpperCamelCase_ ) def lowerCAmelCase__ ( self: Any , UpperCamelCase_: Optional[Any] , *UpperCamelCase_: str ): return self.download_and_extract(UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Tuple , UpperCamelCase_: str ): return self.download_and_extract(UpperCamelCase_ ) def lowerCAmelCase__ ( self: str , UpperCamelCase_: int , *UpperCamelCase_: List[str] , **UpperCamelCase_: str ): return path def lowerCAmelCase__ ( self: Dict ): return {} def lowerCAmelCase__ ( self: str , UpperCamelCase_: List[Any] , UpperCamelCase_: Union[str, Any] ): __lowerCamelCase = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(UpperCamelCase_ , UpperCamelCase_ ): for single_url in single_urls: download_callback(UpperCamelCase_ ) else: __lowerCamelCase = single_urls download_callback(UpperCamelCase_ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(UpperCamelCase_ , UpperCamelCase_ ): __lowerCamelCase = [os.path.join(UpperCamelCase_ , urllib.parse.quote_plus(Path(UpperCamelCase_ ).name ) ) for x in single_urls] else: __lowerCamelCase = single_urls __lowerCamelCase = os.path.join(UpperCamelCase_ , urllib.parse.quote_plus(Path(UpperCamelCase_ ).name ) ) __lowerCamelCase = value # make sure that values are unique if all(isinstance(UpperCamelCase_ , UpperCamelCase_ ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique __lowerCamelCase = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Optional[int] ): __lowerCamelCase = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one __lowerCamelCase = all(bool(re.findall("""[0-9]{3,}-of-[0-9]{3,}""" , UpperCamelCase_ ) ) for url in data_url ) __lowerCamelCase = all( url.startswith("""https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed""" ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): __lowerCamelCase = [data_url[0]] * len(UpperCamelCase_ ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(UpperCamelCase_ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus __lowerCamelCase = os.path.join(UpperCamelCase_ , urllib.parse.quote_plus(single_url.split("""/""" )[-1] ) ) dummy_data_list.append(UpperCamelCase_ ) return dummy_data_list def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: Optional[int] , UpperCamelCase_: Optional[Any] ): for download_callback in self.download_callbacks: download_callback(UpperCamelCase_ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus __lowerCamelCase = os.path.join(UpperCamelCase_ , urllib.parse.quote_plus(data_url.split("""/""" )[-1] ) ) if os.path.exists(UpperCamelCase_ ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def lowerCAmelCase__ ( self: Optional[Any] ): pass def lowerCAmelCase__ ( self: List[Any] ): pass def lowerCAmelCase__ ( self: Any , UpperCamelCase_: Dict ): def _iter_archive_members(UpperCamelCase_: Any ): # this preserves the order of the members inside the ZIP archive __lowerCamelCase = Path(self.dummy_file ).parent __lowerCamelCase = path.relative_to(UpperCamelCase_ ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: __lowerCamelCase = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(UpperCamelCase_ ) __lowerCamelCase = Path(UpperCamelCase_ ) __lowerCamelCase = _iter_archive_members(UpperCamelCase_ ) if self.use_local_dummy_data else path.rglob("""*""" ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith((""".""", """__""") ): yield file_path.relative_to(UpperCamelCase_ ).as_posix(), file_path.open("""rb""" ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: Dict ): if not isinstance(UpperCamelCase_ , UpperCamelCase_ ): __lowerCamelCase = [paths] for path in paths: if os.path.isfile(UpperCamelCase_ ): if os.path.basename(UpperCamelCase_ ).startswith((""".""", """__""") ): return yield path else: for dirpath, dirnames, filenames in os.walk(UpperCamelCase_ ): if os.path.basename(UpperCamelCase_ ).startswith((""".""", """__""") ): continue dirnames.sort() for filename in sorted(UpperCamelCase_ ): if filename.startswith((""".""", """__""") ): continue yield os.path.join(UpperCamelCase_ , UpperCamelCase_ )
12
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPImageProcessor, CLIPVisionConfig, CLIPVisionModel from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEImgaImgPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import floats_tensor, load_image, load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): '''simple docstring''' __UpperCamelCase : Dict = ShapEImgaImgPipeline __UpperCamelCase : int = ['''image'''] __UpperCamelCase : Any = ['''image'''] __UpperCamelCase : List[str] = [ '''num_images_per_prompt''', '''num_inference_steps''', '''generator''', '''latents''', '''guidance_scale''', '''frame_size''', '''output_type''', '''return_dict''', ] __UpperCamelCase : List[str] = False @property def __magic_name__ ( self : str ): """simple docstring""" return 3_2 @property def __magic_name__ ( self : Tuple ): """simple docstring""" return 3_2 @property def __magic_name__ ( self : Any ): """simple docstring""" return self.time_input_dim * 4 @property def __magic_name__ ( self : Union[str, Any] ): """simple docstring""" return 8 @property def __magic_name__ ( self : Optional[Any] ): """simple docstring""" torch.manual_seed(0 ) _A: Optional[int] = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=6_4 , projection_dim=self.text_embedder_hidden_size , intermediate_size=3_7 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=1 , ) _A: str = CLIPVisionModel(lowerCAmelCase_ ) return model @property def __magic_name__ ( self : List[Any] ): """simple docstring""" _A: Any = CLIPImageProcessor( crop_size=2_2_4 , do_center_crop=lowerCAmelCase_ , do_normalize=lowerCAmelCase_ , do_resize=lowerCAmelCase_ , image_mean=[0.48145466, 0.4578275, 0.40821073] , image_std=[0.26862954, 0.26130258, 0.27577711] , resample=3 , size=2_2_4 , ) return image_processor @property def __magic_name__ ( self : int ): """simple docstring""" torch.manual_seed(0 ) _A: Optional[Any] = { '''num_attention_heads''': 2, '''attention_head_dim''': 1_6, '''embedding_dim''': self.time_input_dim, '''num_embeddings''': 3_2, '''embedding_proj_dim''': self.text_embedder_hidden_size, '''time_embed_dim''': self.time_embed_dim, '''num_layers''': 1, '''clip_embed_dim''': self.time_input_dim * 2, '''additional_embeddings''': 0, '''time_embed_act_fn''': '''gelu''', '''norm_in_type''': '''layer''', '''embedding_proj_norm_type''': '''layer''', '''encoder_hid_proj_type''': None, '''added_emb_type''': None, } _A: Optional[Any] = PriorTransformer(**lowerCAmelCase_ ) return model @property def __magic_name__ ( self : Optional[int] ): """simple docstring""" torch.manual_seed(0 ) _A: Dict = { '''param_shapes''': ( (self.renderer_dim, 9_3), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), '''d_latent''': self.time_input_dim, '''d_hidden''': self.renderer_dim, '''n_output''': 1_2, '''background''': ( 0.1, 0.1, 0.1, ), } _A: Dict = ShapERenderer(**lowerCAmelCase_ ) return model def __magic_name__ ( self : List[Any] ): """simple docstring""" _A: Union[str, Any] = self.dummy_prior _A: Tuple = self.dummy_image_encoder _A: Tuple = self.dummy_image_processor _A: List[str] = self.dummy_renderer _A: Union[str, Any] = HeunDiscreteScheduler( beta_schedule='''exp''' , num_train_timesteps=1_0_2_4 , prediction_type='''sample''' , use_karras_sigmas=lowerCAmelCase_ , clip_sample=lowerCAmelCase_ , clip_sample_range=1.0 , ) _A: List[str] = { '''prior''': prior, '''image_encoder''': image_encoder, '''image_processor''': image_processor, '''renderer''': renderer, '''scheduler''': scheduler, } return components def __magic_name__ ( self : Any , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Tuple=0 ): """simple docstring""" _A: Any = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(lowerCAmelCase_ ) ).to(lowerCAmelCase_ ) if str(lowerCAmelCase_ ).startswith('''mps''' ): _A: Union[str, Any] = torch.manual_seed(lowerCAmelCase_ ) else: _A: Optional[int] = torch.Generator(device=lowerCAmelCase_ ).manual_seed(lowerCAmelCase_ ) _A: Optional[int] = { '''image''': input_image, '''generator''': generator, '''num_inference_steps''': 1, '''frame_size''': 3_2, '''output_type''': '''np''', } return inputs def __magic_name__ ( self : str ): """simple docstring""" _A: int = '''cpu''' _A: int = self.get_dummy_components() _A: int = self.pipeline_class(**lowerCAmelCase_ ) _A: Dict = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _A: List[str] = pipe(**self.get_dummy_inputs(lowerCAmelCase_ ) ) _A: Any = output.images[0] _A: str = image[0, -3:, -3:, -1] assert image.shape == (2_0, 3_2, 3_2, 3) _A: Dict = np.array( [ 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def __magic_name__ ( self : Optional[int] ): """simple docstring""" self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def __magic_name__ ( self : Optional[int] ): """simple docstring""" _A: str = torch_device == '''cpu''' _A: int = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=lowerCAmelCase_ , relax_max_difference=lowerCAmelCase_ , ) def __magic_name__ ( self : Union[str, Any] ): """simple docstring""" _A: Dict = self.get_dummy_components() _A: Union[str, Any] = self.pipeline_class(**lowerCAmelCase_ ) _A: Optional[Any] = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _A: Any = 1 _A: Dict = 2 _A: int = self.get_dummy_inputs(lowerCAmelCase_ ) for key in inputs.keys(): if key in self.batch_params: _A: Optional[int] = batch_size * [inputs[key]] _A: str = pipe(**lowerCAmelCase_ , num_images_per_prompt=lowerCAmelCase_ )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __magic_name__ ( self : Union[str, Any] ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __magic_name__ ( self : int ): """simple docstring""" _A: Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/shap_e/corgi.png''' ) _A: List[Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/shap_e/test_shap_e_img2img_out.npy''' ) _A: Dict = ShapEImgaImgPipeline.from_pretrained('''openai/shap-e-img2img''' ) _A: Optional[int] = pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _A: Dict = torch.Generator(device=lowerCAmelCase_ ).manual_seed(0 ) _A: Optional[Any] = pipe( lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=3.0 , num_inference_steps=6_4 , frame_size=6_4 , output_type='''np''' , ).images[0] assert images.shape == (2_0, 6_4, 6_4, 3) assert_mean_pixel_difference(lowerCAmelCase_ , lowerCAmelCase_ )
352
from __future__ import annotations UpperCAmelCase__ : List[str] = list[list[int]] # assigning initial values to the grid UpperCAmelCase__ : Matrix = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution UpperCAmelCase__ : Matrix = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def lowerCamelCase__ ( a , a , a , a ) -> bool: for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def lowerCamelCase__ ( a ) -> tuple[int, int] | None: for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def lowerCamelCase__ ( a ) -> Matrix | None: if location := find_empty_location(a ): _A , _A: Optional[Any] = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 10 ): if is_safe(a , a , a , a ): _A: str = digit if sudoku(a ) is not None: return grid _A: Tuple = 0 return None def lowerCamelCase__ ( a ) -> None: for row in grid: for cell in row: print(a , end=''' ''' ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print('\nExample grid:\n' + '=' * 20) print_solution(example_grid) print('\nExample grid solution:') UpperCAmelCase__ : int = sudoku(example_grid) if solution is not None: print_solution(solution) else: print('Cannot find a solution.')
301
0
'''simple docstring''' from __future__ import annotations from math import pi, sqrt def _A (lowerCAmelCase__ :float , lowerCAmelCase__ :float ) -> tuple: '''simple docstring''' if inductance <= 0: raise ValueError('Inductance cannot be 0 or negative' ) elif capacitance <= 0: raise ValueError('Capacitance cannot be 0 or negative' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
168
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) a_ : Union[str, Any] = { "configuration_layoutlmv3": [ "LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP", "LayoutLMv3Config", "LayoutLMv3OnnxConfig", ], "processing_layoutlmv3": ["LayoutLMv3Processor"], "tokenization_layoutlmv3": ["LayoutLMv3Tokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : Tuple = ["LayoutLMv3TokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : str = [ "LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST", "LayoutLMv3ForQuestionAnswering", "LayoutLMv3ForSequenceClassification", "LayoutLMv3ForTokenClassification", "LayoutLMv3Model", "LayoutLMv3PreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : Any = [ "TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST", "TFLayoutLMv3ForQuestionAnswering", "TFLayoutLMv3ForSequenceClassification", "TFLayoutLMv3ForTokenClassification", "TFLayoutLMv3Model", "TFLayoutLMv3PreTrainedModel", ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : Any = ["LayoutLMv3FeatureExtractor"] a_ : List[str] = ["LayoutLMv3ImageProcessor"] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys a_ : Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
168
1
"""simple docstring""" import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import MaskaFormerConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskaFormerForUniversalSegmentation, MaskaFormerModel if is_vision_available(): from transformers import MaskaFormerImageProcessor if is_vision_available(): from PIL import Image class __a : '''simple docstring''' def __init__( self , _a , _a=2 , _a=True , _a=False , _a=10 , _a=3 , _a=32 * 8 , _a=32 * 8 , _a=4 , _a=64 , ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = parent SCREAMING_SNAKE_CASE__ : str = batch_size SCREAMING_SNAKE_CASE__ : str = is_training SCREAMING_SNAKE_CASE__ : int = use_auxiliary_loss SCREAMING_SNAKE_CASE__ : Union[str, Any] = num_queries SCREAMING_SNAKE_CASE__ : Any = num_channels SCREAMING_SNAKE_CASE__ : Optional[Any] = min_size SCREAMING_SNAKE_CASE__ : Tuple = max_size SCREAMING_SNAKE_CASE__ : Optional[Any] = num_labels SCREAMING_SNAKE_CASE__ : Optional[int] = hidden_dim SCREAMING_SNAKE_CASE__ : List[str] = hidden_dim def _a ( self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[str] = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( _a ) SCREAMING_SNAKE_CASE__ : Dict = torch.ones([self.batch_size, self.min_size, self.max_size] , device=_a ) SCREAMING_SNAKE_CASE__ : Tuple = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=_a ) > 0.5 ).float() SCREAMING_SNAKE_CASE__ : str = (torch.rand((self.batch_size, self.num_labels) , device=_a ) > 0.5).long() SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def _a ( self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = MaskaFormerConfig( hidden_size=self.hidden_dim , ) SCREAMING_SNAKE_CASE__ : Dict = self.num_queries SCREAMING_SNAKE_CASE__ : Tuple = self.num_labels SCREAMING_SNAKE_CASE__ : Dict = [1, 1, 1, 1] SCREAMING_SNAKE_CASE__ : Optional[int] = self.num_channels SCREAMING_SNAKE_CASE__ : List[str] = 64 SCREAMING_SNAKE_CASE__ : Tuple = 128 SCREAMING_SNAKE_CASE__ : List[str] = self.hidden_dim SCREAMING_SNAKE_CASE__ : List[str] = self.hidden_dim SCREAMING_SNAKE_CASE__ : int = self.hidden_dim return config def _a ( self ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE__ : Union[str, Any] = {"""pixel_values""": pixel_values, """pixel_mask""": pixel_mask} return config, inputs_dict def _a ( self , _a , _a ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = output.encoder_hidden_states SCREAMING_SNAKE_CASE__ : Tuple = output.pixel_decoder_hidden_states SCREAMING_SNAKE_CASE__ : Union[str, Any] = output.transformer_decoder_hidden_states self.parent.assertTrue(len(_a ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(_a ) , config.decoder_layers ) def _a ( self , _a , _a , _a , _a=False ) -> List[Any]: """simple docstring""" with torch.no_grad(): SCREAMING_SNAKE_CASE__ : Optional[Any] = MaskaFormerModel(config=_a ) model.to(_a ) model.eval() SCREAMING_SNAKE_CASE__ : Optional[Any] = model(pixel_values=_a , pixel_mask=_a ) SCREAMING_SNAKE_CASE__ : Any = model(_a , output_hidden_states=_a ) self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.hidden_dim) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(_a , _a ) def _a ( self , _a , _a , _a , _a , _a ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = MaskaFormerForUniversalSegmentation(config=_a ) model.to(_a ) model.eval() def comm_check_on_output(_a ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ : Optional[int] = model(pixel_values=_a , pixel_mask=_a ) SCREAMING_SNAKE_CASE__ : List[Any] = model(_a ) comm_check_on_output(_a ) SCREAMING_SNAKE_CASE__ : Optional[Any] = model( pixel_values=_a , pixel_mask=_a , mask_labels=_a , class_labels=_a ) comm_check_on_output(_a ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class __a (UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase): '''simple docstring''' _SCREAMING_SNAKE_CASE :Optional[int] = (MaskaFormerModel, MaskaFormerForUniversalSegmentation) if is_torch_available() else () _SCREAMING_SNAKE_CASE :Tuple = {"""feature-extraction""": MaskaFormerModel} if is_torch_available() else {} _SCREAMING_SNAKE_CASE :Tuple = False _SCREAMING_SNAKE_CASE :Dict = False _SCREAMING_SNAKE_CASE :Dict = False _SCREAMING_SNAKE_CASE :List[str] = False def _a ( self ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = MaskaFormerModelTester(self ) SCREAMING_SNAKE_CASE__ : Optional[int] = ConfigTester(self , config_class=_a , has_text_modality=_a ) def _a ( self ) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() def _a ( self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a , **_a , output_hidden_states=_a ) def _a ( self ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskaformer_instance_segmentation_head_model(*_a ) @unittest.skip(reason="""Mask2Former does not use inputs_embeds""" ) def _a ( self ) -> int: """simple docstring""" pass @unittest.skip(reason="""Mask2Former does not have a get_input_embeddings method""" ) def _a ( self ) -> Any: """simple docstring""" pass @unittest.skip(reason="""Mask2Former is not a generative model""" ) def _a ( self ) -> Tuple: """simple docstring""" pass @unittest.skip(reason="""Mask2Former does not use token embeddings""" ) def _a ( self ) -> List[str]: """simple docstring""" pass @require_torch_multi_gpu @unittest.skip( reason="""Mask2Former has some layers using `add_module` which doesn't work well with `nn.DataParallel`""" ) def _a ( self ) -> Optional[int]: """simple docstring""" pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def _a ( self ) -> List[str]: """simple docstring""" pass def _a ( self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ : str = model_class(_a ) SCREAMING_SNAKE_CASE__ : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE__ : str = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE__ : Optional[int] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _a ) @slow def _a ( self ) -> Union[str, Any]: """simple docstring""" for model_name in ["facebook/mask2former-swin-small-coco-instance"]: SCREAMING_SNAKE_CASE__ : Any = MaskaFormerModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def _a ( self ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = (self.model_tester.min_size,) * 2 SCREAMING_SNAKE_CASE__ : Tuple = { """pixel_values""": torch.randn((2, 3, *size) , device=_a ), """mask_labels""": torch.randn((2, 10, *size) , device=_a ), """class_labels""": torch.zeros(2 , 10 , device=_a ).long(), } SCREAMING_SNAKE_CASE__ : List[str] = self.model_tester.get_config() SCREAMING_SNAKE_CASE__ : Dict = MaskaFormerForUniversalSegmentation(_a ).to(_a ) SCREAMING_SNAKE_CASE__ : Optional[int] = model(**_a ) self.assertTrue(outputs.loss is not None ) def _a ( self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskaformer_model(_a , **_a , output_hidden_states=_a ) def _a ( self ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ : Union[str, Any] = model_class(_a ).to(_a ) SCREAMING_SNAKE_CASE__ : int = model(**_a , output_attentions=_a ) self.assertTrue(outputs.attentions is not None ) def _a ( self ) -> Optional[Any]: """simple docstring""" if not self.model_tester.is_training: return SCREAMING_SNAKE_CASE__ : Dict = self.all_model_classes[1] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Any = self.model_tester.prepare_config_and_inputs() SCREAMING_SNAKE_CASE__ : int = model_class(_a ) model.to(_a ) model.train() SCREAMING_SNAKE_CASE__ : str = model(_a , mask_labels=_a , class_labels=_a ).loss loss.backward() def _a ( self ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = self.all_model_classes[1] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[Any] = self.model_tester.prepare_config_and_inputs() SCREAMING_SNAKE_CASE__ : Tuple = True SCREAMING_SNAKE_CASE__ : Union[str, Any] = True SCREAMING_SNAKE_CASE__ : Optional[int] = model_class(_a ).to(_a ) model.train() SCREAMING_SNAKE_CASE__ : Tuple = model(_a , mask_labels=_a , class_labels=_a ) SCREAMING_SNAKE_CASE__ : Any = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() SCREAMING_SNAKE_CASE__ : int = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() SCREAMING_SNAKE_CASE__ : Tuple = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() SCREAMING_SNAKE_CASE__ : Optional[Any] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=_a ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) a :str = 1e-4 def _lowercase ( ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ : Dict = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_vision @slow class __a (unittest.TestCase): '''simple docstring''' @cached_property def _a ( self ) -> Tuple: """simple docstring""" return "facebook/mask2former-swin-small-coco-instance" @cached_property def _a ( self ) -> Optional[Any]: """simple docstring""" return MaskaFormerImageProcessor.from_pretrained(self.model_checkpoints ) if is_vision_available() else None def _a ( self ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = MaskaFormerModel.from_pretrained(self.model_checkpoints ).to(_a ) SCREAMING_SNAKE_CASE__ : Optional[Any] = self.default_image_processor SCREAMING_SNAKE_CASE__ : Union[str, Any] = prepare_img() SCREAMING_SNAKE_CASE__ : List[Any] = image_processor(_a , return_tensors="""pt""" ).to(_a ) SCREAMING_SNAKE_CASE__ : int = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a , (1, 3, 384, 384) ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ : str = model(**_a ) SCREAMING_SNAKE_CASE__ : int = torch.tensor( [[-0.2_790, -1.0_717, -1.1_668], [-0.5_128, -0.3_128, -0.4_987], [-0.5_832, 0.1_971, -0.0_197]] ).to(_a ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , _a , atol=_a ) ) SCREAMING_SNAKE_CASE__ : int = torch.tensor( [[0.8_973, 1.1_847, 1.1_776], [1.1_934, 1.5_040, 1.5_128], [1.1_153, 1.4_486, 1.4_951]] ).to(_a ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , _a , atol=_a ) ) SCREAMING_SNAKE_CASE__ : int = torch.tensor( [[2.1_152, 1.7_000, -0.8_603], [1.5_808, 1.8_004, -0.9_353], [1.6_043, 1.7_495, -0.5_999]] ).to(_a ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , _a , atol=_a ) ) def _a ( self ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() SCREAMING_SNAKE_CASE__ : List[Any] = self.default_image_processor SCREAMING_SNAKE_CASE__ : Any = prepare_img() SCREAMING_SNAKE_CASE__ : Any = image_processor(_a , return_tensors="""pt""" ).to(_a ) SCREAMING_SNAKE_CASE__ : List[Any] = inputs["""pixel_values"""].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(_a , (1, 3, 384, 384) ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ : Optional[Any] = model(**_a ) # masks_queries_logits SCREAMING_SNAKE_CASE__ : Dict = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) ) SCREAMING_SNAKE_CASE__ : int = [ [-8.7_839, -9.0_056, -8.8_121], [-7.4_104, -7.0_313, -6.5_401], [-6.6_105, -6.3_427, -6.4_675], ] SCREAMING_SNAKE_CASE__ : Tuple = torch.tensor(_a ).to(_a ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , _a , atol=_a ) ) # class_queries_logits SCREAMING_SNAKE_CASE__ : int = outputs.class_queries_logits self.assertEqual(class_queries_logits.shape , (1, model.config.num_queries, model.config.num_labels + 1) ) SCREAMING_SNAKE_CASE__ : List[Any] = torch.tensor( [ [1.8_324, -8.0_835, -4.1_922], [0.8_450, -9.0_050, -3.6_053], [0.3_045, -7.7_293, -3.0_275], ] ).to(_a ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , _a , atol=_a ) ) def _a ( self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = MaskaFormerForUniversalSegmentation.from_pretrained(self.model_checkpoints ).to(_a ).eval() SCREAMING_SNAKE_CASE__ : str = self.default_image_processor SCREAMING_SNAKE_CASE__ : int = image_processor( [np.zeros((3, 800, 1_333) ), np.zeros((3, 800, 1_333) )] , segmentation_maps=[np.zeros((384, 384) ).astype(np.floataa ), np.zeros((384, 384) ).astype(np.floataa )] , return_tensors="""pt""" , ) SCREAMING_SNAKE_CASE__ : Optional[int] = inputs["""pixel_values"""].to(_a ) SCREAMING_SNAKE_CASE__ : int = [el.to(_a ) for el in inputs["""mask_labels"""]] SCREAMING_SNAKE_CASE__ : Union[str, Any] = [el.to(_a ) for el in inputs["""class_labels"""]] with torch.no_grad(): SCREAMING_SNAKE_CASE__ : Optional[Any] = model(**_a ) self.assertTrue(outputs.loss is not None )
56
"""simple docstring""" def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> int: return number | (1 << position) def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> int: return number & ~(1 << position) def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> int: return number ^ (1 << position) def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> bool: return ((number >> position) & 1) == 1 def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> int: return int((number & (1 << position)) != 0 ) if __name__ == "__main__": import doctest doctest.testmod()
56
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowercase = logging.get_logger(__name__) __lowercase = { """roberta-base""": """https://huggingface.co/roberta-base/resolve/main/config.json""", """roberta-large""": """https://huggingface.co/roberta-large/resolve/main/config.json""", """roberta-large-mnli""": """https://huggingface.co/roberta-large-mnli/resolve/main/config.json""", """distilroberta-base""": """https://huggingface.co/distilroberta-base/resolve/main/config.json""", """roberta-base-openai-detector""": """https://huggingface.co/roberta-base-openai-detector/resolve/main/config.json""", """roberta-large-openai-detector""": """https://huggingface.co/roberta-large-openai-detector/resolve/main/config.json""", } class _A ( _a ): """simple docstring""" UpperCAmelCase : str = """roberta""" def __init__( self : Tuple , __UpperCAmelCase : Tuple=50265 , __UpperCAmelCase : Dict=768 , __UpperCAmelCase : Optional[int]=12 , __UpperCAmelCase : Optional[int]=12 , __UpperCAmelCase : Any=3072 , __UpperCAmelCase : Optional[int]="gelu" , __UpperCAmelCase : Optional[Any]=0.1 , __UpperCAmelCase : Union[str, Any]=0.1 , __UpperCAmelCase : Dict=512 , __UpperCAmelCase : Optional[int]=2 , __UpperCAmelCase : List[Any]=0.02 , __UpperCAmelCase : List[str]=1e-12 , __UpperCAmelCase : List[Any]=1 , __UpperCAmelCase : str=0 , __UpperCAmelCase : Optional[int]=2 , __UpperCAmelCase : Optional[int]="absolute" , __UpperCAmelCase : List[str]=True , __UpperCAmelCase : Dict=None , **__UpperCAmelCase : str , ): super().__init__(pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase) a : int = vocab_size a : Optional[Any] = hidden_size a : Any = num_hidden_layers a : Dict = num_attention_heads a : Dict = hidden_act a : int = intermediate_size a : str = hidden_dropout_prob a : List[Any] = attention_probs_dropout_prob a : Any = max_position_embeddings a : str = type_vocab_size a : Optional[int] = initializer_range a : Optional[Any] = layer_norm_eps a : Optional[Any] = position_embedding_type a : Any = use_cache a : List[str] = classifier_dropout class _A ( _a ): """simple docstring""" @property def __snake_case ( self : int): if self.task == "multiple-choice": a : Optional[Any] = {0: "batch", 1: "choice", 2: "sequence"} else: a : Optional[Any] = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ])
40
from __future__ import annotations import unittest import numpy as np from transformers import LayoutLMConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.layoutlm.modeling_tf_layoutlm import ( TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMForMaskedLM, TFLayoutLMForQuestionAnswering, TFLayoutLMForSequenceClassification, TFLayoutLMForTokenClassification, TFLayoutLMModel, ) class _snake_case : def __init__( self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=7 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=99 , _lowerCamelCase=32 , _lowerCamelCase=2 , _lowerCamelCase=4 , _lowerCamelCase=37 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=512 , _lowerCamelCase=16 , _lowerCamelCase=2 , _lowerCamelCase=0.02 , _lowerCamelCase=3 , _lowerCamelCase=4 , _lowerCamelCase=None , _lowerCamelCase=1000 , ): a :str = parent a :str = batch_size a :List[Any] = seq_length a :Union[str, Any] = is_training a :str = use_input_mask a :Tuple = use_token_type_ids a :Optional[int] = use_labels a :Union[str, Any] = vocab_size a :Optional[Any] = hidden_size a :Any = num_hidden_layers a :Optional[int] = num_attention_heads a :Tuple = intermediate_size a :Dict = hidden_act a :str = hidden_dropout_prob a :List[Any] = attention_probs_dropout_prob a :List[Any] = max_position_embeddings a :List[str] = type_vocab_size a :List[Any] = type_sequence_label_size a :Union[str, Any] = initializer_range a :Optional[Any] = num_labels a :Optional[int] = num_choices a :Union[str, Any] = scope a :List[str] = range_bbox def SCREAMING_SNAKE_CASE__ ( self ): a :str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # convert bbox to numpy since TF does not support item assignment a :Union[str, Any] = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ).numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: a :List[Any] = bbox[i, j, 3] a :List[str] = bbox[i, j, 1] a :List[str] = t if bbox[i, j, 2] < bbox[i, j, 0]: a :Dict = bbox[i, j, 2] a :Dict = bbox[i, j, 0] a :Any = t a :Optional[Any] = tf.convert_to_tensor(_lowerCamelCase ) a :int = None if self.use_input_mask: a :List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) a :Optional[int] = None if self.use_token_type_ids: a :Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) a :List[Any] = None a :List[Any] = None a :List[Any] = None if self.use_labels: a :Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a :Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a :List[str] = ids_tensor([self.batch_size] , self.num_choices ) a :List[Any] = LayoutLMConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): a :Optional[int] = TFLayoutLMModel(config=_lowerCamelCase ) a :Dict = model(_lowerCamelCase , _lowerCamelCase , attention_mask=_lowerCamelCase , token_type_ids=_lowerCamelCase ) a :Union[str, Any] = model(_lowerCamelCase , _lowerCamelCase , token_type_ids=_lowerCamelCase ) a :Union[str, Any] = model(_lowerCamelCase , _lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): a :List[str] = TFLayoutLMForMaskedLM(config=_lowerCamelCase ) a :int = model(_lowerCamelCase , _lowerCamelCase , attention_mask=_lowerCamelCase , token_type_ids=_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): a :Optional[int] = self.num_labels a :List[Any] = TFLayoutLMForSequenceClassification(config=_lowerCamelCase ) a :int = model(_lowerCamelCase , _lowerCamelCase , attention_mask=_lowerCamelCase , token_type_ids=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): a :int = self.num_labels a :Optional[int] = TFLayoutLMForTokenClassification(config=_lowerCamelCase ) a :int = model(_lowerCamelCase , _lowerCamelCase , attention_mask=_lowerCamelCase , token_type_ids=_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): a :Optional[Any] = TFLayoutLMForQuestionAnswering(config=_lowerCamelCase ) a :Optional[int] = model(_lowerCamelCase , _lowerCamelCase , attention_mask=_lowerCamelCase , token_type_ids=_lowerCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE__ ( self ): a :List[str] = self.prepare_config_and_inputs() ( ( a ) , ( a ) , ( a ) , ( a ) , ( a ) , ( a ) , ( a ) , ( a ) , ) :List[Any] = config_and_inputs a :Union[str, Any] = { '''input_ids''': input_ids, '''bbox''': bbox, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask, } return config, inputs_dict @require_tf class _snake_case ( _snake_case , _snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE__ = ( ( TFLayoutLMModel, TFLayoutLMForMaskedLM, TFLayoutLMForTokenClassification, TFLayoutLMForSequenceClassification, TFLayoutLMForQuestionAnswering, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE__ = ( { 'feature-extraction': TFLayoutLMModel, 'fill-mask': TFLayoutLMForMaskedLM, 'text-classification': TFLayoutLMForSequenceClassification, 'token-classification': TFLayoutLMForTokenClassification, 'zero-shot': TFLayoutLMForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = 10 def SCREAMING_SNAKE_CASE__ ( self ): a :Dict = TFLayoutLMModelTester(self ) a :Dict = ConfigTester(self , config_class=_lowerCamelCase , hidden_size=37 ) def SCREAMING_SNAKE_CASE__ ( self ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self ): a :str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): a :List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): a :List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): a :int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ): a :Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_lowerCamelCase ) @slow def SCREAMING_SNAKE_CASE__ ( self ): for model_name in TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a :str = TFLayoutLMModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) @unittest.skip('''Onnx compliancy broke with TF 2.10''' ) def SCREAMING_SNAKE_CASE__ ( self ): pass def __lowerCamelCase ( ): """simple docstring""" a :Tuple = tf.convert_to_tensor([[101,1019,1014,1016,1037,1_2849,4747,1004,1_4246,2278,5439,4524,5002,2930,2193,2930,4341,3208,1005,1055,2171,2848,1_1300,3531,102],[101,4070,4034,7020,1024,3058,1015,1013,2861,1013,6070,1_9274,2772,6205,2_7814,1_6147,1_6147,4343,2047,1_0283,1_0969,1_4389,1012,2338,102]] ) # noqa: E231 a :Any = tf.convert_to_tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],] ) # noqa: E231 a :List[str] = tf.convert_to_tensor([[[0,0,0,0],[423,237,440,251],[427,272,441,287],[419,115,437,129],[961,885,992,912],[256,38,330,58],[256,38,330,58],[336,42,353,57],[360,39,401,56],[360,39,401,56],[411,39,471,59],[479,41,528,59],[533,39,630,60],[67,113,134,131],[141,115,209,132],[68,149,133,166],[141,149,187,164],[195,148,287,165],[195,148,287,165],[195,148,287,165],[295,148,349,165],[441,149,492,166],[497,149,546,164],[64,201,125,218],[1000,1000,1000,1000]],[[0,0,0,0],[662,150,754,166],[665,199,742,211],[519,213,554,228],[519,213,554,228],[134,433,187,454],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[314,469,376,482],[504,684,582,706],[941,825,973,900],[941,825,973,900],[941,825,973,900],[941,825,973,900],[610,749,652,765],[130,659,168,672],[176,657,237,672],[238,657,312,672],[443,653,628,672],[443,653,628,672],[716,301,825,317],[1000,1000,1000,1000]]] ) # noqa: E231 a :List[str] = tf.convert_to_tensor([[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]] ) # noqa: E231 # these are sequence labels (i.e. at the token level) a :Any = tf.convert_to_tensor([[-100,10,10,10,9,1,-100,7,7,-100,7,7,4,2,5,2,8,8,-100,-100,5,0,3,2,-100],[-100,12,12,12,-100,12,10,-100,-100,-100,-100,10,12,9,-100,-100,-100,10,10,10,9,12,-100,10,-100]] ) # noqa: E231 # fmt: on return input_ids, attention_mask, bbox, token_type_ids, labels @require_tf class _snake_case ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE__ ( self ): a :List[Any] = TFLayoutLMModel.from_pretrained('''microsoft/layoutlm-base-uncased''' ) a , a , a , a , a :Optional[Any] = prepare_layoutlm_batch_inputs() # forward pass a :Tuple = model(input_ids=_lowerCamelCase , bbox=_lowerCamelCase , attention_mask=_lowerCamelCase , token_type_ids=_lowerCamelCase ) # test the sequence output on [0, :3, :3] a :List[str] = tf.convert_to_tensor( [[0.1785, -0.1947, -0.0425], [-0.3254, -0.2807, 0.2553], [-0.5391, -0.3322, 0.3364]] , ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , _lowerCamelCase , atol=1e-3 ) ) # test the pooled output on [1, :3] a :List[str] = tf.convert_to_tensor([-0.6580, -0.0214, 0.8552] ) self.assertTrue(np.allclose(outputs.pooler_output[1, :3] , _lowerCamelCase , atol=1e-3 ) ) @slow def SCREAMING_SNAKE_CASE__ ( self ): # initialize model with randomly initialized sequence classification head a :str = TFLayoutLMForSequenceClassification.from_pretrained('''microsoft/layoutlm-base-uncased''' , num_labels=2 ) a , a , a , a , a :List[str] = prepare_layoutlm_batch_inputs() # forward pass a :List[Any] = model( input_ids=_lowerCamelCase , bbox=_lowerCamelCase , attention_mask=_lowerCamelCase , token_type_ids=_lowerCamelCase , labels=tf.convert_to_tensor([1, 1] ) , ) # test whether we get a loss as a scalar a :Union[str, Any] = outputs.loss a :Optional[Any] = (2,) self.assertEqual(loss.shape , _lowerCamelCase ) # test the shape of the logits a :Any = outputs.logits a :Tuple = (2, 2) self.assertEqual(logits.shape , _lowerCamelCase ) @slow def SCREAMING_SNAKE_CASE__ ( self ): # initialize model with randomly initialized token classification head a :Dict = TFLayoutLMForTokenClassification.from_pretrained('''microsoft/layoutlm-base-uncased''' , num_labels=13 ) a , a , a , a , a :Dict = prepare_layoutlm_batch_inputs() # forward pass a :List[Any] = model( input_ids=_lowerCamelCase , bbox=_lowerCamelCase , attention_mask=_lowerCamelCase , token_type_ids=_lowerCamelCase , labels=_lowerCamelCase ) # test the shape of the logits a :Optional[Any] = outputs.logits a :List[Any] = tf.convert_to_tensor((2, 25, 13) ) self.assertEqual(logits.shape , _lowerCamelCase ) @slow def SCREAMING_SNAKE_CASE__ ( self ): # initialize model with randomly initialized token classification head a :List[Any] = TFLayoutLMForQuestionAnswering.from_pretrained('''microsoft/layoutlm-base-uncased''' ) a , a , a , a , a :Any = prepare_layoutlm_batch_inputs() # forward pass a :str = model(input_ids=_lowerCamelCase , bbox=_lowerCamelCase , attention_mask=_lowerCamelCase , token_type_ids=_lowerCamelCase ) # test the shape of the logits a :Optional[int] = tf.convert_to_tensor((2, 25) ) self.assertEqual(outputs.start_logits.shape , _lowerCamelCase ) self.assertEqual(outputs.end_logits.shape , _lowerCamelCase )
94
0
'''simple docstring''' import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging _UpperCamelCase = logging.get_logger(__name__) class _A ( __SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Optional[Any] = ["input_features"] def __init__( self , __UpperCAmelCase=80 , __UpperCAmelCase=16_000 , __UpperCAmelCase=160 , __UpperCAmelCase=30 , __UpperCAmelCase=400 , __UpperCAmelCase=0.0 , __UpperCAmelCase=False , **__UpperCAmelCase , ) -> List[str]: '''simple docstring''' super().__init__( feature_size=__UpperCAmelCase , sampling_rate=__UpperCAmelCase , padding_value=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , **__UpperCAmelCase , ) __UpperCAmelCase : int = n_fft __UpperCAmelCase : int = hop_length __UpperCAmelCase : Union[str, Any] = chunk_length __UpperCAmelCase : Optional[int] = chunk_length * sampling_rate __UpperCAmelCase : Tuple = self.n_samples // hop_length __UpperCAmelCase : List[Any] = sampling_rate __UpperCAmelCase : Optional[int] = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=__UpperCAmelCase , min_frequency=0.0 , max_frequency=8000.0 , sampling_rate=__UpperCAmelCase , norm="""slaney""" , mel_scale="""slaney""" , ) def __A ( self , __UpperCAmelCase ) -> np.ndarray: '''simple docstring''' __UpperCAmelCase : Optional[int] = spectrogram( __UpperCAmelCase , window_function(self.n_fft , """hann""" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel="""log10""" , ) __UpperCAmelCase : Any = log_spec[:, :-1] __UpperCAmelCase : Dict = np.maximum(__UpperCAmelCase , log_spec.max() - 8.0 ) __UpperCAmelCase : Union[str, Any] = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def __A ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = 0.0 ) -> List[np.ndarray]: '''simple docstring''' if attention_mask is not None: __UpperCAmelCase : List[str] = np.array(__UpperCAmelCase , np.intaa ) __UpperCAmelCase : int = [] for vector, length in zip(__UpperCAmelCase , attention_mask.sum(-1 ) ): __UpperCAmelCase : List[Any] = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: __UpperCAmelCase : int = padding_value normed_input_values.append(__UpperCAmelCase ) else: __UpperCAmelCase : List[Any] = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def __call__( self , __UpperCAmelCase , __UpperCAmelCase = True , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = "max_length" , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , **__UpperCAmelCase , ) -> BatchFeature: '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a' f' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input' f' was sampled with {self.sampling_rate} and not {sampling_rate}.' ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) __UpperCAmelCase : Dict = isinstance(__UpperCAmelCase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'Only mono-channel audio is supported for input to {self}' ) __UpperCAmelCase : Optional[int] = is_batched_numpy or ( isinstance(__UpperCAmelCase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __UpperCAmelCase : Union[str, Any] = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(__UpperCAmelCase , np.ndarray ): __UpperCAmelCase : Tuple = np.asarray(__UpperCAmelCase , dtype=np.floataa ) elif isinstance(__UpperCAmelCase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __UpperCAmelCase : Any = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __UpperCAmelCase : Dict = [np.asarray([raw_speech] ).T] __UpperCAmelCase : List[Any] = BatchFeature({"""input_features""": raw_speech} ) # convert into correct format for padding __UpperCAmelCase : Optional[int] = self.pad( __UpperCAmelCase , padding=__UpperCAmelCase , max_length=max_length if max_length else self.n_samples , truncation=__UpperCAmelCase , pad_to_multiple_of=__UpperCAmelCase , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: __UpperCAmelCase : List[Any] = self.zero_mean_unit_var_norm( padded_inputs["""input_features"""] , attention_mask=padded_inputs["""attention_mask"""] , padding_value=self.padding_value , ) __UpperCAmelCase : Any = np.stack(padded_inputs["""input_features"""] , axis=0 ) # make sure list is in array format __UpperCAmelCase : str = padded_inputs.get("""input_features""" ).transpose(2 , 0 , 1 ) __UpperCAmelCase : Optional[int] = [self._np_extract_fbank_features(__UpperCAmelCase ) for waveform in input_features[0]] if isinstance(input_features[0] , __UpperCAmelCase ): __UpperCAmelCase : Union[str, Any] = [np.asarray(__UpperCAmelCase , dtype=np.floataa ) for feature in input_features] else: __UpperCAmelCase : Optional[int] = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) __UpperCAmelCase : Dict = padded_inputs["""attention_mask"""][:, :: self.hop_length] if return_tensors is not None: __UpperCAmelCase : Any = padded_inputs.convert_to_tensors(__UpperCAmelCase ) return padded_inputs def __A ( self ) -> Dict[str, Any]: '''simple docstring''' __UpperCAmelCase : Any = copy.deepcopy(self.__dict__ ) __UpperCAmelCase : Any = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
16
'''simple docstring''' from ..utils import DummyObject, requires_backends class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : List[str] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[int]: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : str = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Dict = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> List[str]: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Optional[int] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Dict: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : List[Any] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> str: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Optional[Any] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> str: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Tuple = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Tuple = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Tuple: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Any = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Dict: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : str = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> str: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Optional[int] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Tuple: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Union[str, Any] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : List[Any] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Dict: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Optional[Any] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Optional[int] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> List[str]: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Any = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Dict: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Tuple = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Tuple: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : str = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Dict = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Optional[Any] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[int]: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Union[str, Any] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Tuple: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Union[str, Any] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> int: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Dict = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> int: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : List[str] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> int: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Union[str, Any] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Union[str, Any] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : List[Any] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Any: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Optional[int] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Dict: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Any = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : List[Any] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Any: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] ) class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Optional[Any] = ["sentencepiece"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> List[str]: '''simple docstring''' requires_backends(self , ["""sentencepiece"""] )
16
1
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class lowerCamelCase_ (snake_case__ ): '''simple docstring''' __UpperCamelCase: torch.FloatTensor class lowerCamelCase_ (snake_case__ , snake_case__ ): '''simple docstring''' @register_to_config def __init__( self : Union[str, Any] , A : int = 65536 , A : Optional[int] = None , A : int = 2 , A : int = 2 , A : int = 0 , A : str = "fourier" , A : bool = True , A : bool = False , A : float = 0.0 , A : Tuple[str] = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , A : Tuple[str] = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , A : Tuple[str] = "UNetMidBlock1D" , A : str = None , A : Tuple[int] = (32, 32, 64) , A : str = None , A : int = 8 , A : int = 1 , A : bool = False , ): super().__init__() _UpperCAmelCase : List[str] = sample_size # time if time_embedding_type == "fourier": _UpperCAmelCase : List[str] = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=A , log=A , flip_sin_to_cos=A ) _UpperCAmelCase : Union[str, Any] = 2 * block_out_channels[0] elif time_embedding_type == "positional": _UpperCAmelCase : Dict = Timesteps( block_out_channels[0] , flip_sin_to_cos=A , downscale_freq_shift=A ) _UpperCAmelCase : str = block_out_channels[0] if use_timestep_embedding: _UpperCAmelCase : List[str] = block_out_channels[0] * 4 _UpperCAmelCase : List[str] = TimestepEmbedding( in_channels=A , time_embed_dim=A , act_fn=A , out_dim=block_out_channels[0] , ) _UpperCAmelCase : Optional[int] = nn.ModuleList([] ) _UpperCAmelCase : Tuple = None _UpperCAmelCase : Optional[Any] = nn.ModuleList([] ) _UpperCAmelCase : str = None # down _UpperCAmelCase : Optional[int] = in_channels for i, down_block_type in enumerate(A ): _UpperCAmelCase : Union[str, Any] = output_channel _UpperCAmelCase : Tuple = block_out_channels[i] if i == 0: input_channel += extra_in_channels _UpperCAmelCase : str = i == len(A ) - 1 _UpperCAmelCase : Union[str, Any] = get_down_block( A , num_layers=A , in_channels=A , out_channels=A , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(A ) # mid _UpperCAmelCase : Optional[int] = get_mid_block( A , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=A , add_downsample=A , ) # up _UpperCAmelCase : int = list(reversed(A ) ) _UpperCAmelCase : Union[str, Any] = reversed_block_out_channels[0] if out_block_type is None: _UpperCAmelCase : Optional[Any] = out_channels else: _UpperCAmelCase : Any = block_out_channels[0] for i, up_block_type in enumerate(A ): _UpperCAmelCase : Dict = output_channel _UpperCAmelCase : int = ( reversed_block_out_channels[i + 1] if i < len(A ) - 1 else final_upsample_channels ) _UpperCAmelCase : Optional[Any] = i == len(A ) - 1 _UpperCAmelCase : Optional[Any] = get_up_block( A , num_layers=A , in_channels=A , out_channels=A , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(A ) _UpperCAmelCase : Tuple = output_channel # out _UpperCAmelCase : Optional[int] = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) _UpperCAmelCase : List[str] = get_out_block( out_block_type=A , num_groups_out=A , embed_dim=block_out_channels[0] , out_channels=A , act_fn=A , fc_dim=block_out_channels[-1] // 4 , ) def _A ( self : Dict , A : torch.FloatTensor , A : Union[torch.Tensor, float, int] , A : bool = True , ): _UpperCAmelCase : List[Any] = timestep if not torch.is_tensor(A ): _UpperCAmelCase : Optional[int] = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(A ) and len(timesteps.shape ) == 0: _UpperCAmelCase : Tuple = timesteps[None].to(sample.device ) _UpperCAmelCase : Dict = self.time_proj(A ) if self.config.use_timestep_embedding: _UpperCAmelCase : Union[str, Any] = self.time_mlp(A ) else: _UpperCAmelCase : str = timestep_embed[..., None] _UpperCAmelCase : Tuple = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) _UpperCAmelCase : List[Any] = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down _UpperCAmelCase : Dict = () for downsample_block in self.down_blocks: _UpperCAmelCase , _UpperCAmelCase : List[str] = downsample_block(hidden_states=A , temb=A ) down_block_res_samples += res_samples # 3. mid if self.mid_block: _UpperCAmelCase : Optional[int] = self.mid_block(A , A ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): _UpperCAmelCase : int = down_block_res_samples[-1:] _UpperCAmelCase : Optional[int] = down_block_res_samples[:-1] _UpperCAmelCase : Optional[Any] = upsample_block(A , res_hidden_states_tuple=A , temb=A ) # 5. post-process if self.out_block: _UpperCAmelCase : Optional[Any] = self.out_block(A , A ) if not return_dict: return (sample,) return UNetaDOutput(sample=A )
31
'''simple docstring''' from pathlib import Path from typing import List from transformers import is_torch_available, is_vision_available from transformers.testing_utils import get_tests_dir, is_tool_test from transformers.tools.agent_types import AGENT_TYPE_MAPPING, AgentAudio, AgentImage, AgentText if is_torch_available(): import torch if is_vision_available(): from PIL import Image _lowercase : List[str] = ["text", "image", "audio"] def snake_case_ ( __SCREAMING_SNAKE_CASE : List[str] ): """simple docstring""" lowercase_ : int = [] for input_type in input_types: if input_type == "text": inputs.append('''Text input''' ) elif input_type == "image": inputs.append( Image.open(Path(get_tests_dir('''fixtures/tests_samples/COCO''' ) ) / '''000000039769.png''' ).resize((512, 512) ) ) elif input_type == "audio": inputs.append(torch.ones(3000 ) ) elif isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): inputs.append(create_inputs(__SCREAMING_SNAKE_CASE ) ) else: raise ValueError(F'''Invalid type requested: {input_type}''' ) return inputs def snake_case_ ( __SCREAMING_SNAKE_CASE : List ): """simple docstring""" lowercase_ : Optional[Any] = [] for output in outputs: if isinstance(__SCREAMING_SNAKE_CASE , (str, AgentText) ): output_types.append('''text''' ) elif isinstance(__SCREAMING_SNAKE_CASE , (Image.Image, AgentImage) ): output_types.append('''image''' ) elif isinstance(__SCREAMING_SNAKE_CASE , (torch.Tensor, AgentAudio) ): output_types.append('''audio''' ) else: raise ValueError(F'''Invalid output: {output}''' ) return output_types @is_tool_test class lowerCAmelCase__ : def _snake_case ( self ): """simple docstring""" self.assertTrue(hasattr(self.tool , '''inputs''' ) ) self.assertTrue(hasattr(self.tool , '''outputs''' ) ) lowercase_ : Optional[Any] = self.tool.inputs for _input in inputs: if isinstance(_input , __SCREAMING_SNAKE_CASE ): for __input in _input: self.assertTrue(__input in authorized_types ) else: self.assertTrue(_input in authorized_types ) lowercase_ : int = self.tool.outputs for _output in outputs: self.assertTrue(_output in authorized_types ) def _snake_case ( self ): """simple docstring""" lowercase_ : int = create_inputs(self.tool.inputs ) lowercase_ : Tuple = self.tool(*__SCREAMING_SNAKE_CASE ) # There is a single output if len(self.tool.outputs ) == 1: lowercase_ : Any = [outputs] self.assertListEqual(output_types(__SCREAMING_SNAKE_CASE ) , self.tool.outputs ) def _snake_case ( self ): """simple docstring""" self.assertTrue(hasattr(self.tool , '''description''' ) ) self.assertTrue(hasattr(self.tool , '''default_checkpoint''' ) ) self.assertTrue(self.tool.description.startswith('''This is a tool that''' ) ) def _snake_case ( self ): """simple docstring""" lowercase_ : int = create_inputs(self.tool.inputs ) lowercase_ : int = self.tool(*__SCREAMING_SNAKE_CASE ) if not isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase_ : Optional[Any] = [outputs] self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , len(self.tool.outputs ) ) for output, output_type in zip(__SCREAMING_SNAKE_CASE , self.tool.outputs ): lowercase_ : Optional[int] = AGENT_TYPE_MAPPING[output_type] self.assertTrue(isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) def _snake_case ( self ): """simple docstring""" lowercase_ : Dict = create_inputs(self.tool.inputs ) lowercase_ : int = [] for _input, input_type in zip(__SCREAMING_SNAKE_CASE , self.tool.inputs ): if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): _inputs.append([AGENT_TYPE_MAPPING[_input_type](_input ) for _input_type in input_type] ) else: _inputs.append(AGENT_TYPE_MAPPING[input_type](_input ) ) # Should not raise an error lowercase_ : Optional[Any] = self.tool(*__SCREAMING_SNAKE_CASE ) if not isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase_ : Dict = [outputs] self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , len(self.tool.outputs ) )
93
0
"""simple docstring""" import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class _lowercase ( pl.LightningModule ): """simple docstring""" def __init__( self : Union[str, Any] , UpperCamelCase__ : List[Any] ) -> Dict: '''simple docstring''' super().__init__() __UpperCamelCase =model __UpperCamelCase =2 __UpperCamelCase =nn.Linear(self.model.config.hidden_size , self.num_labels ) def UpperCAmelCase_ ( self : Optional[int] ) -> Dict: '''simple docstring''' pass def lowerCAmelCase (__UpperCamelCase : str , __UpperCamelCase : str , __UpperCamelCase : str ): """simple docstring""" __UpperCamelCase =LongformerModel.from_pretrained(__UpperCamelCase ) __UpperCamelCase =LightningModel(__UpperCamelCase ) __UpperCamelCase =torch.load(__UpperCamelCase , map_location=torch.device('''cpu''' ) ) lightning_model.load_state_dict(ckpt['''state_dict'''] ) # init longformer question answering model __UpperCamelCase =LongformerForQuestionAnswering.from_pretrained(__UpperCamelCase ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(__UpperCamelCase ) print(F"""Conversion successful. Model saved under {pytorch_dump_folder_path}""" ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--longformer_model''', default=None, type=str, required=True, help='''model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.''', ) parser.add_argument( '''--longformer_question_answering_ckpt_path''', default=None, type=str, required=True, help='''Path the official PyTorch Lightning Checkpoint.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __lowercase = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
364
"""simple docstring""" from dataclasses import dataclass from typing import Dict, Optional, Union import torch import torch.nn.functional as F from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .attention_processor import AttentionProcessor, AttnProcessor from .embeddings import TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin @dataclass class _lowercase ( __a ): """simple docstring""" lowercase__ = 42 class _lowercase ( __a , __a ): """simple docstring""" @register_to_config def __init__( self : List[Any] , UpperCamelCase__ : int = 32 , UpperCamelCase__ : int = 64 , UpperCamelCase__ : int = 20 , UpperCamelCase__ : int = 768 , UpperCamelCase__ : str=77 , UpperCamelCase__ : Dict=4 , UpperCamelCase__ : float = 0.0 , UpperCamelCase__ : str = "silu" , UpperCamelCase__ : Optional[str] = None , UpperCamelCase__ : Optional[str] = None , UpperCamelCase__ : Optional[str] = "linear" , UpperCamelCase__ : Optional[str] = "prd" , UpperCamelCase__ : Optional[int] = None , UpperCamelCase__ : Optional[int] = None , UpperCamelCase__ : Optional[int] = None , ) -> Any: '''simple docstring''' super().__init__() __UpperCamelCase =num_attention_heads __UpperCamelCase =attention_head_dim __UpperCamelCase =num_attention_heads * attention_head_dim __UpperCamelCase =additional_embeddings __UpperCamelCase =time_embed_dim or inner_dim __UpperCamelCase =embedding_proj_dim or embedding_dim __UpperCamelCase =clip_embed_dim or embedding_dim __UpperCamelCase =Timesteps(UpperCamelCase__ , UpperCamelCase__ , 0 ) __UpperCamelCase =TimestepEmbedding(UpperCamelCase__ , UpperCamelCase__ , out_dim=UpperCamelCase__ , act_fn=UpperCamelCase__ ) __UpperCamelCase =nn.Linear(UpperCamelCase__ , UpperCamelCase__ ) if embedding_proj_norm_type is None: __UpperCamelCase =None elif embedding_proj_norm_type == "layer": __UpperCamelCase =nn.LayerNorm(UpperCamelCase__ ) else: raise ValueError(f"""unsupported embedding_proj_norm_type: {embedding_proj_norm_type}""" ) __UpperCamelCase =nn.Linear(UpperCamelCase__ , UpperCamelCase__ ) if encoder_hid_proj_type is None: __UpperCamelCase =None elif encoder_hid_proj_type == "linear": __UpperCamelCase =nn.Linear(UpperCamelCase__ , UpperCamelCase__ ) else: raise ValueError(f"""unsupported encoder_hid_proj_type: {encoder_hid_proj_type}""" ) __UpperCamelCase =nn.Parameter(torch.zeros(1 , num_embeddings + additional_embeddings , UpperCamelCase__ ) ) if added_emb_type == "prd": __UpperCamelCase =nn.Parameter(torch.zeros(1 , 1 , UpperCamelCase__ ) ) elif added_emb_type is None: __UpperCamelCase =None else: raise ValueError( f"""`added_emb_type`: {added_emb_type} is not supported. Make sure to choose one of `'prd'` or `None`.""" ) __UpperCamelCase =nn.ModuleList( [ BasicTransformerBlock( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , dropout=UpperCamelCase__ , activation_fn='''gelu''' , attention_bias=UpperCamelCase__ , ) for d in range(UpperCamelCase__ ) ] ) if norm_in_type == "layer": __UpperCamelCase =nn.LayerNorm(UpperCamelCase__ ) elif norm_in_type is None: __UpperCamelCase =None else: raise ValueError(f"""Unsupported norm_in_type: {norm_in_type}.""" ) __UpperCamelCase =nn.LayerNorm(UpperCamelCase__ ) __UpperCamelCase =nn.Linear(UpperCamelCase__ , UpperCamelCase__ ) __UpperCamelCase =torch.full( [num_embeddings + additional_embeddings, num_embeddings + additional_embeddings] , -1_00_00.0 ) causal_attention_mask.triu_(1 ) __UpperCamelCase =causal_attention_mask[None, ...] self.register_buffer('''causal_attention_mask''' , UpperCamelCase__ , persistent=UpperCamelCase__ ) __UpperCamelCase =nn.Parameter(torch.zeros(1 , UpperCamelCase__ ) ) __UpperCamelCase =nn.Parameter(torch.zeros(1 , UpperCamelCase__ ) ) @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def UpperCAmelCase_ ( self : Any ) -> Dict[str, AttentionProcessor]: '''simple docstring''' __UpperCamelCase ={} def fn_recursive_add_processors(UpperCamelCase__ : str , UpperCamelCase__ : torch.nn.Module , UpperCamelCase__ : Dict[str, AttentionProcessor] ): if hasattr(UpperCamelCase__ , '''set_processor''' ): __UpperCamelCase =module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(f"""{name}.{sub_name}""" , UpperCamelCase__ , UpperCamelCase__ ) return processors for name, module in self.named_children(): fn_recursive_add_processors(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) return processors def UpperCAmelCase_ ( self : int , UpperCamelCase__ : Union[AttentionProcessor, Dict[str, AttentionProcessor]] ) -> Optional[int]: '''simple docstring''' __UpperCamelCase =len(self.attn_processors.keys() ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) and len(UpperCamelCase__ ) != count: raise ValueError( f"""A dict of processors was passed, but the number of processors {len(UpperCamelCase__ )} does not match the""" f""" number of attention layers: {count}. Please make sure to pass {count} processor classes.""" ) def fn_recursive_attn_processor(UpperCamelCase__ : str , UpperCamelCase__ : torch.nn.Module , UpperCamelCase__ : int ): if hasattr(UpperCamelCase__ , '''set_processor''' ): if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): module.set_processor(UpperCamelCase__ ) else: module.set_processor(processor.pop(f"""{name}.processor""" ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(f"""{name}.{sub_name}""" , UpperCamelCase__ , UpperCamelCase__ ) for name, module in self.named_children(): fn_recursive_attn_processor(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def UpperCAmelCase_ ( self : Dict ) -> Tuple: '''simple docstring''' self.set_attn_processor(AttnProcessor() ) def UpperCAmelCase_ ( self : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Union[torch.Tensor, float, int] , UpperCamelCase__ : torch.FloatTensor , UpperCamelCase__ : Optional[torch.FloatTensor] = None , UpperCamelCase__ : Optional[torch.BoolTensor] = None , UpperCamelCase__ : bool = True , ) -> Tuple: '''simple docstring''' __UpperCamelCase =hidden_states.shape[0] __UpperCamelCase =timestep if not torch.is_tensor(UpperCamelCase__ ): __UpperCamelCase =torch.tensor([timesteps] , dtype=torch.long , device=hidden_states.device ) elif torch.is_tensor(UpperCamelCase__ ) and len(timesteps.shape ) == 0: __UpperCamelCase =timesteps[None].to(hidden_states.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML __UpperCamelCase =timesteps * torch.ones(UpperCamelCase__ , dtype=timesteps.dtype , device=timesteps.device ) __UpperCamelCase =self.time_proj(UpperCamelCase__ ) # timesteps does not contain any weights and will always return f32 tensors # but time_embedding might be fp16, so we need to cast here. __UpperCamelCase =timesteps_projected.to(dtype=self.dtype ) __UpperCamelCase =self.time_embedding(UpperCamelCase__ ) if self.embedding_proj_norm is not None: __UpperCamelCase =self.embedding_proj_norm(UpperCamelCase__ ) __UpperCamelCase =self.embedding_proj(UpperCamelCase__ ) if self.encoder_hidden_states_proj is not None and encoder_hidden_states is not None: __UpperCamelCase =self.encoder_hidden_states_proj(UpperCamelCase__ ) elif self.encoder_hidden_states_proj is not None and encoder_hidden_states is None: raise ValueError('''`encoder_hidden_states_proj` requires `encoder_hidden_states` to be set''' ) __UpperCamelCase =self.proj_in(UpperCamelCase__ ) __UpperCamelCase =self.positional_embedding.to(hidden_states.dtype ) __UpperCamelCase =[] __UpperCamelCase =0 if encoder_hidden_states is not None: additional_embeds.append(UpperCamelCase__ ) additional_embeddings_len += encoder_hidden_states.shape[1] if len(proj_embeddings.shape ) == 2: __UpperCamelCase =proj_embeddings[:, None, :] if len(hidden_states.shape ) == 2: __UpperCamelCase =hidden_states[:, None, :] __UpperCamelCase =additional_embeds + [ proj_embeddings, time_embeddings[:, None, :], hidden_states, ] if self.prd_embedding is not None: __UpperCamelCase =self.prd_embedding.to(hidden_states.dtype ).expand(UpperCamelCase__ , -1 , -1 ) additional_embeds.append(UpperCamelCase__ ) __UpperCamelCase =torch.cat( UpperCamelCase__ , dim=1 , ) # Allow positional_embedding to not include the `addtional_embeddings` and instead pad it with zeros for these additional tokens __UpperCamelCase =additional_embeddings_len + proj_embeddings.shape[1] + 1 if positional_embeddings.shape[1] < hidden_states.shape[1]: __UpperCamelCase =F.pad( UpperCamelCase__ , ( 0, 0, additional_embeddings_len, self.prd_embedding.shape[1] if self.prd_embedding is not None else 0, ) , value=0.0 , ) __UpperCamelCase =hidden_states + positional_embeddings if attention_mask is not None: __UpperCamelCase =(1 - attention_mask.to(hidden_states.dtype )) * -1_00_00.0 __UpperCamelCase =F.pad(UpperCamelCase__ , (0, self.additional_embeddings) , value=0.0 ) __UpperCamelCase =(attention_mask[:, None, :] + self.causal_attention_mask).to(hidden_states.dtype ) __UpperCamelCase =attention_mask.repeat_interleave(self.config.num_attention_heads , dim=0 ) if self.norm_in is not None: __UpperCamelCase =self.norm_in(UpperCamelCase__ ) for block in self.transformer_blocks: __UpperCamelCase =block(UpperCamelCase__ , attention_mask=UpperCamelCase__ ) __UpperCamelCase =self.norm_out(UpperCamelCase__ ) if self.prd_embedding is not None: __UpperCamelCase =hidden_states[:, -1] else: __UpperCamelCase =hidden_states[:, additional_embeddings_len:] __UpperCamelCase =self.proj_to_clip_embeddings(UpperCamelCase__ ) if not return_dict: return (predicted_image_embedding,) return PriorTransformerOutput(predicted_image_embedding=UpperCamelCase__ ) def UpperCAmelCase_ ( self : Dict , UpperCamelCase__ : int ) -> List[Any]: '''simple docstring''' __UpperCamelCase =(prior_latents * self.clip_std) + self.clip_mean return prior_latents
85
0
"""simple docstring""" from __future__ import annotations from math import ceil, floor, sqrt def _SCREAMING_SNAKE_CASE ( _lowercase : int = 200_0000 ) ->int: '''simple docstring''' a : list[int] = [0] a : int for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target a : int = 0 # the area corresponding to the grid that gives the product closest to target a : int = 0 # an estimate of b, using the quadratic formula a : float # the largest integer less than b_estimate a : int # the largest integer less than b_estimate a : int # the triangle number corresponding to b_floor a : int # the triangle number corresponding to b_ceil a : int for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): a : Union[str, Any] = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 a : Union[str, Any] = floor(_lowercase ) a : str = ceil(_lowercase ) a : Tuple = triangle_numbers[b_floor] a : Tuple = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): a : str = triangle_b_first_guess * triangle_a a : Union[str, Any] = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): a : Dict = triangle_b_second_guess * triangle_a a : str = idx_a * b_ceil return area if __name__ == "__main__": print(F'''{solution() = }''')
105
"""simple docstring""" import json import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from transformers import OneFormerImageProcessor from transformers.models.oneformer.image_processing_oneformer import binary_mask_to_rle from transformers.models.oneformer.modeling_oneformer import OneFormerForUniversalSegmentationOutput if is_vision_available(): from PIL import Image def lowercase (_lowerCAmelCase , _lowerCAmelCase="shi-labs/oneformer_demo" ): with open(hf_hub_download(_lowerCAmelCase , _lowerCAmelCase , repo_type="""dataset""" ) , """r""" ) as f: __lowerCAmelCase = json.load(_lowerCAmelCase ) __lowerCAmelCase = {} __lowerCAmelCase = [] __lowerCAmelCase = [] for key, info in class_info.items(): __lowerCAmelCase = info["""name"""] class_names.append(info["""name"""] ) if info["isthing"]: thing_ids.append(int(_lowerCAmelCase ) ) __lowerCAmelCase = thing_ids __lowerCAmelCase = class_names return metadata class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_=7 , snake_case_=3 , snake_case_=30 , snake_case_=400 , snake_case_=None , snake_case_=True , snake_case_=True , snake_case_=[0.5, 0.5, 0.5] , snake_case_=[0.5, 0.5, 0.5] , snake_case_=10 , snake_case_=False , snake_case_=255 , snake_case_="shi-labs/oneformer_demo" , snake_case_="ade20k_panoptic.json" , snake_case_=10 , ) -> Union[str, Any]: __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = num_channels __lowerCAmelCase = min_resolution __lowerCAmelCase = max_resolution __lowerCAmelCase = do_resize __lowerCAmelCase = {"""shortest_edge""": 32, """longest_edge""": 1_333} if size is None else size __lowerCAmelCase = do_normalize __lowerCAmelCase = image_mean __lowerCAmelCase = image_std __lowerCAmelCase = class_info_file __lowerCAmelCase = prepare_metadata(snake_case_ , snake_case_ ) __lowerCAmelCase = num_text __lowerCAmelCase = repo_path # for the post_process_functions __lowerCAmelCase = 2 __lowerCAmelCase = 10 __lowerCAmelCase = 10 __lowerCAmelCase = 3 __lowerCAmelCase = 4 __lowerCAmelCase = num_labels __lowerCAmelCase = do_reduce_labels __lowerCAmelCase = ignore_index def A__ ( self ) -> Any: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "num_labels": self.num_labels, "do_reduce_labels": self.do_reduce_labels, "ignore_index": self.ignore_index, "class_info_file": self.class_info_file, "metadata": self.metadata, "num_text": self.num_text, } def A__ ( self , snake_case_ , snake_case_=False ) -> Dict: if not batched: __lowerCAmelCase = image_inputs[0] if isinstance(snake_case_ , Image.Image ): __lowerCAmelCase , __lowerCAmelCase = image.size else: __lowerCAmelCase , __lowerCAmelCase = image.shape[1], image.shape[2] if w < h: __lowerCAmelCase = int(self.size["""shortest_edge"""] * h / w ) __lowerCAmelCase = self.size["""shortest_edge"""] elif w > h: __lowerCAmelCase = self.size["""shortest_edge"""] __lowerCAmelCase = int(self.size["""shortest_edge"""] * w / h ) else: __lowerCAmelCase = self.size["""shortest_edge"""] __lowerCAmelCase = self.size["""shortest_edge"""] else: __lowerCAmelCase = [] for image in image_inputs: __lowerCAmelCase , __lowerCAmelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __lowerCAmelCase = max(snake_case_ , key=lambda snake_case_ : item[0] )[0] __lowerCAmelCase = max(snake_case_ , key=lambda snake_case_ : item[1] )[1] return expected_height, expected_width def A__ ( self ) -> Tuple: return OneFormerForUniversalSegmentationOutput( # +1 for null class class_queries_logits=torch.randn((self.batch_size, self.num_queries, self.num_classes + 1) ) , masks_queries_logits=torch.randn((self.batch_size, self.num_queries, self.height, self.width) ) , ) @require_torch @require_vision class lowerCAmelCase_ ( A__ , unittest.TestCase ): '''simple docstring''' _snake_case = OneFormerImageProcessor if (is_vision_available() and is_torch_available()) else None # only for test_image_processing_common.test_image_proc_to_json_string _snake_case = image_processing_class def A__ ( self ) -> str: __lowerCAmelCase = OneFormerImageProcessorTester(self ) @property def A__ ( self ) -> Dict: return self.image_processing_tester.prepare_image_processor_dict() def A__ ( self ) -> Union[str, Any]: __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case_ , """image_mean""" ) ) self.assertTrue(hasattr(snake_case_ , """image_std""" ) ) self.assertTrue(hasattr(snake_case_ , """do_normalize""" ) ) self.assertTrue(hasattr(snake_case_ , """do_resize""" ) ) self.assertTrue(hasattr(snake_case_ , """size""" ) ) self.assertTrue(hasattr(snake_case_ , """ignore_index""" ) ) self.assertTrue(hasattr(snake_case_ , """class_info_file""" ) ) self.assertTrue(hasattr(snake_case_ , """num_text""" ) ) self.assertTrue(hasattr(snake_case_ , """repo_path""" ) ) self.assertTrue(hasattr(snake_case_ , """metadata""" ) ) self.assertTrue(hasattr(snake_case_ , """do_reduce_labels""" ) ) def A__ ( self ) -> List[str]: pass def A__ ( self ) -> Union[str, Any]: # Initialize image_processor __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowerCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , Image.Image ) # Test not batched input __lowerCAmelCase = image_processor(image_inputs[0] , ["""semantic"""] , return_tensors="""pt""" ).pixel_values __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ , batched=snake_case_ ) __lowerCAmelCase = image_processor( snake_case_ , ["""semantic"""] * len(snake_case_ ) , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A__ ( self ) -> List[str]: # Initialize image_processor __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowerCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=snake_case_ , numpify=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , np.ndarray ) # Test not batched input __lowerCAmelCase = image_processor(image_inputs[0] , ["""semantic"""] , return_tensors="""pt""" ).pixel_values __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ , batched=snake_case_ ) __lowerCAmelCase = image_processor( snake_case_ , ["""semantic"""] * len(snake_case_ ) , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A__ ( self ) -> Tuple: # Initialize image_processor __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowerCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=snake_case_ , torchify=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , torch.Tensor ) # Test not batched input __lowerCAmelCase = image_processor(image_inputs[0] , ["""semantic"""] , return_tensors="""pt""" ).pixel_values __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ ) self.assertEqual( encoded_images.shape , (1, self.image_processing_tester.num_channels, expected_height, expected_width) , ) # Test batched __lowerCAmelCase , __lowerCAmelCase = self.image_processing_tester.get_expected_values(snake_case_ , batched=snake_case_ ) __lowerCAmelCase = image_processor( snake_case_ , ["""semantic"""] * len(snake_case_ ) , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processing_tester.batch_size, self.image_processing_tester.num_channels, expected_height, expected_width, ) , ) def A__ ( self , snake_case_=False , snake_case_=False , snake_case_="np" ) -> Optional[Any]: __lowerCAmelCase = self.image_processing_class(**self.image_processor_dict ) # prepare image and target __lowerCAmelCase = self.image_processing_tester.num_labels __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = prepare_image_inputs(self.image_processing_tester , equal_resolution=snake_case_ ) if with_segmentation_maps: __lowerCAmelCase = num_labels if is_instance_map: __lowerCAmelCase = list(range(snake_case_ ) ) * 2 __lowerCAmelCase = dict(enumerate(snake_case_ ) ) __lowerCAmelCase = [ np.random.randint(0 , high * 2 , (img.size[1], img.size[0]) ).astype(np.uinta ) for img in image_inputs ] if segmentation_type == "pil": __lowerCAmelCase = [Image.fromarray(snake_case_ ) for annotation in annotations] __lowerCAmelCase = image_processor( snake_case_ , ["""semantic"""] * len(snake_case_ ) , snake_case_ , return_tensors="""pt""" , instance_id_to_semantic_id=snake_case_ , pad_and_return_pixel_mask=snake_case_ , ) return inputs def A__ ( self ) -> List[str]: pass def A__ ( self ) -> Optional[Any]: def common(snake_case_=False , snake_case_=None ): __lowerCAmelCase = self.comm_get_image_processor_inputs( with_segmentation_maps=snake_case_ , is_instance_map=snake_case_ , segmentation_type=snake_case_ ) __lowerCAmelCase = inputs["""mask_labels"""] __lowerCAmelCase = inputs["""class_labels"""] __lowerCAmelCase = inputs["""pixel_values"""] __lowerCAmelCase = inputs["""text_inputs"""] # check the batch_size for mask_label, class_label, text_input in zip(snake_case_ , snake_case_ , snake_case_ ): self.assertEqual(mask_label.shape[0] , class_label.shape[0] ) # this ensure padding has happened self.assertEqual(mask_label.shape[1:] , pixel_values.shape[2:] ) self.assertEqual(len(snake_case_ ) , self.image_processing_tester.num_text ) common() common(is_instance_map=snake_case_ ) common(is_instance_map=snake_case_ , segmentation_type="""pil""" ) common(is_instance_map=snake_case_ , segmentation_type="""pil""" ) def A__ ( self ) -> Optional[int]: __lowerCAmelCase = np.zeros((20, 50) ) __lowerCAmelCase = 1 __lowerCAmelCase = 1 __lowerCAmelCase = 1 __lowerCAmelCase = binary_mask_to_rle(snake_case_ ) self.assertEqual(len(snake_case_ ) , 4 ) self.assertEqual(rle[0] , 21 ) self.assertEqual(rle[1] , 45 ) def A__ ( self ) -> Optional[Any]: __lowerCAmelCase = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file="""ade20k_panoptic.json""" , num_text=self.image_processing_tester.num_text , repo_path="""shi-labs/oneformer_demo""" , ) __lowerCAmelCase = self.image_processing_tester.get_fake_oneformer_outputs() __lowerCAmelCase = fature_extractor.post_process_semantic_segmentation(snake_case_ ) self.assertEqual(len(snake_case_ ) , self.image_processing_tester.batch_size ) self.assertEqual( segmentation[0].shape , ( self.image_processing_tester.height, self.image_processing_tester.width, ) , ) __lowerCAmelCase = [(1, 4) for i in range(self.image_processing_tester.batch_size )] __lowerCAmelCase = fature_extractor.post_process_semantic_segmentation(snake_case_ , target_sizes=snake_case_ ) self.assertEqual(segmentation[0].shape , target_sizes[0] ) def A__ ( self ) -> Union[str, Any]: __lowerCAmelCase = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file="""ade20k_panoptic.json""" , num_text=self.image_processing_tester.num_text , repo_path="""shi-labs/oneformer_demo""" , ) __lowerCAmelCase = self.image_processing_tester.get_fake_oneformer_outputs() __lowerCAmelCase = image_processor.post_process_instance_segmentation(snake_case_ , threshold=0 ) self.assertTrue(len(snake_case_ ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue("""segmentation""" in el ) self.assertTrue("""segments_info""" in el ) self.assertEqual(type(el["""segments_info"""] ) , snake_case_ ) self.assertEqual( el["""segmentation"""].shape , (self.image_processing_tester.height, self.image_processing_tester.width) ) def A__ ( self ) -> Union[str, Any]: __lowerCAmelCase = self.image_processing_class( num_labels=self.image_processing_tester.num_classes , max_seq_length=77 , task_seq_length=77 , class_info_file="""ade20k_panoptic.json""" , num_text=self.image_processing_tester.num_text , repo_path="""shi-labs/oneformer_demo""" , ) __lowerCAmelCase = self.image_processing_tester.get_fake_oneformer_outputs() __lowerCAmelCase = image_processor.post_process_panoptic_segmentation(snake_case_ , threshold=0 ) self.assertTrue(len(snake_case_ ) == self.image_processing_tester.batch_size ) for el in segmentation: self.assertTrue("""segmentation""" in el ) self.assertTrue("""segments_info""" in el ) self.assertEqual(type(el["""segments_info"""] ) , snake_case_ ) self.assertEqual( el["""segmentation"""].shape , (self.image_processing_tester.height, self.image_processing_tester.width) )
301
0
import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = {"""vocab_file""": """vocab.txt"""} _SCREAMING_SNAKE_CASE = { """vocab_file""": { """openbmb/cpm-ant-10b""": """https://huggingface.co/openbmb/cpm-ant-10b/blob/main/vocab.txt""", }, } _SCREAMING_SNAKE_CASE = { """openbmb/cpm-ant-10b""": 1_0_2_4, } def lowercase( UpperCamelCase_ ) -> Tuple: '''simple docstring''' UpperCamelCase = collections.OrderedDict() with open(UpperCamelCase_ , """r""" , encoding="""utf-8""" ) as reader: UpperCamelCase = reader.readlines() for index, token in enumerate(UpperCamelCase_ ): UpperCamelCase = token.rstrip("""\n""" ) UpperCamelCase = index return vocab class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): def __init__( self : Dict , lowerCamelCase_ : str , lowerCamelCase_ : Union[str, Any]="<unk>" , lowerCamelCase_ : Any=200 ): """simple docstring""" UpperCamelCase = vocab UpperCamelCase = unk_token UpperCamelCase = max_input_chars_per_word def lowerCamelCase_ ( self : Optional[Any] , lowerCamelCase_ : Union[str, Any] ): """simple docstring""" UpperCamelCase = list(lowerCamelCase_ ) if len(lowerCamelCase_ ) > self.max_input_chars_per_word: return [self.unk_token] UpperCamelCase = 0 UpperCamelCase = [] while start < len(lowerCamelCase_ ): UpperCamelCase = len(lowerCamelCase_ ) UpperCamelCase = None while start < end: UpperCamelCase = """""".join(chars[start:end] ) if substr in self.vocab: UpperCamelCase = substr break end -= 1 if cur_substr is None: sub_tokens.append(self.unk_token ) start += 1 else: sub_tokens.append(lowerCamelCase_ ) UpperCamelCase = end return sub_tokens class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): __lowerCAmelCase = VOCAB_FILES_NAMES __lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCAmelCase = ["""input_ids""", """attention_mask"""] __lowerCAmelCase = False def __init__( self : Tuple , lowerCamelCase_ : Any , lowerCamelCase_ : Optional[Any]="<d>" , lowerCamelCase_ : List[Any]="</d>" , lowerCamelCase_ : Optional[Any]="<s>" , lowerCamelCase_ : List[str]="</s>" , lowerCamelCase_ : int="<pad>" , lowerCamelCase_ : List[Any]="<unk>" , lowerCamelCase_ : Optional[Any]="</n>" , lowerCamelCase_ : Tuple="</_>" , lowerCamelCase_ : Any="left" , **lowerCamelCase_ : str , ): """simple docstring""" requires_backends(self , ["""jieba"""] ) super().__init__( bod_token=lowerCamelCase_ , eod_token=lowerCamelCase_ , bos_token=lowerCamelCase_ , eos_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , unk_token=lowerCamelCase_ , line_token=lowerCamelCase_ , space_token=lowerCamelCase_ , padding_side=lowerCamelCase_ , **lowerCamelCase_ , ) UpperCamelCase = bod_token UpperCamelCase = eod_token UpperCamelCase = load_vocab(lowerCamelCase_ ) UpperCamelCase = self.encoder[space_token] UpperCamelCase = self.encoder[line_token] del self.encoder[space_token] del self.encoder[line_token] UpperCamelCase = collections.OrderedDict(sorted(self.encoder.items() , key=lambda lowerCamelCase_ : x[1] ) ) UpperCamelCase = {v: k for k, v in self.encoder.items()} UpperCamelCase = WordpieceTokenizer(vocab=self.encoder , unk_token=self.unk_token ) @property def lowerCamelCase_ ( self : List[str] ): """simple docstring""" return self.encoder[self.bod_token] @property def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" return self.encoder[self.eod_token] @property def lowerCamelCase_ ( self : List[str] ): """simple docstring""" return self.encoder["\n"] @property def lowerCamelCase_ ( self : Dict ): """simple docstring""" return len(self.encoder ) def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def lowerCamelCase_ ( self : int , lowerCamelCase_ : Any ): """simple docstring""" UpperCamelCase = [] for x in jieba.cut(lowerCamelCase_ , cut_all=lowerCamelCase_ ): output_tokens.extend(self.wordpiece_tokenizer.tokenize(lowerCamelCase_ ) ) return output_tokens def lowerCamelCase_ ( self : str , lowerCamelCase_ : Tuple , **lowerCamelCase_ : Tuple ): """simple docstring""" UpperCamelCase = [i for i in token_ids if i >= 0] UpperCamelCase = [ x for x in token_ids if x != self.pad_token_id and x != self.eos_token_id and x != self.bos_token_id ] return super()._decode(lowerCamelCase_ , **lowerCamelCase_ ) def lowerCamelCase_ ( self : int , lowerCamelCase_ : int ): """simple docstring""" return token in self.encoder def lowerCamelCase_ ( self : List[Any] , lowerCamelCase_ : List[str] ): """simple docstring""" return "".join(lowerCamelCase_ ) def lowerCamelCase_ ( self : Optional[Any] , lowerCamelCase_ : List[Any] ): """simple docstring""" return self.encoder.get(lowerCamelCase_ , self.encoder.get(self.unk_token ) ) def lowerCamelCase_ ( self : List[str] , lowerCamelCase_ : Tuple ): """simple docstring""" return self.decoder.get(lowerCamelCase_ , self.unk_token ) def lowerCamelCase_ ( self : Tuple , lowerCamelCase_ : str , lowerCamelCase_ : Optional[str] = None ): """simple docstring""" if os.path.isdir(lowerCamelCase_ ): UpperCamelCase = os.path.join( lowerCamelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) else: UpperCamelCase = (filename_prefix + """-""" if filename_prefix else """""") + save_directory UpperCamelCase = 0 if " " in self.encoder: UpperCamelCase = self.encoder[""" """] del self.encoder[" "] if "\n" in self.encoder: UpperCamelCase = self.encoder["""\n"""] del self.encoder["\n"] UpperCamelCase = collections.OrderedDict(sorted(self.encoder.items() , key=lambda lowerCamelCase_ : x[1] ) ) with open(lowerCamelCase_ , """w""" , encoding="""utf-8""" ) as writer: for token, token_index in self.encoder.items(): if index != token_index: logger.warning( f"""Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.""" """ Please check that the vocabulary is not corrupted!""" ) UpperCamelCase = token_index writer.write(token + """\n""" ) index += 1 return (vocab_file,) def lowerCamelCase_ ( self : Optional[int] , lowerCamelCase_ : List[int] , lowerCamelCase_ : List[int] = None ): """simple docstring""" if token_ids_a is None: return [self.bos_token_id] + token_ids_a return [self.bos_token_id] + token_ids_a + [self.bos_token_id] + token_ids_a def lowerCamelCase_ ( self : str , lowerCamelCase_ : List[int] , lowerCamelCase_ : Optional[List[int]] = None , lowerCamelCase_ : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ , token_ids_a=lowerCamelCase_ , already_has_special_tokens=lowerCamelCase_ ) if token_ids_a is not None: return [1] + ([0] * len(lowerCamelCase_ )) + [1] + ([0] * len(lowerCamelCase_ )) return [1] + ([0] * len(lowerCamelCase_ ))
352
import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = {"""vocab_file""": """vocab.txt"""} _SCREAMING_SNAKE_CASE = { """vocab_file""": { """openbmb/cpm-ant-10b""": """https://huggingface.co/openbmb/cpm-ant-10b/blob/main/vocab.txt""", }, } _SCREAMING_SNAKE_CASE = { """openbmb/cpm-ant-10b""": 1_0_2_4, } def lowercase( UpperCamelCase_ ) -> Tuple: '''simple docstring''' UpperCamelCase = collections.OrderedDict() with open(UpperCamelCase_ , """r""" , encoding="""utf-8""" ) as reader: UpperCamelCase = reader.readlines() for index, token in enumerate(UpperCamelCase_ ): UpperCamelCase = token.rstrip("""\n""" ) UpperCamelCase = index return vocab class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): def __init__( self : Dict , lowerCamelCase_ : str , lowerCamelCase_ : Union[str, Any]="<unk>" , lowerCamelCase_ : Any=200 ): """simple docstring""" UpperCamelCase = vocab UpperCamelCase = unk_token UpperCamelCase = max_input_chars_per_word def lowerCamelCase_ ( self : Optional[Any] , lowerCamelCase_ : Union[str, Any] ): """simple docstring""" UpperCamelCase = list(lowerCamelCase_ ) if len(lowerCamelCase_ ) > self.max_input_chars_per_word: return [self.unk_token] UpperCamelCase = 0 UpperCamelCase = [] while start < len(lowerCamelCase_ ): UpperCamelCase = len(lowerCamelCase_ ) UpperCamelCase = None while start < end: UpperCamelCase = """""".join(chars[start:end] ) if substr in self.vocab: UpperCamelCase = substr break end -= 1 if cur_substr is None: sub_tokens.append(self.unk_token ) start += 1 else: sub_tokens.append(lowerCamelCase_ ) UpperCamelCase = end return sub_tokens class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): __lowerCAmelCase = VOCAB_FILES_NAMES __lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCAmelCase = ["""input_ids""", """attention_mask"""] __lowerCAmelCase = False def __init__( self : Tuple , lowerCamelCase_ : Any , lowerCamelCase_ : Optional[Any]="<d>" , lowerCamelCase_ : List[Any]="</d>" , lowerCamelCase_ : Optional[Any]="<s>" , lowerCamelCase_ : List[str]="</s>" , lowerCamelCase_ : int="<pad>" , lowerCamelCase_ : List[Any]="<unk>" , lowerCamelCase_ : Optional[Any]="</n>" , lowerCamelCase_ : Tuple="</_>" , lowerCamelCase_ : Any="left" , **lowerCamelCase_ : str , ): """simple docstring""" requires_backends(self , ["""jieba"""] ) super().__init__( bod_token=lowerCamelCase_ , eod_token=lowerCamelCase_ , bos_token=lowerCamelCase_ , eos_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , unk_token=lowerCamelCase_ , line_token=lowerCamelCase_ , space_token=lowerCamelCase_ , padding_side=lowerCamelCase_ , **lowerCamelCase_ , ) UpperCamelCase = bod_token UpperCamelCase = eod_token UpperCamelCase = load_vocab(lowerCamelCase_ ) UpperCamelCase = self.encoder[space_token] UpperCamelCase = self.encoder[line_token] del self.encoder[space_token] del self.encoder[line_token] UpperCamelCase = collections.OrderedDict(sorted(self.encoder.items() , key=lambda lowerCamelCase_ : x[1] ) ) UpperCamelCase = {v: k for k, v in self.encoder.items()} UpperCamelCase = WordpieceTokenizer(vocab=self.encoder , unk_token=self.unk_token ) @property def lowerCamelCase_ ( self : List[str] ): """simple docstring""" return self.encoder[self.bod_token] @property def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" return self.encoder[self.eod_token] @property def lowerCamelCase_ ( self : List[str] ): """simple docstring""" return self.encoder["\n"] @property def lowerCamelCase_ ( self : Dict ): """simple docstring""" return len(self.encoder ) def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def lowerCamelCase_ ( self : int , lowerCamelCase_ : Any ): """simple docstring""" UpperCamelCase = [] for x in jieba.cut(lowerCamelCase_ , cut_all=lowerCamelCase_ ): output_tokens.extend(self.wordpiece_tokenizer.tokenize(lowerCamelCase_ ) ) return output_tokens def lowerCamelCase_ ( self : str , lowerCamelCase_ : Tuple , **lowerCamelCase_ : Tuple ): """simple docstring""" UpperCamelCase = [i for i in token_ids if i >= 0] UpperCamelCase = [ x for x in token_ids if x != self.pad_token_id and x != self.eos_token_id and x != self.bos_token_id ] return super()._decode(lowerCamelCase_ , **lowerCamelCase_ ) def lowerCamelCase_ ( self : int , lowerCamelCase_ : int ): """simple docstring""" return token in self.encoder def lowerCamelCase_ ( self : List[Any] , lowerCamelCase_ : List[str] ): """simple docstring""" return "".join(lowerCamelCase_ ) def lowerCamelCase_ ( self : Optional[Any] , lowerCamelCase_ : List[Any] ): """simple docstring""" return self.encoder.get(lowerCamelCase_ , self.encoder.get(self.unk_token ) ) def lowerCamelCase_ ( self : List[str] , lowerCamelCase_ : Tuple ): """simple docstring""" return self.decoder.get(lowerCamelCase_ , self.unk_token ) def lowerCamelCase_ ( self : Tuple , lowerCamelCase_ : str , lowerCamelCase_ : Optional[str] = None ): """simple docstring""" if os.path.isdir(lowerCamelCase_ ): UpperCamelCase = os.path.join( lowerCamelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) else: UpperCamelCase = (filename_prefix + """-""" if filename_prefix else """""") + save_directory UpperCamelCase = 0 if " " in self.encoder: UpperCamelCase = self.encoder[""" """] del self.encoder[" "] if "\n" in self.encoder: UpperCamelCase = self.encoder["""\n"""] del self.encoder["\n"] UpperCamelCase = collections.OrderedDict(sorted(self.encoder.items() , key=lambda lowerCamelCase_ : x[1] ) ) with open(lowerCamelCase_ , """w""" , encoding="""utf-8""" ) as writer: for token, token_index in self.encoder.items(): if index != token_index: logger.warning( f"""Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.""" """ Please check that the vocabulary is not corrupted!""" ) UpperCamelCase = token_index writer.write(token + """\n""" ) index += 1 return (vocab_file,) def lowerCamelCase_ ( self : Optional[int] , lowerCamelCase_ : List[int] , lowerCamelCase_ : List[int] = None ): """simple docstring""" if token_ids_a is None: return [self.bos_token_id] + token_ids_a return [self.bos_token_id] + token_ids_a + [self.bos_token_id] + token_ids_a def lowerCamelCase_ ( self : str , lowerCamelCase_ : List[int] , lowerCamelCase_ : Optional[List[int]] = None , lowerCamelCase_ : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ , token_ids_a=lowerCamelCase_ , already_has_special_tokens=lowerCamelCase_ ) if token_ids_a is not None: return [1] + ([0] * len(lowerCamelCase_ )) + [1] + ([0] * len(lowerCamelCase_ )) return [1] + ([0] * len(lowerCamelCase_ ))
165
0
'''simple docstring''' import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a ( _lowerCamelCase , _lowerCamelCase , unittest.TestCase ): snake_case_ = AutoencoderKL snake_case_ = "sample" snake_case_ = 1e-2 @property def A_ ( self : Dict ): snake_case_ = 4 snake_case_ = 3 snake_case_ = (32, 32) snake_case_ = floats_tensor((batch_size, num_channels) + sizes ).to(lowercase_ ) return {"sample": image} @property def A_ ( self : List[Any] ): return (3, 32, 32) @property def A_ ( self : Dict ): return (3, 32, 32) def A_ ( self : Union[str, Any] ): snake_case_ = { '''block_out_channels''': [32, 64], '''in_channels''': 3, '''out_channels''': 3, '''down_block_types''': ['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''], '''up_block_types''': ['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''], '''latent_channels''': 4, } snake_case_ = self.dummy_input return init_dict, inputs_dict def A_ ( self : Any ): pass def A_ ( self : str ): pass @unittest.skipIf(torch_device == '''mps''' , '''Gradient checkpointing skipped on MPS''' ) def A_ ( self : Dict ): # enable deterministic behavior for gradient checkpointing snake_case_ ,snake_case_ = self.prepare_init_args_and_inputs_for_common() snake_case_ = self.model_class(**lowercase_ ) model.to(lowercase_ ) assert not model.is_gradient_checkpointing and model.training snake_case_ = model(**lowercase_ ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() snake_case_ = torch.randn_like(lowercase_ ) snake_case_ = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing snake_case_ = self.model_class(**lowercase_ ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowercase_ ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training snake_case_ = model_a(**lowercase_ ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() snake_case_ = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1e-5 ) snake_case_ = dict(model.named_parameters() ) snake_case_ = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5e-5 ) ) def A_ ( self : Tuple ): snake_case_ ,snake_case_ = AutoencoderKL.from_pretrained('''fusing/autoencoder-kl-dummy''' , output_loading_info=lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertEqual(len(loading_info['''missing_keys'''] ) , 0 ) model.to(lowercase_ ) snake_case_ = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def A_ ( self : Tuple ): snake_case_ = AutoencoderKL.from_pretrained('''fusing/autoencoder-kl-dummy''' ) snake_case_ = model.to(lowercase_ ) model.eval() if torch_device == "mps": snake_case_ = torch.manual_seed(0 ) else: snake_case_ = torch.Generator(device=lowercase_ ).manual_seed(0 ) snake_case_ = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) snake_case_ = image.to(lowercase_ ) with torch.no_grad(): snake_case_ = model(lowercase_ , sample_posterior=lowercase_ , generator=lowercase_ ).sample snake_case_ = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": snake_case_ = torch.tensor( [ -4.0_078e-01, -3.8_323e-04, -1.2_681e-01, -1.1_462e-01, 2.0_095e-01, 1.0_893e-01, -8.8_247e-02, -3.0_361e-01, -9.8_644e-03, ] ) elif torch_device == "cpu": snake_case_ = torch.tensor( [-0.1352, 0.0878, 0.0419, -0.0818, -0.1069, 0.0688, -0.1458, -0.4446, -0.0026] ) else: snake_case_ = torch.tensor( [-0.2421, 0.4642, 0.2507, -0.0438, 0.0682, 0.3160, -0.2018, -0.0727, 0.2485] ) self.assertTrue(torch_all_close(lowercase_ , lowercase_ , rtol=1e-2 ) ) @slow class a ( unittest.TestCase ): def A_ ( self : Union[str, Any] , lowercase_ : Optional[Any] , lowercase_ : Optional[int] ): return F"gaussian_noise_s={seed}_shape={'_'.join([str(lowercase_ ) for s in shape] )}.npy" def A_ ( self : Any ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A_ ( self : Dict , lowercase_ : List[Any]=0 , lowercase_ : Union[str, Any]=(4, 3, 512, 512) , lowercase_ : Optional[Any]=False ): snake_case_ = torch.floataa if fpaa else torch.floataa snake_case_ = torch.from_numpy(load_hf_numpy(self.get_file_format(lowercase_ , lowercase_ ) ) ).to(lowercase_ ).to(lowercase_ ) return image def A_ ( self : Any , lowercase_ : Dict="CompVis/stable-diffusion-v1-4" , lowercase_ : List[str]=False ): snake_case_ = '''fp16''' if fpaa else None snake_case_ = torch.floataa if fpaa else torch.floataa snake_case_ = AutoencoderKL.from_pretrained( lowercase_ , subfolder='''vae''' , torch_dtype=lowercase_ , revision=lowercase_ , ) model.to(lowercase_ ).eval() return model def A_ ( self : Any , lowercase_ : int=0 ): if torch_device == "mps": return torch.manual_seed(lowercase_ ) return torch.Generator(device=lowercase_ ).manual_seed(lowercase_ ) @parameterized.expand( [ # fmt: off [33, [-0.1603, 0.9878, -0.0495, -0.0790, -0.2709, 0.8375, -0.2060, -0.0824], [-0.2395, 0.0098, 0.0102, -0.0709, -0.2840, -0.0274, -0.0718, -0.1824]], [47, [-0.2376, 0.1168, 0.1332, -0.4840, -0.2508, -0.0791, -0.0493, -0.4089], [0.0350, 0.0847, 0.0467, 0.0344, -0.0842, -0.0547, -0.0633, -0.1131]], # fmt: on ] ) def A_ ( self : Union[str, Any] , lowercase_ : Tuple , lowercase_ : Tuple , lowercase_ : Tuple ): snake_case_ = self.get_sd_vae_model() snake_case_ = self.get_sd_image(lowercase_ ) snake_case_ = self.get_generator(lowercase_ ) with torch.no_grad(): snake_case_ = model(lowercase_ , generator=lowercase_ , sample_posterior=lowercase_ ).sample assert sample.shape == image.shape snake_case_ = sample[-1, -2:, -2:, :2].flatten().float().cpu() snake_case_ = torch.tensor(expected_slice_mps if torch_device == '''mps''' else expected_slice ) assert torch_all_close(lowercase_ , lowercase_ , atol=3e-3 ) @parameterized.expand( [ # fmt: off [33, [-0.0513, 0.0289, 1.3799, 0.2166, -0.2573, -0.0871, 0.5103, -0.0999]], [47, [-0.4128, -0.1320, -0.3704, 0.1965, -0.4116, -0.2332, -0.3340, 0.2247]], # fmt: on ] ) @require_torch_gpu def A_ ( self : Optional[int] , lowercase_ : Union[str, Any] , lowercase_ : Dict ): snake_case_ = self.get_sd_vae_model(fpaa=lowercase_ ) snake_case_ = self.get_sd_image(lowercase_ , fpaa=lowercase_ ) snake_case_ = self.get_generator(lowercase_ ) with torch.no_grad(): snake_case_ = model(lowercase_ , generator=lowercase_ , sample_posterior=lowercase_ ).sample assert sample.shape == image.shape snake_case_ = sample[-1, -2:, :2, -2:].flatten().float().cpu() snake_case_ = torch.tensor(lowercase_ ) assert torch_all_close(lowercase_ , lowercase_ , atol=1e-2 ) @parameterized.expand( [ # fmt: off [33, [-0.1609, 0.9866, -0.0487, -0.0777, -0.2716, 0.8368, -0.2055, -0.0814], [-0.2395, 0.0098, 0.0102, -0.0709, -0.2840, -0.0274, -0.0718, -0.1824]], [47, [-0.2377, 0.1147, 0.1333, -0.4841, -0.2506, -0.0805, -0.0491, -0.4085], [0.0350, 0.0847, 0.0467, 0.0344, -0.0842, -0.0547, -0.0633, -0.1131]], # fmt: on ] ) def A_ ( self : Tuple , lowercase_ : Dict , lowercase_ : str , lowercase_ : Optional[int] ): snake_case_ = self.get_sd_vae_model() snake_case_ = self.get_sd_image(lowercase_ ) with torch.no_grad(): snake_case_ = model(lowercase_ ).sample assert sample.shape == image.shape snake_case_ = sample[-1, -2:, -2:, :2].flatten().float().cpu() snake_case_ = torch.tensor(expected_slice_mps if torch_device == '''mps''' else expected_slice ) assert torch_all_close(lowercase_ , lowercase_ , atol=3e-3 ) @parameterized.expand( [ # fmt: off [13, [-0.2051, -0.1803, -0.2311, -0.2114, -0.3292, -0.3574, -0.2953, -0.3323]], [37, [-0.2632, -0.2625, -0.2199, -0.2741, -0.4539, -0.4990, -0.3720, -0.4925]], # fmt: on ] ) @require_torch_gpu def A_ ( self : Dict , lowercase_ : Tuple , lowercase_ : Optional[int] ): snake_case_ = self.get_sd_vae_model() snake_case_ = self.get_sd_image(lowercase_ , shape=(3, 4, 64, 64) ) with torch.no_grad(): snake_case_ = model.decode(lowercase_ ).sample assert list(sample.shape ) == [3, 3, 512, 512] snake_case_ = sample[-1, -2:, :2, -2:].flatten().cpu() snake_case_ = torch.tensor(lowercase_ ) assert torch_all_close(lowercase_ , lowercase_ , atol=1e-3 ) @parameterized.expand( [ # fmt: off [27, [-0.0369, 0.0207, -0.0776, -0.0682, -0.1747, -0.1930, -0.1465, -0.2039]], [16, [-0.1628, -0.2134, -0.2747, -0.2642, -0.3774, -0.4404, -0.3687, -0.4277]], # fmt: on ] ) @require_torch_gpu def A_ ( self : Union[str, Any] , lowercase_ : Any , lowercase_ : Optional[Any] ): snake_case_ = self.get_sd_vae_model(fpaa=lowercase_ ) snake_case_ = self.get_sd_image(lowercase_ , shape=(3, 4, 64, 64) , fpaa=lowercase_ ) with torch.no_grad(): snake_case_ = model.decode(lowercase_ ).sample assert list(sample.shape ) == [3, 3, 512, 512] snake_case_ = sample[-1, -2:, :2, -2:].flatten().float().cpu() snake_case_ = torch.tensor(lowercase_ ) assert torch_all_close(lowercase_ , lowercase_ , atol=5e-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason='''xformers is not required when using PyTorch 2.0.''' ) def A_ ( self : Optional[Any] , lowercase_ : List[str] ): snake_case_ = self.get_sd_vae_model(fpaa=lowercase_ ) snake_case_ = self.get_sd_image(lowercase_ , shape=(3, 4, 64, 64) , fpaa=lowercase_ ) with torch.no_grad(): snake_case_ = model.decode(lowercase_ ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): snake_case_ = model.decode(lowercase_ ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowercase_ , lowercase_ , atol=1e-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason='''xformers is not required when using PyTorch 2.0.''' ) def A_ ( self : Optional[Any] , lowercase_ : Any ): snake_case_ = self.get_sd_vae_model() snake_case_ = self.get_sd_image(lowercase_ , shape=(3, 4, 64, 64) ) with torch.no_grad(): snake_case_ = model.decode(lowercase_ ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): snake_case_ = model.decode(lowercase_ ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowercase_ , lowercase_ , atol=1e-2 ) @parameterized.expand( [ # fmt: off [33, [-0.3001, 0.0918, -2.6984, -3.9720, -3.2099, -5.0353, 1.7338, -0.2065, 3.4267]], [47, [-1.5030, -4.3871, -6.0355, -9.1157, -1.6661, -2.7853, 2.1607, -5.0823, 2.5633]], # fmt: on ] ) def A_ ( self : str , lowercase_ : Optional[int] , lowercase_ : Tuple ): snake_case_ = self.get_sd_vae_model() snake_case_ = self.get_sd_image(lowercase_ ) snake_case_ = self.get_generator(lowercase_ ) with torch.no_grad(): snake_case_ = model.encode(lowercase_ ).latent_dist snake_case_ = dist.sample(generator=lowercase_ ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] snake_case_ = sample[0, -1, -3:, -3:].flatten().cpu() snake_case_ = torch.tensor(lowercase_ ) snake_case_ = 3e-3 if torch_device != '''mps''' else 1e-2 assert torch_all_close(lowercase_ , lowercase_ , atol=lowercase_ )
56
'''simple docstring''' from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxCrossAttnUpBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, FlaxUpBlockaD, ) @flax.struct.dataclass class a ( _lowerCamelCase ): snake_case_ = 42 @flax_register_to_config class a ( nn.Module , _lowerCamelCase , _lowerCamelCase ): snake_case_ = 32 snake_case_ = 4 snake_case_ = 4 snake_case_ = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) snake_case_ = ("UpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D") snake_case_ = False snake_case_ = (320, 640, 1_280, 1_280) snake_case_ = 2 snake_case_ = 8 snake_case_ = None snake_case_ = 1_280 snake_case_ = 0.0 snake_case_ = False snake_case_ = jnp.floataa snake_case_ = True snake_case_ = 0 snake_case_ = False def A_ ( self : Optional[int] , lowercase_ : jax.random.KeyArray ): # init input tensors snake_case_ = (1, self.in_channels, self.sample_size, self.sample_size) snake_case_ = jnp.zeros(lowercase_ , dtype=jnp.floataa ) snake_case_ = jnp.ones((1,) , dtype=jnp.intaa ) snake_case_ = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) snake_case_ ,snake_case_ = jax.random.split(lowercase_ ) snake_case_ = {'''params''': params_rng, '''dropout''': dropout_rng} return self.init(lowercase_ , lowercase_ , lowercase_ , lowercase_ )["params"] def A_ ( self : List[str] ): snake_case_ = self.block_out_channels snake_case_ = block_out_channels[0] * 4 if self.num_attention_heads is not None: raise ValueError( '''At the moment it is not possible to define the number of attention heads via `num_attention_heads` because of a naming issue as described in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131. Passing `num_attention_heads` will only be supported in diffusers v0.19.''' ) # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. snake_case_ = self.num_attention_heads or self.attention_head_dim # input snake_case_ = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time snake_case_ = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) snake_case_ = FlaxTimestepEmbedding(lowercase_ , dtype=self.dtype ) snake_case_ = self.only_cross_attention if isinstance(lowercase_ , lowercase_ ): snake_case_ = (only_cross_attention,) * len(self.down_block_types ) if isinstance(lowercase_ , lowercase_ ): snake_case_ = (num_attention_heads,) * len(self.down_block_types ) # down snake_case_ = [] snake_case_ = block_out_channels[0] for i, down_block_type in enumerate(self.down_block_types ): snake_case_ = output_channel snake_case_ = block_out_channels[i] snake_case_ = i == len(lowercase_ ) - 1 if down_block_type == "CrossAttnDownBlock2D": snake_case_ = FlaxCrossAttnDownBlockaD( in_channels=lowercase_ , out_channels=lowercase_ , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: snake_case_ = FlaxDownBlockaD( in_channels=lowercase_ , out_channels=lowercase_ , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(lowercase_ ) snake_case_ = down_blocks # mid snake_case_ = FlaxUNetMidBlockaDCrossAttn( in_channels=block_out_channels[-1] , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) # up snake_case_ = [] snake_case_ = list(reversed(lowercase_ ) ) snake_case_ = list(reversed(lowercase_ ) ) snake_case_ = list(reversed(lowercase_ ) ) snake_case_ = reversed_block_out_channels[0] for i, up_block_type in enumerate(self.up_block_types ): snake_case_ = output_channel snake_case_ = reversed_block_out_channels[i] snake_case_ = reversed_block_out_channels[min(i + 1 , len(lowercase_ ) - 1 )] snake_case_ = i == len(lowercase_ ) - 1 if up_block_type == "CrossAttnUpBlock2D": snake_case_ = FlaxCrossAttnUpBlockaD( in_channels=lowercase_ , out_channels=lowercase_ , prev_output_channel=lowercase_ , num_layers=self.layers_per_block + 1 , num_attention_heads=reversed_num_attention_heads[i] , add_upsample=not is_final_block , dropout=self.dropout , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: snake_case_ = FlaxUpBlockaD( in_channels=lowercase_ , out_channels=lowercase_ , prev_output_channel=lowercase_ , num_layers=self.layers_per_block + 1 , add_upsample=not is_final_block , dropout=self.dropout , dtype=self.dtype , ) up_blocks.append(lowercase_ ) snake_case_ = output_channel snake_case_ = up_blocks # out snake_case_ = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) snake_case_ = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : Union[str, Any] , lowercase_ : Union[str, Any] , lowercase_ : int , lowercase_ : Any , lowercase_ : int=None , lowercase_ : Any=None , lowercase_ : bool = True , lowercase_ : bool = False , ): # 1. time if not isinstance(lowercase_ , jnp.ndarray ): snake_case_ = jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(lowercase_ , jnp.ndarray ) and len(timesteps.shape ) == 0: snake_case_ = timesteps.astype(dtype=jnp.floataa ) snake_case_ = jnp.expand_dims(lowercase_ , 0 ) snake_case_ = self.time_proj(lowercase_ ) snake_case_ = self.time_embedding(lowercase_ ) # 2. pre-process snake_case_ = jnp.transpose(lowercase_ , (0, 2, 3, 1) ) snake_case_ = self.conv_in(lowercase_ ) # 3. down snake_case_ = (sample,) for down_block in self.down_blocks: if isinstance(lowercase_ , lowercase_ ): snake_case_ ,snake_case_ = down_block(lowercase_ , lowercase_ , lowercase_ , deterministic=not train ) else: snake_case_ ,snake_case_ = down_block(lowercase_ , lowercase_ , deterministic=not train ) down_block_res_samples += res_samples if down_block_additional_residuals is not None: snake_case_ = () for down_block_res_sample, down_block_additional_residual in zip( lowercase_ , lowercase_ ): down_block_res_sample += down_block_additional_residual new_down_block_res_samples += (down_block_res_sample,) snake_case_ = new_down_block_res_samples # 4. mid snake_case_ = self.mid_block(lowercase_ , lowercase_ , lowercase_ , deterministic=not train ) if mid_block_additional_residual is not None: sample += mid_block_additional_residual # 5. up for up_block in self.up_blocks: snake_case_ = down_block_res_samples[-(self.layers_per_block + 1) :] snake_case_ = down_block_res_samples[: -(self.layers_per_block + 1)] if isinstance(lowercase_ , lowercase_ ): snake_case_ = up_block( lowercase_ , temb=lowercase_ , encoder_hidden_states=lowercase_ , res_hidden_states_tuple=lowercase_ , deterministic=not train , ) else: snake_case_ = up_block(lowercase_ , temb=lowercase_ , res_hidden_states_tuple=lowercase_ , deterministic=not train ) # 6. post-process snake_case_ = self.conv_norm_out(lowercase_ ) snake_case_ = nn.silu(lowercase_ ) snake_case_ = self.conv_out(lowercase_ ) snake_case_ = jnp.transpose(lowercase_ , (0, 3, 1, 2) ) if not return_dict: return (sample,) return FlaxUNetaDConditionOutput(sample=lowercase_ )
56
1
from __future__ import annotations A : List[str] = [-1_0, -5, 0, 5, 5.1, 1_1, 1_3, 2_1, 3, 4, -2_1, -1_0, -5, -1, 0] A : Optional[Any] = [-5, 0, 5, 5.1, 1_1, 1_3, 2_1, -1, 4, -1, -1_0, -5, -1, 0, -1] def __lowerCAmelCase ( a__ ) -> list[float]: __a = [] __a = len(a__ ) for i in range(a__ ): __a = -1 for j in range(i + 1 , a__ ): if arr[i] < arr[j]: __a = arr[j] break result.append(a__ ) return result def __lowerCAmelCase ( a__ ) -> list[float]: __a = [] for i, outer in enumerate(a__ ): __a = -1 for inner in arr[i + 1 :]: if outer < inner: __a = inner break result.append(a__ ) return result def __lowerCAmelCase ( a__ ) -> list[float]: __a = len(a__ ) __a = [] __a = [-1] * arr_size for index in reversed(range(a__ ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: __a = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) A : str = ( 'from __main__ import arr, next_greatest_element_slow, ' 'next_greatest_element_fast, next_greatest_element' ) print( 'next_greatest_element_slow():', timeit('next_greatest_element_slow(arr)', setup=setup), ) print( 'next_greatest_element_fast():', timeit('next_greatest_element_fast(arr)', setup=setup), ) print( ' next_greatest_element():', timeit('next_greatest_element(arr)', setup=setup), )
367
import sys def __lowerCAmelCase ( a__ ) -> Optional[int]: __a = len(a__ ) __a = [[0 for x in range(a__ )] for x in range(a__ )] __a = [[0 for x in range(a__ )] for x in range(a__ )] for chain_length in range(2 , a__ ): for a in range(1 , n - chain_length + 1 ): __a = a + chain_length - 1 __a = sys.maxsize for c in range(a__ , a__ ): __a = ( matrix[a][c] + matrix[c + 1][b] + array[a - 1] * array[c] * array[b] ) if cost < matrix[a][b]: __a = cost __a = c return matrix, sol def __lowerCAmelCase ( a__ , a__ , a__ ) -> Any: if i == j: print('''A''' + str(a__ ) , end=''' ''' ) else: print('''(''' , end=''' ''' ) print_optiomal_solution(a__ , a__ , optimal_solution[i][j] ) print_optiomal_solution(a__ , optimal_solution[i][j] + 1 , a__ ) print(''')''' , end=''' ''' ) def __lowerCAmelCase ( ) -> int: __a = [30, 35, 15, 5, 10, 20, 25] __a = len(a__ ) # Size of matrix created from above array will be # 30*35 35*15 15*5 5*10 10*20 20*25 __a , __a = matrix_chain_order(a__ ) print('''No. of Operation required: ''' + str(matrix[1][n - 1] ) ) print_optiomal_solution(a__ , 1 , n - 1 ) if __name__ == "__main__": main()
33
0
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging lowerCAmelCase_ = logging.get_logger(__name__) class __A ( A_ ): '''simple docstring''' lowerCAmelCase : Tuple = ["input_features"] def __init__( self : Optional[Any] ,_snake_case : List[Any]=80 ,_snake_case : Dict=16_000 ,_snake_case : Any=160 ,_snake_case : Tuple=30 ,_snake_case : List[str]=400 ,_snake_case : Dict=0.0 ,_snake_case : List[Any]=False ,**_snake_case : str ,) -> Any: """simple docstring""" super().__init__( feature_size=_snake_case ,sampling_rate=_snake_case ,padding_value=_snake_case ,return_attention_mask=_snake_case ,**_snake_case ,) lowercase__ : int = n_fft lowercase__ : List[str] = hop_length lowercase__ : int = chunk_length lowercase__ : Union[str, Any] = chunk_length * sampling_rate lowercase__ : Union[str, Any] = self.n_samples // hop_length lowercase__ : Tuple = sampling_rate lowercase__ : List[Any] = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 ,num_mel_filters=_snake_case ,min_frequency=0.0 ,max_frequency=8000.0 ,sampling_rate=_snake_case ,norm='''slaney''' ,mel_scale='''slaney''' ,) def UpperCAmelCase ( self : Optional[Any] ,_snake_case : np.array ) -> np.ndarray: """simple docstring""" lowercase__ : List[str] = spectrogram( _snake_case ,window_function(self.n_fft ,'''hann''' ) ,frame_length=self.n_fft ,hop_length=self.hop_length ,power=2.0 ,mel_filters=self.mel_filters ,log_mel='''log10''' ,) lowercase__ : Tuple = log_spec[:, :-1] lowercase__ : Optional[Any] = np.maximum(_snake_case ,log_spec.max() - 8.0 ) lowercase__ : Optional[int] = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def UpperCAmelCase ( _snake_case : List[np.ndarray] ,_snake_case : List[np.ndarray] ,_snake_case : float = 0.0 ) -> List[np.ndarray]: """simple docstring""" if attention_mask is not None: lowercase__ : Optional[Any] = np.array(_snake_case ,np.intaa ) lowercase__ : Tuple = [] for vector, length in zip(_snake_case ,attention_mask.sum(-1 ) ): lowercase__ : Optional[int] = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1e-7 ) if length < normed_slice.shape[0]: lowercase__ : List[Any] = padding_value normed_input_values.append(_snake_case ) else: lowercase__ : Optional[Any] = [(x - x.mean()) / np.sqrt(x.var() + 1e-7 ) for x in input_values] return normed_input_values def __call__( self : Optional[int] ,_snake_case : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,_snake_case : bool = True ,_snake_case : Optional[int] = None ,_snake_case : Optional[Union[str, TensorType]] = None ,_snake_case : Optional[bool] = None ,_snake_case : Optional[str] = "max_length" ,_snake_case : Optional[int] = None ,_snake_case : Optional[int] = None ,_snake_case : Optional[bool] = None ,**_snake_case : Tuple ,) -> BatchFeature: """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"""The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a""" f""" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input""" f""" was sampled with {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( '''It is strongly recommended to pass the `sampling_rate` argument to this function. ''' '''Failing to do so can result in silent errors that might be hard to debug.''' ) lowercase__ : Dict = isinstance(_snake_case ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"""Only mono-channel audio is supported for input to {self}""" ) lowercase__ : List[Any] = is_batched_numpy or ( isinstance(_snake_case ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: lowercase__ : Any = [np.asarray([speech] ,dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(_snake_case ,np.ndarray ): lowercase__ : List[Any] = np.asarray(_snake_case ,dtype=np.floataa ) elif isinstance(_snake_case ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowercase__ : Optional[int] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowercase__ : str = [np.asarray([raw_speech] ).T] lowercase__ : int = BatchFeature({'''input_features''': raw_speech} ) # convert into correct format for padding lowercase__ : Optional[Any] = self.pad( _snake_case ,padding=_snake_case ,max_length=max_length if max_length else self.n_samples ,truncation=_snake_case ,pad_to_multiple_of=_snake_case ,return_attention_mask=return_attention_mask or do_normalize ,) # zero-mean and unit-variance normalization if do_normalize: lowercase__ : int = self.zero_mean_unit_var_norm( padded_inputs['''input_features'''] ,attention_mask=padded_inputs['''attention_mask'''] ,padding_value=self.padding_value ,) lowercase__ : Dict = np.stack(padded_inputs['''input_features'''] ,axis=0 ) # make sure list is in array format lowercase__ : int = padded_inputs.get('''input_features''' ).transpose(2 ,0 ,1 ) lowercase__ : Optional[int] = [self._np_extract_fbank_features(_snake_case ) for waveform in input_features[0]] if isinstance(input_features[0] ,_snake_case ): lowercase__ : Optional[int] = [np.asarray(_snake_case ,dtype=np.floataa ) for feature in input_features] else: lowercase__ : Tuple = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) lowercase__ : Tuple = padded_inputs['''attention_mask'''][:, :: self.hop_length] if return_tensors is not None: lowercase__ : int = padded_inputs.convert_to_tensors(_snake_case ) return padded_inputs def UpperCAmelCase ( self : Optional[int] ) -> Dict[str, Any]: """simple docstring""" lowercase__ : List[Any] = copy.deepcopy(self.__dict__ ) lowercase__ : Dict = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
16
"""simple docstring""" import unittest from accelerate import debug_launcher from accelerate.test_utils import require_cpu, test_ops, test_script @require_cpu class __A ( unittest.TestCase ): '''simple docstring''' def UpperCAmelCase ( self : Optional[int] ) -> str: """simple docstring""" debug_launcher(test_script.main ) def UpperCAmelCase ( self : Dict ) -> Tuple: """simple docstring""" debug_launcher(test_ops.main )
16
1
import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class A_ ( snake_case__ ): _UpperCAmelCase : Optional[int] = ["""image_processor""", """tokenizer"""] _UpperCAmelCase : List[Any] = """BlipImageProcessor""" _UpperCAmelCase : Optional[Any] = """AutoTokenizer""" def __init__( self : str ,SCREAMING_SNAKE_CASE__ : Union[str, Any] ,SCREAMING_SNAKE_CASE__ : Dict ,SCREAMING_SNAKE_CASE__ : List[str]): super().__init__(_A ,_A) # add QFormer tokenizer __lowerCamelCase : List[str] = qformer_tokenizer def __call__( self : Optional[int] ,SCREAMING_SNAKE_CASE__ : Any = None ,SCREAMING_SNAKE_CASE__ : Optional[Any] = None ,SCREAMING_SNAKE_CASE__ : Union[str, Any] = True ,SCREAMING_SNAKE_CASE__ : Tuple = False ,SCREAMING_SNAKE_CASE__ : List[Any] = None ,SCREAMING_SNAKE_CASE__ : Any = None ,SCREAMING_SNAKE_CASE__ : Union[str, Any] = 0 ,SCREAMING_SNAKE_CASE__ : Optional[int] = None ,SCREAMING_SNAKE_CASE__ : Dict = None ,SCREAMING_SNAKE_CASE__ : str = False ,SCREAMING_SNAKE_CASE__ : List[str] = False ,SCREAMING_SNAKE_CASE__ : List[str] = False ,SCREAMING_SNAKE_CASE__ : Tuple = False ,SCREAMING_SNAKE_CASE__ : str = False ,SCREAMING_SNAKE_CASE__ : int = True ,SCREAMING_SNAKE_CASE__ : Optional[int] = None ,**SCREAMING_SNAKE_CASE__ : str ,): if images is None and text is None: raise ValueError('You have to specify at least images or text.') __lowerCamelCase : Optional[int] = BatchFeature() if text is not None: __lowerCamelCase : Dict = self.tokenizer( text=_A ,add_special_tokens=_A ,padding=_A ,truncation=_A ,max_length=_A ,stride=_A ,pad_to_multiple_of=_A ,return_attention_mask=_A ,return_overflowing_tokens=_A ,return_special_tokens_mask=_A ,return_offsets_mapping=_A ,return_token_type_ids=_A ,return_length=_A ,verbose=_A ,return_tensors=_A ,**_A ,) encoding.update(_A) __lowerCamelCase : Optional[Any] = self.qformer_tokenizer( text=_A ,add_special_tokens=_A ,padding=_A ,truncation=_A ,max_length=_A ,stride=_A ,pad_to_multiple_of=_A ,return_attention_mask=_A ,return_overflowing_tokens=_A ,return_special_tokens_mask=_A ,return_offsets_mapping=_A ,return_token_type_ids=_A ,return_length=_A ,verbose=_A ,return_tensors=_A ,**_A ,) __lowerCamelCase : Dict = qformer_text_encoding.pop('input_ids') __lowerCamelCase : Optional[int] = qformer_text_encoding.pop('attention_mask') if images is not None: __lowerCamelCase : int = self.image_processor(_A ,return_tensors=_A) encoding.update(_A) return encoding def lowerCAmelCase ( self : Any ,*SCREAMING_SNAKE_CASE__ : int ,**SCREAMING_SNAKE_CASE__ : Union[str, Any]): return self.tokenizer.batch_decode(*_A ,**_A) def lowerCAmelCase ( self : Union[str, Any] ,*SCREAMING_SNAKE_CASE__ : Union[str, Any] ,**SCREAMING_SNAKE_CASE__ : Tuple): return self.tokenizer.decode(*_A ,**_A) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def lowerCAmelCase ( self : int): __lowerCamelCase : Dict = self.tokenizer.model_input_names __lowerCamelCase : Tuple = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names)) def lowerCAmelCase ( self : int ,SCREAMING_SNAKE_CASE__ : List[Any] ,**SCREAMING_SNAKE_CASE__ : Any): if os.path.isfile(_A): raise ValueError(F"Provided path ({save_directory}) should be a directory, not a file") os.makedirs(_A ,exist_ok=_A) __lowerCamelCase : int = os.path.join(_A ,'qformer_tokenizer') self.qformer_tokenizer.save_pretrained(_A) return super().save_pretrained(_A ,**_A) @classmethod def lowerCAmelCase ( cls : List[str] ,SCREAMING_SNAKE_CASE__ : Optional[Any] ,**SCREAMING_SNAKE_CASE__ : Union[str, Any]): __lowerCamelCase : Optional[int] = AutoTokenizer.from_pretrained(_A ,subfolder='qformer_tokenizer') __lowerCamelCase : Union[str, Any] = cls._get_arguments_from_pretrained(_A ,**_A) args.append(_A) return cls(*_A)
364
from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> List[Any]: # A local function to see if a dot lands in the circle. def is_in_circle(lowerCamelCase__ , lowerCamelCase__ ) -> bool: __lowerCamelCase : Dict = sqrt((x**2) + (y**2) ) # Our circle has a radius of 1, so a distance # greater than 1 would land outside the circle. return distance_from_centre <= 1 # The proportion of guesses that landed in the circle __lowerCamelCase : Optional[int] = mean( int(is_in_circle(uniform(-1.0 , 1.0 ) , uniform(-1.0 , 1.0 ) ) ) for _ in range(lowerCamelCase__ ) ) # The ratio of the area for circle to square is pi/4. __lowerCamelCase : Any = proportion * 4 print(F"The estimated value of pi is {pi_estimate}" ) print(F"The numpy value of pi is {pi}" ) print(F"The total error is {abs(pi - pi_estimate )}" ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = 0.0 , lowerCamelCase__ = 1.0 , ) -> float: return mean( function_to_integrate(uniform(lowerCamelCase__ , lowerCamelCase__ ) ) for _ in range(lowerCamelCase__ ) ) * (max_value - min_value) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ = 0.0 , lowerCamelCase__ = 1.0 ) -> None: def identity_function(lowerCamelCase__ ) -> float: return x __lowerCamelCase : str = area_under_curve_estimator( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) __lowerCamelCase : int = (max_value * max_value - min_value * min_value) / 2 print('******************' ) print(F"Estimating area under y=x where x varies from {min_value} to {max_value}" ) print(F"Estimated value is {estimated_value}" ) print(F"Expected value is {expected_value}" ) print(F"Total error is {abs(estimated_value - expected_value )}" ) print('******************' ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> None: def function_to_integrate(lowerCamelCase__ ) -> float: return sqrt(4.0 - x * x ) __lowerCamelCase : Any = area_under_curve_estimator( lowerCamelCase__ , lowerCamelCase__ , 0.0 , 2.0 ) print('******************' ) print('Estimating pi using area_under_curve_estimator' ) print(F"Estimated value is {estimated_value}" ) print(F"Expected value is {pi}" ) print(F"Total error is {abs(estimated_value - pi )}" ) print('******************' ) if __name__ == "__main__": import doctest doctest.testmod()
113
0
"""simple docstring""" import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class A_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self :List[str] ) -> Dict: # A mock response for an HTTP head request to emulate server down UpperCAmelCase = mock.Mock() UpperCAmelCase = 5_00 UpperCAmelCase = {} UpperCAmelCase = HTTPError UpperCAmelCase = {} # Download this model to make sure it's in the cache. UpperCAmelCase = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' , return_value=lowercase_ ) as mock_head: UpperCAmelCase = BertTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def UpperCAmelCase__ ( self :List[Any] ) -> Dict: # A mock response for an HTTP head request to emulate server down UpperCAmelCase = mock.Mock() UpperCAmelCase = 5_00 UpperCAmelCase = {} UpperCAmelCase = HTTPError UpperCAmelCase = {} # Download this model to make sure it's in the cache. UpperCAmelCase = GPTaTokenizerFast.from_pretrained('gpt2' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('requests.Session.request' , return_value=lowercase_ ) as mock_head: UpperCAmelCase = GPTaTokenizerFast.from_pretrained('gpt2' ) # This check we did call the fake head request mock_head.assert_called() def UpperCAmelCase__ ( self :Any ) -> Union[str, Any]: # This test is for deprecated behavior and can be removed in v5 try: UpperCAmelCase = tempfile.mktemp() with open(lowercase_ , 'wb' ) as f: http_get('https://huggingface.co/albert-base-v1/resolve/main/spiece.model' , lowercase_ ) UpperCAmelCase = AlbertTokenizer.from_pretrained(lowercase_ ) finally: os.remove(lowercase_ ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile('tokenizer.json' ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open('tokenizer.json' , 'wb' ) as f: http_get('https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json' , lowercase_ ) UpperCAmelCase = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 10_00 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove('tokenizer.json' ) def UpperCAmelCase__ ( self :Tuple ) -> Optional[int]: # This test is for deprecated behavior and can be removed in v5 UpperCAmelCase = AlbertTokenizer.from_pretrained('https://huggingface.co/albert-base-v1/resolve/main/spiece.model' ) @is_staging_test class A_ ( unittest.TestCase ): """simple docstring""" __UpperCamelCase = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """bla""", """blou"""] @classmethod def UpperCAmelCase__ ( cls :Tuple ) -> List[Any]: UpperCAmelCase = TOKEN HfFolder.save_token(lowercase_ ) @classmethod def UpperCAmelCase__ ( cls :List[str] ) -> Optional[int]: try: delete_repo(token=cls._token , repo_id='test-tokenizer' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-tokenizer-org' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-tokenizer' ) except HTTPError: pass def UpperCAmelCase__ ( self :Any ) -> str: with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase = os.path.join(lowercase_ , 'vocab.txt' ) with open(lowercase_ , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) UpperCAmelCase = BertTokenizer(lowercase_ ) tokenizer.push_to_hub('test-tokenizer' , use_auth_token=self._token ) UpperCAmelCase = BertTokenizer.from_pretrained(f"""{USER}/test-tokenizer""" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='test-tokenizer' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(lowercase_ , repo_id='test-tokenizer' , push_to_hub=lowercase_ , use_auth_token=self._token ) UpperCAmelCase = BertTokenizer.from_pretrained(f"""{USER}/test-tokenizer""" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def UpperCAmelCase__ ( self :int ) -> Union[str, Any]: with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase = os.path.join(lowercase_ , 'vocab.txt' ) with open(lowercase_ , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) UpperCAmelCase = BertTokenizer(lowercase_ ) tokenizer.push_to_hub('valid_org/test-tokenizer-org' , use_auth_token=self._token ) UpperCAmelCase = BertTokenizer.from_pretrained('valid_org/test-tokenizer-org' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='valid_org/test-tokenizer-org' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( lowercase_ , repo_id='valid_org/test-tokenizer-org' , push_to_hub=lowercase_ , use_auth_token=self._token ) UpperCAmelCase = BertTokenizer.from_pretrained('valid_org/test-tokenizer-org' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def UpperCAmelCase__ ( self :int ) -> Tuple: CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase = os.path.join(lowercase_ , 'vocab.txt' ) with open(lowercase_ , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) UpperCAmelCase = CustomTokenizer(lowercase_ ) # No fast custom tokenizer tokenizer.push_to_hub('test-dynamic-tokenizer' , use_auth_token=self._token ) UpperCAmelCase = AutoTokenizer.from_pretrained(f"""{USER}/test-dynamic-tokenizer""" , trust_remote_code=lowercase_ ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizer' ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase = os.path.join(lowercase_ , 'vocab.txt' ) with open(lowercase_ , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) UpperCAmelCase = BertTokenizerFast.from_pretrained(lowercase_ ) bert_tokenizer.save_pretrained(lowercase_ ) UpperCAmelCase = CustomTokenizerFast.from_pretrained(lowercase_ ) tokenizer.push_to_hub('test-dynamic-tokenizer' , use_auth_token=self._token ) UpperCAmelCase = AutoTokenizer.from_pretrained(f"""{USER}/test-dynamic-tokenizer""" , trust_remote_code=lowercase_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizerFast' ) UpperCAmelCase = AutoTokenizer.from_pretrained( f"""{USER}/test-dynamic-tokenizer""" , use_fast=lowercase_ , trust_remote_code=lowercase_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , 'CustomTokenizer' ) class A_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self :List[str] ) -> Tuple: UpperCAmelCase = Trie() trie.add('Hello 友達' ) self.assertEqual(trie.data , {'H': {'e': {'l': {'l': {'o': {' ': {'友': {'達': {'': 1}}}}}}}}} ) trie.add('Hello' ) trie.data self.assertEqual(trie.data , {'H': {'e': {'l': {'l': {'o': {'': 1, ' ': {'友': {'達': {'': 1}}}}}}}}} ) def UpperCAmelCase__ ( self :Tuple ) -> str: UpperCAmelCase = Trie() self.assertEqual(trie.split('[CLS] This is a extra_id_100' ) , ['[CLS] This is a extra_id_100'] ) trie.add('[CLS]' ) trie.add('extra_id_1' ) trie.add('extra_id_100' ) self.assertEqual(trie.split('[CLS] This is a extra_id_100' ) , ['[CLS]', ' This is a ', 'extra_id_100'] ) def UpperCAmelCase__ ( self :Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase = Trie() trie.add('A' ) self.assertEqual(trie.split('ABC' ) , ['A', 'BC'] ) self.assertEqual(trie.split('BCA' ) , ['BC', 'A'] ) def UpperCAmelCase__ ( self :Any ) -> int: UpperCAmelCase = Trie() trie.add('TOKEN]' ) trie.add('[SPECIAL_TOKEN]' ) self.assertEqual(trie.split('This is something [SPECIAL_TOKEN]' ) , ['This is something ', '[SPECIAL_TOKEN]'] ) def UpperCAmelCase__ ( self :Any ) -> Optional[int]: UpperCAmelCase = Trie() trie.add('A' ) trie.add('P' ) trie.add('[SPECIAL_TOKEN]' ) self.assertEqual(trie.split('This is something [SPECIAL_TOKEN]' ) , ['This is something ', '[SPECIAL_TOKEN]'] ) def UpperCAmelCase__ ( self :int ) -> Union[str, Any]: UpperCAmelCase = Trie() trie.add('AB' ) trie.add('B' ) trie.add('C' ) self.assertEqual(trie.split('ABC' ) , ['AB', 'C'] ) def UpperCAmelCase__ ( self :Optional[int] ) -> List[Any]: UpperCAmelCase = Trie() trie.add('ABC' ) trie.add('B' ) trie.add('CD' ) self.assertEqual(trie.split('ABCD' ) , ['ABC', 'D'] ) def UpperCAmelCase__ ( self :Optional[int] ) -> Optional[Any]: # Even if the offsets are wrong, we necessarily output correct string # parts. UpperCAmelCase = Trie() UpperCAmelCase = trie.cut_text('ABC' , [0, 0, 2, 1, 2, 3] ) self.assertEqual(lowercase_ , ['AB', 'C'] )
78
'''simple docstring''' import argparse import json from typing import List from ltp import LTP from transformers.models.bert.tokenization_bert import BertTokenizer def UpperCamelCase_( snake_case : Any ): '''simple docstring''' if ( (cp >= 0X4E00 and cp <= 0X9FFF) or (cp >= 0X3400 and cp <= 0X4DBF) # or (cp >= 0X2_0000 and cp <= 0X2_A6DF) # or (cp >= 0X2_A700 and cp <= 0X2_B73F) # or (cp >= 0X2_B740 and cp <= 0X2_B81F) # or (cp >= 0X2_B820 and cp <= 0X2_CEAF) # or (cp >= 0XF900 and cp <= 0XFAFF) or (cp >= 0X2_F800 and cp <= 0X2_FA1F) # ): # return True return False def UpperCamelCase_( snake_case : str ): '''simple docstring''' for char in word: snake_case_ = ord(snake_case ) if not _is_chinese_char(snake_case ): return 0 return 1 def UpperCamelCase_( snake_case : List[str] ): '''simple docstring''' snake_case_ = set() for token in tokens: snake_case_ = len(snake_case ) > 1 and is_chinese(snake_case ) if chinese_word: word_set.add(snake_case ) snake_case_ = list(snake_case ) return word_list def UpperCamelCase_( snake_case : List[str] , snake_case : set() ): '''simple docstring''' if not chinese_word_set: return bert_tokens snake_case_ = max([len(snake_case ) for w in chinese_word_set] ) snake_case_ = bert_tokens snake_case_ , snake_case_ = 0, len(snake_case ) while start < end: snake_case_ = True if is_chinese(bert_word[start] ): snake_case_ = min(end - start , snake_case ) for i in range(snake_case , 1 , -1 ): snake_case_ = "".join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1 , start + i ): snake_case_ = "##" + bert_word[j] snake_case_ = start + i snake_case_ = False break if single_word: start += 1 return bert_word def UpperCamelCase_( snake_case : List[str] , snake_case : LTP , snake_case : BertTokenizer ): '''simple docstring''' snake_case_ = [] for i in range(0 , len(snake_case ) , 1_0_0 ): snake_case_ = ltp_tokenizer.pipeline(lines[i : i + 1_0_0] , tasks=["cws"] ).cws snake_case_ = [get_chinese_word(snake_case ) for r in res] ltp_res.extend(snake_case ) assert len(snake_case ) == len(snake_case ) snake_case_ = [] for i in range(0 , len(snake_case ) , 1_0_0 ): snake_case_ = bert_tokenizer(lines[i : i + 1_0_0] , add_special_tokens=snake_case , truncation=snake_case , max_length=5_1_2 ) bert_res.extend(res["input_ids"] ) assert len(snake_case ) == len(snake_case ) snake_case_ = [] for input_ids, chinese_word in zip(snake_case , snake_case ): snake_case_ = [] for id in input_ids: snake_case_ = bert_tokenizer._convert_id_to_token(snake_case ) input_tokens.append(snake_case ) snake_case_ = add_sub_symbol(snake_case , snake_case ) snake_case_ = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(snake_case ): if token[:2] == "##": snake_case_ = token[2:] # save chinese tokens' pos if len(snake_case ) == 1 and _is_chinese_char(ord(snake_case ) ): ref_id.append(snake_case ) ref_ids.append(snake_case ) assert len(snake_case ) == len(snake_case ) return ref_ids def UpperCamelCase_( snake_case : Any ): '''simple docstring''' with open(args.file_name , "r" , encoding="utf-8" ) as f: snake_case_ = f.readlines() snake_case_ = [line.strip() for line in data if len(snake_case ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' snake_case_ = LTP(args.ltp ) # faster in GPU device snake_case_ = BertTokenizer.from_pretrained(args.bert ) snake_case_ = prepare_ref(snake_case , snake_case , snake_case ) with open(args.save_path , "w" , encoding="utf-8" ) as f: snake_case_ = [json.dumps(snake_case ) + "\n" for ref in ref_ids] f.writelines(snake_case ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE : List[str] = argparse.ArgumentParser(description="prepare_chinese_ref") parser.add_argument( "--file_name", required=False, type=str, default="./resources/chinese-demo.txt", help="file need process, same as training data in lm", ) parser.add_argument( "--ltp", required=False, type=str, default="./resources/ltp", help="resources for LTP tokenizer, usually a path", ) parser.add_argument( "--bert", required=False, type=str, default="./resources/robert", help="resources for Bert tokenizer", ) parser.add_argument( "--save_path", required=False, type=str, default="./resources/ref.txt", help="path to save res", ) _SCREAMING_SNAKE_CASE : Union[str, Any] = parser.parse_args() main(args)
85
0
'''simple docstring''' from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_tf_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_tf_available(): import tensorflow as tf a : Union[str, Any] = logging.get_logger(__name__) @dataclass class UpperCamelCase_ ( __magic_name__ ): lowercase = [ 'no_inference', 'no_cuda', 'no_tpu', 'no_speed', 'no_memory', 'no_env_print', 'no_multi_process', ] def __init__( self , **A ) -> Union[str, Any]: for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: UpperCAmelCase : Tuple = deprecated_arg[3:] UpperCAmelCase : Union[str, Any] = not kwargs.pop(A ) logger.warning( f'''{deprecated_arg} is depreciated. Please use --no-{positive_arg} or''' f''' {positive_arg}={kwargs[positive_arg]}''' ) UpperCAmelCase : List[Any] = kwargs.pop("""tpu_name""" , self.tpu_name ) UpperCAmelCase : List[str] = kwargs.pop("""device_idx""" , self.device_idx ) UpperCAmelCase : Any = kwargs.pop("""eager_mode""" , self.eager_mode ) UpperCAmelCase : Union[str, Any] = kwargs.pop("""use_xla""" , self.use_xla ) super().__init__(**A ) lowercase = field( default=__magic_name__ , metadata={'help': 'Name of TPU'} , ) lowercase = field( default=0 , metadata={'help': 'CPU / GPU device index. Defaults to 0.'} , ) lowercase = field(default=__magic_name__ , metadata={'help': 'Benchmark models in eager model.'} ) lowercase = field( default=__magic_name__ , metadata={ 'help': 'Benchmark models using XLA JIT compilation. Note that `eager_model` has to be set to `False`.' } , ) @cached_property def _lowercase( self ) -> Tuple["tf.distribute.cluster_resolver.TPUClusterResolver"]: requires_backends(self , ["""tf"""] ) UpperCAmelCase : Tuple = None if self.tpu: try: if self.tpu_name: UpperCAmelCase : Optional[int] = tf.distribute.cluster_resolver.TPUClusterResolver(self.tpu_name ) else: UpperCAmelCase : Optional[Any] = tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: UpperCAmelCase : Tuple = None return tpu @cached_property def _lowercase( self ) -> Tuple["tf.distribute.Strategy", "tf.distribute.cluster_resolver.TPUClusterResolver"]: requires_backends(self , ["""tf"""] ) if self.is_tpu: tf.config.experimental_connect_to_cluster(self._setup_tpu ) tf.tpu.experimental.initialize_tpu_system(self._setup_tpu ) UpperCAmelCase : List[Any] = tf.distribute.TPUStrategy(self._setup_tpu ) else: # currently no multi gpu is allowed if self.is_gpu: # TODO: Currently only single GPU is supported tf.config.set_visible_devices(self.gpu_list[self.device_idx] , """GPU""" ) UpperCAmelCase : Tuple = tf.distribute.OneDeviceStrategy(device=f'''/gpu:{self.device_idx}''' ) else: tf.config.set_visible_devices([] , """GPU""" ) # disable GPU UpperCAmelCase : Union[str, Any] = tf.distribute.OneDeviceStrategy(device=f'''/cpu:{self.device_idx}''' ) return strategy @property def _lowercase( self ) -> bool: requires_backends(self , ["""tf"""] ) return self._setup_tpu is not None @property def _lowercase( self ) -> "tf.distribute.Strategy": requires_backends(self , ["""tf"""] ) return self._setup_strategy @property def _lowercase( self ) -> Dict: requires_backends(self , ["""tf"""] ) return tf.config.list_physical_devices("""GPU""" ) @property def _lowercase( self ) -> int: requires_backends(self , ["""tf"""] ) if self.cuda: return len(self.gpu_list ) return 0 @property def _lowercase( self ) -> bool: return self.n_gpu > 0
360
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType a : int = logging.get_logger(__name__) a : int = { """openai/whisper-base""": """https://huggingface.co/openai/whisper-base/resolve/main/config.json""", } # fmt: off a : Tuple = [ 1, 2, 7, 8, 9, 1_0, 1_4, 2_5, 2_6, 2_7, 2_8, 2_9, 3_1, 5_8, 5_9, 6_0, 6_1, 6_2, 6_3, 9_0, 9_1, 9_2, 9_3, 3_5_7, 3_6_6, 4_3_8, 5_3_2, 6_8_5, 7_0_5, 7_9_6, 9_3_0, 1_0_5_8, 1_2_2_0, 1_2_6_7, 1_2_7_9, 1_3_0_3, 1_3_4_3, 1_3_7_7, 1_3_9_1, 1_6_3_5, 1_7_8_2, 1_8_7_5, 2_1_6_2, 2_3_6_1, 2_4_8_8, 3_4_6_7, 4_0_0_8, 4_2_1_1, 4_6_0_0, 4_8_0_8, 5_2_9_9, 5_8_5_5, 6_3_2_9, 7_2_0_3, 9_6_0_9, 9_9_5_9, 1_0_5_6_3, 1_0_7_8_6, 1_1_4_2_0, 1_1_7_0_9, 1_1_9_0_7, 1_3_1_6_3, 1_3_6_9_7, 1_3_7_0_0, 1_4_8_0_8, 1_5_3_0_6, 1_6_4_1_0, 1_6_7_9_1, 1_7_9_9_2, 1_9_2_0_3, 1_9_5_1_0, 2_0_7_2_4, 2_2_3_0_5, 2_2_9_3_5, 2_7_0_0_7, 3_0_1_0_9, 3_0_4_2_0, 3_3_4_0_9, 3_4_9_4_9, 4_0_2_8_3, 4_0_4_9_3, 4_0_5_4_9, 4_7_2_8_2, 4_9_1_4_6, 5_0_2_5_7, 5_0_3_5_9, 5_0_3_6_0, 5_0_3_6_1 ] a : Optional[int] = [ 1, 2, 7, 8, 9, 1_0, 1_4, 2_5, 2_6, 2_7, 2_8, 2_9, 3_1, 5_8, 5_9, 6_0, 6_1, 6_2, 6_3, 9_0, 9_1, 9_2, 9_3, 3_5_9, 5_0_3, 5_2_2, 5_4_2, 8_7_3, 8_9_3, 9_0_2, 9_1_8, 9_2_2, 9_3_1, 1_3_5_0, 1_8_5_3, 1_9_8_2, 2_4_6_0, 2_6_2_7, 3_2_4_6, 3_2_5_3, 3_2_6_8, 3_5_3_6, 3_8_4_6, 3_9_6_1, 4_1_8_3, 4_6_6_7, 6_5_8_5, 6_6_4_7, 7_2_7_3, 9_0_6_1, 9_3_8_3, 1_0_4_2_8, 1_0_9_2_9, 1_1_9_3_8, 1_2_0_3_3, 1_2_3_3_1, 1_2_5_6_2, 1_3_7_9_3, 1_4_1_5_7, 1_4_6_3_5, 1_5_2_6_5, 1_5_6_1_8, 1_6_5_5_3, 1_6_6_0_4, 1_8_3_6_2, 1_8_9_5_6, 2_0_0_7_5, 2_1_6_7_5, 2_2_5_2_0, 2_6_1_3_0, 2_6_1_6_1, 2_6_4_3_5, 2_8_2_7_9, 2_9_4_6_4, 3_1_6_5_0, 3_2_3_0_2, 3_2_4_7_0, 3_6_8_6_5, 4_2_8_6_3, 4_7_4_2_5, 4_9_8_7_0, 5_0_2_5_4, 5_0_2_5_8, 5_0_3_6_0, 5_0_3_6_1, 5_0_3_6_2 ] class UpperCamelCase_ ( __magic_name__ ): lowercase = 'whisper' lowercase = ['past_key_values'] lowercase = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , A=51865 , A=80 , A=6 , A=4 , A=6 , A=4 , A=1536 , A=1536 , A=0.0 , A=0.0 , A=50257 , A=True , A=True , A="gelu" , A=256 , A=0.0 , A=0.0 , A=0.0 , A=0.0_2 , A=False , A=1500 , A=448 , A=50256 , A=50256 , A=50256 , A=None , A=[220, 50256] , A=False , A=256 , A=False , A=0.0_5 , A=10 , A=2 , A=0.0 , A=10 , A=0 , A=7 , **A , ) -> Optional[Any]: UpperCAmelCase : str = vocab_size UpperCAmelCase : Union[str, Any] = num_mel_bins UpperCAmelCase : Tuple = d_model UpperCAmelCase : Optional[int] = encoder_layers UpperCAmelCase : List[str] = encoder_attention_heads UpperCAmelCase : Optional[int] = decoder_layers UpperCAmelCase : int = decoder_attention_heads UpperCAmelCase : Optional[int] = decoder_ffn_dim UpperCAmelCase : Union[str, Any] = encoder_ffn_dim UpperCAmelCase : List[str] = dropout UpperCAmelCase : Optional[Any] = attention_dropout UpperCAmelCase : Optional[Any] = activation_dropout UpperCAmelCase : Optional[Any] = activation_function UpperCAmelCase : Optional[Any] = init_std UpperCAmelCase : int = encoder_layerdrop UpperCAmelCase : Dict = decoder_layerdrop UpperCAmelCase : Optional[int] = use_cache UpperCAmelCase : List[str] = encoder_layers UpperCAmelCase : Optional[int] = scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase : Union[str, Any] = max_source_positions UpperCAmelCase : Tuple = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. UpperCAmelCase : List[str] = classifier_proj_size UpperCAmelCase : Optional[Any] = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCAmelCase : Optional[Any] = apply_spec_augment UpperCAmelCase : int = mask_time_prob UpperCAmelCase : int = mask_time_length UpperCAmelCase : Dict = mask_time_min_masks UpperCAmelCase : List[str] = mask_feature_prob UpperCAmelCase : Optional[int] = mask_feature_length UpperCAmelCase : int = mask_feature_min_masks UpperCAmelCase : List[Any] = median_filter_width super().__init__( pad_token_id=A , bos_token_id=A , eos_token_id=A , is_encoder_decoder=A , decoder_start_token_id=A , suppress_tokens=A , begin_suppress_tokens=A , **A , ) class UpperCamelCase_ ( __magic_name__ ): @property def _lowercase( self ) -> Mapping[str, Mapping[int, str]]: UpperCAmelCase : str = OrderedDict( [ ("""input_features""", {0: """batch""", 1: """feature_size""", 2: """encoder_sequence"""}), ] ) if self.use_past: UpperCAmelCase : List[Any] = {0: """batch"""} else: UpperCAmelCase : Dict = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(A , direction="""inputs""" ) return common_inputs def _lowercase( self , A , A = -1 , A = -1 , A = False , A = None , A = 22050 , A = 5.0 , A = 220 , ) -> Mapping[str, Any]: UpperCAmelCase : Optional[int] = OrderedDict() UpperCAmelCase : Any = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=A , framework=A , sampling_rate=A , time_duration=A , frequency=A , ) UpperCAmelCase : List[str] = encoder_inputs["""input_features"""].shape[2] UpperCAmelCase : List[Any] = encoder_sequence_length // 2 if self.use_past else seq_length UpperCAmelCase : Any = super().generate_dummy_inputs( preprocessor.tokenizer , A , A , A , A ) UpperCAmelCase : List[str] = encoder_inputs.pop("""input_features""" ) UpperCAmelCase : Any = decoder_inputs.pop("""decoder_input_ids""" ) if "past_key_values" in decoder_inputs: UpperCAmelCase : Union[str, Any] = decoder_inputs.pop("""past_key_values""" ) return dummy_inputs @property def _lowercase( self ) -> float: return 1e-3
338
0
'''simple docstring''' import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def snake_case_ ( lowerCAmelCase_ )-> Any: '''simple docstring''' _UpperCAmelCase ,_UpperCAmelCase : Tuple = image.size _UpperCAmelCase ,_UpperCAmelCase : Dict = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 _UpperCAmelCase : int = image.resize((w, h) , resample=PIL_INTERPOLATION["""lanczos"""] ) _UpperCAmelCase : Any = np.array(snake_case__ ).astype(np.floataa ) / 255.0 _UpperCAmelCase : Optional[int] = image[None].transpose(0 , 3 , 1 , 2 ) _UpperCAmelCase : List[str] = torch.from_numpy(snake_case__ ) return 2.0 * image - 1.0 class lowercase ( A__ ): """simple docstring""" def __init__( self ,a_ ,a_ ,a_ ,) -> Optional[int]: super().__init__() self.register_modules(vqvae=__UpperCAmelCase ,unet=__UpperCAmelCase ,scheduler=__UpperCAmelCase ) @torch.no_grad() def __call__( self ,a_ = None ,a_ = 1 ,a_ = 100 ,a_ = 0.0 ,a_ = None ,a_ = "pil" ,a_ = True ,) -> Union[Tuple, ImagePipelineOutput]: if isinstance(__UpperCAmelCase ,PIL.Image.Image ): _UpperCAmelCase : str = 1 elif isinstance(__UpperCAmelCase ,torch.Tensor ): _UpperCAmelCase : str = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(__UpperCAmelCase )}''' ) if isinstance(__UpperCAmelCase ,PIL.Image.Image ): _UpperCAmelCase : Tuple = preprocess(__UpperCAmelCase ) _UpperCAmelCase ,_UpperCAmelCase : str = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image _UpperCAmelCase : Optional[int] = (batch_size, self.unet.config.in_channels // 2, height, width) _UpperCAmelCase : List[Any] = next(self.unet.parameters() ).dtype _UpperCAmelCase : List[Any] = randn_tensor(__UpperCAmelCase ,generator=__UpperCAmelCase ,device=self.device ,dtype=__UpperCAmelCase ) _UpperCAmelCase : Optional[int] = image.to(device=self.device ,dtype=__UpperCAmelCase ) # set timesteps and move to the correct device self.scheduler.set_timesteps(__UpperCAmelCase ,device=self.device ) _UpperCAmelCase : Tuple = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler _UpperCAmelCase : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] _UpperCAmelCase : Dict = """eta""" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) _UpperCAmelCase : int = {} if accepts_eta: _UpperCAmelCase : List[Any] = eta for t in self.progress_bar(__UpperCAmelCase ): # concat latents and low resolution image in the channel dimension. _UpperCAmelCase : List[str] = torch.cat([latents, image] ,dim=1 ) _UpperCAmelCase : Optional[int] = self.scheduler.scale_model_input(__UpperCAmelCase ,__UpperCAmelCase ) # predict the noise residual _UpperCAmelCase : int = self.unet(__UpperCAmelCase ,__UpperCAmelCase ).sample # compute the previous noisy sample x_t -> x_t-1 _UpperCAmelCase : Optional[Any] = self.scheduler.step(__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase ,**__UpperCAmelCase ).prev_sample # decode the image latents with the VQVAE _UpperCAmelCase : Union[str, Any] = self.vqvae.decode(__UpperCAmelCase ).sample _UpperCAmelCase : str = torch.clamp(__UpperCAmelCase ,-1.0 ,1.0 ) _UpperCAmelCase : Union[str, Any] = image / 2 + 0.5 _UpperCAmelCase : int = image.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": _UpperCAmelCase : Union[str, Any] = self.numpy_to_pil(__UpperCAmelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__UpperCAmelCase )
215
"""simple docstring""" from math import ceil def A ( snake_case__ , snake_case__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = list(range(0 , snake_case__ ) ) SCREAMING_SNAKE_CASE__ = [item for sublist in list(device_map.values() ) for item in sublist] # Duplicate check SCREAMING_SNAKE_CASE__ = [] for i in device_map_blocks: if device_map_blocks.count(snake_case__ ) > 1 and i not in duplicate_blocks: duplicate_blocks.append(snake_case__ ) # Missing blocks SCREAMING_SNAKE_CASE__ = [i for i in blocks if i not in device_map_blocks] SCREAMING_SNAKE_CASE__ = [i for i in device_map_blocks if i not in blocks] if len(snake_case__ ) != 0: raise ValueError( """Duplicate attention blocks specified in device_map. Attention blocks must be specified to one device.""" """ These attention blocks were specified more than once: """ + str(snake_case__ ) ) if len(snake_case__ ) != 0: raise ValueError( """There are attention blocks for this model that are not specified in the device_map. Add these attention """ """blocks to a device on the device_map: """ + str(snake_case__ ) ) if len(snake_case__ ) != 0: raise ValueError( """The device_map contains more attention blocks than this model has. Remove these from the device_map:""" + str(snake_case__ ) ) def A ( snake_case__ , snake_case__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = list(range(snake_case__ ) ) SCREAMING_SNAKE_CASE__ = int(ceil(n_layers / len(snake_case__ ) ) ) SCREAMING_SNAKE_CASE__ = [layers[i : i + n_blocks] for i in range(0 , snake_case__ , snake_case__ )] return dict(zip(snake_case__ , snake_case__ ) )
165
0
"""simple docstring""" def UpperCAmelCase__ ( lowerCAmelCase__ :Optional[int] = 1_0 , lowerCAmelCase__ :Tuple = 1_0_0_0 , lowerCAmelCase__ :str = True ) -> int: '''simple docstring''' assert ( isinstance(__snake_case , __snake_case ) and isinstance(__snake_case , __snake_case ) and isinstance(__snake_case , __snake_case ) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError("""Invalid value for min_val or max_val (min_value < max_value)""" ) return min_val if option else max_val def UpperCAmelCase__ ( lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Optional[Any] ) -> int: '''simple docstring''' return int((number_a + number_a) / 2 ) def UpperCAmelCase__ ( lowerCAmelCase__ :int , lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Dict ) -> None: '''simple docstring''' assert ( isinstance(__snake_case , __snake_case ) and isinstance(__snake_case , __snake_case ) and isinstance(__snake_case , __snake_case ) ), 'argument values must be type of "int"' if lower > higher: raise ValueError("""argument value for lower and higher must be(lower > higher)""" ) if not lower < to_guess < higher: raise ValueError( """guess value must be within the range of lower and higher value""" ) def answer(lowerCAmelCase__ :List[str] ) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print("""started...""" ) lowercase = lower lowercase = higher lowercase = [] while True: lowercase = get_avg(__snake_case , __snake_case ) last_numbers.append(__snake_case ) if answer(__snake_case ) == "low": lowercase = number elif answer(__snake_case ) == "high": lowercase = number else: break print(f'guess the number : {last_numbers[-1]}' ) print(f'details : {last_numbers!s}' ) def UpperCAmelCase__ ( ) -> None: '''simple docstring''' lowercase = int(input("""Enter lower value : """ ).strip() ) lowercase = int(input("""Enter high value : """ ).strip() ) lowercase = int(input("""Enter value to guess : """ ).strip() ) guess_the_number(__snake_case , __snake_case , __snake_case ) if __name__ == "__main__": main()
363
"""simple docstring""" import copy from typing import Dict, List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __lowerCAmelCase : Tuple ={ """facebook/mask2former-swin-small-coco-instance""": ( """https://huggingface.co/facebook/mask2former-swin-small-coco-instance/blob/main/config.json""" ) # See all Mask2Former models at https://huggingface.co/models?filter=mask2former } __lowerCAmelCase : Optional[Any] =logging.get_logger(__name__) class _A ( lowerCAmelCase ): snake_case__ : Dict = 'mask2former' snake_case__ : Union[str, Any] = ['swin'] snake_case__ : Any = {'hidden_size': 'hidden_dim'} def __init__( self , __lowerCAmelCase = None , __lowerCAmelCase = 256 , __lowerCAmelCase = 256 , __lowerCAmelCase = 256 , __lowerCAmelCase = 1024 , __lowerCAmelCase = "relu" , __lowerCAmelCase = 6 , __lowerCAmelCase = 10 , __lowerCAmelCase = 8 , __lowerCAmelCase = 0.0 , __lowerCAmelCase = 2048 , __lowerCAmelCase = False , __lowerCAmelCase = False , __lowerCAmelCase = 4 , __lowerCAmelCase = 255 , __lowerCAmelCase = 100 , __lowerCAmelCase = 0.1 , __lowerCAmelCase = 2.0 , __lowerCAmelCase = 5.0 , __lowerCAmelCase = 5.0 , __lowerCAmelCase = 1_2544 , __lowerCAmelCase = 3.0 , __lowerCAmelCase = 0.7_5 , __lowerCAmelCase = 0.0_2 , __lowerCAmelCase = 1.0 , __lowerCAmelCase = True , __lowerCAmelCase = [4, 8, 16, 32] , __lowerCAmelCase = None , **__lowerCAmelCase , ): """simple docstring""" if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `Swin` backbone.""" ) lowercase = CONFIG_MAPPING["""swin"""]( image_size=224 , in_channels=3 , patch_size=4 , embed_dim=96 , depths=[2, 2, 18, 2] , num_heads=[3, 6, 12, 24] , window_size=7 , drop_path_rate=0.3 , use_absolute_embeddings=__lowerCAmelCase , out_features=["""stage1""", """stage2""", """stage3""", """stage4"""] , ) if isinstance(__lowerCAmelCase , __lowerCAmelCase ): lowercase = backbone_config.pop("""model_type""" ) lowercase = CONFIG_MAPPING[backbone_model_type] lowercase = config_class.from_dict(__lowerCAmelCase ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( f'Backbone {backbone_config.model_type} is not a supported model and may not be compatible with Mask2Former. ' f'Supported model types: {",".join(self.backbones_supported )}' ) lowercase = backbone_config lowercase = feature_size lowercase = mask_feature_size lowercase = hidden_dim lowercase = encoder_feedforward_dim lowercase = activation_function lowercase = encoder_layers lowercase = decoder_layers lowercase = num_attention_heads lowercase = dropout lowercase = dim_feedforward lowercase = pre_norm lowercase = enforce_input_projection lowercase = common_stride lowercase = ignore_value lowercase = num_queries lowercase = no_object_weight lowercase = class_weight lowercase = mask_weight lowercase = dice_weight lowercase = train_num_points lowercase = oversample_ratio lowercase = importance_sample_ratio lowercase = init_std lowercase = init_xavier_std lowercase = use_auxiliary_loss lowercase = feature_strides lowercase = output_auxiliary_logits lowercase = decoder_layers super().__init__(**__lowerCAmelCase ) @classmethod def A__ ( cls , __lowerCAmelCase , **__lowerCAmelCase ): """simple docstring""" return cls( backbone_config=__lowerCAmelCase , **__lowerCAmelCase , ) def A__ ( self ): """simple docstring""" lowercase = copy.deepcopy(self.__dict__ ) lowercase = self.backbone_config.to_dict() lowercase = self.__class__.model_type return output
32
0
"""simple docstring""" import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets __lowerCAmelCase : List[Any] ='''\ @inproceedings{lin-2004-rouge, title = "{ROUGE}: A Package for Automatic Evaluation of Summaries", author = "Lin, Chin-Yew", booktitle = "Text Summarization Branches Out", month = jul, year = "2004", address = "Barcelona, Spain", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W04-1013", pages = "74--81", } ''' __lowerCAmelCase : Dict ='''\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge ''' __lowerCAmelCase : Any =''' Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring, `"rougeL"`: Longest common subsequence based scoring. `"rougeLSum"`: rougeLsum splits text using `"\n"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric(\'rouge\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\'] >>> print(results["rouge1"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results["rouge1"].mid.fmeasure) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _A ( datasets.Metric ): def A__ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/google-research/google-research/tree/master/rouge"""] , reference_urls=[ """https://en.wikipedia.org/wiki/ROUGE_(metric)""", """https://github.com/google-research/google-research/tree/master/rouge""", ] , ) def A__ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None , __lowerCAmelCase=True , __lowerCAmelCase=False ): """simple docstring""" if rouge_types is None: lowercase = ['''rouge1''', '''rouge2''', '''rougeL''', '''rougeLsum'''] lowercase = rouge_scorer.RougeScorer(rouge_types=__lowerCAmelCase , use_stemmer=__lowerCAmelCase ) if use_aggregator: lowercase = scoring.BootstrapAggregator() else: lowercase = [] for ref, pred in zip(__lowerCAmelCase , __lowerCAmelCase ): lowercase = scorer.score(__lowerCAmelCase , __lowerCAmelCase ) if use_aggregator: aggregator.add_scores(__lowerCAmelCase ) else: scores.append(__lowerCAmelCase ) if use_aggregator: lowercase = aggregator.aggregate() else: lowercase = {} for key in scores[0]: lowercase = [score[key] for score in scores] return result
197
"""simple docstring""" def lowercase ( __snake_case : Optional[int] ): lowercase_ : int = 0 lowercase_ : Optional[Any] = len(__snake_case ) for i in range(n - 1 ): for j in range(i + 1 , __snake_case ): if arr[i] > arr[j]: num_inversions += 1 return num_inversions def lowercase ( __snake_case : str ): if len(__snake_case ) <= 1: return arr, 0 lowercase_ : Optional[Any] = len(__snake_case ) // 2 lowercase_ : List[Any] = arr[0:mid] lowercase_ : Union[str, Any] = arr[mid:] lowercase_ , lowercase_ : Tuple = count_inversions_recursive(__snake_case ) lowercase_ , lowercase_ : List[Any] = count_inversions_recursive(__snake_case ) lowercase_ , lowercase_ : List[Any] = _count_cross_inversions(__snake_case , __snake_case ) lowercase_ : List[Any] = inversion_p + inversions_q + cross_inversions return c, num_inversions def lowercase ( __snake_case : str , __snake_case : Optional[int] ): lowercase_ : Optional[Any] = [] lowercase_ : Any = 0 while i < len(__snake_case ) and j < len(__snake_case ): if p[i] > q[j]: # if P[1] > Q[j], then P[k] > Q[k] for all i < k <= len(P) # These are all inversions. The claim emerges from the # property that P is sorted. num_inversion += len(__snake_case ) - i r.append(q[j] ) j += 1 else: r.append(p[i] ) i += 1 if i < len(__snake_case ): r.extend(p[i:] ) else: r.extend(q[j:] ) return r, num_inversion def lowercase ( ): lowercase_ : Union[str, Any] = [1_0, 2, 1, 5, 5, 2, 1_1] # this arr has 8 inversions: # (10, 2), (10, 1), (10, 5), (10, 5), (10, 2), (2, 1), (5, 2), (5, 2) lowercase_ : int = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : Dict = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 8 print('''number of inversions = ''' , __snake_case ) # testing an array with zero inversion (a sorted arr_1) arr_a.sort() lowercase_ : Dict = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : Dict = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 0 print('''number of inversions = ''' , __snake_case ) # an empty list should also have zero inversions lowercase_ : List[Any] = [] lowercase_ : Any = count_inversions_bf(__snake_case ) lowercase_ , lowercase_ : List[str] = count_inversions_recursive(__snake_case ) assert num_inversions_bf == num_inversions_recursive == 0 print('''number of inversions = ''' , __snake_case ) if __name__ == "__main__": main()
33
0
from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, is_vision_available, logging if is_vision_available(): import PIL if is_torch_available(): import torch UpperCamelCase__ = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( _a ): snake_case : List[Any] = ["""pixel_values"""] def __init__( self , __lowerCAmelCase = True , __lowerCAmelCase = None , __lowerCAmelCase = PILImageResampling.BILINEAR , __lowerCAmelCase = True , __lowerCAmelCase = 1 / 255 , __lowerCAmelCase = True , __lowerCAmelCase = None , __lowerCAmelCase = True , **__lowerCAmelCase , ): super().__init__(**__lowerCAmelCase ) UpperCamelCase__ = size if size is not None else {"""shortest_edge""": 224} UpperCamelCase__ = get_size_dict(__lowerCAmelCase , default_to_square=__lowerCAmelCase ) UpperCamelCase__ = crop_size if crop_size is not None else {"""height""": 256, """width""": 256} UpperCamelCase__ = get_size_dict(__lowerCAmelCase , param_name="""crop_size""" ) UpperCamelCase__ = do_resize UpperCamelCase__ = size UpperCamelCase__ = resample UpperCamelCase__ = do_rescale UpperCamelCase__ = rescale_factor UpperCamelCase__ = do_center_crop UpperCamelCase__ = crop_size UpperCamelCase__ = do_flip_channel_order def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = PIL.Image.BILINEAR , __lowerCAmelCase = None , **__lowerCAmelCase , ): UpperCamelCase__ = get_size_dict(__lowerCAmelCase , default_to_square=__lowerCAmelCase ) if "shortest_edge" not in size: raise ValueError(f"""The `size` dictionary must contain the key `shortest_edge`. Got {size.keys()}""" ) UpperCamelCase__ = get_resize_output_image_size(__lowerCAmelCase , size=size["""shortest_edge"""] , default_to_square=__lowerCAmelCase ) return resize(__lowerCAmelCase , size=__lowerCAmelCase , resample=__lowerCAmelCase , data_format=__lowerCAmelCase , **__lowerCAmelCase ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = None , **__lowerCAmelCase , ): UpperCamelCase__ = get_size_dict(__lowerCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(__lowerCAmelCase , size=(size["""height"""], size["""width"""]) , data_format=__lowerCAmelCase , **__lowerCAmelCase ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = None , **__lowerCAmelCase , ): return rescale(__lowerCAmelCase , scale=__lowerCAmelCase , data_format=__lowerCAmelCase , **__lowerCAmelCase ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase = None ): return flip_channel_order(__lowerCAmelCase , data_format=__lowerCAmelCase ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = ChannelDimension.FIRST , **__lowerCAmelCase , ): UpperCamelCase__ = do_resize if do_resize is not None else self.do_resize UpperCamelCase__ = resample if resample is not None else self.resample UpperCamelCase__ = do_rescale if do_rescale is not None else self.do_rescale UpperCamelCase__ = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCamelCase__ = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCamelCase__ = ( do_flip_channel_order if do_flip_channel_order is not None else self.do_flip_channel_order ) UpperCamelCase__ = size if size is not None else self.size UpperCamelCase__ = get_size_dict(__lowerCAmelCase , default_to_square=__lowerCAmelCase ) UpperCamelCase__ = crop_size if crop_size is not None else self.crop_size UpperCamelCase__ = get_size_dict(__lowerCAmelCase , param_name="""crop_size""" ) UpperCamelCase__ = make_list_of_images(__lowerCAmelCase ) if not valid_images(__lowerCAmelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) # All transformations expect numpy arrays. UpperCamelCase__ = [to_numpy_array(__lowerCAmelCase ) for image in images] if do_resize: UpperCamelCase__ = [self.resize(image=__lowerCAmelCase , size=__lowerCAmelCase , resample=__lowerCAmelCase ) for image in images] if do_center_crop: UpperCamelCase__ = [self.center_crop(image=__lowerCAmelCase , size=__lowerCAmelCase ) for image in images] if do_rescale: UpperCamelCase__ = [self.rescale(image=__lowerCAmelCase , scale=__lowerCAmelCase ) for image in images] # the pretrained checkpoints assume images are BGR, not RGB if do_flip_channel_order: UpperCamelCase__ = [self.flip_channel_order(image=__lowerCAmelCase ) for image in images] UpperCamelCase__ = [to_channel_dimension_format(__lowerCAmelCase , __lowerCAmelCase ) for image in images] UpperCamelCase__ = {"""pixel_values""": images} return BatchFeature(data=__lowerCAmelCase , tensor_type=__lowerCAmelCase ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase = None ): UpperCamelCase__ = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(__lowerCAmelCase ) != len(__lowerCAmelCase ): raise ValueError( """Make sure that you pass in as many target sizes as the batch dimension of the logits""" ) if is_torch_tensor(__lowerCAmelCase ): UpperCamelCase__ = target_sizes.numpy() UpperCamelCase__ = [] for idx in range(len(__lowerCAmelCase ) ): UpperCamelCase__ = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="""bilinear""" , align_corners=__lowerCAmelCase ) UpperCamelCase__ = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(__lowerCAmelCase ) else: UpperCamelCase__ = logits.argmax(dim=1 ) UpperCamelCase__ = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
87
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process UpperCamelCase__ = logging.getLogger(__name__) def _UpperCamelCase (a__ :Union[str, Any] , a__ :Optional[Any] ): """simple docstring""" return (preds == labels).mean() @dataclass class __SCREAMING_SNAKE_CASE : snake_case : str = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) snake_case : Optional[str] = field( default=_a , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) snake_case : Optional[str] = field( default=_a , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) snake_case : Optional[str] = field( default=_a , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) @dataclass class __SCREAMING_SNAKE_CASE : snake_case : str = field(metadata={"""help""": """The name of the task to train on: """ + """, """.join(processors.keys() )} ) snake_case : str = field(metadata={"""help""": """Should contain the data files for the task."""} ) snake_case : int = field( default=1_28 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) snake_case : bool = field( default=_a , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) def _UpperCamelCase (): """simple docstring""" UpperCamelCase__ = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" """ --overwrite_output_dir to overcome.""" ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( """Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s""" , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("""Training/evaluation parameters %s""" , a__ ) # Set seed set_seed(training_args.seed ) try: UpperCamelCase__ = processors[data_args.task_name]() UpperCamelCase__ = processor.get_labels() UpperCamelCase__ = len(a__ ) except KeyError: raise ValueError("""Task not found: %s""" % (data_args.task_name) ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. UpperCamelCase__ = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=a__ , finetuning_task=data_args.task_name , cache_dir=model_args.cache_dir , ) UpperCamelCase__ = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) UpperCamelCase__ = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=a__ , cache_dir=model_args.cache_dir , ) # Get datasets UpperCamelCase__ = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=a__ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) UpperCamelCase__ = ( MultipleChoiceDataset( data_dir=data_args.data_dir , tokenizer=a__ , task=data_args.task_name , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def compute_metrics(a__ :EvalPrediction ) -> Dict: UpperCamelCase__ = np.argmax(p.predictions , axis=1 ) return {"acc": simple_accuracy(a__ , p.label_ids )} # Data collator UpperCamelCase__ = DataCollatorWithPadding(a__ , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer UpperCamelCase__ = Trainer( model=a__ , args=a__ , train_dataset=a__ , eval_dataset=a__ , compute_metrics=a__ , data_collator=a__ , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation UpperCamelCase__ = {} if training_args.do_eval: logger.info("""*** Evaluate ***""" ) UpperCamelCase__ = trainer.evaluate() UpperCamelCase__ = os.path.join(training_args.output_dir , """eval_results.txt""" ) if trainer.is_world_master(): with open(a__ , """w""" ) as writer: logger.info("""***** Eval results *****""" ) for key, value in result.items(): logger.info(""" %s = %s""" , a__ , a__ ) writer.write("""%s = %s\n""" % (key, value) ) results.update(a__ ) return results def _UpperCamelCase (a__ :Union[str, Any] ): """simple docstring""" main() if __name__ == "__main__": main()
87
1
lowercase : dict[tuple[int, int, int], int] = {} def A_ ( A__ , A__ , A__ ) -> int: # if we are absent twice, or late 3 consecutive days, # no further prize strings are possible if late == 3 or absent == 2: return 0 # if we have no days left, and have not failed any other rules, # we have a prize string if days == 0: return 1 # No easy solution, so now we need to do the recursive calculation # First, check if the combination is already in the cache, and # if yes, return the stored value from there since we already # know the number of possible prize strings from this point on a__ : Any = (days, absent, late) if key in cache: return cache[key] # now we calculate the three possible ways that can unfold from # this point on, depending on our attendance today # 1) if we are late (but not absent), the "absent" counter stays as # it is, but the "late" counter increases by one a__ : List[str] = _calculate(days - 1 , A__ , late + 1 ) # 2) if we are absent, the "absent" counter increases by 1, and the # "late" counter resets to 0 a__ : int = _calculate(days - 1 , absent + 1 , 0 ) # 3) if we are on time, this resets the "late" counter and keeps the # absent counter a__ : List[str] = _calculate(days - 1 , A__ , 0 ) a__ : Dict = state_late + state_absent + state_ontime a__ : Tuple = prizestrings return prizestrings def A_ ( A__ = 30 ) -> int: return _calculate(A__ , absent=0 , late=0 ) if __name__ == "__main__": print(solution())
99
"""simple docstring""" import argparse import math import os from copy import deepcopy import torch from audio_diffusion.models import DiffusionAttnUnetaD from diffusion import sampling from torch import nn from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel __UpperCamelCase = { '''gwf-440k''': { '''url''': '''https://model-server.zqevans2.workers.dev/gwf-440k.ckpt''', '''sample_rate''': 48000, '''sample_size''': 65536, }, '''jmann-small-190k''': { '''url''': '''https://model-server.zqevans2.workers.dev/jmann-small-190k.ckpt''', '''sample_rate''': 48000, '''sample_size''': 65536, }, '''jmann-large-580k''': { '''url''': '''https://model-server.zqevans2.workers.dev/jmann-large-580k.ckpt''', '''sample_rate''': 48000, '''sample_size''': 131072, }, '''maestro-uncond-150k''': { '''url''': '''https://model-server.zqevans2.workers.dev/maestro-uncond-150k.ckpt''', '''sample_rate''': 16000, '''sample_size''': 65536, }, '''unlocked-uncond-250k''': { '''url''': '''https://model-server.zqevans2.workers.dev/unlocked-uncond-250k.ckpt''', '''sample_rate''': 16000, '''sample_size''': 65536, }, '''honk-140k''': { '''url''': '''https://model-server.zqevans2.workers.dev/honk-140k.ckpt''', '''sample_rate''': 16000, '''sample_size''': 65536, }, } def lowercase (SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Optional[int] ) -> Optional[Any]: return torch.atana(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) / math.pi * 2 def lowercase (SCREAMING_SNAKE_CASE_ : Optional[int] ) -> int: SCREAMING_SNAKE_CASE = torch.sin(t * math.pi / 2 ) ** 2 SCREAMING_SNAKE_CASE = (1 - sigma**2) ** 0.5 return alpha_sigma_to_t(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) class lowerCAmelCase ( lowerCamelCase_ ): '''simple docstring''' pass class lowerCAmelCase ( nn.Module ): '''simple docstring''' def __init__( self , lowerCAmelCase__ ) -> int: super().__init__() SCREAMING_SNAKE_CASE = DiffusionAttnUnetaD(lowerCAmelCase__ , n_attn_layers=4 ) SCREAMING_SNAKE_CASE = deepcopy(self.diffusion ) SCREAMING_SNAKE_CASE = torch.quasirandom.SobolEngine(1 , scramble=lowerCAmelCase__ ) def lowercase (SCREAMING_SNAKE_CASE_ : Union[str, Any] ) -> Any: SCREAMING_SNAKE_CASE = MODELS_MAP[model_name]['url'] os.system(F'wget {url} ./' ) return F'./{model_name}.ckpt' __UpperCamelCase = { '''1''': '''resnets.0''', '''2''': '''attentions.0''', '''3''': '''resnets.1''', '''4''': '''attentions.1''', '''5''': '''resnets.2''', '''6''': '''attentions.2''', } __UpperCamelCase = { '''8''': '''resnets.0''', '''9''': '''attentions.0''', '''10''': '''resnets.1''', '''11''': '''attentions.1''', '''12''': '''resnets.2''', '''13''': '''attentions.2''', } __UpperCamelCase = { '''1''': '''resnets.0''', '''2''': '''attentions.0''', '''3''': '''resnets.1''', '''4''': '''attentions.1''', '''5''': '''resnets.2''', '''6''': '''attentions.2''', '''8''': '''resnets.3''', '''9''': '''attentions.3''', '''10''': '''resnets.4''', '''11''': '''attentions.4''', '''12''': '''resnets.5''', '''13''': '''attentions.5''', } __UpperCamelCase = { '''0''': '''resnets.0''', '''1''': '''resnets.1''', '''2''': '''resnets.2''', '''4''': '''resnets.0''', '''5''': '''resnets.1''', '''6''': '''resnets.2''', } __UpperCamelCase = { '''skip''': '''conv_skip''', '''main.0''': '''conv_1''', '''main.1''': '''group_norm_1''', '''main.3''': '''conv_2''', '''main.4''': '''group_norm_2''', } __UpperCamelCase = { '''norm''': '''group_norm''', '''qkv_proj''': ['''query''', '''key''', '''value'''], '''out_proj''': ['''proj_attn'''], } def lowercase (SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> List[Any]: if name.startswith('skip' ): return name.replace('skip' , RES_CONV_MAP['skip'] ) # name has to be of format main.{digit} if not name.startswith('main.' ): raise ValueError(F'ResConvBlock error with {name}' ) return name.replace(name[:6] , RES_CONV_MAP[name[:6]] ) def lowercase (SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> Dict: for key, value in ATTN_MAP.items(): if name.startswith(SCREAMING_SNAKE_CASE_ ) and not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): return name.replace(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) elif name.startswith(SCREAMING_SNAKE_CASE_ ): return [name.replace(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for v in value] raise ValueError(F'Attn error with {name}' ) def lowercase (SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : Tuple=13 ) -> Union[str, Any]: SCREAMING_SNAKE_CASE = input_string if string.split('.' )[0] == "timestep_embed": return string.replace('timestep_embed' , 'time_proj' ) SCREAMING_SNAKE_CASE = 0 if string.startswith('net.3.' ): depth += 1 SCREAMING_SNAKE_CASE = string[6:] elif string.startswith('net.' ): SCREAMING_SNAKE_CASE = string[4:] while string.startswith('main.7.' ): depth += 1 SCREAMING_SNAKE_CASE = string[7:] if string.startswith('main.' ): SCREAMING_SNAKE_CASE = string[5:] # mid block if string[:2].isdigit(): SCREAMING_SNAKE_CASE = string[:2] SCREAMING_SNAKE_CASE = string[2:] else: SCREAMING_SNAKE_CASE = string[0] SCREAMING_SNAKE_CASE = string[1:] if depth == max_depth: SCREAMING_SNAKE_CASE = MID_NUM_TO_LAYER[layer_num] SCREAMING_SNAKE_CASE = 'mid_block' elif depth > 0 and int(SCREAMING_SNAKE_CASE_ ) < 7: SCREAMING_SNAKE_CASE = DOWN_NUM_TO_LAYER[layer_num] SCREAMING_SNAKE_CASE = F'down_blocks.{depth}' elif depth > 0 and int(SCREAMING_SNAKE_CASE_ ) > 7: SCREAMING_SNAKE_CASE = UP_NUM_TO_LAYER[layer_num] SCREAMING_SNAKE_CASE = F'up_blocks.{max_depth - depth - 1}' elif depth == 0: SCREAMING_SNAKE_CASE = DEPTH_0_TO_LAYER[layer_num] SCREAMING_SNAKE_CASE = F'up_blocks.{max_depth - 1}' if int(SCREAMING_SNAKE_CASE_ ) > 3 else 'down_blocks.0' if not string_left.startswith('.' ): raise ValueError(F'Naming error with {input_string} and string_left: {string_left}.' ) SCREAMING_SNAKE_CASE = string_left[1:] if "resnets" in new_layer: SCREAMING_SNAKE_CASE = convert_resconv_naming(SCREAMING_SNAKE_CASE_ ) elif "attentions" in new_layer: SCREAMING_SNAKE_CASE = convert_attn_naming(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = new_string_left if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): SCREAMING_SNAKE_CASE = prefix + '.' + new_layer + '.' + string_left else: SCREAMING_SNAKE_CASE = [prefix + '.' + new_layer + '.' + s for s in string_left] return new_string def lowercase (SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> Dict: SCREAMING_SNAKE_CASE = {} for k, v in state_dict.items(): if k.endswith('kernel' ): # up- and downsample layers, don't have trainable weights continue SCREAMING_SNAKE_CASE = rename(SCREAMING_SNAKE_CASE_ ) # check if we need to transform from Conv => Linear for attention if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): SCREAMING_SNAKE_CASE = transform_conv_attns(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) else: SCREAMING_SNAKE_CASE = v return new_state_dict def lowercase (SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Union[str, Any] ) -> str: if len(SCREAMING_SNAKE_CASE_ ) == 1: if len(v.shape ) == 3: # weight SCREAMING_SNAKE_CASE = v[:, :, 0] else: # bias SCREAMING_SNAKE_CASE = v else: # qkv matrices SCREAMING_SNAKE_CASE = v.shape[0] SCREAMING_SNAKE_CASE = trippled_shape // 3 for i in range(3 ): if len(v.shape ) == 3: SCREAMING_SNAKE_CASE = v[i * single_shape : (i + 1) * single_shape, :, 0] else: SCREAMING_SNAKE_CASE = v[i * single_shape : (i + 1) * single_shape] return new_state_dict def lowercase (SCREAMING_SNAKE_CASE_ : Union[str, Any] ) -> List[Any]: SCREAMING_SNAKE_CASE = torch.device('cuda' if torch.cuda.is_available() else 'cpu' ) SCREAMING_SNAKE_CASE = args.model_path.split('/' )[-1].split('.' )[0] if not os.path.isfile(args.model_path ): assert ( model_name == args.model_path ), F'Make sure to provide one of the official model names {MODELS_MAP.keys()}' SCREAMING_SNAKE_CASE = download(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = MODELS_MAP[model_name]['sample_rate'] SCREAMING_SNAKE_CASE = MODELS_MAP[model_name]['sample_size'] SCREAMING_SNAKE_CASE = Object() SCREAMING_SNAKE_CASE = sample_size SCREAMING_SNAKE_CASE = sample_rate SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = UNetaDModel(sample_size=SCREAMING_SNAKE_CASE_ , sample_rate=SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = diffusers_model.state_dict() SCREAMING_SNAKE_CASE = DiffusionUncond(SCREAMING_SNAKE_CASE_ ) orig_model.load_state_dict(torch.load(args.model_path , map_location=SCREAMING_SNAKE_CASE_ )['state_dict'] ) SCREAMING_SNAKE_CASE = orig_model.diffusion_ema.eval() SCREAMING_SNAKE_CASE = orig_model.state_dict() SCREAMING_SNAKE_CASE = rename_orig_weights(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = set(renamed_state_dict.keys() ) - set(diffusers_state_dict.keys() ) SCREAMING_SNAKE_CASE = set(diffusers_state_dict.keys() ) - set(renamed_state_dict.keys() ) assert len(SCREAMING_SNAKE_CASE_ ) == 0, F'Problem with {renamed_minus_diffusers}' assert all(k.endswith('kernel' ) for k in list(SCREAMING_SNAKE_CASE_ ) ), F'Problem with {diffusers_minus_renamed}' for key, value in renamed_state_dict.items(): assert ( diffusers_state_dict[key].squeeze().shape == value.squeeze().shape ), F'Shape for {key} doesn\'t match. Diffusers: {diffusers_state_dict[key].shape} vs. {value.shape}' if key == "time_proj.weight": SCREAMING_SNAKE_CASE = value.squeeze() SCREAMING_SNAKE_CASE = value diffusers_model.load_state_dict(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = 1_00 SCREAMING_SNAKE_CASE = 33 SCREAMING_SNAKE_CASE = IPNDMScheduler(num_train_timesteps=SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = torch.randn([1, 2, config.sample_size] , generator=SCREAMING_SNAKE_CASE_ ).to(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = torch.linspace(1 , 0 , steps + 1 , device=SCREAMING_SNAKE_CASE_ )[:-1] SCREAMING_SNAKE_CASE = get_crash_schedule(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = DanceDiffusionPipeline(unet=SCREAMING_SNAKE_CASE_ , scheduler=SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = torch.manual_seed(33 ) SCREAMING_SNAKE_CASE = pipe(num_inference_steps=SCREAMING_SNAKE_CASE_ , generator=SCREAMING_SNAKE_CASE_ ).audios SCREAMING_SNAKE_CASE = sampling.iplms_sample(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , {} ) SCREAMING_SNAKE_CASE = generated.clamp(-1 , 1 ) SCREAMING_SNAKE_CASE = (generated - audio).abs().sum() SCREAMING_SNAKE_CASE = (generated - audio).abs().max() if args.save: pipe.save_pretrained(args.checkpoint_path ) print('Diff sum' , SCREAMING_SNAKE_CASE_ ) print('Diff max' , SCREAMING_SNAKE_CASE_ ) assert diff_max < 1E-3, F'Diff max: {diff_max} is too much :-/' print(F'Conversion for {model_name} successful!' ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''--model_path''', default=None, type=str, required=True, help='''Path to the model to convert.''') parser.add_argument( '''--save''', default=True, type=bool, required=False, help='''Whether to save the converted model or not.''' ) parser.add_argument('''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the output model.''') __UpperCamelCase = parser.parse_args() main(args)
113
0
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake _SCREAMING_SNAKE_CASE = numpy.array([0, 0]) _SCREAMING_SNAKE_CASE = numpy.array([0.5, 0.8_6_6_0_2_5_4]) _SCREAMING_SNAKE_CASE = numpy.array([1, 0]) _SCREAMING_SNAKE_CASE = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def _lowerCAmelCase ( lowerCamelCase_ : list[numpy.ndarray] , lowerCamelCase_ : int ): __lowercase = initial_vectors for _ in range(lowerCamelCase_ ): __lowercase = iteration_step(lowerCamelCase_ ) return vectors def _lowerCAmelCase ( lowerCamelCase_ : list[numpy.ndarray] ): __lowercase = [] for i, start_vector in enumerate(vectors[:-1] ): __lowercase = vectors[i + 1] new_vectors.append(lowerCamelCase_ ) __lowercase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 6_0 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def _lowerCAmelCase ( lowerCamelCase_ : numpy.ndarray , lowerCamelCase_ : float ): __lowercase = numpy.radians(lowerCamelCase_ ) __lowercase , __lowercase = numpy.cos(lowerCamelCase_ ), numpy.sin(lowerCamelCase_ ) __lowercase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(lowerCamelCase_ , lowerCamelCase_ ) def _lowerCAmelCase ( lowerCamelCase_ : list[numpy.ndarray] ): __lowercase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() __lowercase , __lowercase = zip(*lowerCamelCase_ ) plt.plot(lowerCamelCase_ , lowerCamelCase_ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() _SCREAMING_SNAKE_CASE = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
217
'''simple docstring''' import argparse import os from pathlib import Path import fairseq import torch from packaging import version from torch import nn from transformers import ( BartConfig, BartForConditionalGeneration, BartForSequenceClassification, BartModel, BartTokenizer, ) from transformers.utils import logging _SCREAMING_SNAKE_CASE = ['''bart.large''', '''bart.large.mnli''', '''bart.large.cnn''', '''bart_xsum/model.pt'''] _SCREAMING_SNAKE_CASE = {'''bart.large''': BartModel, '''bart.large.mnli''': BartForSequenceClassification} if version.parse(fairseq.__version__) < version.parse('''0.9.0'''): raise Exception('''requires fairseq >= 0.9.0''') logging.set_verbosity_info() _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = ''' Hello world! cécé herlolip''' _SCREAMING_SNAKE_CASE = [ ('''model.classification_heads.mnli.dense.weight''', '''classification_head.dense.weight'''), ('''model.classification_heads.mnli.dense.bias''', '''classification_head.dense.bias'''), ('''model.classification_heads.mnli.out_proj.weight''', '''classification_head.out_proj.weight'''), ('''model.classification_heads.mnli.out_proj.bias''', '''classification_head.out_proj.bias'''), ] def _lowerCAmelCase ( lowerCamelCase_ : int ): __lowercase = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''_float_tensor''', ] for k in ignore_keys: state_dict.pop(lowerCamelCase_ , lowerCamelCase_ ) def _lowerCAmelCase ( lowerCamelCase_ : Tuple , lowerCamelCase_ : str , lowerCamelCase_ : Union[str, Any] ): __lowercase = dct.pop(lowerCamelCase_ ) __lowercase = val def _lowerCAmelCase ( lowerCamelCase_ : Any ): __lowercase = torch.load(lowerCamelCase_ , map_location='''cpu''' ) __lowercase = torch.hub.load('''pytorch/fairseq''' , '''bart.large.cnn''' ).eval() hub_interface.model.load_state_dict(sd['''model'''] ) return hub_interface def _lowerCAmelCase ( lowerCamelCase_ : List[str] ): __lowercase , __lowercase = emb.weight.shape __lowercase = nn.Linear(lowerCamelCase_ , lowerCamelCase_ , bias=lowerCamelCase_ ) __lowercase = emb.weight.data return lin_layer @torch.no_grad() def _lowerCAmelCase ( lowerCamelCase_ : Union[str, Any] , lowerCamelCase_ : Optional[Any] , lowerCamelCase_ : List[Any]=None ): if not os.path.exists(lowerCamelCase_ ): __lowercase = torch.hub.load('''pytorch/fairseq''' , lowerCamelCase_ ).eval() else: __lowercase = load_xsum_checkpoint(lowerCamelCase_ ) bart.model.upgrade_state_dict(bart.model.state_dict() ) if hf_checkpoint_name is None: __lowercase = checkpoint_path.replace('''.''' , '''-''' ) __lowercase = BartConfig.from_pretrained(lowerCamelCase_ ) __lowercase = bart.encode(lowerCamelCase_ ).unsqueeze(0 ) __lowercase = BartTokenizer.from_pretrained(lowerCamelCase_ ).encode(lowerCamelCase_ , return_tensors='''pt''' ).unsqueeze(0 ) if not torch.eq(lowerCamelCase_ , lowerCamelCase_ ).all(): raise ValueError( f"converted tokenizer and pretrained tokenizer returned different output: {tokens} != {tokensa}" ) if checkpoint_path == "bart.large.mnli": __lowercase = bart.state_dict() remove_ignore_keys_(lowerCamelCase_ ) __lowercase = state_dict['''model.decoder.embed_tokens.weight'''] for src, dest in mnli_rename_keys: rename_key(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) __lowercase = BartForSequenceClassification(lowerCamelCase_ ).eval() model.load_state_dict(lowerCamelCase_ ) __lowercase = bart.predict('''mnli''' , lowerCamelCase_ , return_logits=lowerCamelCase_ ) __lowercase = model(lowerCamelCase_ )[0] # logits else: # no classification heads to worry about __lowercase = bart.model.state_dict() remove_ignore_keys_(lowerCamelCase_ ) __lowercase = state_dict['''decoder.embed_tokens.weight'''] __lowercase = bart.extract_features(lowerCamelCase_ ) if hf_checkpoint_name == "facebook/bart-large": __lowercase = BartModel(lowerCamelCase_ ).eval() model.load_state_dict(lowerCamelCase_ ) __lowercase = model(lowerCamelCase_ ).model[0] else: __lowercase = BartForConditionalGeneration(lowerCamelCase_ ).eval() # an existing summarization ckpt model.model.load_state_dict(lowerCamelCase_ ) if hasattr(lowerCamelCase_ , '''lm_head''' ): __lowercase = make_linear_from_emb(model.model.shared ) __lowercase = model.model(lowerCamelCase_ )[0] # Check results if fairseq_output.shape != new_model_outputs.shape: raise ValueError( f"`fairseq_output` shape and `new_model_output` shape are different: {fairseq_output.shape=}, {new_model_outputs.shape}" ) if (fairseq_output != new_model_outputs).any().item(): raise ValueError('''Some values in `fairseq_output` are different from `new_model_outputs`''' ) Path(lowerCamelCase_ ).mkdir(exist_ok=lowerCamelCase_ ) model.save_pretrained(lowerCamelCase_ ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( '''fairseq_path''', type=str, help='''bart.large, bart.large.cnn or a path to a model.pt on local filesystem.''' ) parser.add_argument('''pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--hf_config''', default=None, type=str, help='''Which huggingface architecture to use: bart-large-xsum''' ) _SCREAMING_SNAKE_CASE = parser.parse_args() convert_bart_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, hf_checkpoint_name=args.hf_config)
217
1
'''simple docstring''' import warnings from ...utils import logging from .image_processing_segformer import SegformerImageProcessor _lowerCAmelCase = logging.get_logger(__name__) class A ( UpperCamelCase_ ): '''simple docstring''' def __init__(self , *_UpperCAmelCase , **_UpperCAmelCase ) -> None: warnings.warn( "The class SegformerFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use SegformerImageProcessor instead." , __SCREAMING_SNAKE_CASE , ) super().__init__(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE )
298
import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class lowercase_ ( unittest.TestCase ): """simple docstring""" @property def SCREAMING_SNAKE_CASE_ ( self ) ->List[str]: torch.manual_seed(0 ) lowerCAmelCase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) return model def SCREAMING_SNAKE_CASE_ ( self ) ->int: lowerCAmelCase = self.dummy_uncond_unet lowerCAmelCase = KarrasVeScheduler() lowerCAmelCase = KarrasVePipeline(unet=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = pipe(num_inference_steps=2 , generator=__SCREAMING_SNAKE_CASE , output_type='''numpy''' ).images lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = pipe(num_inference_steps=2 , generator=__SCREAMING_SNAKE_CASE , output_type='''numpy''' , return_dict=__SCREAMING_SNAKE_CASE )[0] lowerCAmelCase = image[0, -3:, -3:, -1] lowerCAmelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCAmelCase = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class lowercase_ ( unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE_ ( self ) ->List[Any]: lowerCAmelCase = '''google/ncsnpp-celebahq-256''' lowerCAmelCase = UNetaDModel.from_pretrained(__SCREAMING_SNAKE_CASE ) lowerCAmelCase = KarrasVeScheduler() lowerCAmelCase = KarrasVePipeline(unet=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = pipe(num_inference_steps=20 , generator=__SCREAMING_SNAKE_CASE , output_type='''numpy''' ).images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) lowerCAmelCase = np.array([0.5_7_8, 0.5_8_1_1, 0.5_9_2_4, 0.5_8_0_9, 0.5_8_7, 0.5_8_8_6, 0.5_8_6_1, 0.5_8_0_2, 0.5_8_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
338
0
import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) lowerCamelCase : List[str] = 'hf-internal-testing/tiny-random-bert' lowerCamelCase : Tuple = os.path.join(TRANSFORMERS_CACHE, 'models--hf-internal-testing--tiny-random-bert') lowerCamelCase : Union[str, Any] = '9b8c223d42b2188cb49d29af482996f9d0f3e5a6' class __lowercase (unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> Optional[Any]: snake_case : int = cached_file(A , A ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(A ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(A , A ) ) ) with open(os.path.join(A , """refs""" , """main""" ) ) as f: snake_case : Any = f.read() self.assertEqual(A , os.path.join(A , """snapshots""" , A , A ) ) self.assertTrue(os.path.isfile(A ) ) # File is cached at the same place the second time. snake_case : Tuple = cached_file(A , A ) self.assertEqual(A , A ) # Using a specific revision to test the full commit hash. snake_case : Union[str, Any] = cached_file(A , A , revision="""9b8c223""" ) self.assertEqual(A , os.path.join(A , """snapshots""" , A , A ) ) def UpperCAmelCase ( self ) -> List[Any]: with self.assertRaisesRegex(A , """is not a valid model identifier""" ): snake_case : List[Any] = cached_file("""tiny-random-bert""" , A ) with self.assertRaisesRegex(A , """is not a valid git identifier""" ): snake_case : List[str] = cached_file(A , A , revision="""aaaa""" ) with self.assertRaisesRegex(A , """does not appear to have a file named""" ): snake_case : Union[str, Any] = cached_file(A , """conf""" ) def UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex(A , """does not appear to have a file named""" ): snake_case : List[Any] = cached_file(A , """conf""" ) with open(os.path.join(A , """refs""" , """main""" ) ) as f: snake_case : List[str] = f.read() self.assertTrue(os.path.isfile(os.path.join(A , """.no_exist""" , A , """conf""" ) ) ) snake_case : Optional[int] = cached_file(A , """conf""" , _raise_exceptions_for_missing_entries=A ) self.assertIsNone(A ) snake_case : List[Any] = cached_file(A , """conf""" , local_files_only=A , _raise_exceptions_for_missing_entries=A ) self.assertIsNone(A ) snake_case : Optional[Any] = mock.Mock() snake_case : int = 5_0_0 snake_case : int = {} snake_case : Any = HTTPError snake_case : List[Any] = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("""requests.Session.request""" , return_value=A ) as mock_head: snake_case : Tuple = cached_file(A , """conf""" , _raise_exceptions_for_connection_errors=A ) self.assertIsNone(A ) # This check we did call the fake head request mock_head.assert_called() def UpperCAmelCase ( self ) -> int: self.assertTrue(has_file("""hf-internal-testing/tiny-bert-pt-only""" , A ) ) self.assertFalse(has_file("""hf-internal-testing/tiny-bert-pt-only""" , A ) ) self.assertFalse(has_file("""hf-internal-testing/tiny-bert-pt-only""" , A ) ) def UpperCAmelCase ( self ) -> List[Any]: # `get_file_from_repo` returns None if the file does not exist self.assertIsNone(get_file_from_repo("""bert-base-cased""" , """ahah.txt""" ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(A , """is not a valid model identifier""" ): get_file_from_repo("""bert-base-case""" , A ) # The function raises if the revision does not exist. with self.assertRaisesRegex(A , """is not a valid git identifier""" ): get_file_from_repo("""bert-base-cased""" , A , revision="""ahaha""" ) snake_case : Any = get_file_from_repo("""bert-base-cased""" , A ) # The name is the cached name which is not very easy to test, so instead we load the content. snake_case : List[Any] = json.loads(open(A , """r""" ).read() ) self.assertEqual(config["""hidden_size"""] , 7_6_8 ) def UpperCAmelCase ( self ) -> int: with tempfile.TemporaryDirectory() as tmp_dir: snake_case : Union[str, Any] = Path(A ) / """a.txt""" filename.touch() self.assertEqual(get_file_from_repo(A , """a.txt""" ) , str(A ) ) self.assertIsNone(get_file_from_repo(A , """b.txt""" ) )
176
from math import pow def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ,lowercase ,lowercase ,lowercase ,) -> tuple[int, int]: if current_sum == needed_sum: # If the sum of the powers is equal to needed_sum, then we have a solution. solutions_count += 1 return current_sum, solutions_count snake_case : Union[str, Any] = int(pow(lowercase ,lowercase ) ) if current_sum + i_to_n <= needed_sum: # If the sum of the powers is less than needed_sum, then continue adding powers. current_sum += i_to_n snake_case , snake_case : List[Any] = backtrack( lowercase ,lowercase ,current_number + 1 ,lowercase ,lowercase ) current_sum -= i_to_n if i_to_n < needed_sum: # If the power of i is less than needed_sum, then try with the next power. snake_case , snake_case : str = backtrack( lowercase ,lowercase ,current_number + 1 ,lowercase ,lowercase ) return current_sum, solutions_count def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ) -> int: if not (1 <= needed_sum <= 1000 and 2 <= power <= 10): raise ValueError( """Invalid input\n""" """needed_sum must be between 1 and 1000, power between 2 and 10.""" ) return backtrack(lowercase ,lowercase ,1 ,0 ,0 )[1] # Return the solutions_count if __name__ == "__main__": import doctest doctest.testmod()
176
1
"""simple docstring""" from math import sqrt def _lowerCamelCase ( _UpperCamelCase = 100_0000 ): '''simple docstring''' __lowerCAmelCase = 0 __lowerCAmelCase = 0 __lowerCAmelCase = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(_UpperCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
57
def SCREAMING_SNAKE_CASE_ ( __A : list[int] , __A : str ) -> list[int]: """simple docstring""" a_ : Any = int(__A ) # Initialize Result a_ : Tuple = [] # Traverse through all denomination for denomination in reversed(__A ): # Find denominations while int(__A ) >= int(__A ): total_value -= int(__A ) answer.append(__A ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": UpperCAmelCase_ : Union[str, Any] = [] UpperCAmelCase_ : Union[str, Any] = '0' if ( input('Do you want to enter your denominations ? (yY/n): ').strip().lower() == "y" ): UpperCAmelCase_ : List[Any] = int(input('Enter the number of denominations you want to add: ').strip()) for i in range(0, n): denominations.append(int(input(F'Denomination {i}: ').strip())) UpperCAmelCase_ : str = input('Enter the change you want to make in Indian Currency: ').strip() else: # All denominations of Indian Currency if user does not enter UpperCAmelCase_ : List[Any] = [1, 2, 5, 10, 20, 50, 100, 500, 2000] UpperCAmelCase_ : str = input('Enter the change you want to make: ').strip() if int(value) == 0 or int(value) < 0: print('The total value cannot be zero or negative.') else: print(F'Following is minimal change for {value}: ') UpperCAmelCase_ : Optional[Any] = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=' ')
32
0
import os from argparse import ArgumentParser from typing import List import torch.utils.data from datasets import Dataset, IterableDataset from datasets.distributed import split_dataset_by_node UpperCamelCase = 4 UpperCamelCase = 3 class snake_case_ ( __A ): pass def lowercase_ ( _lowerCamelCase : List[str]): for shard in shards: for i in range(_lowerCamelCase): yield {"i": i, "shard": shard} def lowercase_ ( ): lowercase__ : List[str] = int(os.environ["RANK"]) lowercase__ : Union[str, Any] = int(os.environ["WORLD_SIZE"]) lowercase__ : Union[str, Any] = ArgumentParser() parser.add_argument("--streaming" , type=_lowerCamelCase) parser.add_argument("--local_rank" , type=_lowerCamelCase) parser.add_argument("--num_workers" , type=_lowerCamelCase , default=0) lowercase__ : int = parser.parse_args() lowercase__ : Union[str, Any] = args.streaming lowercase__ : List[Any] = args.num_workers lowercase__ : Dict = {"shards": [f'''shard_{shard_idx}''' for shard_idx in range(_lowerCamelCase)]} lowercase__ : int = IterableDataset.from_generator(_lowerCamelCase , gen_kwargs=_lowerCamelCase) if not streaming: lowercase__ : str = Dataset.from_list(list(_lowerCamelCase)) lowercase__ : List[str] = split_dataset_by_node(_lowerCamelCase , rank=_lowerCamelCase , world_size=_lowerCamelCase) lowercase__ : Any = torch.utils.data.DataLoader(_lowerCamelCase , num_workers=_lowerCamelCase) lowercase__ : Dict = NUM_SHARDS * NUM_ITEMS_PER_SHARD lowercase__ : Any = full_size // world_size expected_local_size += int(rank < (full_size % world_size)) lowercase__ : List[str] = sum(1 for _ in dataloader) if local_size != expected_local_size: raise FailedTestError(f'''local_size {local_size} != expected_local_size {expected_local_size}''') if __name__ == "__main__": main()
357
def lowercase_ ( _lowerCamelCase : int = 10 , _lowerCamelCase : int = 1000 , _lowerCamelCase : bool = True): assert ( isinstance(_lowerCamelCase , _lowerCamelCase) and isinstance(_lowerCamelCase , _lowerCamelCase) and isinstance(_lowerCamelCase , _lowerCamelCase) ), "Invalid type of value(s) specified to function!" if min_val > max_val: raise ValueError("Invalid value for min_val or max_val (min_value < max_value)") return min_val if option else max_val def lowercase_ ( _lowerCamelCase : int , _lowerCamelCase : int): return int((number_a + number_a) / 2) def lowercase_ ( _lowerCamelCase : int , _lowerCamelCase : int , _lowerCamelCase : int): assert ( isinstance(_lowerCamelCase , _lowerCamelCase) and isinstance(_lowerCamelCase , _lowerCamelCase) and isinstance(_lowerCamelCase , _lowerCamelCase) ), 'argument values must be type of "int"' if lower > higher: raise ValueError("argument value for lower and higher must be(lower > higher)") if not lower < to_guess < higher: raise ValueError( "guess value must be within the range of lower and higher value") def answer(_lowerCamelCase : int) -> str: if number > to_guess: return "high" elif number < to_guess: return "low" else: return "same" print("started...") lowercase__ : Optional[int] = lower lowercase__ : List[Any] = higher lowercase__ : Dict = [] while True: lowercase__ : Any = get_avg(_lowerCamelCase , _lowerCamelCase) last_numbers.append(_lowerCamelCase) if answer(_lowerCamelCase) == "low": lowercase__ : List[str] = number elif answer(_lowerCamelCase) == "high": lowercase__ : Optional[int] = number else: break print(f'''guess the number : {last_numbers[-1]}''') print(f'''details : {last_numbers!s}''') def lowercase_ ( ): lowercase__ : Tuple = int(input("Enter lower value : ").strip()) lowercase__ : Optional[int] = int(input("Enter high value : ").strip()) lowercase__ : Optional[Any] = int(input("Enter value to guess : ").strip()) guess_the_number(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase) if __name__ == "__main__": main()
333
0
import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class snake_case_ : def __init__( self : str , lowercase_ : str , lowercase_ : Union[str, Any]=13 , lowercase_ : List[Any]=7 , lowercase_ : Dict=True , lowercase_ : Any=True , lowercase_ : Optional[Any]=True , lowercase_ : int=True , lowercase_ : str=99 , lowercase_ : int=64 , lowercase_ : Tuple=5 , lowercase_ : Any=4 , lowercase_ : Tuple=37 , lowercase_ : List[str]="gelu" , lowercase_ : Any=0.1 , lowercase_ : Optional[Any]=0.1 , lowercase_ : List[str]=5_12 , lowercase_ : Dict=16 , lowercase_ : List[str]=2 , lowercase_ : List[Any]=0.02 , lowercase_ : Tuple=3 , lowercase_ : str=4 , lowercase_ : str=None , ) -> str: lowercase__ : Optional[Any] = parent lowercase__ : Tuple = batch_size lowercase__ : Optional[int] = seq_length lowercase__ : List[str] = is_training lowercase__ : Dict = use_input_mask lowercase__ : str = use_token_type_ids lowercase__ : Optional[Any] = use_labels lowercase__ : Optional[int] = vocab_size lowercase__ : List[str] = hidden_size lowercase__ : Tuple = num_hidden_layers lowercase__ : Optional[int] = num_attention_heads lowercase__ : str = intermediate_size lowercase__ : Dict = hidden_act lowercase__ : Tuple = hidden_dropout_prob lowercase__ : List[str] = attention_probs_dropout_prob lowercase__ : Any = max_position_embeddings lowercase__ : List[str] = type_vocab_size lowercase__ : Union[str, Any] = type_sequence_label_size lowercase__ : Tuple = initializer_range lowercase__ : int = num_labels lowercase__ : Tuple = num_choices lowercase__ : Dict = scope lowercase__ : List[Any] = vocab_size - 1 def __UpperCamelCase ( self : Any ) -> Optional[Any]: lowercase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ : Optional[int] = None if self.use_input_mask: lowercase__ : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) lowercase__ : Dict = None if self.use_labels: lowercase__ : int = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase__ : List[Any] = self.get_config() return config, input_ids, input_mask, token_labels def __UpperCamelCase ( self : Optional[int] ) -> List[Any]: return GPTNeoXConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase_ , initializer_range=self.initializer_range , pad_token_id=self.pad_token_id , ) def __UpperCamelCase ( self : str ) -> Optional[Any]: lowercase__ , lowercase__ , lowercase__ , lowercase__ : int = self.prepare_config_and_inputs() lowercase__ : Optional[Any] = True return config, input_ids, input_mask, token_labels def __UpperCamelCase ( self : Dict , lowercase_ : List[Any] , lowercase_ : int , lowercase_ : str ) -> Any: lowercase__ : str = GPTNeoXModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() lowercase__ : Union[str, Any] = model(lowercase_ , attention_mask=lowercase_ ) lowercase__ : Dict = model(lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCamelCase ( self : str , lowercase_ : List[Any] , lowercase_ : Optional[int] , lowercase_ : Optional[Any] ) -> str: lowercase__ : List[str] = True lowercase__ : List[Any] = GPTNeoXModel(lowercase_ ) model.to(lowercase_ ) model.eval() lowercase__ : int = model(lowercase_ , attention_mask=lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCamelCase ( self : List[str] , lowercase_ : int , lowercase_ : Tuple , lowercase_ : List[Any] , lowercase_ : int ) -> List[str]: lowercase__ : List[str] = GPTNeoXForCausalLM(config=lowercase_ ) model.to(lowercase_ ) model.eval() lowercase__ : Any = model(lowercase_ , attention_mask=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCamelCase ( self : str , lowercase_ : List[str] , lowercase_ : Tuple , lowercase_ : Optional[int] , lowercase_ : List[str] ) -> Tuple: lowercase__ : Optional[Any] = self.num_labels lowercase__ : List[str] = GPTNeoXForQuestionAnswering(lowercase_ ) model.to(lowercase_ ) model.eval() lowercase__ : Any = model(lowercase_ , attention_mask=lowercase_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __UpperCamelCase ( self : int , lowercase_ : List[Any] , lowercase_ : str , lowercase_ : Union[str, Any] , lowercase_ : Optional[Any] ) -> Tuple: lowercase__ : Union[str, Any] = self.num_labels lowercase__ : Optional[Any] = GPTNeoXForSequenceClassification(lowercase_ ) model.to(lowercase_ ) model.eval() lowercase__ : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase__ : Any = model(lowercase_ , attention_mask=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCamelCase ( self : Optional[int] , lowercase_ : Union[str, Any] , lowercase_ : int , lowercase_ : Tuple , lowercase_ : List[str] ) -> Dict: lowercase__ : Dict = self.num_labels lowercase__ : Any = GPTNeoXForTokenClassification(lowercase_ ) model.to(lowercase_ ) model.eval() lowercase__ : Optional[int] = model(lowercase_ , attention_mask=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCamelCase ( self : int , lowercase_ : str , lowercase_ : Optional[int] , lowercase_ : Dict ) -> Union[str, Any]: lowercase__ : Tuple = True lowercase__ : List[Any] = GPTNeoXForCausalLM(config=lowercase_ ) model.to(lowercase_ ) model.eval() # first forward pass lowercase__ : Union[str, Any] = model(lowercase_ , attention_mask=lowercase_ , use_cache=lowercase_ ) lowercase__ : Dict = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids lowercase__ : Tuple = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowercase__ : Optional[Any] = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and lowercase__ : Dict = torch.cat([input_ids, next_tokens] , dim=-1 ) lowercase__ : Dict = torch.cat([input_mask, next_mask] , dim=-1 ) lowercase__ : List[Any] = model(lowercase_ , attention_mask=lowercase_ , output_hidden_states=lowercase_ ) lowercase__ : Optional[Any] = output_from_no_past["hidden_states"][0] lowercase__ : Union[str, Any] = model( lowercase_ , attention_mask=lowercase_ , past_key_values=lowercase_ , output_hidden_states=lowercase_ , )["hidden_states"][0] # select random slice lowercase__ : str = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowercase__ : Any = output_from_no_past[:, -3:, random_slice_idx].detach() lowercase__ : List[str] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase_ , lowercase_ , atol=1E-3 ) ) def __UpperCamelCase ( self : str ) -> str: lowercase__ : Tuple = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ , lowercase__ : Optional[int] = config_and_inputs lowercase__ : List[str] = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class snake_case_ ( __A ,__A ,__A ,unittest.TestCase ): __A : Union[str, Any] = ( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) __A : List[str] = (GPTNeoXForCausalLM,) if is_torch_available() else () __A : Any = ( { "feature-extraction": GPTNeoXModel, "question-answering": GPTNeoXForQuestionAnswering, "text-classification": GPTNeoXForSequenceClassification, "text-generation": GPTNeoXForCausalLM, "token-classification": GPTNeoXForTokenClassification, "zero-shot": GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) __A : int = False __A : List[Any] = False __A : Optional[Any] = False __A : Dict = False def __UpperCamelCase ( self : str ) -> Tuple: lowercase__ : int = GPTNeoXModelTester(self ) lowercase__ : Union[str, Any] = ConfigTester(self , config_class=lowercase_ , hidden_size=64 , num_attention_heads=8 ) def __UpperCamelCase ( self : List[str] ) -> Union[str, Any]: self.config_tester.run_common_tests() def __UpperCamelCase ( self : int ) -> str: lowercase__ , lowercase__ , lowercase__ , lowercase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(lowercase_ , lowercase_ , lowercase_ ) def __UpperCamelCase ( self : Union[str, Any] ) -> int: lowercase__ , lowercase__ , lowercase__ , lowercase__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(lowercase_ , lowercase_ , lowercase_ ) def __UpperCamelCase ( self : List[Any] ) -> List[str]: # This regression test was failing with PyTorch < 1.3 lowercase__ , lowercase__ , lowercase__ , lowercase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_decoder() lowercase__ : Tuple = None self.model_tester.create_and_check_model_as_decoder(lowercase_ , lowercase_ , lowercase_ ) def __UpperCamelCase ( self : Dict ) -> Dict: lowercase__ , lowercase__ , lowercase__ , lowercase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(lowercase_ , lowercase_ , lowercase_ ) def __UpperCamelCase ( self : Dict ) -> List[Any]: lowercase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*lowercase_ ) def __UpperCamelCase ( self : Optional[int] ) -> Optional[int]: lowercase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase_ ) def __UpperCamelCase ( self : List[Any] ) -> Optional[int]: lowercase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase_ ) def __UpperCamelCase ( self : Any ) -> Any: lowercase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase_ ) @unittest.skip(reason="Feed forward chunking is not implemented" ) def __UpperCamelCase ( self : Dict ) -> Tuple: pass @parameterized.expand([("linear",), ("dynamic",)] ) def __UpperCamelCase ( self : int , lowercase_ : str ) -> int: lowercase__ , lowercase__ : str = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ : str = ids_tensor([1, 10] , config.vocab_size ) lowercase__ : Tuple = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights lowercase__ : List[str] = GPTNeoXModel(lowercase_ ) original_model.to(lowercase_ ) original_model.eval() lowercase__ : Dict = original_model(lowercase_ ).last_hidden_state lowercase__ : Dict = original_model(lowercase_ ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights lowercase__ : Union[str, Any] = {"type": scaling_type, "factor": 10.0} lowercase__ : str = GPTNeoXModel(lowercase_ ) scaled_model.to(lowercase_ ) scaled_model.eval() lowercase__ : Optional[int] = scaled_model(lowercase_ ).last_hidden_state lowercase__ : Any = scaled_model(lowercase_ ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(lowercase_ , lowercase_ , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(lowercase_ , lowercase_ , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(lowercase_ , lowercase_ , atol=1E-5 ) ) @require_torch class snake_case_ ( unittest.TestCase ): @slow def __UpperCamelCase ( self : Optional[Any] ) -> int: lowercase__ : str = AutoTokenizer.from_pretrained("EleutherAI/pythia-410m-deduped" ) for checkpointing in [True, False]: lowercase__ : Union[str, Any] = GPTNeoXForCausalLM.from_pretrained("EleutherAI/pythia-410m-deduped" ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(lowercase_ ) lowercase__ : Any = tokenizer("My favorite food is" , return_tensors="pt" ).to(lowercase_ ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 lowercase__ : int = "My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI'm not sure" lowercase__ : Optional[Any] = model.generate(**lowercase_ , do_sample=lowercase_ , max_new_tokens=20 ) lowercase__ : str = tokenizer.batch_decode(lowercase_ )[0] self.assertEqual(lowercase_ , lowercase_ )
87
import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def lowercase_ ( _lowerCamelCase : Optional[Any] , _lowerCamelCase : int): assert isinstance(_lowerCamelCase , _lowerCamelCase) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True]) def lowercase_ ( _lowerCamelCase : List[Any] , _lowerCamelCase : Any , _lowerCamelCase : str): lowercase__ : Optional[int] = tmp_path / "cache" lowercase__ : str = {"col_1": "string", "col_2": "int64", "col_3": "float64"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): lowercase__ : Union[str, Any] = JsonDatasetReader(_lowerCamelCase , cache_dir=_lowerCamelCase , keep_in_memory=_lowerCamelCase).read() _check_json_dataset(_lowerCamelCase , _lowerCamelCase) @pytest.mark.parametrize( "features" , [ None, {"col_1": "string", "col_2": "int64", "col_3": "float64"}, {"col_1": "string", "col_2": "string", "col_3": "string"}, {"col_1": "int32", "col_2": "int32", "col_3": "int32"}, {"col_1": "float32", "col_2": "float32", "col_3": "float32"}, ] , ) def lowercase_ ( _lowerCamelCase : List[str] , _lowerCamelCase : Dict , _lowerCamelCase : Dict): lowercase__ : List[Any] = tmp_path / "cache" lowercase__ : Union[str, Any] = {"col_1": "string", "col_2": "int64", "col_3": "float64"} lowercase__ : List[Any] = features.copy() if features else default_expected_features lowercase__ : List[Any] = ( Features({feature: Value(_lowerCamelCase) for feature, dtype in features.items()}) if features is not None else None ) lowercase__ : Any = JsonDatasetReader(_lowerCamelCase , features=_lowerCamelCase , cache_dir=_lowerCamelCase).read() _check_json_dataset(_lowerCamelCase , _lowerCamelCase) @pytest.mark.parametrize( "features" , [ None, {"col_3": "float64", "col_1": "string", "col_2": "int64"}, ] , ) def lowercase_ ( _lowerCamelCase : Any , _lowerCamelCase : Any , _lowerCamelCase : List[str]): lowercase__ : Optional[Any] = tmp_path / "cache" lowercase__ : Tuple = {"col_3": "float64", "col_1": "string", "col_2": "int64"} lowercase__ : List[Any] = features.copy() if features else default_expected_features lowercase__ : int = ( Features({feature: Value(_lowerCamelCase) for feature, dtype in features.items()}) if features is not None else None ) lowercase__ : Any = JsonDatasetReader(_lowerCamelCase , features=_lowerCamelCase , cache_dir=_lowerCamelCase).read() assert isinstance(_lowerCamelCase , _lowerCamelCase) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def lowercase_ ( _lowerCamelCase : List[Any] , _lowerCamelCase : Optional[int]): # jsonl_312_path features are {"col_3": "float64", "col_1": "string", "col_2": "int64"} lowercase__ : Any = {"col_2": "int64", "col_3": "float64", "col_1": "string"} lowercase__ : str = features.copy() lowercase__ : str = ( Features({feature: Value(_lowerCamelCase) for feature, dtype in features.items()}) if features is not None else None ) lowercase__ : Optional[int] = tmp_path / "cache" lowercase__ : Any = JsonDatasetReader(_lowerCamelCase , features=_lowerCamelCase , cache_dir=_lowerCamelCase).read() assert isinstance(_lowerCamelCase , _lowerCamelCase) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("split" , [None, NamedSplit("train"), "train", "test"]) def lowercase_ ( _lowerCamelCase : Dict , _lowerCamelCase : Optional[int] , _lowerCamelCase : List[str]): lowercase__ : Union[str, Any] = tmp_path / "cache" lowercase__ : List[Any] = {"col_1": "string", "col_2": "int64", "col_3": "float64"} lowercase__ : Union[str, Any] = JsonDatasetReader(_lowerCamelCase , cache_dir=_lowerCamelCase , split=_lowerCamelCase).read() _check_json_dataset(_lowerCamelCase , _lowerCamelCase) assert dataset.split == split if split else "train" @pytest.mark.parametrize("path_type" , [str, list]) def lowercase_ ( _lowerCamelCase : Union[str, Any] , _lowerCamelCase : str , _lowerCamelCase : int): if issubclass(_lowerCamelCase , _lowerCamelCase): lowercase__ : Tuple = jsonl_path elif issubclass(_lowerCamelCase , _lowerCamelCase): lowercase__ : str = [jsonl_path] lowercase__ : str = tmp_path / "cache" lowercase__ : Optional[Any] = {"col_1": "string", "col_2": "int64", "col_3": "float64"} lowercase__ : Tuple = JsonDatasetReader(_lowerCamelCase , cache_dir=_lowerCamelCase).read() _check_json_dataset(_lowerCamelCase , _lowerCamelCase) def lowercase_ ( _lowerCamelCase : List[Any] , _lowerCamelCase : Optional[Any] , _lowerCamelCase : Optional[int]=("train",)): assert isinstance(_lowerCamelCase , _lowerCamelCase) for split in splits: lowercase__ : Optional[Any] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory" , [False, True]) def lowercase_ ( _lowerCamelCase : Union[str, Any] , _lowerCamelCase : str , _lowerCamelCase : str): lowercase__ : List[str] = tmp_path / "cache" lowercase__ : str = {"col_1": "string", "col_2": "int64", "col_3": "float64"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): lowercase__ : Optional[Any] = JsonDatasetReader({"train": jsonl_path} , cache_dir=_lowerCamelCase , keep_in_memory=_lowerCamelCase).read() _check_json_datasetdict(_lowerCamelCase , _lowerCamelCase) @pytest.mark.parametrize( "features" , [ None, {"col_1": "string", "col_2": "int64", "col_3": "float64"}, {"col_1": "string", "col_2": "string", "col_3": "string"}, {"col_1": "int32", "col_2": "int32", "col_3": "int32"}, {"col_1": "float32", "col_2": "float32", "col_3": "float32"}, ] , ) def lowercase_ ( _lowerCamelCase : Any , _lowerCamelCase : List[str] , _lowerCamelCase : List[str]): lowercase__ : str = tmp_path / "cache" lowercase__ : Tuple = {"col_1": "string", "col_2": "int64", "col_3": "float64"} lowercase__ : Tuple = features.copy() if features else default_expected_features lowercase__ : Union[str, Any] = ( Features({feature: Value(_lowerCamelCase) for feature, dtype in features.items()}) if features is not None else None ) lowercase__ : Tuple = JsonDatasetReader({"train": jsonl_path} , features=_lowerCamelCase , cache_dir=_lowerCamelCase).read() _check_json_datasetdict(_lowerCamelCase , _lowerCamelCase) @pytest.mark.parametrize("split" , [None, NamedSplit("train"), "train", "test"]) def lowercase_ ( _lowerCamelCase : str , _lowerCamelCase : Dict , _lowerCamelCase : Tuple): if split: lowercase__ : Tuple = {split: jsonl_path} else: lowercase__ : Tuple = "train" lowercase__ : int = {"train": jsonl_path, "test": jsonl_path} lowercase__ : Dict = tmp_path / "cache" lowercase__ : Union[str, Any] = {"col_1": "string", "col_2": "int64", "col_3": "float64"} lowercase__ : Union[str, Any] = JsonDatasetReader(_lowerCamelCase , cache_dir=_lowerCamelCase).read() _check_json_datasetdict(_lowerCamelCase , _lowerCamelCase , splits=list(path.keys())) assert all(dataset[split].split == split for split in path.keys()) def lowercase_ ( _lowerCamelCase : Union[str, Any]): return json.load(_lowerCamelCase) def lowercase_ ( _lowerCamelCase : Optional[int]): return [json.loads(_lowerCamelCase) for line in buffer] class snake_case_ : @pytest.mark.parametrize("lines, load_json_function" , [(True, load_json_lines), (False, load_json)] ) def __UpperCamelCase ( self : List[Any] , lowercase_ : Any , lowercase_ : Optional[Any] , lowercase_ : Dict ) -> Optional[Any]: with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , lines=lowercase_ ).write() buffer.seek(0 ) lowercase__ : Optional[int] = load_json_function(lowercase_ ) assert isinstance(lowercase_ , lowercase_ ) assert isinstance(exported_content[0] , lowercase_ ) assert len(lowercase_ ) == 10 @pytest.mark.parametrize( "orient, container, keys, len_at" , [ ("records", list, {"tokens", "labels", "answers", "id"}, None), ("split", dict, {"columns", "data"}, "data"), ("index", dict, set("0123456789" ), None), ("columns", dict, {"tokens", "labels", "answers", "id"}, "tokens"), ("values", list, None, None), ("table", dict, {"schema", "data"}, "data"), ] , ) def __UpperCamelCase ( self : str , lowercase_ : int , lowercase_ : str , lowercase_ : Union[str, Any] , lowercase_ : Dict , lowercase_ : Tuple ) -> List[str]: with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , lines=lowercase_ , orient=lowercase_ ).write() buffer.seek(0 ) lowercase__ : str = load_json(lowercase_ ) assert isinstance(lowercase_ , lowercase_ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(lowercase_ , "keys" ) and not hasattr(exported_content[0] , "keys" ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(lowercase_ ) == 10 @pytest.mark.parametrize("lines, load_json_function" , [(True, load_json_lines), (False, load_json)] ) def __UpperCamelCase ( self : List[Any] , lowercase_ : int , lowercase_ : Optional[Any] , lowercase_ : Dict ) -> Optional[int]: with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , lines=lowercase_ , num_proc=2 ).write() buffer.seek(0 ) lowercase__ : str = load_json_function(lowercase_ ) assert isinstance(lowercase_ , lowercase_ ) assert isinstance(exported_content[0] , lowercase_ ) assert len(lowercase_ ) == 10 @pytest.mark.parametrize( "orient, container, keys, len_at" , [ ("records", list, {"tokens", "labels", "answers", "id"}, None), ("split", dict, {"columns", "data"}, "data"), ("index", dict, set("0123456789" ), None), ("columns", dict, {"tokens", "labels", "answers", "id"}, "tokens"), ("values", list, None, None), ("table", dict, {"schema", "data"}, "data"), ] , ) def __UpperCamelCase ( self : Union[str, Any] , lowercase_ : Dict , lowercase_ : Dict , lowercase_ : Optional[int] , lowercase_ : Optional[Any] , lowercase_ : Dict ) -> Any: with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , lines=lowercase_ , orient=lowercase_ , num_proc=2 ).write() buffer.seek(0 ) lowercase__ : Optional[Any] = load_json(lowercase_ ) assert isinstance(lowercase_ , lowercase_ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(lowercase_ , "keys" ) and not hasattr(exported_content[0] , "keys" ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(lowercase_ ) == 10 def __UpperCamelCase ( self : Dict , lowercase_ : List[str] ) -> str: with pytest.raises(lowercase_ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , num_proc=0 ) @pytest.mark.parametrize("compression, extension" , [("gzip", "gz"), ("bz2", "bz2"), ("xz", "xz")] ) def __UpperCamelCase ( self : List[Any] , lowercase_ : Tuple , lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : str , lowercase_ : List[Any] ) -> Any: lowercase__ : Dict = tmp_path_factory.mktemp("data" ) / F'''test.json.{extension}''' lowercase__ : Optional[int] = str(shared_datadir / F'''test_file.json.{extension}''' ) JsonDatasetWriter(lowercase_ , lowercase_ , compression=lowercase_ ).write() with fsspec.open(lowercase_ , "rb" , compression="infer" ) as f: lowercase__ : List[Any] = f.read() with fsspec.open(lowercase_ , "rb" , compression="infer" ) as f: lowercase__ : str = f.read() assert exported_content == original_content
87
1
"""simple docstring""" import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal _UpperCamelCase = datasets.utils.logging.get_logger(__name__) _UpperCamelCase = ["""names""", """prefix"""] _UpperCamelCase = ["""warn_bad_lines""", """error_bad_lines""", """mangle_dupe_cols"""] _UpperCamelCase = ["""encoding_errors""", """on_bad_lines"""] _UpperCamelCase = ["""date_format"""] @dataclass class lowerCamelCase__ ( datasets.BuilderConfig ): SCREAMING_SNAKE_CASE = ''',''' SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = '''infer''' SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = '''.''' SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = '''"''' SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = 10_000 SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = '''strict''' SCREAMING_SNAKE_CASE = '''error''' SCREAMING_SNAKE_CASE = None def _UpperCamelCase ( self ): if self.delimiter is not None: UpperCAmelCase = self.delimiter if self.column_names is not None: UpperCAmelCase = self.column_names @property def _UpperCamelCase ( self ): UpperCAmelCase = { """sep""": self.sep, """header""": self.header, """names""": self.names, """index_col""": self.index_col, """usecols""": self.usecols, """prefix""": self.prefix, """mangle_dupe_cols""": self.mangle_dupe_cols, """engine""": self.engine, """converters""": self.converters, """true_values""": self.true_values, """false_values""": self.false_values, """skipinitialspace""": self.skipinitialspace, """skiprows""": self.skiprows, """nrows""": self.nrows, """na_values""": self.na_values, """keep_default_na""": self.keep_default_na, """na_filter""": self.na_filter, """verbose""": self.verbose, """skip_blank_lines""": self.skip_blank_lines, """thousands""": self.thousands, """decimal""": self.decimal, """lineterminator""": self.lineterminator, """quotechar""": self.quotechar, """quoting""": self.quoting, """escapechar""": self.escapechar, """comment""": self.comment, """encoding""": self.encoding, """dialect""": self.dialect, """error_bad_lines""": self.error_bad_lines, """warn_bad_lines""": self.warn_bad_lines, """skipfooter""": self.skipfooter, """doublequote""": self.doublequote, """memory_map""": self.memory_map, """float_precision""": self.float_precision, """chunksize""": self.chunksize, """encoding_errors""": self.encoding_errors, """on_bad_lines""": self.on_bad_lines, """date_format""": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() ,A ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class lowerCamelCase__ ( datasets.ArrowBasedBuilder ): SCREAMING_SNAKE_CASE = CsvConfig def _UpperCamelCase ( self ): return datasets.DatasetInfo(features=self.config.features ) def _UpperCamelCase ( self ,A ): if not self.config.data_files: raise ValueError(F'''At least one data file must be specified, but got data_files={self.config.data_files}''' ) UpperCAmelCase = dl_manager.download_and_extract(self.config.data_files ) if isinstance(A ,(str, list, tuple) ): UpperCAmelCase = data_files if isinstance(A ,A ): UpperCAmelCase = [files] UpperCAmelCase = [dl_manager.iter_files(A ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN ,gen_kwargs={"""files""": files} )] UpperCAmelCase = [] for split_name, files in data_files.items(): if isinstance(A ,A ): UpperCAmelCase = [files] UpperCAmelCase = [dl_manager.iter_files(A ) for file in files] splits.append(datasets.SplitGenerator(name=A ,gen_kwargs={"""files""": files} ) ) return splits def _UpperCamelCase ( self ,A ): if self.config.features is not None: UpperCAmelCase = self.config.features.arrow_schema if all(not require_storage_cast(A ) for feature in self.config.features.values() ): # cheaper cast UpperCAmelCase = pa.Table.from_arrays([pa_table[field.name] for field in schema] ,schema=A ) else: # more expensive cast; allows str <-> int/float or str to Audio for example UpperCAmelCase = table_cast(A ,A ) return pa_table def _UpperCamelCase ( self ,A ): UpperCAmelCase = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str UpperCAmelCase = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(A ) else object for name, dtype, feature in zip(schema.names ,schema.types ,self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(A ) ): UpperCAmelCase = pd.read_csv(A ,iterator=A ,dtype=A ,**self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(A ): UpperCAmelCase = pa.Table.from_pandas(A ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(A ) except ValueError as e: logger.error(F'''Failed to read file \'{file}\' with error {type(A )}: {e}''' ) raise
364
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _UpperCamelCase = {"""configuration_glpn""": ["""GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GLPNConfig"""]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = ["""GLPNFeatureExtractor"""] _UpperCamelCase = ["""GLPNImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ """GLPN_PRETRAINED_MODEL_ARCHIVE_LIST""", """GLPNForDepthEstimation""", """GLPNLayer""", """GLPNModel""", """GLPNPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
234
0
"""simple docstring""" from math import factorial __A = {str(d): factorial(d) for d in range(10)} def a__ ( __SCREAMING_SNAKE_CASE ) -> int: return sum(DIGIT_FACTORIAL[d] for d in str(__SCREAMING_SNAKE_CASE ) ) def a__ ( ) -> int: __lowerCAmelCase: Dict = 7 * factorial(9 ) + 1 return sum(i for i in range(3 , __SCREAMING_SNAKE_CASE ) if sum_of_digit_factorial(__SCREAMING_SNAKE_CASE ) == i ) if __name__ == "__main__": print(F'''{solution() = }''')
217
"""simple docstring""" def a__ ( __SCREAMING_SNAKE_CASE ) -> int: __lowerCAmelCase: Optional[Any] = 1 for i in range(1 , num + 1 ): fact *= i return fact def a__ ( __SCREAMING_SNAKE_CASE ) -> int: __lowerCAmelCase: List[str] = 0 while number > 0: __lowerCAmelCase: Any = number % 1_0 sum_of_digits += last_digit __lowerCAmelCase: List[Any] = number // 1_0 # Removing the last_digit from the given number return sum_of_digits def a__ ( __SCREAMING_SNAKE_CASE = 1_0_0 ) -> int: __lowerCAmelCase: Tuple = factorial(__SCREAMING_SNAKE_CASE ) __lowerCAmelCase: Optional[int] = split_and_add(__SCREAMING_SNAKE_CASE ) return result if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
217
1
UpperCAmelCase_ = '''Tobias Carryer''' from time import time class lowerCamelCase__: def __init__( self: List[str] , UpperCamelCase_: List[str] , UpperCamelCase_: List[Any] , UpperCamelCase_: int , UpperCamelCase_: Any=int(time() ) ): # noqa: B008 __lowerCamelCase = multiplier __lowerCamelCase = increment __lowerCamelCase = modulo __lowerCamelCase = seed def lowerCAmelCase__ ( self: Union[str, Any] ): __lowerCamelCase = (self.multiplier * self.seed + self.increment) % self.modulo return self.seed if __name__ == "__main__": # Show the LCG in action. UpperCAmelCase_ = LinearCongruentialGenerator(1_664_525, 1_013_904_223, 2 << 31) while True: print(lcg.next_number())
369
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'nvidia/segformer-b0-finetuned-ade-512-512': ( 'https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json' ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : str = 'segformer' def __init__( self: Union[str, Any] , UpperCamelCase_: Optional[int]=3 , UpperCamelCase_: Any=4 , UpperCamelCase_: int=[2, 2, 2, 2] , UpperCamelCase_: Optional[Any]=[8, 4, 2, 1] , UpperCamelCase_: Union[str, Any]=[32, 64, 1_60, 2_56] , UpperCamelCase_: int=[7, 3, 3, 3] , UpperCamelCase_: Dict=[4, 2, 2, 2] , UpperCamelCase_: str=[1, 2, 5, 8] , UpperCamelCase_: List[str]=[4, 4, 4, 4] , UpperCamelCase_: Optional[int]="gelu" , UpperCamelCase_: List[Any]=0.0 , UpperCamelCase_: List[Any]=0.0 , UpperCamelCase_: Tuple=0.1 , UpperCamelCase_: Optional[int]=0.02 , UpperCamelCase_: List[Any]=0.1 , UpperCamelCase_: Optional[int]=1E-6 , UpperCamelCase_: Optional[int]=2_56 , UpperCamelCase_: Optional[Any]=2_55 , **UpperCamelCase_: List[Any] , ): super().__init__(**UpperCamelCase_ ) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( """Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be""" """ removed, as the behaviour will default to that of reshape_last_stage = True.""" , UpperCamelCase_ , ) __lowerCamelCase = num_channels __lowerCamelCase = num_encoder_blocks __lowerCamelCase = depths __lowerCamelCase = sr_ratios __lowerCamelCase = hidden_sizes __lowerCamelCase = patch_sizes __lowerCamelCase = strides __lowerCamelCase = mlp_ratios __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = classifier_dropout_prob __lowerCamelCase = initializer_range __lowerCamelCase = drop_path_rate __lowerCamelCase = layer_norm_eps __lowerCamelCase = decoder_hidden_size __lowerCamelCase = kwargs.get("""reshape_last_stage""" , UpperCamelCase_ ) __lowerCamelCase = semantic_loss_ignore_index class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Any = version.parse('1.11') @property def lowerCAmelCase__ ( self: Any ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def lowerCAmelCase__ ( self: Union[str, Any] ): return 1E-4 @property def lowerCAmelCase__ ( self: Dict ): return 12
29
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case = logging.get_logger(__name__) __snake_case = { """transfo-xl-wt103""": """https://huggingface.co/transfo-xl-wt103/resolve/main/config.json""", } class lowercase__ ( _UpperCAmelCase ): A__ : int ="""transfo-xl""" A__ : Optional[Any] =["""mems"""] A__ : Union[str, Any] ={ """n_token""": """vocab_size""", """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : Optional[Any] , UpperCAmelCase_ : Optional[Any]=267735 , UpperCAmelCase_ : Tuple=[20000, 40000, 200000] , UpperCAmelCase_ : List[Any]=1024 , UpperCAmelCase_ : Optional[Any]=1024 , UpperCAmelCase_ : List[str]=16 , UpperCAmelCase_ : Dict=64 , UpperCAmelCase_ : int=4096 , UpperCAmelCase_ : List[Any]=4 , UpperCAmelCase_ : str=False , UpperCAmelCase_ : List[Any]=18 , UpperCAmelCase_ : Dict=1600 , UpperCAmelCase_ : Optional[int]=1000 , UpperCAmelCase_ : Optional[int]=True , UpperCAmelCase_ : Tuple=True , UpperCAmelCase_ : int=0 , UpperCAmelCase_ : str=-1 , UpperCAmelCase_ : Optional[int]=True , UpperCAmelCase_ : Union[str, Any]=0.1 , UpperCAmelCase_ : List[Any]=0.0 , UpperCAmelCase_ : Optional[int]=True , UpperCAmelCase_ : Optional[Any]="normal" , UpperCAmelCase_ : List[Any]=0.01 , UpperCAmelCase_ : Union[str, Any]=0.01 , UpperCAmelCase_ : Optional[int]=0.02 , UpperCAmelCase_ : Union[str, Any]=1e-5 , UpperCAmelCase_ : str=0 , **UpperCAmelCase_ : Optional[int] , ): SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = [] self.cutoffs.extend(UpperCAmelCase_ ) if proj_share_all_but_first: SCREAMING_SNAKE_CASE__ = [False] + [True] * len(self.cutoffs ) else: SCREAMING_SNAKE_CASE__ = [False] + [False] * len(self.cutoffs ) SCREAMING_SNAKE_CASE__ = d_model SCREAMING_SNAKE_CASE__ = d_embed SCREAMING_SNAKE_CASE__ = d_head SCREAMING_SNAKE_CASE__ = d_inner SCREAMING_SNAKE_CASE__ = div_val SCREAMING_SNAKE_CASE__ = pre_lnorm SCREAMING_SNAKE_CASE__ = n_layer SCREAMING_SNAKE_CASE__ = n_head SCREAMING_SNAKE_CASE__ = mem_len SCREAMING_SNAKE_CASE__ = same_length SCREAMING_SNAKE_CASE__ = attn_type SCREAMING_SNAKE_CASE__ = clamp_len SCREAMING_SNAKE_CASE__ = sample_softmax SCREAMING_SNAKE_CASE__ = adaptive SCREAMING_SNAKE_CASE__ = dropout SCREAMING_SNAKE_CASE__ = dropatt SCREAMING_SNAKE_CASE__ = untie_r SCREAMING_SNAKE_CASE__ = init SCREAMING_SNAKE_CASE__ = init_range SCREAMING_SNAKE_CASE__ = proj_init_std SCREAMING_SNAKE_CASE__ = init_std SCREAMING_SNAKE_CASE__ = layer_norm_epsilon super().__init__(eos_token_id=UpperCAmelCase_ , **UpperCAmelCase_ ) @property def A_ ( self : List[Any] ): # Message copied from Transformer-XL documentation logger.info(F'The model {self.model_type} is one of the few models that has no sequence length limit.' ) return -1 @max_position_embeddings.setter def A_ ( self : Tuple , UpperCAmelCase_ : Dict ): # Message copied from Transformer-XL documentation raise NotImplementedError( F'The model {self.model_type} is one of the few models that has no sequence length limit.' )
176
import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() __snake_case = logging.get_logger(__name__) __snake_case = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """lm_head""", """mask_emb""": """masked_spec_embed""", } __snake_case = [ """lm_head""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> List[str]: '''simple docstring''' for attribute in key.split('.' ): SCREAMING_SNAKE_CASE__ = getattr(UpperCamelCase_ , UpperCamelCase_ ) if weight_type is not None: SCREAMING_SNAKE_CASE__ = getattr(UpperCamelCase_ , UpperCamelCase_ ).shape else: SCREAMING_SNAKE_CASE__ = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": SCREAMING_SNAKE_CASE__ = value elif weight_type == "weight_g": SCREAMING_SNAKE_CASE__ = value elif weight_type == "weight_v": SCREAMING_SNAKE_CASE__ = value elif weight_type == "bias": SCREAMING_SNAKE_CASE__ = value else: SCREAMING_SNAKE_CASE__ = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = fairseq_model.state_dict() SCREAMING_SNAKE_CASE__ = hf_model.feature_extractor SCREAMING_SNAKE_CASE__ = hf_model.adapter for name, value in fairseq_dict.items(): SCREAMING_SNAKE_CASE__ = False if "conv_layers" in name: load_conv_layer( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , hf_model.config.feat_extract_norm == 'group' , ) SCREAMING_SNAKE_CASE__ = True elif any(x in name for x in ['adaptor', 'w2v_encoder.proj.', 'w2v_proj_ln.'] ): load_adapter(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: SCREAMING_SNAKE_CASE__ = True if "*" in mapped_key: SCREAMING_SNAKE_CASE__ = name.split(UpperCamelCase_ )[0].split('.' )[-2] SCREAMING_SNAKE_CASE__ = mapped_key.replace('*' , UpperCamelCase_ ) if "weight_g" in name: SCREAMING_SNAKE_CASE__ = 'weight_g' elif "weight_v" in name: SCREAMING_SNAKE_CASE__ = 'weight_v' elif "bias" in name: SCREAMING_SNAKE_CASE__ = 'bias' elif "weight" in name: SCREAMING_SNAKE_CASE__ = 'weight' else: SCREAMING_SNAKE_CASE__ = None set_recursively(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) continue if not is_used: unused_weights.append(UpperCamelCase_ ) logger.warning(F'Unused weights: {unused_weights}' ) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE__ = full_name.split('conv_layers.' )[-1] SCREAMING_SNAKE_CASE__ = name.split('.' ) SCREAMING_SNAKE_CASE__ = int(items[0] ) SCREAMING_SNAKE_CASE__ = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) SCREAMING_SNAKE_CASE__ = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) SCREAMING_SNAKE_CASE__ = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) SCREAMING_SNAKE_CASE__ = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) SCREAMING_SNAKE_CASE__ = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(UpperCamelCase_ ) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE__ = full_name.split('adaptor.' )[-1] SCREAMING_SNAKE_CASE__ = name.split('.' ) if items[1].isdigit(): SCREAMING_SNAKE_CASE__ = int(items[1] ) else: SCREAMING_SNAKE_CASE__ = None if "adaptor" not in full_name: if "proj_ln" in full_name: # has to be layer norm if "bias" in name: assert ( value.shape == adapter.proj_layer_norm.bias.data.shape ), F'{full_name} has size {value.shape}, but {adapter.proj_layer_norm.bias.data.shape} was found.' SCREAMING_SNAKE_CASE__ = value logger.info(F'Adapter proj layer norm bias was initialized from {full_name}.' ) if "weight" in name: assert ( value.shape == adapter.proj_layer_norm.weight.data.shape ), F'{full_name} has size {value.shape}, but {adapter.proj_layer_norm.weight.data.shape} was found.' SCREAMING_SNAKE_CASE__ = value else: # has to be projection layer if "bias" in name: assert ( value.shape == adapter.proj.bias.data.shape ), F'{full_name} has size {value.shape}, but {adapter.proj.bias.data.shape} was found.' SCREAMING_SNAKE_CASE__ = value logger.info(F'Adapter proj layer bias was initialized from {full_name}.' ) if "weight" in name: assert ( value.shape == adapter.proj.weight.data.shape ), F'{full_name} has size {value.shape}, but {adapter.proj.weight.data.shape} was found.' SCREAMING_SNAKE_CASE__ = value logger.info(F'Adapter proj layer weight was initialized from {full_name}.' ) elif isinstance(UpperCamelCase_ , UpperCamelCase_ ): if "bias" in name: assert ( value.shape == adapter.layers[layer_id].conv.bias.data.shape ), F'{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.bias.data.shape} was found.' SCREAMING_SNAKE_CASE__ = value logger.info(F'Adapter layer {layer_id} bias was initialized from {full_name}.' ) elif "weight" in name: assert ( value.shape == adapter.layers[layer_id].conv.weight.data.shape ), F'{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.weight.data.shape} was found.' SCREAMING_SNAKE_CASE__ = value logger.info(F'Adapter layer {layer_id} bias was initialized from {full_name}.' ) else: unused_weights.append(UpperCamelCase_ ) def _lowercase ( UpperCamelCase_ ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = emb.weight.shape SCREAMING_SNAKE_CASE__ = nn.Linear(UpperCamelCase_ , UpperCamelCase_ , bias=UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = emb.weight.data return lin_layer @torch.no_grad() def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , ) -> Tuple: '''simple docstring''' SCREAMING_SNAKE_CASE__ = WavaVecaConfig.from_pretrained( UpperCamelCase_ , add_adapter=UpperCamelCase_ , adapter_stride=UpperCamelCase_ , adapter_kernel_size=UpperCamelCase_ , use_auth_token=UpperCamelCase_ , output_hidden_size=UpperCamelCase_ , ) SCREAMING_SNAKE_CASE__ = MBartConfig.from_pretrained(UpperCamelCase_ ) # load model SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={ 'config_yaml': config_yaml_path, 'data': '/'.join(dict_path.split('/' )[:-1] ), 'w2v_path': checkpoint_path, 'load_pretrained_decoder_from': None, } , ) SCREAMING_SNAKE_CASE__ = model[0].eval() # load feature extractor SCREAMING_SNAKE_CASE__ = WavaVecaFeatureExtractor.from_pretrained(UpperCamelCase_ , use_auth_token=UpperCamelCase_ ) # set weights for wav2vec2 encoder SCREAMING_SNAKE_CASE__ = WavaVecaModel(UpperCamelCase_ ) recursively_load_weights_wavaveca(model.encoder , UpperCamelCase_ ) # load decoder weights SCREAMING_SNAKE_CASE__ = MBartForCausalLM(UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=UpperCamelCase_ ) logger.warning(F'The following keys are missing when loading the decoder weights: {missing_keys}' ) logger.warning(F'The following keys are unexpected when loading the decoder weights: {unexpected_keys}' ) SCREAMING_SNAKE_CASE__ = SpeechEncoderDecoderModel(encoder=UpperCamelCase_ , decoder=UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = MBartaaTokenizer(UpperCamelCase_ ) tokenizer.save_pretrained(UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = hf_wavavec.config.to_dict() SCREAMING_SNAKE_CASE__ = tokenizer.pad_token_id SCREAMING_SNAKE_CASE__ = tokenizer.bos_token_id SCREAMING_SNAKE_CASE__ = tokenizer.eos_token_id SCREAMING_SNAKE_CASE__ = 'mbart50' SCREAMING_SNAKE_CASE__ = 'wav2vec2' SCREAMING_SNAKE_CASE__ = tokenizer.eos_token_id SCREAMING_SNAKE_CASE__ = 250004 SCREAMING_SNAKE_CASE__ = tokenizer.eos_token_id SCREAMING_SNAKE_CASE__ = SpeechEncoderDecoderConfig.from_dict(UpperCamelCase_ ) hf_wavavec.save_pretrained(UpperCamelCase_ ) feature_extractor.save_pretrained(UpperCamelCase_ ) if __name__ == "__main__": __snake_case = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_yaml_path""", default=None, type=str, help="""Path to yaml file of fine-tuned model""") parser.add_argument( """--encoder_config_path""", default="""facebook/wav2vec2-xls-r-1b""", type=str, help="""Path to hf encoder wav2vec2 checkpoint config""", ) parser.add_argument( """--decoder_config_path""", default="""facebook/mbart-large-50-one-to-many-mmt""", type=str, help="""Path to hf decoder checkpoint config""", ) parser.add_argument("""--add_adapter""", default=True, type=bool, help="""whethere to add model adapter layers""") parser.add_argument("""--adapter_stride""", default=2, type=int, help="""stride of adapter layers""") parser.add_argument("""--adapter_kernel_size""", default=3, type=int, help="""kernel size of adapter layers""") parser.add_argument("""--encoder_output_dim""", default=10_24, type=int, help="""encoder output dim""") parser.add_argument("""--start_token_id""", default=25_00_04, type=int, help="""`decoder_start_token_id` of model config""") __snake_case = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, args.config_yaml_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, add_adapter=args.add_adapter, adapter_kernel_size=args.adapter_kernel_size, adapter_stride=args.adapter_stride, decoder_start_token_id=args.start_token_id, encoder_output_dim=args.encoder_output_dim, )
176
1
'''simple docstring''' import argparse from typing import List import evaluate import numpy as np import torch from datasets import DatasetDict, load_dataset # New Code # # We'll be using StratifiedKFold for this example from sklearn.model_selection import StratifiedKFold from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to perform Cross Validation, # and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To help focus on the differences in the code, building `DataLoaders` # was refactored into its own function. # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## UpperCamelCase_ = 1_6 UpperCamelCase_ = 3_2 def lowercase__( __UpperCamelCase: Accelerator ,__UpperCamelCase: DatasetDict ,__UpperCamelCase: List[int] ,__UpperCamelCase: List[int] ,__UpperCamelCase: int = 16 ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = AutoTokenizer.from_pretrained('bert-base-cased' ) SCREAMING_SNAKE_CASE : Any = DatasetDict( { 'train': dataset['train'].select(__UpperCamelCase ), 'validation': dataset['train'].select(__UpperCamelCase ), 'test': dataset['validation'], } ) def tokenize_function(__UpperCamelCase: Union[str, Any] ): # max_length=None => use the model max length (it's actually the default) SCREAMING_SNAKE_CASE : int = tokenizer(examples['sentence1'] ,examples['sentence2'] ,truncation=__UpperCamelCase ,max_length=__UpperCamelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): SCREAMING_SNAKE_CASE : List[str] = datasets.map( __UpperCamelCase ,batched=__UpperCamelCase ,remove_columns=['idx', 'sentence1', 'sentence2'] ,) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library SCREAMING_SNAKE_CASE : Dict = tokenized_datasets.rename_column('label' ,'labels' ) def collate_fn(__UpperCamelCase: Optional[Any] ): # On TPU it's best to pad everything to the same length or training will be very slow. SCREAMING_SNAKE_CASE : int = 1_28 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": SCREAMING_SNAKE_CASE : Optional[Any] = 16 elif accelerator.mixed_precision != "no": SCREAMING_SNAKE_CASE : Dict = 8 else: SCREAMING_SNAKE_CASE : Dict = None return tokenizer.pad( __UpperCamelCase ,padding='longest' ,max_length=__UpperCamelCase ,pad_to_multiple_of=__UpperCamelCase ,return_tensors='pt' ,) # Instantiate dataloaders. SCREAMING_SNAKE_CASE : Optional[int] = DataLoader( tokenized_datasets['train'] ,shuffle=__UpperCamelCase ,collate_fn=__UpperCamelCase ,batch_size=__UpperCamelCase ) SCREAMING_SNAKE_CASE : str = DataLoader( tokenized_datasets['validation'] ,shuffle=__UpperCamelCase ,collate_fn=__UpperCamelCase ,batch_size=__UpperCamelCase ) SCREAMING_SNAKE_CASE : Dict = DataLoader( tokenized_datasets['test'] ,shuffle=__UpperCamelCase ,collate_fn=__UpperCamelCase ,batch_size=__UpperCamelCase ) return train_dataloader, eval_dataloader, test_dataloader def lowercase__( __UpperCamelCase: Any ,__UpperCamelCase: Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE : str = [] # Download the dataset SCREAMING_SNAKE_CASE : str = load_dataset('glue' ,'mrpc' ) # Create our splits SCREAMING_SNAKE_CASE : Any = StratifiedKFold(n_splits=int(args.num_folds ) ) # Initialize accelerator SCREAMING_SNAKE_CASE : Tuple = Accelerator(cpu=args.cpu ,mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs SCREAMING_SNAKE_CASE : List[str] = config['lr'] SCREAMING_SNAKE_CASE : Optional[int] = int(config['num_epochs'] ) SCREAMING_SNAKE_CASE : Tuple = int(config['seed'] ) SCREAMING_SNAKE_CASE : str = int(config['batch_size'] ) SCREAMING_SNAKE_CASE : List[str] = evaluate.load('glue' ,'mrpc' ) # If the batch size is too big we use gradient accumulation SCREAMING_SNAKE_CASE : List[str] = 1 if batch_size > MAX_GPU_BATCH_SIZE and accelerator.distributed_type != DistributedType.TPU: SCREAMING_SNAKE_CASE : int = batch_size // MAX_GPU_BATCH_SIZE SCREAMING_SNAKE_CASE : Optional[int] = MAX_GPU_BATCH_SIZE set_seed(__UpperCamelCase ) # New Code # # Create our folds: SCREAMING_SNAKE_CASE : Optional[Any] = kfold.split(np.zeros(datasets['train'].num_rows ) ,datasets['train']['label'] ) SCREAMING_SNAKE_CASE : Dict = [] # Iterate over them for i, (train_idxs, valid_idxs) in enumerate(__UpperCamelCase ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[Any] = get_fold_dataloaders( __UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,) # Instantiate the model (we build the model here so that the seed also control new weights initialization) SCREAMING_SNAKE_CASE : Tuple = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' ,return_dict=__UpperCamelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). SCREAMING_SNAKE_CASE : Optional[Any] = model.to(accelerator.device ) # Instantiate optimizer SCREAMING_SNAKE_CASE : str = AdamW(params=model.parameters() ,lr=__UpperCamelCase ) # Instantiate scheduler SCREAMING_SNAKE_CASE : Optional[Any] = get_linear_schedule_with_warmup( optimizer=__UpperCamelCase ,num_warmup_steps=1_00 ,num_training_steps=(len(__UpperCamelCase ) * num_epochs) // gradient_accumulation_steps ,) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[Any] = accelerator.prepare( __UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) # Now we train the model for epoch in range(__UpperCamelCase ): model.train() for step, batch in enumerate(__UpperCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) SCREAMING_SNAKE_CASE : Optional[int] = model(**__UpperCamelCase ) SCREAMING_SNAKE_CASE : Union[str, Any] = outputs.loss SCREAMING_SNAKE_CASE : List[Any] = loss / gradient_accumulation_steps accelerator.backward(__UpperCamelCase ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__UpperCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): SCREAMING_SNAKE_CASE : Tuple = model(**__UpperCamelCase ) SCREAMING_SNAKE_CASE : int = outputs.logits.argmax(dim=-1 ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[Any] = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=__UpperCamelCase ,references=__UpperCamelCase ,) SCREAMING_SNAKE_CASE : Optional[int] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"epoch {epoch}:" ,__UpperCamelCase ) # New Code # # We also run predictions on the test set at the very end SCREAMING_SNAKE_CASE : int = [] for step, batch in enumerate(__UpperCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): SCREAMING_SNAKE_CASE : str = model(**__UpperCamelCase ) SCREAMING_SNAKE_CASE : Tuple = outputs.logits SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[Any] = accelerator.gather_for_metrics((predictions, batch['labels']) ) fold_predictions.append(predictions.cpu() ) if i == 0: # We need all of the test predictions test_references.append(references.cpu() ) # Use accelerator.print to print only on the main process. test_predictions.append(torch.cat(__UpperCamelCase ,dim=0 ) ) # We now need to release all our memory and get rid of the current model, optimizer, etc accelerator.free_memory() # New Code # # Finally we check the accuracy of our folded results: SCREAMING_SNAKE_CASE : Optional[int] = torch.cat(__UpperCamelCase ,dim=0 ) SCREAMING_SNAKE_CASE : Any = torch.stack(__UpperCamelCase ,dim=0 ).sum(dim=0 ).div(int(args.num_folds ) ).argmax(dim=-1 ) SCREAMING_SNAKE_CASE : Any = metric.compute(predictions=__UpperCamelCase ,references=__UpperCamelCase ) accelerator.print('Average test metrics from all folds:' ,__UpperCamelCase ) def lowercase__( ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' ,type=__UpperCamelCase ,default=__UpperCamelCase ,choices=['no', 'fp16', 'bf16', 'fp8'] ,help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' ,) parser.add_argument('--cpu' ,action='store_true' ,help='If passed, will train on the CPU.' ) # New Code # parser.add_argument('--num_folds' ,type=__UpperCamelCase ,default=3 ,help='The number of splits to perform across the dataset' ) SCREAMING_SNAKE_CASE : int = parser.parse_args() SCREAMING_SNAKE_CASE : int = {'lr': 2e-5, 'num_epochs': 3, 'seed': 42, 'batch_size': 16} training_function(__UpperCamelCase ,__UpperCamelCase ) if __name__ == "__main__": main()
246
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator class _a : '''simple docstring''' def __init__( self, A ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = value SCREAMING_SNAKE_CASE : Node | None = None SCREAMING_SNAKE_CASE : Node | None = None class _a : '''simple docstring''' def __init__( self, A ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[Any] = tree def UpperCamelCase_ ( self, A ): '''simple docstring''' if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self ): '''simple docstring''' yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
246
1
__A : Dict = { 0: '0', 1: '1', 2: '2', 3: '3', 4: '4', 5: '5', 6: '6', 7: '7', 8: '8', 9: '9', 10: 'a', 11: 'b', 12: 'c', 13: 'd', 14: 'e', 15: 'f', } def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> str: '''simple docstring''' assert type(_UpperCAmelCase ) in (int, float) and decimal == int(_UpperCAmelCase ) lowerCAmelCase : Optional[Any] = int(_UpperCAmelCase ) lowerCAmelCase : List[str] = '' lowerCAmelCase : Optional[int] = False if decimal < 0: lowerCAmelCase : Optional[int] = True decimal *= -1 while decimal > 0: lowerCAmelCase , lowerCAmelCase : List[Any] = divmod(_UpperCAmelCase, 16 ) lowerCAmelCase : Optional[Any] = values[remainder] + hexadecimal lowerCAmelCase : Tuple = '0x' + hexadecimal if negative: lowerCAmelCase : List[Any] = '-' + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
138
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A_ ( _a , unittest.TestCase ): '''simple docstring''' a__ = LongformerTokenizer a__ = True a__ = LongformerTokenizerFast a__ = True def lowerCAmelCase_ (self ) -> Any: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __UpperCAmelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] __UpperCAmelCase = dict(zip(lowercase__ , range(len(lowercase__ ) ) ) ) __UpperCAmelCase = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] __UpperCAmelCase = {'''unk_token''': '''<unk>'''} __UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) __UpperCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(lowercase__ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(lowercase__ ) ) def lowerCAmelCase_ (self , **lowercase__ ) -> int: kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **lowercase__ ) def lowerCAmelCase_ (self , **lowercase__ ) -> Tuple: kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **lowercase__ ) def lowerCAmelCase_ (self , lowercase__ ) -> Dict: __UpperCAmelCase = '''lower newer''' __UpperCAmelCase = '''lower newer''' return input_text, output_text def lowerCAmelCase_ (self ) -> Optional[Any]: __UpperCAmelCase = self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) __UpperCAmelCase = '''lower newer''' __UpperCAmelCase = ['''l''', '''o''', '''w''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] __UpperCAmelCase = tokenizer.tokenize(lowercase__ ) # , add_prefix_space=True) self.assertListEqual(lowercase__ , lowercase__ ) __UpperCAmelCase = tokens + [tokenizer.unk_token] __UpperCAmelCase = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowercase__ ) , lowercase__ ) def lowerCAmelCase_ (self ) -> int: __UpperCAmelCase = self.get_tokenizer() self.assertListEqual(tokenizer.encode('''Hello world!''' , add_special_tokens=lowercase__ ) , [0, 31_414, 232, 328, 2] ) self.assertListEqual( tokenizer.encode('''Hello world! cécé herlolip 418''' , add_special_tokens=lowercase__ ) , [0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2] , ) @slow def lowerCAmelCase_ (self ) -> int: __UpperCAmelCase = self.tokenizer_class.from_pretrained('''allenai/longformer-base-4096''' ) __UpperCAmelCase = tokenizer.encode('''sequence builders''' , add_special_tokens=lowercase__ ) __UpperCAmelCase = tokenizer.encode('''multi-sequence build''' , add_special_tokens=lowercase__ ) __UpperCAmelCase = tokenizer.encode( '''sequence builders''' , add_special_tokens=lowercase__ , add_prefix_space=lowercase__ ) __UpperCAmelCase = tokenizer.encode( '''sequence builders''' , '''multi-sequence build''' , add_special_tokens=lowercase__ , add_prefix_space=lowercase__ ) __UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(lowercase__ ) __UpperCAmelCase = tokenizer.build_inputs_with_special_tokens(lowercase__ , lowercase__ ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def lowerCAmelCase_ (self ) -> Any: __UpperCAmelCase = self.get_tokenizer() __UpperCAmelCase = '''Encode this sequence.''' __UpperCAmelCase = tokenizer.byte_encoder[''' '''.encode('''utf-8''' )[0]] # Testing encoder arguments __UpperCAmelCase = tokenizer.encode(lowercase__ , add_special_tokens=lowercase__ , add_prefix_space=lowercase__ ) __UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(lowercase__ , lowercase__ ) __UpperCAmelCase = tokenizer.encode(lowercase__ , add_special_tokens=lowercase__ , add_prefix_space=lowercase__ ) __UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(lowercase__ , lowercase__ ) tokenizer.add_special_tokens({'''bos_token''': '''<s>'''} ) __UpperCAmelCase = tokenizer.encode(lowercase__ , add_special_tokens=lowercase__ ) __UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(lowercase__ , lowercase__ ) # Testing spaces after special tokens __UpperCAmelCase = '''<mask>''' tokenizer.add_special_tokens( {'''mask_token''': AddedToken(lowercase__ , lstrip=lowercase__ , rstrip=lowercase__ )} ) # mask token has a left space __UpperCAmelCase = tokenizer.convert_tokens_to_ids(lowercase__ ) __UpperCAmelCase = '''Encode <mask> sequence''' __UpperCAmelCase = '''Encode <mask>sequence''' __UpperCAmelCase = tokenizer.encode(lowercase__ ) __UpperCAmelCase = encoded.index(lowercase__ ) __UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(lowercase__ , lowercase__ ) __UpperCAmelCase = tokenizer.encode(lowercase__ ) __UpperCAmelCase = encoded.index(lowercase__ ) __UpperCAmelCase = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(lowercase__ , lowercase__ ) def lowerCAmelCase_ (self ) -> Tuple: pass def lowerCAmelCase_ (self ) -> int: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(lowercase__ , **lowercase__ ) __UpperCAmelCase = self.tokenizer_class.from_pretrained(lowercase__ , **lowercase__ ) __UpperCAmelCase = '''A, <mask> AllenNLP sentence.''' __UpperCAmelCase = tokenizer_r.encode_plus(lowercase__ , add_special_tokens=lowercase__ , return_token_type_ids=lowercase__ ) __UpperCAmelCase = tokenizer_p.encode_plus(lowercase__ , add_special_tokens=lowercase__ , return_token_type_ids=lowercase__ ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r['''token_type_ids'''] ) , sum(tokens_p['''token_type_ids'''] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r['''attention_mask'''] ) / len(tokens_r['''attention_mask'''] ) , sum(tokens_p['''attention_mask'''] ) / len(tokens_p['''attention_mask'''] ) , ) __UpperCAmelCase = tokenizer_r.convert_ids_to_tokens(tokens_r['''input_ids'''] ) __UpperCAmelCase = tokenizer_p.convert_ids_to_tokens(tokens_p['''input_ids'''] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p['''input_ids'''] , [0, 250, 6, 50_264, 3_823, 487, 21_992, 3_645, 4, 2] ) self.assertSequenceEqual(tokens_r['''input_ids'''] , [0, 250, 6, 50_264, 3_823, 487, 21_992, 3_645, 4, 2] ) self.assertSequenceEqual( lowercase__ , ['''<s>''', '''A''', ''',''', '''<mask>''', '''ĠAllen''', '''N''', '''LP''', '''Ġsentence''', '''.''', '''</s>'''] ) self.assertSequenceEqual( lowercase__ , ['''<s>''', '''A''', ''',''', '''<mask>''', '''ĠAllen''', '''N''', '''LP''', '''Ġsentence''', '''.''', '''</s>'''] ) def lowerCAmelCase_ (self ) -> Optional[int]: for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=lowercase__ , add_prefix_space=lowercase__ , trim_offsets=lowercase__ ) __UpperCAmelCase = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) __UpperCAmelCase = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state['''add_prefix_space'''] , lowercase__ ) self.assertEqual(post_processor_state['''add_prefix_space'''] , lowercase__ ) self.assertEqual(post_processor_state['''trim_offsets'''] , lowercase__ ) def lowerCAmelCase_ (self ) -> Union[str, Any]: # Test which aims to verify that the offsets are well adapted to the argument `add_prefix_space` and # `trim_offsets` for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __UpperCAmelCase = '''hello''' # `hello` is a token in the vocabulary of `pretrained_name` __UpperCAmelCase = F'''{text_of_1_token} {text_of_1_token}''' __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( lowercase__ , use_fast=lowercase__ , add_prefix_space=lowercase__ , trim_offsets=lowercase__ ) __UpperCAmelCase = tokenizer_r(lowercase__ , return_offsets_mapping=lowercase__ , add_special_tokens=lowercase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(lowercase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(lowercase__ ) + 1, len(lowercase__ ) + 1 + len(lowercase__ )) , ) __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( lowercase__ , use_fast=lowercase__ , add_prefix_space=lowercase__ , trim_offsets=lowercase__ ) __UpperCAmelCase = tokenizer_r(lowercase__ , return_offsets_mapping=lowercase__ , add_special_tokens=lowercase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(lowercase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(lowercase__ ) + 1, len(lowercase__ ) + 1 + len(lowercase__ )) , ) __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( lowercase__ , use_fast=lowercase__ , add_prefix_space=lowercase__ , trim_offsets=lowercase__ ) __UpperCAmelCase = tokenizer_r(lowercase__ , return_offsets_mapping=lowercase__ , add_special_tokens=lowercase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(lowercase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(lowercase__ ), len(lowercase__ ) + 1 + len(lowercase__ )) , ) __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( lowercase__ , use_fast=lowercase__ , add_prefix_space=lowercase__ , trim_offsets=lowercase__ ) __UpperCAmelCase = tokenizer_r(lowercase__ , return_offsets_mapping=lowercase__ , add_special_tokens=lowercase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(lowercase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(lowercase__ ), len(lowercase__ ) + 1 + len(lowercase__ )) , ) __UpperCAmelCase = F''' {text}''' # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( lowercase__ , use_fast=lowercase__ , add_prefix_space=lowercase__ , trim_offsets=lowercase__ ) __UpperCAmelCase = tokenizer_r(lowercase__ , return_offsets_mapping=lowercase__ , add_special_tokens=lowercase__ ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(lowercase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(lowercase__ ) + 1, 1 + len(lowercase__ ) + 1 + len(lowercase__ )) , ) __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( lowercase__ , use_fast=lowercase__ , add_prefix_space=lowercase__ , trim_offsets=lowercase__ ) __UpperCAmelCase = tokenizer_r(lowercase__ , return_offsets_mapping=lowercase__ , add_special_tokens=lowercase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(lowercase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(lowercase__ ), 1 + len(lowercase__ ) + 1 + len(lowercase__ )) , ) __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained( lowercase__ , use_fast=lowercase__ , add_prefix_space=lowercase__ , trim_offsets=lowercase__ ) __UpperCAmelCase = tokenizer_r(lowercase__ , return_offsets_mapping=lowercase__ , add_special_tokens=lowercase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(lowercase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(lowercase__ ), 1 + len(lowercase__ ) + 1 + len(lowercase__ )) , )
333
0
'''simple docstring''' from __future__ import annotations __UpperCAmelCase :Tuple = "Muhammad Umer Farooq" __UpperCAmelCase :Tuple = "MIT" __UpperCAmelCase :Union[str, Any] = "1.0.0" __UpperCAmelCase :Optional[int] = "Muhammad Umer Farooq" __UpperCAmelCase :Optional[Any] = "contact@muhammadumerfarooq.me" __UpperCAmelCase :Any = "Alpha" import re from html.parser import HTMLParser from urllib import parse import requests class a ( _a ): """simple docstring""" def __init__( self : Tuple , snake_case : str ) -> None: super().__init__() __UpperCAmelCase : list[str] = [] __UpperCAmelCase : Optional[int] = domain def lowerCamelCase__ ( self : Union[str, Any] , snake_case : str , snake_case : list[tuple[str, str | None]] ) -> None: # Only parse the 'anchor' tag. if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: __UpperCAmelCase : Optional[Any] = parse.urljoin(self.domain , snake_case ) self.urls.append(snake_case ) def _a ( _lowercase : str ): '''simple docstring''' return ".".join(get_sub_domain_name(_lowercase ).split('''.''' )[-2:] ) def _a ( _lowercase : str ): '''simple docstring''' return parse.urlparse(_lowercase ).netloc def _a ( _lowercase : str = "https://github.com" ): '''simple docstring''' __UpperCAmelCase : Union[str, Any] = get_domain_name(_lowercase ) # Initialize the parser __UpperCAmelCase : Dict = Parser(_lowercase ) try: # Open URL __UpperCAmelCase : Dict = requests.get(_lowercase ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through __UpperCAmelCase : str = set() for link in parser.urls: # open URL. # read = requests.get(link) try: __UpperCAmelCase : Tuple = requests.get(_lowercase ) # Get the valid email. __UpperCAmelCase : Dict = re.findall('''[a-zA-Z0-9]+@''' + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(_lowercase ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(_lowercase ) if __name__ == "__main__": __UpperCAmelCase :List[str] = emails_from_url("https://github.com") print(f"""{len(emails)} emails found:""") print("\n".join(sorted(emails)))
240
'''simple docstring''' import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, ) from transformers.models.albert.modeling_albert import ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST class a : """simple docstring""" def __init__( self : Any , snake_case : Any , snake_case : Optional[int]=13 , snake_case : List[str]=7 , snake_case : List[str]=True , snake_case : List[Any]=True , snake_case : int=True , snake_case : Tuple=True , snake_case : int=99 , snake_case : Any=16 , snake_case : Dict=36 , snake_case : Any=6 , snake_case : Dict=6 , snake_case : Dict=6 , snake_case : int=37 , snake_case : int="gelu" , snake_case : str=0.1 , snake_case : Any=0.1 , snake_case : Dict=512 , snake_case : List[Any]=16 , snake_case : Any=2 , snake_case : Any=0.02 , snake_case : Optional[int]=3 , snake_case : List[Any]=4 , snake_case : List[str]=None , ) -> Union[str, Any]: __UpperCAmelCase : str = parent __UpperCAmelCase : List[str] = batch_size __UpperCAmelCase : int = seq_length __UpperCAmelCase : Optional[Any] = is_training __UpperCAmelCase : List[str] = use_input_mask __UpperCAmelCase : List[Any] = use_token_type_ids __UpperCAmelCase : Dict = use_labels __UpperCAmelCase : int = vocab_size __UpperCAmelCase : Optional[int] = embedding_size __UpperCAmelCase : str = hidden_size __UpperCAmelCase : Union[str, Any] = num_hidden_layers __UpperCAmelCase : List[Any] = num_hidden_groups __UpperCAmelCase : Any = num_attention_heads __UpperCAmelCase : int = intermediate_size __UpperCAmelCase : Optional[Any] = hidden_act __UpperCAmelCase : Tuple = hidden_dropout_prob __UpperCAmelCase : Union[str, Any] = attention_probs_dropout_prob __UpperCAmelCase : Optional[int] = max_position_embeddings __UpperCAmelCase : List[str] = type_vocab_size __UpperCAmelCase : Any = type_sequence_label_size __UpperCAmelCase : List[Any] = initializer_range __UpperCAmelCase : Dict = num_labels __UpperCAmelCase : str = num_choices __UpperCAmelCase : Union[str, Any] = scope def lowerCamelCase__ ( self : List[Any] ) -> Optional[Any]: __UpperCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCAmelCase : int = None if self.use_input_mask: __UpperCAmelCase : Tuple = random_attention_mask([self.batch_size, self.seq_length] ) __UpperCAmelCase : Dict = None if self.use_token_type_ids: __UpperCAmelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __UpperCAmelCase : int = None __UpperCAmelCase : List[str] = None __UpperCAmelCase : Optional[Any] = None if self.use_labels: __UpperCAmelCase : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __UpperCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __UpperCAmelCase : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) __UpperCAmelCase : List[str] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase__ ( self : List[str] ) -> Optional[Any]: return AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , num_hidden_groups=self.num_hidden_groups , ) def lowerCamelCase__ ( self : Tuple , snake_case : Optional[Any] , snake_case : Optional[int] , snake_case : List[Any] , snake_case : Optional[Any] , snake_case : Optional[int] , snake_case : Union[str, Any] , snake_case : int ) -> Optional[int]: __UpperCAmelCase : List[Any] = AlbertModel(config=snake_case ) model.to(snake_case ) model.eval() __UpperCAmelCase : Tuple = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case ) __UpperCAmelCase : List[str] = model(snake_case , token_type_ids=snake_case ) __UpperCAmelCase : str = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowerCamelCase__ ( self : List[str] , snake_case : str , snake_case : Optional[int] , snake_case : List[Any] , snake_case : Any , snake_case : Dict , snake_case : Dict , snake_case : Optional[int] ) -> Optional[int]: __UpperCAmelCase : str = AlbertForPreTraining(config=snake_case ) model.to(snake_case ) model.eval() __UpperCAmelCase : Union[str, Any] = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case , sentence_order_label=snake_case , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.sop_logits.shape , (self.batch_size, config.num_labels) ) def lowerCamelCase__ ( self : Dict , snake_case : Union[str, Any] , snake_case : Dict , snake_case : Any , snake_case : Tuple , snake_case : Union[str, Any] , snake_case : Any , snake_case : Tuple ) -> Union[str, Any]: __UpperCAmelCase : Dict = AlbertForMaskedLM(config=snake_case ) model.to(snake_case ) model.eval() __UpperCAmelCase : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase__ ( self : str , snake_case : Tuple , snake_case : List[str] , snake_case : Union[str, Any] , snake_case : List[str] , snake_case : Tuple , snake_case : Optional[Any] , snake_case : Tuple ) -> int: __UpperCAmelCase : Optional[Any] = AlbertForQuestionAnswering(config=snake_case ) model.to(snake_case ) model.eval() __UpperCAmelCase : Optional[Any] = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , start_positions=snake_case , end_positions=snake_case , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase__ ( self : Tuple , snake_case : List[str] , snake_case : Dict , snake_case : Optional[int] , snake_case : Dict , snake_case : int , snake_case : Optional[int] , snake_case : Optional[Any] ) -> Any: __UpperCAmelCase : Optional[int] = self.num_labels __UpperCAmelCase : Any = AlbertForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() __UpperCAmelCase : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase__ ( self : Tuple , snake_case : Tuple , snake_case : List[Any] , snake_case : Optional[int] , snake_case : str , snake_case : Dict , snake_case : Union[str, Any] , snake_case : List[str] ) -> int: __UpperCAmelCase : Optional[int] = self.num_labels __UpperCAmelCase : Optional[int] = AlbertForTokenClassification(config=snake_case ) model.to(snake_case ) model.eval() __UpperCAmelCase : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase__ ( self : Tuple , snake_case : Tuple , snake_case : List[Any] , snake_case : Dict , snake_case : int , snake_case : List[Any] , snake_case : List[Any] , snake_case : Optional[Any] ) -> Tuple: __UpperCAmelCase : Optional[int] = self.num_choices __UpperCAmelCase : List[Any] = AlbertForMultipleChoice(config=snake_case ) model.to(snake_case ) model.eval() __UpperCAmelCase : Dict = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __UpperCAmelCase : List[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __UpperCAmelCase : List[Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __UpperCAmelCase : List[str] = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase__ ( self : Union[str, Any] ) -> Any: __UpperCAmelCase : List[Any] = self.prepare_config_and_inputs() ( ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ) : List[Any] = config_and_inputs __UpperCAmelCase : List[str] = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class a ( _a , _a , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = ( ( AlbertModel, AlbertForPreTraining, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertForQuestionAnswering, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE : Any = ( { "feature-extraction": AlbertModel, "fill-mask": AlbertForMaskedLM, "question-answering": AlbertForQuestionAnswering, "text-classification": AlbertForSequenceClassification, "token-classification": AlbertForTokenClassification, "zero-shot": AlbertForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Any = True def lowerCamelCase__ ( self : Optional[int] , snake_case : Any , snake_case : Dict , snake_case : Tuple=False ) -> Optional[Any]: __UpperCAmelCase : Any = super()._prepare_for_class(snake_case , snake_case , return_labels=snake_case ) if return_labels: if model_class in get_values(snake_case ): __UpperCAmelCase : Dict = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=snake_case ) __UpperCAmelCase : Union[str, Any] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=snake_case ) return inputs_dict def lowerCamelCase__ ( self : Dict ) -> int: __UpperCAmelCase : List[Any] = AlbertModelTester(self ) __UpperCAmelCase : Any = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def lowerCamelCase__ ( self : Optional[Any] ) -> Optional[Any]: self.config_tester.run_common_tests() def lowerCamelCase__ ( self : Any ) -> Any: __UpperCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def lowerCamelCase__ ( self : Any ) -> Optional[Any]: __UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*snake_case ) def lowerCamelCase__ ( self : Optional[int] ) -> Optional[Any]: __UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case ) def lowerCamelCase__ ( self : Union[str, Any] ) -> Tuple: __UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*snake_case ) def lowerCamelCase__ ( self : Dict ) -> str: __UpperCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) def lowerCamelCase__ ( self : List[str] ) -> Tuple: __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def lowerCamelCase__ ( self : str ) -> Any: __UpperCAmelCase : int = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __UpperCAmelCase : Tuple = type self.model_tester.create_and_check_model(*snake_case ) @slow def lowerCamelCase__ ( self : Tuple ) -> Optional[int]: for model_name in ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCAmelCase : List[Any] = AlbertModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @require_torch class a ( unittest.TestCase ): """simple docstring""" @slow def lowerCamelCase__ ( self : Optional[Any] ) -> Union[str, Any]: __UpperCAmelCase : Optional[int] = AlbertModel.from_pretrained('''albert-base-v2''' ) __UpperCAmelCase : str = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) __UpperCAmelCase : List[Any] = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __UpperCAmelCase : Optional[int] = model(snake_case , attention_mask=snake_case )[0] __UpperCAmelCase : Any = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , snake_case ) __UpperCAmelCase : int = torch.tensor( [[[-0.6_513, 1.5_035, -0.2_766], [-0.6_515, 1.5_046, -0.2_780], [-0.6_512, 1.5_049, -0.2_784]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , snake_case , atol=1E-4 ) )
240
1
"""simple docstring""" import numpy # List of input, output pairs lowercase__ : List[str] = ( ((5, 2, 3), 1_5), ((6, 5, 9), 2_5), ((1_1, 1_2, 1_3), 4_1), ((1, 1, 1), 8), ((1_1, 1_2, 1_3), 4_1), ) lowercase__ : Optional[int] = (((5_1_5, 2_2, 1_3), 5_5_5), ((6_1, 3_5, 4_9), 1_5_0)) lowercase__ : Optional[int] = [2, 4, 1, 5] lowercase__ : Union[str, Any] = len(train_data) lowercase__ : List[Any] = 0.009 def UpperCamelCase_ ( lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : str="train" ) -> int: """simple docstring""" return calculate_hypothesis_value(__lowerCAmelCase , __lowerCAmelCase ) - output( __lowerCAmelCase , __lowerCAmelCase ) def UpperCamelCase_ ( lowerCAmelCase__ : str ) -> Dict: """simple docstring""" lowerCAmelCase_ : Any = 0 for i in range(len(__lowerCAmelCase ) - 1 ): hyp_val += data_input_tuple[i] * parameter_vector[i + 1] hyp_val += parameter_vector[0] return hyp_val def UpperCamelCase_ ( lowerCAmelCase__ : int , lowerCAmelCase__ : int ) -> Dict: """simple docstring""" if data_set == "train": return train_data[example_no][1] elif data_set == "test": return test_data[example_no][1] return None def UpperCamelCase_ ( lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Union[str, Any] ) -> List[Any]: """simple docstring""" if data_set == "train": return _hypothesis_value(train_data[example_no][0] ) elif data_set == "test": return _hypothesis_value(test_data[example_no][0] ) return None def UpperCamelCase_ ( lowerCAmelCase__ : str , lowerCAmelCase__ : Any=m ) -> List[Any]: """simple docstring""" lowerCAmelCase_ : Optional[Any] = 0 for i in range(__lowerCAmelCase ): if index == -1: summation_value += _error(__lowerCAmelCase ) else: summation_value += _error(__lowerCAmelCase ) * train_data[i][0][index] return summation_value def UpperCamelCase_ ( lowerCAmelCase__ : str ) -> Optional[Any]: """simple docstring""" lowerCAmelCase_ : List[str] = summation_of_cost_derivative(__lowerCAmelCase , __lowerCAmelCase ) / m return cost_derivative_value def UpperCamelCase_ ( ) -> Dict: """simple docstring""" global parameter_vector # Tune these values to set a tolerance value for predicted output lowerCAmelCase_ : Any = 0.000002 lowerCAmelCase_ : str = 0 lowerCAmelCase_ : Optional[Any] = 0 while True: j += 1 lowerCAmelCase_ : Dict = [0, 0, 0, 0] for i in range(0 , len(__lowerCAmelCase ) ): lowerCAmelCase_ : Optional[int] = get_cost_derivative(i - 1 ) lowerCAmelCase_ : Union[str, Any] = ( parameter_vector[i] - LEARNING_RATE * cost_derivative ) if numpy.allclose( __lowerCAmelCase , __lowerCAmelCase , atol=__lowerCAmelCase , rtol=__lowerCAmelCase , ): break lowerCAmelCase_ : Union[str, Any] = temp_parameter_vector print(('Number of iterations:', j) ) def UpperCamelCase_ ( ) -> Any: """simple docstring""" for i in range(len(__lowerCAmelCase ) ): print(('Actual output value:', output(__lowerCAmelCase , 'test' )) ) print(('Hypothesis output:', calculate_hypothesis_value(__lowerCAmelCase , 'test' )) ) if __name__ == "__main__": run_gradient_descent() print("""\nTesting gradient descent for a linear hypothesis function.\n""") test_gradient_descent()
224
'''simple docstring''' import json import logging import os import sys from pathlib import Path import finetune_rag from transformers.file_utils import is_apex_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, require_ray, require_torch_gpu, require_torch_multi_gpu, ) logging.basicConfig(level=logging.DEBUG) lowerCamelCase__ = logging.getLogger() lowerCamelCase__ = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class lowerCAmelCase__ ( UpperCAmelCase__ ): def lowerCAmelCase__ ( self : List[str] , lowerCamelCase__ : Optional[int] ) ->Tuple: '''simple docstring''' os.makedirs(lowerCamelCase__ , exist_ok=lowerCamelCase__ ) _UpperCAmelCase : List[Any] = {"source": "What is love ?", "target": "life"} _UpperCAmelCase : Any = {"train": 12, "val": 2, "test": 2} for split in ["train", "test", "val"]: for field in ["source", "target"]: _UpperCAmelCase : Dict = "\n".join([contents[field]] * n_lines[split] ) with open(os.path.join(lowerCamelCase__ , F"""{split}.{field}""" ) , "w" ) as f: f.write(lowerCamelCase__ ) def lowerCAmelCase__ ( self : int , lowerCamelCase__ : int , lowerCamelCase__ : str = "pytorch" ) ->Any: '''simple docstring''' _UpperCAmelCase : Any = self.get_auto_remove_tmp_dir() _UpperCAmelCase : int = os.path.join(lowerCamelCase__ , "output" ) _UpperCAmelCase : Tuple = os.path.join(lowerCamelCase__ , "data" ) self._create_dummy_data(data_dir=lowerCamelCase__ ) _UpperCAmelCase : str = F""" --data_dir {data_dir} \ --output_dir {output_dir} \ --model_name_or_path facebook/rag-sequence-base \ --model_type rag_sequence \ --do_train \ --do_predict \ --n_val -1 \ --val_check_interval 1.0 \ --train_batch_size 2 \ --eval_batch_size 1 \ --max_source_length 25 \ --max_target_length 25 \ --val_max_target_length 25 \ --test_max_target_length 25 \ --label_smoothing 0.1 \ --dropout 0.1 \ --attention_dropout 0.1 \ --weight_decay 0.001 \ --adam_epsilon 1e-08 \ --max_grad_norm 0.1 \ --lr_scheduler polynomial \ --learning_rate 3e-04 \ --num_train_epochs 1 \ --warmup_steps 4 \ --gradient_accumulation_steps 1 \ --distributed-port 8787 \ --use_dummy_dataset 1 \ --distributed_retriever {distributed_retriever} \ """.split() if gpus > 0: testargs.append(F"""--gpus={gpus}""" ) if is_apex_available(): testargs.append("--fp16" ) else: testargs.append("--gpus=0" ) testargs.append("--distributed_backend=ddp_cpu" ) testargs.append("--num_processes=2" ) _UpperCAmelCase : str = [sys.executable, str(Path(finetune_rag.__file__ ).resolve() )] + testargs execute_subprocess_async(lowerCamelCase__ , env=self.get_env() ) _UpperCAmelCase : Optional[int] = os.path.join(lowerCamelCase__ , "metrics.json" ) with open(lowerCamelCase__ ) as f: _UpperCAmelCase : Dict = json.load(lowerCamelCase__ ) return result @require_torch_gpu def lowerCAmelCase__ ( self : Dict ) ->Dict: '''simple docstring''' _UpperCAmelCase : Optional[Any] = self._run_finetune(gpus=1 ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 ) @require_torch_multi_gpu def lowerCAmelCase__ ( self : List[Any] ) ->List[str]: '''simple docstring''' _UpperCAmelCase : List[str] = self._run_finetune(gpus=2 ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 ) @require_torch_gpu @require_ray def lowerCAmelCase__ ( self : int ) ->str: '''simple docstring''' _UpperCAmelCase : Any = self._run_finetune(gpus=1 , distributed_retriever="ray" ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 ) @require_torch_multi_gpu @require_ray def lowerCAmelCase__ ( self : int ) ->Any: '''simple docstring''' _UpperCAmelCase : str = self._run_finetune(gpus=1 , distributed_retriever="ray" ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 )
234
0
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCamelCase : str = {"""configuration_focalnet""": ["""FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """FocalNetConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Dict = [ """FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """FocalNetForImageClassification""", """FocalNetForMaskedImageModeling""", """FocalNetBackbone""", """FocalNetModel""", """FocalNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_focalnet import FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FocalNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_focalnet import ( FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST, FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, FocalNetPreTrainedModel, ) else: import sys _lowerCamelCase : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
231
from __future__ import annotations import queue class UpperCamelCase_ : '''simple docstring''' def __init__( self : Optional[Any] , UpperCAmelCase__ : Dict) ->Any: '''simple docstring''' A__ = data A__ = None A__ = None def SCREAMING_SNAKE_CASE ( ) -> TreeNode: """simple docstring""" print('''\n********Press N to stop entering at any point of time********\n''' ) A__ = input('''Enter the value of the root node: ''' ).strip().lower() A__ = queue.Queue() A__ = TreeNode(int(lowercase_ ) ) q.put(lowercase_ ) while not q.empty(): A__ = q.get() A__ = f"""Enter the left node of {node_found.data}: """ A__ = input(lowercase_ ).strip().lower() or '''n''' if check == "n": return tree_node A__ = TreeNode(int(lowercase_ ) ) A__ = left_node q.put(lowercase_ ) A__ = f"""Enter the right node of {node_found.data}: """ A__ = input(lowercase_ ).strip().lower() or '''n''' if check == "n": return tree_node A__ = TreeNode(int(lowercase_ ) ) A__ = right_node q.put(lowercase_ ) raise def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return print(node.data , end=''',''' ) pre_order(node.left ) pre_order(node.right ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return in_order(node.left ) print(node.data , end=''',''' ) in_order(node.right ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return post_order(node.left ) post_order(node.right ) print(node.data , end=''',''' ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ = queue.Queue() q.put(lowercase_ ) while not q.empty(): A__ = q.get() print(node_dequeued.data , end=''',''' ) if node_dequeued.left: q.put(node_dequeued.left ) if node_dequeued.right: q.put(node_dequeued.right ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ = queue.Queue() q.put(lowercase_ ) while not q.empty(): A__ = [] while not q.empty(): A__ = q.get() print(node_dequeued.data , end=''',''' ) if node_dequeued.left: list_.append(node_dequeued.left ) if node_dequeued.right: list_.append(node_dequeued.right ) print() for node in list_: q.put(lowercase_ ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ = [] A__ = node while n or stack: while n: # start from root node, find its left child print(n.data , end=''',''' ) stack.append(lowercase_ ) A__ = n.left # end of while means current node doesn't have left child A__ = stack.pop() # start to traverse its right child A__ = n.right def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ = [] A__ = node while n or stack: while n: stack.append(lowercase_ ) A__ = n.left A__ = stack.pop() print(n.data , end=''',''' ) A__ = n.right def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ , A__ = [], [] A__ = node stacka.append(lowercase_ ) while stacka: # to find the reversed order of post order, store it in stack2 A__ = stacka.pop() if n.left: stacka.append(n.left ) if n.right: stacka.append(n.right ) stacka.append(lowercase_ ) while stacka: # pop up from stack2 will be the post order print(stacka.pop().data , end=''',''' ) def SCREAMING_SNAKE_CASE ( lowercase_ = "" , lowercase_=50 , lowercase_="*" ) -> str: """simple docstring""" if not s: return "\n" + width * char A__ , A__ = divmod(width - len(lowercase_ ) - 2 , 2 ) return f"""{left * char} {s} {(left + extra) * char}""" if __name__ == "__main__": import doctest doctest.testmod() print(prompt("""Binary Tree Traversals""")) _lowerCamelCase : TreeNode = build_tree() print(prompt("""Pre Order Traversal""")) pre_order(node) print(prompt() + """\n""") print(prompt("""In Order Traversal""")) in_order(node) print(prompt() + """\n""") print(prompt("""Post Order Traversal""")) post_order(node) print(prompt() + """\n""") print(prompt("""Level Order Traversal""")) level_order(node) print(prompt() + """\n""") print(prompt("""Actual Level Order Traversal""")) level_order_actual(node) print("""*""" * 50 + """\n""") print(prompt("""Pre Order Traversal - Iteration Version""")) pre_order_iter(node) print(prompt() + """\n""") print(prompt("""In Order Traversal - Iteration Version""")) in_order_iter(node) print(prompt() + """\n""") print(prompt("""Post Order Traversal - Iteration Version""")) post_order_iter(node) print(prompt())
231
1
import unittest import numpy as np import torch from torch import nn from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import KandinskyVaaPriorPipeline, PriorTransformer, UnCLIPScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import enable_full_determinism, skip_mps from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class a__ ( A__ , unittest.TestCase ): A = KandinskyVaaPriorPipeline A = ['prompt'] A = ['prompt', 'negative_prompt'] A = [ 'num_images_per_prompt', 'generator', 'num_inference_steps', 'latents', 'negative_prompt', 'guidance_scale', 'output_type', 'return_dict', ] A = False @property def __UpperCamelCase ( self : int ): """simple docstring""" return 32 @property def __UpperCamelCase ( self : int ): """simple docstring""" return 32 @property def __UpperCamelCase ( self : List[Any] ): """simple docstring""" return self.time_input_dim @property def __UpperCamelCase ( self : Dict ): """simple docstring""" return self.time_input_dim * 4 @property def __UpperCamelCase ( self : List[Any] ): """simple docstring""" return 100 @property def __UpperCamelCase ( self : int ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) return tokenizer @property def __UpperCamelCase ( self : List[str] ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ : str = CLIPTextConfig( bos_token_id=0,eos_token_id=2,hidden_size=self.text_embedder_hidden_size,projection_dim=self.text_embedder_hidden_size,intermediate_size=37,layer_norm_eps=1E-05,num_attention_heads=4,num_hidden_layers=5,pad_token_id=1,vocab_size=1000,) return CLIPTextModelWithProjection(_A ) @property def __UpperCamelCase ( self : Optional[Any] ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ : Tuple = { "num_attention_heads": 2, "attention_head_dim": 12, "embedding_dim": self.text_embedder_hidden_size, "num_layers": 1, } SCREAMING_SNAKE_CASE_ : List[Any] = PriorTransformer(**_A ) # clip_std and clip_mean is initialized to be 0 so PriorTransformer.post_process_latents will always return 0 - set clip_std to be 1 so it won't return 0 SCREAMING_SNAKE_CASE_ : Dict = nn.Parameter(torch.ones(model.clip_std.shape ) ) return model @property def __UpperCamelCase ( self : Any ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ : Tuple = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size,image_size=224,projection_dim=self.text_embedder_hidden_size,intermediate_size=37,num_attention_heads=4,num_channels=3,num_hidden_layers=5,patch_size=14,) SCREAMING_SNAKE_CASE_ : Optional[Any] = CLIPVisionModelWithProjection(_A ) return model @property def __UpperCamelCase ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = CLIPImageProcessor( crop_size=224,do_center_crop=_A,do_normalize=_A,do_resize=_A,image_mean=[0.48145466, 0.4578275, 0.40821073],image_std=[0.26862954, 0.26130258, 0.27577711],resample=3,size=224,) return image_processor def __UpperCamelCase ( self : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = self.dummy_prior SCREAMING_SNAKE_CASE_ : List[str] = self.dummy_image_encoder SCREAMING_SNAKE_CASE_ : Any = self.dummy_text_encoder SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.dummy_tokenizer SCREAMING_SNAKE_CASE_ : Optional[int] = self.dummy_image_processor SCREAMING_SNAKE_CASE_ : Any = UnCLIPScheduler( variance_type="fixed_small_log",prediction_type="sample",num_train_timesteps=1000,clip_sample=_A,clip_sample_range=10.0,) SCREAMING_SNAKE_CASE_ : str = { "prior": prior, "image_encoder": image_encoder, "text_encoder": text_encoder, "tokenizer": tokenizer, "scheduler": scheduler, "image_processor": image_processor, } return components def __UpperCamelCase ( self : Tuple,_A : int,_A : Dict=0 ): """simple docstring""" if str(_A ).startswith("mps" ): SCREAMING_SNAKE_CASE_ : Tuple = torch.manual_seed(_A ) else: SCREAMING_SNAKE_CASE_ : Union[str, Any] = torch.Generator(device=_A ).manual_seed(_A ) SCREAMING_SNAKE_CASE_ : str = { "prompt": "horse", "generator": generator, "guidance_scale": 4.0, "num_inference_steps": 2, "output_type": "np", } return inputs def __UpperCamelCase ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = "cpu" SCREAMING_SNAKE_CASE_ : List[Any] = self.get_dummy_components() SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.pipeline_class(**_A ) SCREAMING_SNAKE_CASE_ : List[str] = pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) SCREAMING_SNAKE_CASE_ : List[str] = pipe(**self.get_dummy_inputs(_A ) ) SCREAMING_SNAKE_CASE_ : int = output.image_embeds SCREAMING_SNAKE_CASE_ : Any = pipe( **self.get_dummy_inputs(_A ),return_dict=_A,)[0] SCREAMING_SNAKE_CASE_ : Any = image[0, -10:] SCREAMING_SNAKE_CASE_ : List[Any] = image_from_tuple[0, -10:] assert image.shape == (1, 32) SCREAMING_SNAKE_CASE_ : Optional[int] = np.array( [-0.0532, 1.7120, 0.3656, -1.0852, -0.8946, -1.1756, 0.4348, 0.2482, 0.5146, -0.1156] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def __UpperCamelCase ( self : Any ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = torch_device == "cpu" SCREAMING_SNAKE_CASE_ : str = True SCREAMING_SNAKE_CASE_ : List[str] = False self._test_inference_batch_single_identical( test_max_difference=_A,relax_max_difference=_A,test_mean_pixel_difference=_A,) @skip_mps def __UpperCamelCase ( self : str ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = torch_device == "cpu" SCREAMING_SNAKE_CASE_ : Dict = False self._test_attention_slicing_forward_pass( test_max_difference=_A,test_mean_pixel_difference=_A,)
18
from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING __UpperCAmelCase = logging.get_logger(__name__) @add_end_docstrings(_snake_case ) class lowerCamelCase (_snake_case ): '''simple docstring''' def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: super().__init__(*_UpperCamelCase , **_UpperCamelCase ) self.check_model_type(_UpperCamelCase ) def __UpperCAmelCase ( self , _UpperCamelCase=None , _UpperCamelCase=None , _UpperCamelCase=None , **_UpperCamelCase ) -> List[Any]: UpperCAmelCase_ , UpperCAmelCase_ : Tuple = {}, {} if padding is not None: UpperCAmelCase_ : List[str] = padding if truncation is not None: UpperCAmelCase_ : Tuple = truncation if top_k is not None: UpperCAmelCase_ : Dict = top_k return preprocess_params, {}, postprocess_params def __call__( self , _UpperCamelCase , _UpperCamelCase = None , **_UpperCamelCase ) -> int: if isinstance(_UpperCamelCase , (Image.Image, str) ) and isinstance(_UpperCamelCase , _UpperCamelCase ): UpperCAmelCase_ : Optional[Any] = {'image': image, 'question': question} else: UpperCAmelCase_ : List[str] = image UpperCAmelCase_ : Optional[Any] = super().__call__(_UpperCamelCase , **_UpperCamelCase ) return results def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase=False , _UpperCamelCase=False ) -> Optional[Any]: UpperCAmelCase_ : List[Any] = load_image(inputs['image'] ) UpperCAmelCase_ : Dict = self.tokenizer( inputs['question'] , return_tensors=self.framework , padding=_UpperCamelCase , truncation=_UpperCamelCase ) UpperCAmelCase_ : int = self.image_processor(images=_UpperCamelCase , return_tensors=self.framework ) model_inputs.update(_UpperCamelCase ) return model_inputs def __UpperCAmelCase ( self , _UpperCamelCase ) -> Optional[int]: UpperCAmelCase_ : Any = self.model(**_UpperCamelCase ) return model_outputs def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase=5 ) -> str: if top_k > self.model.config.num_labels: UpperCAmelCase_ : Union[str, Any] = self.model.config.num_labels if self.framework == "pt": UpperCAmelCase_ : List[str] = model_outputs.logits.sigmoid()[0] UpperCAmelCase_ , UpperCAmelCase_ : str = probs.topk(_UpperCamelCase ) else: raise ValueError(f"Unsupported framework: {self.framework}" ) UpperCAmelCase_ : Optional[Any] = scores.tolist() UpperCAmelCase_ : Tuple = ids.tolist() return [{"score": score, "answer": self.model.config.idalabel[_id]} for score, _id in zip(_UpperCamelCase , _UpperCamelCase )]
29
0
import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets __A = """\ @inproceedings{lin-2004-rouge, title = \"{ROUGE}: A Package for Automatic Evaluation of Summaries\", author = \"Lin, Chin-Yew\", booktitle = \"Text Summarization Branches Out\", month = jul, year = \"2004\", address = \"Barcelona, Spain\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/W04-1013\", pages = \"74--81\", } """ __A = """\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge """ __A = """ Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `\"rouge{n}\"` (e.g. `\"rouge1\"`, `\"rouge2\"`) where: {n} is the n-gram based scoring, `\"rougeL\"`: Longest common subsequence based scoring. `\"rougeLSum\"`: rougeLsum splits text using `\"\n\"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric('rouge') >>> predictions = [\"hello there\", \"general kenobi\"] >>> references = [\"hello there\", \"general kenobi\"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) ['rouge1', 'rouge2', 'rougeL', 'rougeLsum'] >>> print(results[\"rouge1\"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results[\"rouge1\"].mid.fmeasure) 1.0 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE ( datasets.Metric ): """simple docstring""" def __A ( self: List[Any] ) -> int: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/google-research/google-research/tree/master/rouge'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/ROUGE_(metric)''', '''https://github.com/google-research/google-research/tree/master/rouge''', ] , ) def __A ( self: Dict , __A: Optional[int] , __A: int , __A: int=None , __A: Optional[Any]=True , __A: Optional[Any]=False ) -> int: if rouge_types is None: _A = ['''rouge1''', '''rouge2''', '''rougeL''', '''rougeLsum'''] _A = rouge_scorer.RougeScorer(rouge_types=__A , use_stemmer=__A ) if use_aggregator: _A = scoring.BootstrapAggregator() else: _A = [] for ref, pred in zip(__A , __A ): _A = scorer.score(__A , __A ) if use_aggregator: aggregator.add_scores(__A ) else: scores.append(__A ) if use_aggregator: _A = aggregator.aggregate() else: _A = {} for key in scores[0]: _A = [score[key] for score in scores] return result
352
import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor __A = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE ( snake_case ): """simple docstring""" def __init__( self: List[Any] , *__A: Union[str, Any] , **__A: Optional[Any] ) -> None: warnings.warn( '''The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use DPTImageProcessor instead.''' , __A , ) super().__init__(*__A , **__A )
75
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer import diffusers from diffusers import ( AutoencoderKL, EulerDiscreteScheduler, StableDiffusionLatentUpscalePipeline, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.schedulers import KarrasDiffusionSchedulers from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() def UpperCamelCase ( _lowerCAmelCase : Any ) -> List[Any]: _UpperCAmelCase : int = [tensor.shape for tensor in tensor_list] return all(shape == shapes[0] for shape in shapes[1:] ) class _UpperCAmelCase ( __a , __a , __a , unittest.TestCase): __a : List[str] = StableDiffusionLatentUpscalePipeline __a : str = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { """height""", """width""", """cross_attention_kwargs""", """negative_prompt_embeds""", """prompt_embeds""", } __a : Optional[int] = PipelineTesterMixin.required_optional_params - {"""num_images_per_prompt"""} __a : List[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS __a : str = frozenset( []) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __a : Optional[Any] = frozenset([]) __a : Union[str, Any] = True @property def __snake_case ( self ) -> Optional[int]: '''simple docstring''' _UpperCAmelCase : Optional[Any] = 1 _UpperCAmelCase : Optional[Any] = 4 _UpperCAmelCase : int = (16, 16) _UpperCAmelCase : List[str] = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(_A ) return image def __snake_case ( self ) -> List[Any]: '''simple docstring''' torch.manual_seed(0 ) _UpperCAmelCase : str = UNetaDConditionModel( act_fn="""gelu""" , attention_head_dim=8 , norm_num_groups=_A , block_out_channels=[32, 32, 64, 64] , time_cond_proj_dim=1_60 , conv_in_kernel=1 , conv_out_kernel=1 , cross_attention_dim=32 , down_block_types=( """KDownBlock2D""", """KCrossAttnDownBlock2D""", """KCrossAttnDownBlock2D""", """KCrossAttnDownBlock2D""", ) , in_channels=8 , mid_block_type=_A , only_cross_attention=_A , out_channels=5 , resnet_time_scale_shift="""scale_shift""" , time_embedding_type="""fourier""" , timestep_post_act="""gelu""" , up_block_types=("""KCrossAttnUpBlock2D""", """KCrossAttnUpBlock2D""", """KCrossAttnUpBlock2D""", """KUpBlock2D""") , ) _UpperCAmelCase : Optional[int] = AutoencoderKL( block_out_channels=[32, 32, 64, 64] , in_channels=3 , out_channels=3 , down_block_types=[ """DownEncoderBlock2D""", """DownEncoderBlock2D""", """DownEncoderBlock2D""", """DownEncoderBlock2D""", ] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D""", """UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) _UpperCAmelCase : Any = EulerDiscreteScheduler(prediction_type="""sample""" ) _UpperCAmelCase : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="""quick_gelu""" , projection_dim=5_12 , ) _UpperCAmelCase : Any = CLIPTextModel(_A ) _UpperCAmelCase : str = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _UpperCAmelCase : Optional[int] = { """unet""": model.eval(), """vae""": vae.eval(), """scheduler""": scheduler, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def __snake_case ( self , _A , _A=0 ) -> Dict: '''simple docstring''' if str(_A ).startswith("""mps""" ): _UpperCAmelCase : Tuple = torch.manual_seed(_A ) else: _UpperCAmelCase : Dict = torch.Generator(device=_A ).manual_seed(_A ) _UpperCAmelCase : str = { """prompt""": """A painting of a squirrel eating a burger""", """image""": self.dummy_image.cpu(), """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs def __snake_case ( self ) -> str: '''simple docstring''' _UpperCAmelCase : int = """cpu""" _UpperCAmelCase : str = self.get_dummy_components() _UpperCAmelCase : Optional[Any] = self.pipeline_class(**_A ) pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) _UpperCAmelCase : Optional[int] = self.get_dummy_inputs(_A ) _UpperCAmelCase : Optional[int] = pipe(**_A ).images _UpperCAmelCase : int = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 2_56, 2_56, 3) ) _UpperCAmelCase : List[str] = np.array( [0.47222412, 0.41921633, 0.44717434, 0.46874192, 0.42588258, 0.46150726, 0.4677534, 0.45583832, 0.48579055] ) _UpperCAmelCase : Optional[Any] = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_A , 1e-3 ) def __snake_case ( self ) -> Tuple: '''simple docstring''' super().test_attention_slicing_forward_pass(expected_max_diff=7e-3 ) def __snake_case ( self ) -> Tuple: '''simple docstring''' super().test_cpu_offload_forward_pass(expected_max_diff=3e-3 ) def __snake_case ( self ) -> Any: '''simple docstring''' super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def __snake_case ( self ) -> Any: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=7e-3 ) def __snake_case ( self ) -> Optional[Any]: '''simple docstring''' super().test_pt_np_pil_outputs_equivalent(expected_max_diff=3e-3 ) def __snake_case ( self ) -> Dict: '''simple docstring''' super().test_save_load_local(expected_max_difference=3e-3 ) def __snake_case ( self ) -> List[str]: '''simple docstring''' super().test_save_load_optional_components(expected_max_difference=3e-3 ) def __snake_case ( self ) -> str: '''simple docstring''' _UpperCAmelCase : List[str] = [ """DDIMScheduler""", """DDPMScheduler""", """PNDMScheduler""", """HeunDiscreteScheduler""", """EulerAncestralDiscreteScheduler""", """KDPM2DiscreteScheduler""", """KDPM2AncestralDiscreteScheduler""", """DPMSolverSDEScheduler""", ] _UpperCAmelCase : str = self.get_dummy_components() _UpperCAmelCase : Optional[Any] = self.pipeline_class(**_A ) # make sure that PNDM does not need warm-up pipe.scheduler.register_to_config(skip_prk_steps=_A ) pipe.to(_A ) pipe.set_progress_bar_config(disable=_A ) _UpperCAmelCase : Optional[Any] = self.get_dummy_inputs(_A ) _UpperCAmelCase : List[Any] = 2 _UpperCAmelCase : Optional[int] = [] for scheduler_enum in KarrasDiffusionSchedulers: if scheduler_enum.name in skip_schedulers: # no sigma schedulers are not supported # no schedulers continue _UpperCAmelCase : Union[str, Any] = getattr(_A , scheduler_enum.name ) _UpperCAmelCase : Union[str, Any] = scheduler_cls.from_config(pipe.scheduler.config ) _UpperCAmelCase : Any = pipe(**_A )[0] outputs.append(_A ) assert check_same_shape(_A ) @require_torch_gpu @slow class _UpperCAmelCase ( unittest.TestCase): def __snake_case ( self ) -> Optional[int]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self ) -> Optional[int]: '''simple docstring''' _UpperCAmelCase : Optional[int] = torch.manual_seed(33 ) _UpperCAmelCase : Any = StableDiffusionPipeline.from_pretrained("""CompVis/stable-diffusion-v1-4""" , torch_dtype=torch.floataa ) pipe.to("""cuda""" ) _UpperCAmelCase : Dict = StableDiffusionLatentUpscalePipeline.from_pretrained( """stabilityai/sd-x2-latent-upscaler""" , torch_dtype=torch.floataa ) upscaler.to("""cuda""" ) _UpperCAmelCase : Any = """a photo of an astronaut high resolution, unreal engine, ultra realistic""" _UpperCAmelCase : Optional[int] = pipe(_A , generator=_A , output_type="""latent""" ).images _UpperCAmelCase : Union[str, Any] = upscaler( prompt=_A , image=_A , num_inference_steps=20 , guidance_scale=0 , generator=_A , output_type="""np""" , ).images[0] _UpperCAmelCase : List[str] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/astronaut_1024.npy""" ) assert np.abs((expected_image - image).mean() ) < 5e-2 def __snake_case ( self ) -> Dict: '''simple docstring''' _UpperCAmelCase : Tuple = torch.manual_seed(33 ) _UpperCAmelCase : Dict = StableDiffusionLatentUpscalePipeline.from_pretrained( """stabilityai/sd-x2-latent-upscaler""" , torch_dtype=torch.floataa ) upscaler.to("""cuda""" ) _UpperCAmelCase : Optional[int] = """the temple of fire by Ross Tran and Gerardo Dottori, oil on canvas""" _UpperCAmelCase : Tuple = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_512.png""" ) _UpperCAmelCase : str = upscaler( prompt=_A , image=_A , num_inference_steps=20 , guidance_scale=0 , generator=_A , output_type="""np""" , ).images[0] _UpperCAmelCase : str = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_1024.npy""" ) assert np.abs((expected_image - image).max() ) < 5e-2
246
"""simple docstring""" from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) lowerCamelCase__ : Any = logging.get_logger(__name__) # pylint: disable=invalid-name lowerCamelCase__ : Union[str, Any] = ''' Examples: ```py >>> from diffusers import KandinskyV22Img2ImgPipeline, KandinskyV22PriorPipeline >>> from diffusers.utils import load_image >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16 ... ) >>> pipe_prior.to("cuda") >>> prompt = "A red cartoon frog, 4k" >>> image_emb, zero_image_emb = pipe_prior(prompt, return_dict=False) >>> pipe = KandinskyV22Img2ImgPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-decoder", torch_dtype=torch.float16 ... ) >>> pipe.to("cuda") >>> init_image = load_image( ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" ... "/kandinsky/frog.png" ... ) >>> image = pipe( ... image=init_image, ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=100, ... strength=0.2, ... ).images >>> image[0].save("red_frog.png") ``` ''' def UpperCamelCase ( _lowerCAmelCase : int, _lowerCAmelCase : List[Any], _lowerCAmelCase : Optional[Any]=8 ) -> Union[str, Any]: _UpperCAmelCase : Dict = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 _UpperCAmelCase : Tuple = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor def UpperCamelCase ( _lowerCAmelCase : Union[str, Any], _lowerCAmelCase : Any=512, _lowerCAmelCase : Optional[Any]=512 ) -> Optional[Any]: _UpperCAmelCase : Optional[int] = pil_image.resize((w, h), resample=Image.BICUBIC, reducing_gap=1 ) _UpperCAmelCase : List[Any] = np.array(pil_image.convert("""RGB""" ) ) _UpperCAmelCase : str = arr.astype(np.floataa ) / 127.5 - 1 _UpperCAmelCase : Dict = np.transpose(_lowerCAmelCase, [2, 0, 1] ) _UpperCAmelCase : Union[str, Any] = torch.from_numpy(_lowerCAmelCase ).unsqueeze(0 ) return image class _UpperCAmelCase ( __a): def __init__( self , _A , _A , _A , ) -> int: '''simple docstring''' super().__init__() self.register_modules( unet=_A , scheduler=_A , movq=_A , ) _UpperCAmelCase : Optional[Any] = 2 ** (len(self.movq.config.block_out_channels ) - 1) def __snake_case ( self , _A , _A , _A ) -> List[Any]: '''simple docstring''' _UpperCAmelCase : int = min(int(num_inference_steps * strength ) , _A ) _UpperCAmelCase : Dict = max(num_inference_steps - init_timestep , 0 ) _UpperCAmelCase : Tuple = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def __snake_case ( self , _A , _A , _A , _A , _A , _A , _A=None ) -> List[Any]: '''simple docstring''' if not isinstance(_A , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( f'''`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(_A )}''' ) _UpperCAmelCase : Any = image.to(device=_A , dtype=_A ) _UpperCAmelCase : Optional[int] = batch_size * num_images_per_prompt if image.shape[1] == 4: _UpperCAmelCase : Dict = image else: if isinstance(_A , _A ) and len(_A ) != batch_size: raise ValueError( f'''You have passed a list of generators of length {len(_A )}, but requested an effective batch''' f''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) elif isinstance(_A , _A ): _UpperCAmelCase : Any = [ self.movq.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(_A ) ] _UpperCAmelCase : List[Any] = torch.cat(_A , dim=0 ) else: _UpperCAmelCase : str = self.movq.encode(_A ).latent_dist.sample(_A ) _UpperCAmelCase : Any = self.movq.config.scaling_factor * init_latents _UpperCAmelCase : List[Any] = torch.cat([init_latents] , dim=0 ) _UpperCAmelCase : Union[str, Any] = init_latents.shape _UpperCAmelCase : List[Any] = randn_tensor(_A , generator=_A , device=_A , dtype=_A ) # get latents _UpperCAmelCase : Optional[int] = self.scheduler.add_noise(_A , _A , _A ) _UpperCAmelCase : Optional[int] = init_latents return latents def __snake_case ( self , _A=0 ) -> Optional[Any]: '''simple docstring''' if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("""Please install accelerate via `pip install accelerate`""" ) _UpperCAmelCase : List[Any] = torch.device(f'''cuda:{gpu_id}''' ) _UpperCAmelCase : Tuple = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_A , _A ) def __snake_case ( self , _A=0 ) -> int: '''simple docstring''' if is_accelerate_available() and is_accelerate_version(""">=""" , """0.17.0.dev0""" ): from accelerate import cpu_offload_with_hook else: raise ImportError("""`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.""" ) _UpperCAmelCase : int = torch.device(f'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to("""cpu""" , silence_dtype_warnings=_A ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) _UpperCAmelCase : Any = None for cpu_offloaded_model in [self.unet, self.movq]: _UpperCAmelCase , _UpperCAmelCase : Tuple = cpu_offload_with_hook(_A , _A , prev_module_hook=_A ) # We'll offload the last model manually. _UpperCAmelCase : Optional[int] = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self ) -> List[str]: '''simple docstring''' if not hasattr(self.unet , """_hf_hook""" ): return self.device for module in self.unet.modules(): if ( hasattr(_A , """_hf_hook""" ) and hasattr(module._hf_hook , """execution_device""" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_A ) def __call__( self , _A , _A , _A , _A = 5_12 , _A = 5_12 , _A = 1_00 , _A = 4.0 , _A = 0.3 , _A = 1 , _A = None , _A = "pil" , _A = True , ) -> Any: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = self._execution_device _UpperCAmelCase : List[str] = guidance_scale > 1.0 if isinstance(_A , _A ): _UpperCAmelCase : Dict = torch.cat(_A , dim=0 ) _UpperCAmelCase : Any = image_embeds.shape[0] if isinstance(_A , _A ): _UpperCAmelCase : Any = torch.cat(_A , dim=0 ) if do_classifier_free_guidance: _UpperCAmelCase : str = image_embeds.repeat_interleave(_A , dim=0 ) _UpperCAmelCase : Optional[int] = negative_image_embeds.repeat_interleave(_A , dim=0 ) _UpperCAmelCase : str = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_A ) if not isinstance(_A , _A ): _UpperCAmelCase : str = [image] if not all(isinstance(_A , (PIL.Image.Image, torch.Tensor) ) for i in image ): raise ValueError( f'''Input is in incorrect format: {[type(_A ) for i in image]}. Currently, we only support PIL image and pytorch tensor''' ) _UpperCAmelCase : Union[str, Any] = torch.cat([prepare_image(_A , _A , _A ) for i in image] , dim=0 ) _UpperCAmelCase : List[Any] = image.to(dtype=image_embeds.dtype , device=_A ) _UpperCAmelCase : int = self.movq.encode(_A )["""latents"""] _UpperCAmelCase : Dict = latents.repeat_interleave(_A , dim=0 ) self.scheduler.set_timesteps(_A , device=_A ) _UpperCAmelCase , _UpperCAmelCase : Any = self.get_timesteps(_A , _A , _A ) _UpperCAmelCase : Dict = timesteps[:1].repeat(batch_size * num_images_per_prompt ) _UpperCAmelCase , _UpperCAmelCase : str = downscale_height_and_width(_A , _A , self.movq_scale_factor ) _UpperCAmelCase : List[Any] = self.prepare_latents( _A , _A , _A , _A , image_embeds.dtype , _A , _A ) for i, t in enumerate(self.progress_bar(_A ) ): # expand the latents if we are doing classifier free guidance _UpperCAmelCase : Union[str, Any] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents _UpperCAmelCase : Union[str, Any] = {"""image_embeds""": image_embeds} _UpperCAmelCase : str = self.unet( sample=_A , timestep=_A , encoder_hidden_states=_A , added_cond_kwargs=_A , return_dict=_A , )[0] if do_classifier_free_guidance: _UpperCAmelCase , _UpperCAmelCase : Any = noise_pred.split(latents.shape[1] , dim=1 ) _UpperCAmelCase , _UpperCAmelCase : Any = noise_pred.chunk(2 ) _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = variance_pred.chunk(2 ) _UpperCAmelCase : List[str] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) _UpperCAmelCase : Optional[int] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , """variance_type""" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): _UpperCAmelCase , _UpperCAmelCase : Dict = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 _UpperCAmelCase : List[Any] = self.scheduler.step( _A , _A , _A , generator=_A , )[0] # post-processing _UpperCAmelCase : Optional[int] = self.movq.decode(_A , force_not_quantize=_A )["""sample"""] if output_type not in ["pt", "np", "pil"]: raise ValueError(f'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: _UpperCAmelCase : Any = image * 0.5 + 0.5 _UpperCAmelCase : Dict = image.clamp(0 , 1 ) _UpperCAmelCase : str = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": _UpperCAmelCase : List[str] = self.numpy_to_pil(_A ) if not return_dict: return (image,) return ImagePipelineOutput(images=_A )
246
1
'''simple docstring''' from typing import Optional, Tuple, Union import torch from einops import rearrange, reduce from diffusers import DDIMScheduler, DDPMScheduler, DiffusionPipeline, ImagePipelineOutput, UNetaDConditionModel from diffusers.schedulers.scheduling_ddim import DDIMSchedulerOutput from diffusers.schedulers.scheduling_ddpm import DDPMSchedulerOutput UpperCamelCase_ : Optional[Any] = 8 def __a ( _UpperCamelCase: Optional[int] , _UpperCamelCase: List[str]=BITS ) -> List[Any]: """simple docstring""" _snake_case = x.device _snake_case = (x * 255).int().clamp(0 , 255 ) _snake_case = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_UpperCamelCase ) _snake_case = rearrange(_UpperCamelCase , "d -> d 1 1" ) _snake_case = rearrange(_UpperCamelCase , "b c h w -> b c 1 h w" ) _snake_case = ((x & mask) != 0).float() _snake_case = rearrange(_UpperCamelCase , "b c d h w -> b (c d) h w" ) _snake_case = bits * 2 - 1 return bits def __a ( _UpperCamelCase: Optional[int] , _UpperCamelCase: Dict=BITS ) -> Optional[Any]: """simple docstring""" _snake_case = x.device _snake_case = (x > 0).int() _snake_case = 2 ** torch.arange(bits - 1 , -1 , -1 , device=_UpperCamelCase , dtype=torch.intaa ) _snake_case = rearrange(_UpperCamelCase , "d -> d 1 1" ) _snake_case = rearrange(_UpperCamelCase , "b (c d) h w -> b c d h w" , d=8 ) _snake_case = reduce(x * mask , "b c d h w -> b c h w" , "sum" ) return (dec / 255).clamp(0.0 , 1.0 ) def __a ( self: Any , _UpperCamelCase: torch.FloatTensor , _UpperCamelCase: int , _UpperCamelCase: torch.FloatTensor , _UpperCamelCase: float = 0.0 , _UpperCamelCase: bool = True , _UpperCamelCase: Any=None , _UpperCamelCase: bool = True , ) -> Union[DDIMSchedulerOutput, Tuple]: """simple docstring""" if self.num_inference_steps is None: raise ValueError( "Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler" ) # See formulas (12) and (16) of DDIM paper https://arxiv.org/pdf/2010.02502.pdf # Ideally, read DDIM paper in-detail understanding # Notation (<variable name> -> <name in paper> # - pred_noise_t -> e_theta(x_t, t) # - pred_original_sample -> f_theta(x_t, t) or x_0 # - std_dev_t -> sigma_t # - eta -> η # - pred_sample_direction -> "direction pointing to x_t" # - pred_prev_sample -> "x_t-1" # 1. get previous step value (=t-1) _snake_case = timestep - self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas _snake_case = self.alphas_cumprod[timestep] _snake_case = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.final_alpha_cumprod _snake_case = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 # 4. Clip "predicted x_0" _snake_case = self.bit_scale if self.config.clip_sample: _snake_case = torch.clamp(_UpperCamelCase , -scale , _UpperCamelCase ) # 5. compute variance: "sigma_t(η)" -> see formula (16) # σ_t = sqrt((1 − α_t−1)/(1 − α_t)) * sqrt(1 − α_t/α_t−1) _snake_case = self._get_variance(_UpperCamelCase , _UpperCamelCase ) _snake_case = eta * variance ** 0.5 if use_clipped_model_output: # the model_output is always re-derived from the clipped x_0 in Glide _snake_case = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 # 6. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = (1 - alpha_prod_t_prev - std_dev_t**2) ** 0.5 * model_output # 7. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _snake_case = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if eta > 0: # randn_like does not support generator https://github.com/pytorch/pytorch/issues/27072 _snake_case = model_output.device if torch.is_tensor(_UpperCamelCase ) else "cpu" _snake_case = torch.randn(model_output.shape , dtype=model_output.dtype , generator=_UpperCamelCase ).to(_UpperCamelCase ) _snake_case = self._get_variance(_UpperCamelCase , _UpperCamelCase ) ** 0.5 * eta * noise _snake_case = prev_sample + variance if not return_dict: return (prev_sample,) return DDIMSchedulerOutput(prev_sample=_UpperCamelCase , pred_original_sample=_UpperCamelCase ) def __a ( self: Optional[Any] , _UpperCamelCase: torch.FloatTensor , _UpperCamelCase: int , _UpperCamelCase: torch.FloatTensor , _UpperCamelCase: Dict="epsilon" , _UpperCamelCase: Dict=None , _UpperCamelCase: bool = True , ) -> Union[DDPMSchedulerOutput, Tuple]: """simple docstring""" _snake_case = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type in ["learned", "learned_range"]: _snake_case , _snake_case = torch.split(_UpperCamelCase , sample.shape[1] , dim=1 ) else: _snake_case = None # 1. compute alphas, betas _snake_case = self.alphas_cumprod[t] _snake_case = self.alphas_cumprod[t - 1] if t > 0 else self.one _snake_case = 1 - alpha_prod_t _snake_case = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if prediction_type == "epsilon": _snake_case = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif prediction_type == "sample": _snake_case = model_output else: raise ValueError(F"""Unsupported prediction_type {prediction_type}.""" ) # 3. Clip "predicted x_0" _snake_case = self.bit_scale if self.config.clip_sample: _snake_case = torch.clamp(_UpperCamelCase , -scale , _UpperCamelCase ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _snake_case = (alpha_prod_t_prev ** 0.5 * self.betas[t]) / beta_prod_t _snake_case = self.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _snake_case = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise _snake_case = 0 if t > 0: _snake_case = torch.randn( model_output.size() , dtype=model_output.dtype , layout=model_output.layout , generator=_UpperCamelCase ).to(model_output.device ) _snake_case = (self._get_variance(_UpperCamelCase , predicted_variance=_UpperCamelCase ) ** 0.5) * noise _snake_case = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return DDPMSchedulerOutput(prev_sample=_UpperCamelCase , pred_original_sample=_UpperCamelCase ) class _a ( __lowerCAmelCase ): def __init__( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = 1.0 ,) -> Tuple: super().__init__() _snake_case = bit_scale _snake_case = ( ddim_bit_scheduler_step if isinstance(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) else ddpm_bit_scheduler_step ) self.register_modules(unet=_SCREAMING_SNAKE_CASE ,scheduler=_SCREAMING_SNAKE_CASE ) @torch.no_grad() def __call__( self ,_SCREAMING_SNAKE_CASE = 256 ,_SCREAMING_SNAKE_CASE = 256 ,_SCREAMING_SNAKE_CASE = 50 ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = 1 ,_SCREAMING_SNAKE_CASE = "pil" ,_SCREAMING_SNAKE_CASE = True ,**_SCREAMING_SNAKE_CASE ,) -> Union[Tuple, ImagePipelineOutput]: _snake_case = torch.randn( (batch_size, self.unet.config.in_channels, height, width) ,generator=_SCREAMING_SNAKE_CASE ,) _snake_case = decimal_to_bits(_SCREAMING_SNAKE_CASE ) * self.bit_scale _snake_case = latents.to(self.device ) self.scheduler.set_timesteps(_SCREAMING_SNAKE_CASE ) for t in self.progress_bar(self.scheduler.timesteps ): # predict the noise residual _snake_case = self.unet(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ).sample # compute the previous noisy sample x_t -> x_t-1 _snake_case = self.scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ).prev_sample _snake_case = bits_to_decimal(_SCREAMING_SNAKE_CASE ) if output_type == "pil": _snake_case = self.numpy_to_pil(_SCREAMING_SNAKE_CASE ) if not return_dict: return (image,) return ImagePipelineOutput(images=_SCREAMING_SNAKE_CASE )
142
'''simple docstring''' from manim import * class _a ( __lowerCAmelCase ): def _lowercase ( self ) -> Optional[int]: _snake_case = Rectangle(height=0.5 ,width=0.5 ) _snake_case = Rectangle(height=0.4_6 ,width=0.4_6 ).set_stroke(width=0 ) _snake_case = [mem.copy() for i in range(6 )] _snake_case = [mem.copy() for i in range(6 )] _snake_case = VGroup(*_SCREAMING_SNAKE_CASE ).arrange(_SCREAMING_SNAKE_CASE ,buff=0 ) _snake_case = VGroup(*_SCREAMING_SNAKE_CASE ).arrange(_SCREAMING_SNAKE_CASE ,buff=0 ) _snake_case = VGroup(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ).arrange(_SCREAMING_SNAKE_CASE ,buff=0 ) _snake_case = Text("CPU" ,font_size=24 ) _snake_case = Group(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ).arrange(_SCREAMING_SNAKE_CASE ,buff=0.5 ,aligned_edge=_SCREAMING_SNAKE_CASE ) cpu.move_to([-2.5, -0.5, 0] ) self.add(_SCREAMING_SNAKE_CASE ) _snake_case = [mem.copy() for i in range(4 )] _snake_case = VGroup(*_SCREAMING_SNAKE_CASE ).arrange(_SCREAMING_SNAKE_CASE ,buff=0 ) _snake_case = Text("GPU" ,font_size=24 ) _snake_case = Group(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ).arrange(_SCREAMING_SNAKE_CASE ,buff=0.5 ,aligned_edge=_SCREAMING_SNAKE_CASE ) gpu.move_to([-1, -1, 0] ) self.add(_SCREAMING_SNAKE_CASE ) _snake_case = [mem.copy() for i in range(6 )] _snake_case = VGroup(*_SCREAMING_SNAKE_CASE ).arrange(_SCREAMING_SNAKE_CASE ,buff=0 ) _snake_case = Text("Model" ,font_size=24 ) _snake_case = Group(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ).arrange(_SCREAMING_SNAKE_CASE ,buff=0.5 ,aligned_edge=_SCREAMING_SNAKE_CASE ) model.move_to([3, -1.0, 0] ) self.add(_SCREAMING_SNAKE_CASE ) _snake_case = [] for i, rect in enumerate(_SCREAMING_SNAKE_CASE ): rect.set_stroke(_SCREAMING_SNAKE_CASE ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) _snake_case = Rectangle(height=0.4_6 / 4 ,width=0.4_6 / 3 ).set_stroke(width=0.0 ).set_fill(_SCREAMING_SNAKE_CASE ,opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) ,buff=0.0_2 ,direction=_SCREAMING_SNAKE_CASE ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] ,direction=_SCREAMING_SNAKE_CASE ,buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] ,direction=_SCREAMING_SNAKE_CASE ,buff=0.0 ) self.add(_SCREAMING_SNAKE_CASE ) cpu_targs.append(_SCREAMING_SNAKE_CASE ) _snake_case = [mem.copy() for i in range(6 )] _snake_case = VGroup(*_SCREAMING_SNAKE_CASE ).arrange(_SCREAMING_SNAKE_CASE ,buff=0 ) _snake_case = Text("Loaded Checkpoint" ,font_size=24 ) _snake_case = Group(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ).arrange(_SCREAMING_SNAKE_CASE ,aligned_edge=_SCREAMING_SNAKE_CASE ,buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) _snake_case = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _snake_case = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" ,font_size=18 ,) key_text.move_to([-5, 2.4, 0] ) self.add(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) _snake_case = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" ,font_size=18 ,) blue_text.next_to(_SCREAMING_SNAKE_CASE ,DOWN * 2.4 ,aligned_edge=key_text.get_left() ) _snake_case = MarkupText( f"""Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>.""" ,font_size=24 ,) step_a.move_to([2, 2, 0] ) self.play(Write(_SCREAMING_SNAKE_CASE ) ,Write(_SCREAMING_SNAKE_CASE ) ) self.play(Write(_SCREAMING_SNAKE_CASE ,run_time=1 ) ,Create(_SCREAMING_SNAKE_CASE ,run_time=1 ) ) _snake_case = [] _snake_case = [] for i, rect in enumerate(_SCREAMING_SNAKE_CASE ): _snake_case = fill.copy().set_fill(_SCREAMING_SNAKE_CASE ,opacity=0.7 ) target.move_to(_SCREAMING_SNAKE_CASE ) first_animations.append(GrowFromCenter(_SCREAMING_SNAKE_CASE ,run_time=1 ) ) _snake_case = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(_SCREAMING_SNAKE_CASE ,run_time=1.5 ) ) self.play(*_SCREAMING_SNAKE_CASE ) self.play(*_SCREAMING_SNAKE_CASE ) self.wait()
142
1
from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case : Union[str, Any] = logging.get_logger(__name__) snake_case : Union[str, Any] = { '''huggingface/autoformer-tourism-monthly''': '''https://huggingface.co/huggingface/autoformer-tourism-monthly/resolve/main/config.json''', } class snake_case_ (lowerCamelCase_ ): UpperCAmelCase__ : List[str] = '''autoformer''' UpperCAmelCase__ : Union[str, Any] = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', '''num_hidden_layers''': '''encoder_layers''', } def __init__( self :Tuple ,__snake_case :Optional[int] = None ,__snake_case :Optional[int] = None ,__snake_case :str = "student_t" ,__snake_case :str = "nll" ,__snake_case :int = 1 ,__snake_case :List[int] = [1, 2, 3, 4, 5, 6, 7] ,__snake_case :bool = True ,__snake_case :int = 0 ,__snake_case :int = 0 ,__snake_case :int = 0 ,__snake_case :int = 0 ,__snake_case :Optional[List[int]] = None ,__snake_case :Optional[List[int]] = None ,__snake_case :int = 64 ,__snake_case :int = 2 ,__snake_case :int = 2 ,__snake_case :int = 2 ,__snake_case :int = 2 ,__snake_case :int = 32 ,__snake_case :int = 32 ,__snake_case :str = "gelu" ,__snake_case :float = 0.1 ,__snake_case :float = 0.1 ,__snake_case :float = 0.1 ,__snake_case :float = 0.1 ,__snake_case :float = 0.1 ,__snake_case :int = 1_00 ,__snake_case :float = 0.02 ,__snake_case :bool = True ,__snake_case :Dict=True ,__snake_case :int = 10 ,__snake_case :int = 25 ,__snake_case :int = 3 ,**__snake_case :Optional[int] ,) -> Any: # time series specific configuration a__ = prediction_length a__ = context_length if context_length is not None else prediction_length a__ = distribution_output a__ = loss a__ = input_size a__ = num_time_features a__ = lags_sequence a__ = scaling a__ = num_dynamic_real_features a__ = num_static_real_features a__ = num_static_categorical_features if cardinality is not None and num_static_categorical_features > 0: if len(__snake_case ) != num_static_categorical_features: raise ValueError( 'The cardinality should be a list of the same length as `num_static_categorical_features`' ) a__ = cardinality else: a__ = [0] if embedding_dimension is not None and num_static_categorical_features > 0: if len(__snake_case ) != num_static_categorical_features: raise ValueError( 'The embedding dimension should be a list of the same length as `num_static_categorical_features`' ) a__ = embedding_dimension else: a__ = [min(50 ,(cat + 1) // 2 ) for cat in self.cardinality] a__ = num_parallel_samples # Transformer architecture configuration a__ = input_size * len(self.lags_sequence ) + self._number_of_features a__ = d_model a__ = encoder_attention_heads a__ = decoder_attention_heads a__ = encoder_ffn_dim a__ = decoder_ffn_dim a__ = encoder_layers a__ = decoder_layers a__ = dropout a__ = attention_dropout a__ = activation_dropout a__ = encoder_layerdrop a__ = decoder_layerdrop a__ = activation_function a__ = init_std a__ = use_cache # Autoformer a__ = label_length a__ = moving_average a__ = autocorrelation_factor super().__init__(is_encoder_decoder=__snake_case ,**__snake_case ) @property def lowerCamelCase__( self :int ) -> int: return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
240
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets snake_case : List[Any] = datasets.logging.get_logger(__name__) snake_case : List[str] = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' snake_case : Tuple = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' snake_case : str = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def __lowercase ( __lowerCAmelCase : List[Any] , __lowerCAmelCase : int , __lowerCAmelCase : Dict=False , __lowerCAmelCase : List[Any]=False , __lowerCAmelCase : Optional[Any]=True , __lowerCAmelCase : Dict=False , __lowerCAmelCase : Any="dummy_doc" ): a__ = {doc: key_lines} a__ = {doc: sys_lines} a__ = {} a__ = 0 a__ = 0 a__ = 0 a__ = 0 a__ = 0 a__ = 0 a__ , a__ = reader.get_doc_mentions(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase ) key_singletons_num += singletons_num if NP_only or min_span: a__ = reader.set_annotated_parse_trees(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase , __lowerCAmelCase ) a__ , a__ = reader.get_doc_mentions(__lowerCAmelCase , sys_doc_lines[doc] , __lowerCAmelCase ) sys_singletons_num += singletons_num if NP_only or min_span: a__ = reader.set_annotated_parse_trees(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase , __lowerCAmelCase ) if remove_nested: a__ , a__ = reader.remove_nested_coref_mentions(__lowerCAmelCase , __lowerCAmelCase ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters a__ , a__ = reader.remove_nested_coref_mentions(__lowerCAmelCase , __lowerCAmelCase ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters a__ = reader.get_mention_assignments(__lowerCAmelCase , __lowerCAmelCase ) a__ = reader.get_mention_assignments(__lowerCAmelCase , __lowerCAmelCase ) a__ = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( 'Number of removed nested coreferring mentions in the key ' F'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( 'Number of resulting singleton clusters in the key ' F'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( F'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' 'files, respectively' ) return doc_coref_infos def __lowercase ( __lowerCAmelCase : Any , __lowerCAmelCase : Tuple , __lowerCAmelCase : str , __lowerCAmelCase : Dict , __lowerCAmelCase : List[Any] , __lowerCAmelCase : List[str] , __lowerCAmelCase : Any ): a__ = get_coref_infos(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) a__ = {} a__ = 0 a__ = 0 for name, metric in metrics: a__ , a__ , a__ = evaluator.evaluate_documents(__lowerCAmelCase , __lowerCAmelCase , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F'{name}/recall': recall, F'{name}/precision': precision, F'{name}/f1': fa} ) logger.info( name.ljust(1_0 ) , F'Recall: {recall * 1_0_0:.2f}' , F' Precision: {precision * 1_0_0:.2f}' , F' F1: {fa * 1_0_0:.2f}' , ) if conll_subparts_num == 3: a__ = (conll / 3) * 1_0_0 logger.info(F'CoNLL score: {conll:.2f}' ) output_scores.update({'conll_score': conll} ) return output_scores def __lowercase ( __lowerCAmelCase : str ): a__ = False for line in key_lines: if not line.startswith('#' ): if len(line.split() ) > 6: a__ = line.split()[5] if not parse_col == "-": a__ = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class snake_case_ (datasets.Metric ): def lowerCamelCase__( self :Optional[int] ) -> List[str]: return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('string' ) ), 'references': datasets.Sequence(datasets.Value('string' ) ), } ) ,codebase_urls=['https://github.com/ns-moosavi/coval'] ,reference_urls=[ 'https://github.com/ns-moosavi/coval', 'https://www.aclweb.org/anthology/P16-1060', 'http://www.conll.cemantix.org/2012/data.html', ] ,) def lowerCamelCase__( self :List[str] ,__snake_case :Optional[int] ,__snake_case :int ,__snake_case :Dict=True ,__snake_case :Union[str, Any]=False ,__snake_case :Dict=False ,__snake_case :Union[str, Any]=False ) -> Tuple: a__ = [ ('mentions', evaluator.mentions), ('muc', evaluator.muc), ('bcub', evaluator.b_cubed), ('ceafe', evaluator.ceafe), ('lea', evaluator.lea), ] if min_span: a__ = util.check_gold_parse_annotation(__snake_case ) if not has_gold_parse: raise NotImplementedError('References should have gold parse annotation to use \'min_span\'.' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" a__ = evaluate( key_lines=__snake_case ,sys_lines=__snake_case ,metrics=__snake_case ,NP_only=__snake_case ,remove_nested=__snake_case ,keep_singletons=__snake_case ,min_span=__snake_case ,) return score
240
1
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class lowerCamelCase__ ( unittest.TestCase ): '''simple docstring''' def _lowerCAmelCase ( self ) -> List[Any]: _lowerCAmelCase =[ """safety_checker/pytorch_model.bin""", """safety_checker/model.safetensors""", """vae/diffusion_pytorch_model.bin""", """vae/diffusion_pytorch_model.safetensors""", """text_encoder/pytorch_model.bin""", """text_encoder/model.safetensors""", """unet/diffusion_pytorch_model.bin""", """unet/diffusion_pytorch_model.safetensors""", ] self.assertTrue(is_safetensors_compatible(__UpperCAmelCase ) ) def _lowerCAmelCase ( self ) -> Union[str, Any]: _lowerCAmelCase =[ """unet/diffusion_pytorch_model.bin""", """unet/diffusion_pytorch_model.safetensors""", ] self.assertTrue(is_safetensors_compatible(__UpperCAmelCase ) ) def _lowerCAmelCase ( self ) -> List[Any]: _lowerCAmelCase =[ """safety_checker/pytorch_model.bin""", """safety_checker/model.safetensors""", """vae/diffusion_pytorch_model.bin""", """vae/diffusion_pytorch_model.safetensors""", """text_encoder/pytorch_model.bin""", """text_encoder/model.safetensors""", """unet/diffusion_pytorch_model.bin""", # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(__UpperCAmelCase ) ) def _lowerCAmelCase ( self ) -> Optional[int]: _lowerCAmelCase =[ """text_encoder/pytorch_model.bin""", """text_encoder/model.safetensors""", ] self.assertTrue(is_safetensors_compatible(__UpperCAmelCase ) ) def _lowerCAmelCase ( self ) -> Tuple: _lowerCAmelCase =[ """safety_checker/pytorch_model.bin""", """safety_checker/model.safetensors""", """vae/diffusion_pytorch_model.bin""", """vae/diffusion_pytorch_model.safetensors""", """text_encoder/pytorch_model.bin""", # Removed: 'text_encoder/model.safetensors', """unet/diffusion_pytorch_model.bin""", """unet/diffusion_pytorch_model.safetensors""", ] self.assertFalse(is_safetensors_compatible(__UpperCAmelCase ) ) def _lowerCAmelCase ( self ) -> Optional[Any]: _lowerCAmelCase =[ """safety_checker/pytorch_model.fp16.bin""", """safety_checker/model.fp16.safetensors""", """vae/diffusion_pytorch_model.fp16.bin""", """vae/diffusion_pytorch_model.fp16.safetensors""", """text_encoder/pytorch_model.fp16.bin""", """text_encoder/model.fp16.safetensors""", """unet/diffusion_pytorch_model.fp16.bin""", """unet/diffusion_pytorch_model.fp16.safetensors""", ] _lowerCAmelCase ="""fp16""" self.assertTrue(is_safetensors_compatible(__UpperCAmelCase , variant=__UpperCAmelCase ) ) def _lowerCAmelCase ( self ) -> Tuple: _lowerCAmelCase =[ """unet/diffusion_pytorch_model.fp16.bin""", """unet/diffusion_pytorch_model.fp16.safetensors""", ] _lowerCAmelCase ="""fp16""" self.assertTrue(is_safetensors_compatible(__UpperCAmelCase , variant=__UpperCAmelCase ) ) def _lowerCAmelCase ( self ) -> List[str]: _lowerCAmelCase =[ """unet/diffusion_pytorch_model.bin""", """unet/diffusion_pytorch_model.safetensors""", ] _lowerCAmelCase ="""fp16""" self.assertTrue(is_safetensors_compatible(__UpperCAmelCase , variant=__UpperCAmelCase ) ) def _lowerCAmelCase ( self ) -> int: _lowerCAmelCase =[ """safety_checker/pytorch_model.fp16.bin""", """safety_checker/model.fp16.safetensors""", """vae/diffusion_pytorch_model.fp16.bin""", """vae/diffusion_pytorch_model.fp16.safetensors""", """text_encoder/pytorch_model.fp16.bin""", """text_encoder/model.fp16.safetensors""", """unet/diffusion_pytorch_model.fp16.bin""", # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] _lowerCAmelCase ="""fp16""" self.assertFalse(is_safetensors_compatible(__UpperCAmelCase , variant=__UpperCAmelCase ) ) def _lowerCAmelCase ( self ) -> int: _lowerCAmelCase =[ """text_encoder/pytorch_model.fp16.bin""", """text_encoder/model.fp16.safetensors""", ] _lowerCAmelCase ="""fp16""" self.assertTrue(is_safetensors_compatible(__UpperCAmelCase , variant=__UpperCAmelCase ) ) def _lowerCAmelCase ( self ) -> List[str]: _lowerCAmelCase =[ """text_encoder/pytorch_model.bin""", """text_encoder/model.safetensors""", ] _lowerCAmelCase ="""fp16""" self.assertTrue(is_safetensors_compatible(__UpperCAmelCase , variant=__UpperCAmelCase ) ) def _lowerCAmelCase ( self ) -> Tuple: _lowerCAmelCase =[ """safety_checker/pytorch_model.fp16.bin""", """safety_checker/model.fp16.safetensors""", """vae/diffusion_pytorch_model.fp16.bin""", """vae/diffusion_pytorch_model.fp16.safetensors""", """text_encoder/pytorch_model.fp16.bin""", # 'text_encoder/model.fp16.safetensors', """unet/diffusion_pytorch_model.fp16.bin""", """unet/diffusion_pytorch_model.fp16.safetensors""", ] _lowerCAmelCase ="""fp16""" self.assertFalse(is_safetensors_compatible(__UpperCAmelCase , variant=__UpperCAmelCase ) )
353
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) __A = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = ['PerceiverFeatureExtractor'] __A = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
import argparse import json import os import re import torch from transformers import BloomConfig, BloomModel from transformers.file_utils import CONFIG_NAME, WEIGHTS_NAME from transformers.utils import logging logging.set_verbosity_info() _A = [ "word_embeddings_layernorm.weight", "word_embeddings_layernorm.bias", "input_layernorm.weight", "input_layernorm.bias", "post_attention_layernorm.weight", "post_attention_layernorm.bias", "self_attention.dense.bias", "mlp.dense_4h_to_h.bias", "ln_f.weight", "ln_f.bias", ] _A = [ "mlp.dense_4h_to_h.weight", "self_attention.dense.weight", ] def lowerCamelCase__ ( __lowerCAmelCase : Dict , __lowerCAmelCase : Dict ): """simple docstring""" lowerCAmelCase_ = { "word_embeddings.weight": "word_embeddings.weight", "word_embeddings.norm.weight": "word_embeddings_layernorm.weight", "word_embeddings.norm.bias": "word_embeddings_layernorm.bias", "weight": "ln_f.weight", "bias": "ln_f.bias", } if key in layer_rename_map: return layer_rename_map[key] # Handle transformer blocks lowerCAmelCase_ = int(re.match(r".*layer_(\d*).*" , __lowerCAmelCase )[1] ) layer_number -= 3 return F"""h.{layer_number}.""" + key def lowerCamelCase__ ( __lowerCAmelCase : Dict ): """simple docstring""" if dtype == torch.bool: return 1 / 8 lowerCAmelCase_ = re.search(r"[^\d](\d+)$" , str(__lowerCAmelCase ) ) if bit_search is None: raise ValueError(F"""`dtype` is not a valid dtype: {dtype}.""" ) lowerCAmelCase_ = int(bit_search.groups()[0] ) return bit_size // 8 def lowerCamelCase__ ( __lowerCAmelCase : List[str] , __lowerCAmelCase : Tuple , __lowerCAmelCase : str , __lowerCAmelCase : List[str] , __lowerCAmelCase : Dict ): """simple docstring""" if bloom_config_file == "": lowerCAmelCase_ = BloomConfig() else: lowerCAmelCase_ = BloomConfig.from_json_file(__lowerCAmelCase ) if shard_model: lowerCAmelCase_ = os.listdir(__lowerCAmelCase ) lowerCAmelCase_ = sorted(filter(lambda __lowerCAmelCase : s.startswith("layer" ) and "model_00" in s , __lowerCAmelCase ) ) lowerCAmelCase_ = {"weight_map": {}, "metadata": {}} lowerCAmelCase_ = 0 lowerCAmelCase_ = None lowerCAmelCase_ = BloomConfig() for j, file in enumerate(__lowerCAmelCase ): print("Processing file: {}".format(__lowerCAmelCase ) ) lowerCAmelCase_ = None for i in range(__lowerCAmelCase ): # load all TP files lowerCAmelCase_ = file.replace("model_00" , F"""model_0{i}""" ) lowerCAmelCase_ = torch.load(os.path.join(__lowerCAmelCase , __lowerCAmelCase ) , map_location="cpu" ) # Rename keys in the transformers names lowerCAmelCase_ = list(temp.keys() ) for key in keys: lowerCAmelCase_ = temp.pop(__lowerCAmelCase ) if tensors is None: lowerCAmelCase_ = temp else: for key in tensors.keys(): if any(key.endswith(__lowerCAmelCase ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel lowerCAmelCase_ = 1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks lowerCAmelCase_ = torch.cat([tensors[key], temp[key]] , dim=__lowerCAmelCase ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(__lowerCAmelCase ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): lowerCAmelCase_ = tensors[key] / pretraining_tp torch.save( __lowerCAmelCase , os.path.join( __lowerCAmelCase , "pytorch_model_{}-of-{}.bin".format(str(j + 1 ).zfill(5 ) , str(len(__lowerCAmelCase ) ).zfill(5 ) ) , ) , ) for key in tensors.keys(): lowerCAmelCase_ = tensors[key] total_size += value.numel() * get_dtype_size(value.dtype ) if key not in index_dict["weight_map"]: lowerCAmelCase_ = "pytorch_model_{}-of-{}.bin".format( str(j + 1 ).zfill(5 ) , str(len(__lowerCAmelCase ) ).zfill(5 ) ) lowerCAmelCase_ = BloomConfig() lowerCAmelCase_ = pytorch_dump_folder_path + "/" + CONFIG_NAME lowerCAmelCase_ = total_size with open(__lowerCAmelCase , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) with open(os.path.join(__lowerCAmelCase , WEIGHTS_NAME + ".index.json" ) , "w" , encoding="utf-8" ) as f: lowerCAmelCase_ = json.dumps(__lowerCAmelCase , indent=2 , sort_keys=__lowerCAmelCase ) + "\n" f.write(__lowerCAmelCase ) else: lowerCAmelCase_ = BloomModel(__lowerCAmelCase ) lowerCAmelCase_ = os.listdir(__lowerCAmelCase ) lowerCAmelCase_ = sorted(filter(lambda __lowerCAmelCase : s.startswith("layer" ) and "model_00" in s , __lowerCAmelCase ) ) lowerCAmelCase_ = None for i, file in enumerate(__lowerCAmelCase ): lowerCAmelCase_ = None for i in range(__lowerCAmelCase ): # load all TP files lowerCAmelCase_ = file.replace("model_00" , F"""model_0{i}""" ) lowerCAmelCase_ = torch.load(os.path.join(__lowerCAmelCase , __lowerCAmelCase ) , map_location="cpu" ) # Rename keys in the transformers names lowerCAmelCase_ = list(temp.keys() ) for key in keys: lowerCAmelCase_ = temp.pop(__lowerCAmelCase ) if tensors is None: lowerCAmelCase_ = temp else: for key in tensors.keys(): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) if any(key.endswith(__lowerCAmelCase ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel lowerCAmelCase_ = 1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks lowerCAmelCase_ = torch.cat([tensors[key], temp[key]] , dim=__lowerCAmelCase ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(__lowerCAmelCase ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): lowerCAmelCase_ = tensors[key] / pretraining_tp lowerCAmelCase_ = model.load_state_dict(__lowerCAmelCase , strict=__lowerCAmelCase ) assert not other_keys.unexpected_keys, F"""The keys {other_keys.unexpected_keys} are unexpected""" if missing_keys is None: lowerCAmelCase_ = set(other_keys.missing_keys ) else: lowerCAmelCase_ = missing_keys.intersection(set(other_keys.missing_keys ) ) assert not missing_keys, F"""The keys {missing_keys} are missing""" # Save pytorch-model os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) lowerCAmelCase_ = pytorch_dump_folder_path + "/" + WEIGHTS_NAME lowerCAmelCase_ = pytorch_dump_folder_path + "/" + CONFIG_NAME print(F"""Save PyTorch model to {pytorch_weights_dump_path} with dtype {config.torch_dtype}""" ) if config.torch_dtype is not None: lowerCAmelCase_ = model.to(config.torch_dtype ) torch.save(model.state_dict() , __lowerCAmelCase ) print(F"""Save configuration file to {pytorch_config_dump_path}""" ) with open(__lowerCAmelCase , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _A = argparse.ArgumentParser() # Required parameters parser.add_argument( "--bloom_checkpoint_path", default=None, type=str, required=True, help="Path to the Megatron-LM checkpoint path.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--bloom_config_file", default="", type=str, help=( "An optional config json file corresponding to the pre-trained model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--shard_model", action="store_true", help="An optional setting to shard the output model \nThis enables sharding the converted checkpoint", ) parser.add_argument( "--pretraining_tp", default=4, type=int, help="Pretraining TP rank that has been used when training the model in Megatron-LM \n", ) _A = parser.parse_args() convert_bloom_checkpoint_to_pytorch( args.bloom_checkpoint_path, args.bloom_config_file, args.pytorch_dump_folder_path, args.shard_model, args.pretraining_tp, )
231
import unittest from datasets import load_dataset from transformers import BloomTokenizerFast from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowerCAmelCase ( __a , unittest.TestCase ): _lowercase =None _lowercase =BloomTokenizerFast _lowercase =BloomTokenizerFast _lowercase =True _lowercase =False _lowercase ='''tokenizer_file''' _lowercase ={'''bos_token''': '''<s>''', '''eos_token''': '''</s>''', '''unk_token''': '''<unk>''', '''pad_token''': '''<pad>'''} def __a ( self ) -> Dict: super().setUp() lowerCAmelCase_ = BloomTokenizerFast.from_pretrained("bigscience/tokenizer" ) tokenizer.save_pretrained(self.tmpdirname ) def __a ( self , **_UpperCamelCase ) -> Tuple: kwargs.update(self.special_tokens_map ) return BloomTokenizerFast.from_pretrained(self.tmpdirname , **_UpperCamelCase ) def __a ( self ) -> List[str]: lowerCAmelCase_ = self.get_rust_tokenizer() lowerCAmelCase_ = ["The quick brown fox</s>", "jumps over the lazy dog</s>"] lowerCAmelCase_ = [[2_175, 23_714, 73_173, 144_252, 2], [77, 132_619, 3_478, 368, 109_586, 35_433, 2]] lowerCAmelCase_ = tokenizer.batch_encode_plus(_UpperCamelCase )["input_ids"] self.assertListEqual(_UpperCamelCase , _UpperCamelCase ) lowerCAmelCase_ = tokenizer.batch_decode(_UpperCamelCase ) self.assertListEqual(_UpperCamelCase , _UpperCamelCase ) def __a ( self , _UpperCamelCase=6 ) -> Tuple: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase_ = self.rust_tokenizer_class.from_pretrained(_UpperCamelCase , **_UpperCamelCase ) # tokenizer_r.pad_token = None # Hotfixing padding = None # Simple input lowerCAmelCase_ = "This is a simple input" lowerCAmelCase_ = ["This is a simple input 1", "This is a simple input 2"] lowerCAmelCase_ = ("This is a simple input", "This is a pair") lowerCAmelCase_ = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests try: tokenizer_r.encode(_UpperCamelCase , max_length=_UpperCamelCase ) tokenizer_r.encode_plus(_UpperCamelCase , max_length=_UpperCamelCase ) tokenizer_r.batch_encode_plus(_UpperCamelCase , max_length=_UpperCamelCase ) tokenizer_r.encode(_UpperCamelCase , max_length=_UpperCamelCase ) tokenizer_r.batch_encode_plus(_UpperCamelCase , max_length=_UpperCamelCase ) except ValueError: self.fail("Bloom Tokenizer should be able to deal with padding" ) lowerCAmelCase_ = None # Hotfixing padding = None self.assertRaises(_UpperCamelCase , tokenizer_r.encode , _UpperCamelCase , max_length=_UpperCamelCase , padding="max_length" ) # Simple input self.assertRaises(_UpperCamelCase , tokenizer_r.encode_plus , _UpperCamelCase , max_length=_UpperCamelCase , padding="max_length" ) # Simple input self.assertRaises( _UpperCamelCase , tokenizer_r.batch_encode_plus , _UpperCamelCase , max_length=_UpperCamelCase , padding="max_length" , ) # Pair input self.assertRaises(_UpperCamelCase , tokenizer_r.encode , _UpperCamelCase , max_length=_UpperCamelCase , padding="max_length" ) # Pair input self.assertRaises(_UpperCamelCase , tokenizer_r.encode_plus , _UpperCamelCase , max_length=_UpperCamelCase , padding="max_length" ) # Pair input self.assertRaises( _UpperCamelCase , tokenizer_r.batch_encode_plus , _UpperCamelCase , max_length=_UpperCamelCase , padding="max_length" , ) def __a ( self ) -> Any: lowerCAmelCase_ = self.get_rust_tokenizer() lowerCAmelCase_ = load_dataset("xnli" , "all_languages" , split="test" , streaming=_UpperCamelCase ) lowerCAmelCase_ = next(iter(_UpperCamelCase ) )["premise"] # pick up one data lowerCAmelCase_ = list(sample_data.values() ) lowerCAmelCase_ = list(map(tokenizer.encode , _UpperCamelCase ) ) lowerCAmelCase_ = [tokenizer.decode(_UpperCamelCase , clean_up_tokenization_spaces=_UpperCamelCase ) for x in output_tokens] self.assertListEqual(_UpperCamelCase , _UpperCamelCase ) def __a ( self ) -> List[Any]: # The test has to be overriden because BLOOM uses ALiBi positional embeddings that does not have # any sequence length constraints. This test of the parent class will fail since it relies on the # maximum sequence length of the positoonal embeddings. self.assertGreaterEqual(len(self.tokenizer_class.pretrained_vocab_files_map ) , 1 ) self.assertGreaterEqual(len(list(self.tokenizer_class.pretrained_vocab_files_map.values() )[0] ) , 1 )
231
1
"""simple docstring""" import itertools import os import re a : str = re.compile(R'''([A-Z]+)([A-Z][a-z])''') a : Any = re.compile(R'''([a-z\d])([A-Z])''') a : str = re.compile(R'''(?<!_)_(?!_)''') a : Optional[int] = re.compile(R'''(_{2,})''') a : Union[str, Any] = R'''^\w+(\.\w+)*$''' a : Optional[int] = R'''<>:/\|?*''' def _SCREAMING_SNAKE_CASE ( _lowercase : int ) ->Union[str, Any]: '''simple docstring''' a : Optional[Any] = _uppercase_uppercase_re.sub(R"\1_\2" , lowercase__ ) a : Dict = _lowercase_uppercase_re.sub(R"\1_\2" , lowercase__ ) return name.lower() def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] ) ->Any: '''simple docstring''' a : Optional[int] = _single_underscore_re.split(lowercase__ ) a : Any = [_multiple_underscores_re.split(lowercase__ ) for n in name] return "".join(n.capitalize() for n in itertools.chain.from_iterable(lowercase__ ) if n != "" ) def _SCREAMING_SNAKE_CASE ( _lowercase : Dict ) ->Tuple: '''simple docstring''' if os.path.basename(lowercase__ ) != name: raise ValueError(F"""Should be a dataset name, not a path: {name}""" ) return camelcase_to_snakecase(lowercase__ ) def _SCREAMING_SNAKE_CASE ( _lowercase : Optional[Any] , _lowercase : Dict ) ->Dict: '''simple docstring''' if os.path.basename(lowercase__ ) != name: raise ValueError(F"""Should be a dataset name, not a path: {name}""" ) if not re.match(_split_re , lowercase__ ): raise ValueError(F"""Split name should match \'{_split_re}\'\' but got \'{split}\'.""" ) return F"""{filename_prefix_for_name(lowercase__ )}-{split}""" def _SCREAMING_SNAKE_CASE ( _lowercase : List[str] , _lowercase : str , _lowercase : Any , _lowercase : Tuple=None ) ->Optional[int]: '''simple docstring''' a : str = filename_prefix_for_split(lowercase__ , lowercase__ ) if filetype_suffix: prefix += F""".{filetype_suffix}""" a : str = os.path.join(lowercase__ , lowercase__ ) return F"""{filepath}*""" def _SCREAMING_SNAKE_CASE ( _lowercase : Union[str, Any] , _lowercase : List[str] , _lowercase : Any , _lowercase : Dict=None , _lowercase : Tuple=None ) ->Any: '''simple docstring''' a : List[Any] = filename_prefix_for_split(lowercase__ , lowercase__ ) a : Union[str, Any] = os.path.join(lowercase__ , lowercase__ ) if shard_lengths: a : int = len(lowercase__ ) a : Optional[int] = [F"""{prefix}-{shard_id:05d}-of-{num_shards:05d}""" for shard_id in range(lowercase__ )] if filetype_suffix: a : str = [filename + F""".{filetype_suffix}""" for filename in filenames] return filenames else: a : Dict = prefix if filetype_suffix: filename += F""".{filetype_suffix}""" return [filename]
362
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : int = 50 ) ->int: '''simple docstring''' a : Any = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(F'''{solution() = }''')
79
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyImgaImgPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __magic_name__ ( _UpperCamelCase , unittest.TestCase ): lowerCAmelCase : str = KandinskyImgaImgPipeline lowerCAmelCase : List[Any] = ['prompt', 'image_embeds', 'negative_image_embeds', 'image'] lowerCAmelCase : List[str] = [ 'prompt', 'negative_prompt', 'image_embeds', 'negative_image_embeds', 'image', ] lowerCAmelCase : Union[str, Any] = [ 'generator', 'height', 'width', 'strength', 'guidance_scale', 'negative_prompt', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] lowerCAmelCase : str = False @property def __lowercase ( self : str ): return 32 @property def __lowercase ( self : Union[str, Any] ): return 32 @property def __lowercase ( self : Optional[int] ): return self.time_input_dim @property def __lowercase ( self : List[Any] ): return self.time_input_dim * 4 @property def __lowercase ( self : int ): return 100 @property def __lowercase ( self : Any ): _a : Optional[Any] = XLMRobertaTokenizerFast.from_pretrained('YiYiXu/tiny-random-mclip-base' ) return tokenizer @property def __lowercase ( self : List[Any] ): torch.manual_seed(0 ) _a : Union[str, Any] = MCLIPConfig( numDims=self.cross_attention_dim ,transformerDimensions=self.text_embedder_hidden_size ,hidden_size=self.text_embedder_hidden_size ,intermediate_size=37 ,num_attention_heads=4 ,num_hidden_layers=5 ,vocab_size=1005 ,) _a : int = MultilingualCLIP(_UpperCAmelCase ) _a : List[Any] = text_encoder.eval() return text_encoder @property def __lowercase ( self : Optional[Any] ): torch.manual_seed(0 ) _a : List[Any] = { 'in_channels': 4, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'text_image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'text_image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } _a : Union[str, Any] = UNetaDConditionModel(**_UpperCAmelCase ) return model @property def __lowercase ( self : Union[str, Any] ): return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __lowercase ( self : Optional[int] ): torch.manual_seed(0 ) _a : Any = VQModel(**self.dummy_movq_kwargs ) return model def __lowercase ( self : int ): _a : List[Any] = self.dummy_text_encoder _a : Optional[int] = self.dummy_tokenizer _a : List[Any] = self.dummy_unet _a : Any = self.dummy_movq _a : List[str] = { 'num_train_timesteps': 1000, 'beta_schedule': 'linear', 'beta_start': 0.0_00_85, 'beta_end': 0.0_12, 'clip_sample': False, 'set_alpha_to_one': False, 'steps_offset': 0, 'prediction_type': 'epsilon', 'thresholding': False, } _a : List[str] = DDIMScheduler(**_UpperCAmelCase ) _a : Any = { 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def __lowercase ( self : Union[str, Any] ,_UpperCAmelCase : Dict ,_UpperCAmelCase : Optional[int]=0 ): _a : Any = floats_tensor((1, self.cross_attention_dim) ,rng=random.Random(_UpperCAmelCase ) ).to(_UpperCAmelCase ) _a : Union[str, Any] = floats_tensor((1, self.cross_attention_dim) ,rng=random.Random(seed + 1 ) ).to(_UpperCAmelCase ) # create init_image _a : Optional[int] = floats_tensor((1, 3, 64, 64) ,rng=random.Random(_UpperCAmelCase ) ).to(_UpperCAmelCase ) _a : Optional[int] = image.cpu().permute(0 ,2 ,3 ,1 )[0] _a : Optional[int] = Image.fromarray(np.uinta(_UpperCAmelCase ) ).convert('RGB' ).resize((256, 256) ) if str(_UpperCAmelCase ).startswith('mps' ): _a : Tuple = torch.manual_seed(_UpperCAmelCase ) else: _a : int = torch.Generator(device=_UpperCAmelCase ).manual_seed(_UpperCAmelCase ) _a : List[Any] = { 'prompt': 'horse', 'image': init_image, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 64, 'width': 64, 'num_inference_steps': 10, 'guidance_scale': 7.0, 'strength': 0.2, 'output_type': 'np', } return inputs def __lowercase ( self : Union[str, Any] ): _a : Any = 'cpu' _a : List[Any] = self.get_dummy_components() _a : Tuple = self.pipeline_class(**_UpperCAmelCase ) _a : Optional[Any] = pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) _a : Tuple = pipe(**self.get_dummy_inputs(_UpperCAmelCase ) ) _a : Union[str, Any] = output.images _a : Any = pipe( **self.get_dummy_inputs(_UpperCAmelCase ) ,return_dict=_UpperCAmelCase ,)[0] _a : List[Any] = image[0, -3:, -3:, -1] _a : int = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _a : Tuple = np.array( [0.61_47_49_43, 0.6_07_35_39, 0.43_30_85_44, 0.5_92_82_69, 0.47_49_35_95, 0.46_75_59_73, 0.4_61_38_38, 0.45_36_87_97, 0.50_11_92_33] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" @slow @require_torch_gpu class __magic_name__ ( unittest.TestCase ): def __lowercase ( self : str ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowercase ( self : List[str] ): _a : Optional[Any] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/kandinsky_img2img_frog.npy' ) _a : Any = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png' ) _a : Optional[int] = 'A red cartoon frog, 4k' _a : List[Any] = KandinskyPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-prior' ,torch_dtype=torch.floataa ) pipe_prior.to(_UpperCAmelCase ) _a : List[Any] = KandinskyImgaImgPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1' ,torch_dtype=torch.floataa ) _a : List[Any] = pipeline.to(_UpperCAmelCase ) pipeline.set_progress_bar_config(disable=_UpperCAmelCase ) _a : Any = torch.Generator(device='cpu' ).manual_seed(0 ) _a , _a : Dict = pipe_prior( _UpperCAmelCase ,generator=_UpperCAmelCase ,num_inference_steps=5 ,negative_prompt='' ,).to_tuple() _a : int = pipeline( _UpperCAmelCase ,image=_UpperCAmelCase ,image_embeds=_UpperCAmelCase ,negative_image_embeds=_UpperCAmelCase ,generator=_UpperCAmelCase ,num_inference_steps=100 ,height=768 ,width=768 ,strength=0.2 ,output_type='np' ,) _a : str = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(_UpperCAmelCase ,_UpperCAmelCase )
89
'''simple docstring''' from collections import UserDict from typing import Union import numpy as np import requests from ..utils import ( add_end_docstrings, logging, ) from .audio_classification import ffmpeg_read from .base import PIPELINE_INIT_ARGS, Pipeline a_ : Any = logging.get_logger(__name__) @add_end_docstrings(lowerCamelCase__ ) class __UpperCamelCase ( lowerCamelCase__ ): def __init__( self, **lowerCAmelCase ): """simple docstring""" super().__init__(**lowerCAmelCase ) if self.framework != "pt": raise ValueError(f'''The {self.__class__} is only available in PyTorch.''' ) # No specific FOR_XXX available yet def __call__( self, lowerCAmelCase, **lowerCAmelCase ): """simple docstring""" return super().__call__(lowerCAmelCase, **lowerCAmelCase ) def lowercase__ ( self, **lowerCAmelCase ): """simple docstring""" lowerCamelCase_ ={} if "candidate_labels" in kwargs: lowerCamelCase_ =kwargs['''candidate_labels'''] if "hypothesis_template" in kwargs: lowerCamelCase_ =kwargs['''hypothesis_template'''] return preprocess_params, {}, {} def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase=None, lowerCAmelCase="This is a sound of {}." ): """simple docstring""" if isinstance(lowerCAmelCase, lowerCAmelCase ): if audio.startswith('''http://''' ) or audio.startswith('''https://''' ): # We need to actually check for a real protocol, otherwise it's impossible to use a local file # like http_huggingface_co.png lowerCamelCase_ =requests.get(lowerCAmelCase ).content else: with open(lowerCAmelCase, '''rb''' ) as f: lowerCamelCase_ =f.read() if isinstance(lowerCAmelCase, lowerCAmelCase ): lowerCamelCase_ =ffmpeg_read(lowerCAmelCase, self.feature_extractor.sampling_rate ) if not isinstance(lowerCAmelCase, np.ndarray ): raise ValueError('''We expect a numpy ndarray as input''' ) if len(audio.shape ) != 1: raise ValueError('''We expect a single channel audio input for ZeroShotAudioClassificationPipeline''' ) lowerCamelCase_ =self.feature_extractor( [audio], sampling_rate=self.feature_extractor.sampling_rate, return_tensors='''pt''' ) lowerCamelCase_ =candidate_labels lowerCamelCase_ =[hypothesis_template.format(lowerCAmelCase ) for x in candidate_labels] lowerCamelCase_ =self.tokenizer(lowerCAmelCase, return_tensors=self.framework, padding=lowerCAmelCase ) lowerCamelCase_ =[text_inputs] return inputs def lowercase__ ( self, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =model_inputs.pop('''candidate_labels''' ) lowerCamelCase_ =model_inputs.pop('''text_inputs''' ) if isinstance(text_inputs[0], lowerCAmelCase ): lowerCamelCase_ =text_inputs[0] else: # Batching case. lowerCamelCase_ =text_inputs[0][0] lowerCamelCase_ =self.model(**lowerCAmelCase, **lowerCAmelCase ) lowerCamelCase_ ={ '''candidate_labels''': candidate_labels, '''logits''': outputs.logits_per_audio, } return model_outputs def lowercase__ ( self, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =model_outputs.pop('''candidate_labels''' ) lowerCamelCase_ =model_outputs['''logits'''][0] if self.framework == "pt": lowerCamelCase_ =logits.softmax(dim=0 ) lowerCamelCase_ =probs.tolist() else: raise ValueError('''`tf` framework not supported.''' ) lowerCamelCase_ =[ {'''score''': score, '''label''': candidate_label} for score, candidate_label in sorted(zip(lowerCAmelCase, lowerCAmelCase ), key=lambda lowerCAmelCase : -x[0] ) ] return result
75
0
"""simple docstring""" def lowercase (SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int ) -> float: SCREAMING_SNAKE_CASE = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def lowercase () -> Union[str, Any]: print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
38
"""simple docstring""" from __future__ import annotations def lowercase (SCREAMING_SNAKE_CASE_ : list[int] ) -> bool: return len(set(SCREAMING_SNAKE_CASE_ ) ) == len(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": import doctest doctest.testmod()
38
1
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): def __lowerCamelCase ( self : Any ) ->Optional[Any]: lowerCamelCase__ : Optional[int] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(A , '''tf_padding''' ) ) self.parent.assertTrue(hasattr(A , '''depth_multiplier''' ) ) class __SCREAMING_SNAKE_CASE : def __init__( self : Dict , A : Optional[int] , A : List[Any]=1_3 , A : int=3 , A : List[Any]=3_2 , A : Optional[Any]=0.25 , A : Optional[int]=8 , A : Union[str, Any]=True , A : Any=1_0_2_4 , A : Tuple=3_2 , A : Dict="relu6" , A : Optional[Any]=0.1 , A : Any=0.02 , A : List[Any]=True , A : Optional[Any]=True , A : str=1_0 , A : Optional[int]=None , ) ->int: lowerCamelCase__ : List[Any] = parent lowerCamelCase__ : Any = batch_size lowerCamelCase__ : Union[str, Any] = num_channels lowerCamelCase__ : Dict = image_size lowerCamelCase__ : Dict = depth_multiplier lowerCamelCase__ : str = min_depth lowerCamelCase__ : str = tf_padding lowerCamelCase__ : Union[str, Any] = int(last_hidden_size * depth_multiplier ) lowerCamelCase__ : Tuple = output_stride lowerCamelCase__ : List[str] = hidden_act lowerCamelCase__ : int = classifier_dropout_prob lowerCamelCase__ : Optional[int] = use_labels lowerCamelCase__ : Dict = is_training lowerCamelCase__ : Any = num_labels lowerCamelCase__ : Tuple = initializer_range lowerCamelCase__ : str = scope def __lowerCamelCase ( self : List[Any] ) ->Dict: lowerCamelCase__ : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : Tuple = None lowerCamelCase__ : Any = None if self.use_labels: lowerCamelCase__ : Dict = ids_tensor([self.batch_size] , self.num_labels ) lowerCamelCase__ : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowerCamelCase__ : Tuple = self.get_config() return config, pixel_values, labels, pixel_labels def __lowerCamelCase ( self : Dict ) ->int: return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , min_depth=self.min_depth , tf_padding=self.tf_padding , hidden_act=self.hidden_act , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __lowerCamelCase ( self : Union[str, Any] , A : Optional[Any] , A : Union[str, Any] , A : str , A : Union[str, Any] ) ->int: lowerCamelCase__ : int = MobileNetVaModel(config=A ) model.to(A ) model.eval() lowerCamelCase__ : Tuple = model(A ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __lowerCamelCase ( self : Any , A : Optional[int] , A : List[str] , A : Optional[int] , A : str ) ->Optional[int]: lowerCamelCase__ : Dict = self.num_labels lowerCamelCase__ : Tuple = MobileNetVaForImageClassification(A ) model.to(A ) model.eval() lowerCamelCase__ : int = model(A , labels=A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCamelCase ( self : Tuple ) ->Dict: lowerCamelCase__ : Union[str, Any] = self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = config_and_inputs lowerCamelCase__ : Tuple = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ,lowerCAmelCase_ ,unittest.TestCase ): _UpperCAmelCase : List[Any] = (MobileNetVaModel, MobileNetVaForImageClassification) if is_torch_available() else () _UpperCAmelCase : str = ( {"feature-extraction": MobileNetVaModel, "image-classification": MobileNetVaForImageClassification} if is_torch_available() else {} ) _UpperCAmelCase : Optional[Any] = False _UpperCAmelCase : Tuple = False _UpperCAmelCase : Union[str, Any] = False _UpperCAmelCase : str = False def __lowerCamelCase ( self : str ) ->str: lowerCamelCase__ : Dict = MobileNetVaModelTester(self ) lowerCamelCase__ : Optional[int] = MobileNetVaConfigTester(self , config_class=A , has_text_modality=A ) def __lowerCamelCase ( self : List[Any] ) ->int: self.config_tester.run_common_tests() @unittest.skip(reason='''MobileNetV1 does not use inputs_embeds''' ) def __lowerCamelCase ( self : List[Any] ) ->Tuple: pass @unittest.skip(reason='''MobileNetV1 does not support input and output embeddings''' ) def __lowerCamelCase ( self : Optional[Any] ) ->Optional[int]: pass @unittest.skip(reason='''MobileNetV1 does not output attentions''' ) def __lowerCamelCase ( self : Union[str, Any] ) ->Dict: pass def __lowerCamelCase ( self : str ) ->Any: lowerCamelCase__ , lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] = model_class(A ) lowerCamelCase__ : Tuple = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : str = [*signature.parameters.keys()] lowerCamelCase__ : Any = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A ) def __lowerCamelCase ( self : Union[str, Any] ) ->Any: lowerCamelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def __lowerCamelCase ( self : int ) ->Optional[Any]: def check_hidden_states_output(A : Union[str, Any] , A : List[str] , A : Dict ): lowerCamelCase__ : List[Any] = model_class(A ) model.to(A ) model.eval() with torch.no_grad(): lowerCamelCase__ : Dict = model(**self._prepare_for_class(A , A ) ) lowerCamelCase__ : List[str] = outputs.hidden_states lowerCamelCase__ : Tuple = 2_6 self.assertEqual(len(A ) , A ) lowerCamelCase__ , lowerCamelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Tuple = True check_hidden_states_output(A , A , A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase__ : Optional[Any] = True check_hidden_states_output(A , A , A ) def __lowerCamelCase ( self : int ) ->Union[str, Any]: lowerCamelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A ) @slow def __lowerCamelCase ( self : Any ) ->str: for model_name in MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : int = MobileNetVaModel.from_pretrained(A ) self.assertIsNotNone(A ) def _a ( ) -> str: """simple docstring""" lowerCamelCase__ : Optional[int] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @cached_property def __lowerCamelCase ( self : Dict ) ->str: return ( MobileNetVaImageProcessor.from_pretrained('''google/mobilenet_v1_1.0_224''' ) if is_vision_available() else None ) @slow def __lowerCamelCase ( self : Optional[Any] ) ->Optional[Any]: lowerCamelCase__ : int = MobileNetVaForImageClassification.from_pretrained('''google/mobilenet_v1_1.0_224''' ).to(A ) lowerCamelCase__ : Any = self.default_image_processor lowerCamelCase__ : Tuple = prepare_img() lowerCamelCase__ : List[Any] = image_processor(images=A , return_tensors='''pt''' ).to(A ) # forward pass with torch.no_grad(): lowerCamelCase__ : Any = model(**A ) # verify the logits lowerCamelCase__ : Dict = torch.Size((1, 1_0_0_1) ) self.assertEqual(outputs.logits.shape , A ) lowerCamelCase__ : str = torch.tensor([-4.17_39, -1.12_33, 3.12_05] ).to(A ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , A , atol=1e-4 ) )
142
import warnings from ...utils import logging from .image_processing_segformer import SegformerImageProcessor _A : str = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): def __init__( self : Dict , *A : Any , **A : List[Any] ) ->None: warnings.warn( '''The class SegformerFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use SegformerImageProcessor instead.''' , A , ) super().__init__(*A , **A )
142
1
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _a : Tuple= datasets.logging.get_logger(__name__) _a : Optional[int]= '\\n@InProceedings{moosavi2019minimum,\n author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube},\n title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection},\n year = {2019},\n booktitle = {Proceedings of the 57th Annual Meeting of\n the Association for Computational Linguistics (Volume 1: Long Papers)},\n publisher = {Association for Computational Linguistics},\n address = {Florence, Italy},\n}\n\n@inproceedings{10.3115/1072399.1072405,\nauthor = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette},\ntitle = {A Model-Theoretic Coreference Scoring Scheme},\nyear = {1995},\nisbn = {1558604022},\npublisher = {Association for Computational Linguistics},\naddress = {USA},\nurl = {https://doi.org/10.3115/1072399.1072405},\ndoi = {10.3115/1072399.1072405},\nbooktitle = {Proceedings of the 6th Conference on Message Understanding},\npages = {45–52},\nnumpages = {8},\nlocation = {Columbia, Maryland},\nseries = {MUC6 ’95}\n}\n\n@INPROCEEDINGS{Bagga98algorithmsfor,\n author = {Amit Bagga and Breck Baldwin},\n title = {Algorithms for Scoring Coreference Chains},\n booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference},\n year = {1998},\n pages = {563--566}\n}\n\n@INPROCEEDINGS{Luo05oncoreference,\n author = {Xiaoqiang Luo},\n title = {On coreference resolution performance metrics},\n booktitle = {In Proc. of HLT/EMNLP},\n year = {2005},\n pages = {25--32},\n publisher = {URL}\n}\n\n@inproceedings{moosavi-strube-2016-coreference,\n title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric",\n author = "Moosavi, Nafise Sadat and\n Strube, Michael",\n booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",\n month = aug,\n year = "2016",\n address = "Berlin, Germany",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/P16-1060",\n doi = "10.18653/v1/P16-1060",\n pages = "632--642",\n}\n\n' _a : Tuple= '\\nCoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which\nimplements of the common evaluation metrics including MUC [Vilain et al, 1995],\nB-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005],\nLEA [Moosavi and Strube, 2016] and the averaged CoNLL score\n(the average of the F1 values of MUC, B-cubed and CEAFe)\n[Denis and Baldridge, 2009a; Pradhan et al., 2011].\n\nThis wrapper of CoVal currently only work with CoNLL line format:\nThe CoNLL format has one word per line with all the annotation for this word in column separated by spaces:\nColumn Type Description\n1 Document ID This is a variation on the document filename\n2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n3 Word number\n4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release.\n5 Part-of-Speech\n6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column.\n7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-"\n8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7.\n9 Word sense This is the word sense of the word in Column 3.\n10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data.\n11 Named Entities These columns identifies the spans representing various named entities.\n12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7.\nN Coreference Coreference chain information encoded in a parenthesis structure.\nMore informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html\n\nDetails on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md\n\nCoVal code was written by @ns-moosavi.\nSome parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py\nThe test suite is taken from https://github.com/conll/reference-coreference-scorers/\nMention evaluation and the test suite are added by @andreasvc.\nParsing CoNLL files is developed by Leo Born.\n' _a : Optional[int]= '\nCalculates coreference evaluation metrics.\nArgs:\n predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format.\n Each prediction is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format.\n Each reference is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n keep_singletons: After extracting all mentions of key or system files,\n mentions whose corresponding coreference chain is of size one,\n are considered as singletons. The default evaluation mode will include\n singletons in evaluations if they are included in the key or the system files.\n By setting \'keep_singletons=False\', all singletons in the key and system files\n will be excluded from the evaluation.\n NP_only: Most of the recent coreference resolvers only resolve NP mentions and\n leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs.\n min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans.\n Minimum spans are determined using the MINA algorithm.\n\nReturns:\n \'mentions\': mentions\n \'muc\': MUC metric [Vilain et al, 1995]\n \'bcub\': B-cubed [Bagga and Baldwin, 1998]\n \'ceafe\': CEAFe [Luo et al., 2005]\n \'lea\': LEA [Moosavi and Strube, 2016]\n \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe)\n\nExamples:\n\n >>> coval = datasets.load_metric(\'coval\')\n >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\',\n ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\',\n ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\',\n ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\',\n ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\',\n ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\']\n >>> references = [words]\n >>> predictions = [words]\n >>> results = coval.compute(predictions=predictions, references=references)\n >>> print(results) # doctest:+ELLIPSIS\n {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0}\n' def __UpperCAmelCase ( UpperCAmelCase_ : Dict , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Dict=False , UpperCAmelCase_ : List[Any]=False , UpperCAmelCase_ : Tuple=True , UpperCAmelCase_ : List[Any]=False , UpperCAmelCase_ : List[Any]="dummy_doc" ) -> List[str]: '''simple docstring''' __snake_case : Any = {doc: key_lines} __snake_case : Tuple = {doc: sys_lines} __snake_case : List[Any] = {} __snake_case : Union[str, Any] = 0 __snake_case : List[str] = 0 __snake_case : List[str] = 0 __snake_case : Optional[int] = 0 __snake_case : Optional[Any] = 0 __snake_case : str = 0 __snake_case , __snake_case : Tuple = reader.get_doc_mentions(UpperCamelCase__ , key_doc_lines[doc] , UpperCamelCase__ ) key_singletons_num += singletons_num if NP_only or min_span: __snake_case : Any = reader.set_annotated_parse_trees(UpperCamelCase__ , key_doc_lines[doc] , UpperCamelCase__ , UpperCamelCase__ ) __snake_case , __snake_case : List[str] = reader.get_doc_mentions(UpperCamelCase__ , sys_doc_lines[doc] , UpperCamelCase__ ) sys_singletons_num += singletons_num if NP_only or min_span: __snake_case : List[Any] = reader.set_annotated_parse_trees(UpperCamelCase__ , key_doc_lines[doc] , UpperCamelCase__ , UpperCamelCase__ ) if remove_nested: __snake_case , __snake_case : Tuple = reader.remove_nested_coref_mentions(UpperCamelCase__ , UpperCamelCase__ ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters __snake_case , __snake_case : Any = reader.remove_nested_coref_mentions(UpperCamelCase__ , UpperCamelCase__ ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters __snake_case : Any = reader.get_mention_assignments(UpperCamelCase__ , UpperCamelCase__ ) __snake_case : List[Any] = reader.get_mention_assignments(UpperCamelCase__ , UpperCamelCase__ ) __snake_case : Tuple = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( 'Number of removed nested coreferring mentions in the key ' F"annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}" ) logger.info( 'Number of resulting singleton clusters in the key ' F"annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}" ) if not keep_singletons: logger.info( F"{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system " 'files, respectively' ) return doc_coref_infos def __UpperCAmelCase ( UpperCAmelCase_ : Tuple , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : str , UpperCAmelCase_ : Optional[int] ) -> List[str]: '''simple docstring''' __snake_case : Any = get_coref_infos(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) __snake_case : int = {} __snake_case : int = 0 __snake_case : Union[str, Any] = 0 for name, metric in metrics: __snake_case , __snake_case , __snake_case : Tuple = evaluator.evaluate_documents(UpperCamelCase__ , UpperCamelCase__ , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({F"{name}/recall": recall, F"{name}/precision": precision, F"{name}/f1": fa} ) logger.info( name.ljust(10 ) , F"Recall: {recall * 1_00:.2f}" , F" Precision: {precision * 1_00:.2f}" , F" F1: {fa * 1_00:.2f}" , ) if conll_subparts_num == 3: __snake_case : Any = (conll / 3) * 1_00 logger.info(F"CoNLL score: {conll:.2f}" ) output_scores.update({'conll_score': conll} ) return output_scores def __UpperCAmelCase ( UpperCAmelCase_ : Tuple ) -> Tuple: '''simple docstring''' __snake_case : Union[str, Any] = False for line in key_lines: if not line.startswith('#' ): if len(line.split() ) > 6: __snake_case : List[Any] = line.split()[5] if not parse_col == "-": __snake_case : str = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase ( datasets.Metric ): def _lowercase (self : Dict) -> Tuple: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('string')), 'references': datasets.Sequence(datasets.Value('string')), }) , codebase_urls=['https://github.com/ns-moosavi/coval'] , reference_urls=[ 'https://github.com/ns-moosavi/coval', 'https://www.aclweb.org/anthology/P16-1060', 'http://www.conll.cemantix.org/2012/data.html', ] , ) def _lowercase (self : List[Any] , _A : int , _A : Union[str, Any] , _A : List[Any]=True , _A : Optional[int]=False , _A : Optional[Any]=False , _A : List[str]=False) -> Any: __snake_case : Union[str, Any] = [ ('mentions', evaluator.mentions), ('muc', evaluator.muc), ('bcub', evaluator.b_cubed), ('ceafe', evaluator.ceafe), ('lea', evaluator.lea), ] if min_span: __snake_case : Optional[Any] = util.check_gold_parse_annotation(__lowerCamelCase) if not has_gold_parse: raise NotImplementedError('References should have gold parse annotation to use \'min_span\'.') # util.parse_key_file(key_file) # key_file = key_file + ".parsed" __snake_case : Optional[Any] = evaluate( key_lines=__lowerCamelCase , sys_lines=__lowerCamelCase , metrics=__lowerCamelCase , NP_only=__lowerCamelCase , remove_nested=__lowerCamelCase , keep_singletons=__lowerCamelCase , min_span=__lowerCamelCase , ) return score
367
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from tokenizers.pre_tokenizers import BertPreTokenizer, PreTokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roformer import RoFormerTokenizer from .tokenization_utils import JiebaPreTokenizer _a : List[Any]= logging.get_logger(__name__) _a : Any= {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} _a : int= { "vocab_file": { "junnyu/roformer_chinese_small": "https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/vocab.txt", "junnyu/roformer_chinese_base": "https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/vocab.txt", "junnyu/roformer_chinese_char_small": ( "https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/vocab.txt" ), "junnyu/roformer_chinese_char_base": ( "https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/vocab.txt" ), "junnyu/roformer_small_discriminator": ( "https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/vocab.txt" ), "junnyu/roformer_small_generator": ( "https://huggingface.co/junnyu/roformer_small_generator/resolve/main/vocab.txt" ), } } _a : Optional[Any]= { "junnyu/roformer_chinese_small": 1_536, "junnyu/roformer_chinese_base": 1_536, "junnyu/roformer_chinese_char_small": 512, "junnyu/roformer_chinese_char_base": 512, "junnyu/roformer_small_discriminator": 128, "junnyu/roformer_small_generator": 128, } _a : str= { "junnyu/roformer_chinese_small": {"do_lower_case": True}, "junnyu/roformer_chinese_base": {"do_lower_case": True}, "junnyu/roformer_chinese_char_small": {"do_lower_case": True}, "junnyu/roformer_chinese_char_base": {"do_lower_case": True}, "junnyu/roformer_small_discriminator": {"do_lower_case": True}, "junnyu/roformer_small_generator": {"do_lower_case": True}, } class UpperCamelCase ( lowercase ): UpperCAmelCase : Union[str, Any] = VOCAB_FILES_NAMES UpperCAmelCase : Optional[int] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase : int = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase : Dict = RoFormerTokenizer def __init__(self : List[Any] , _A : Any=None , _A : int=None , _A : Dict=True , _A : List[Any]="[UNK]" , _A : Tuple="[SEP]" , _A : List[Any]="[PAD]" , _A : str="[CLS]" , _A : int="[MASK]" , _A : Optional[int]=True , _A : List[str]=None , **_A : int , ) -> Dict: super().__init__( _A , tokenizer_file=_A , do_lower_case=_A , unk_token=_A , sep_token=_A , pad_token=_A , cls_token=_A , mask_token=_A , tokenize_chinese_chars=_A , strip_accents=_A , **_A , ) __snake_case : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( pre_tok_state.get('lowercase' , _A) != do_lower_case or pre_tok_state.get('strip_accents' , _A) != strip_accents ): __snake_case : Union[str, Any] = getattr(_A , pre_tok_state.pop('type')) __snake_case : Union[str, Any] = do_lower_case __snake_case : str = strip_accents __snake_case : Optional[int] = pre_tok_class(**_A) __snake_case : int = do_lower_case def __getstate__(self : Optional[Any]) -> Dict: __snake_case : Optional[int] = self.__dict__.copy() __snake_case : int = BertPreTokenizer() return state def __setstate__(self : Optional[Any] , _A : Optional[Any]) -> Dict: __snake_case : List[str] = d __snake_case : str = self.__dict__['_tokenizer'].get_vocab() __snake_case : int = PreTokenizer.custom(JiebaPreTokenizer(_A)) def _lowercase (self : int , _A : Tuple , _A : Any=None) -> str: __snake_case : Dict = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase (self : List[str] , _A : List[int] , _A : Optional[List[int]] = None) -> List[int]: __snake_case : Tuple = [self.sep_token_id] __snake_case : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def _lowercase (self : List[Any] , _A : str , _A : Optional[str] = None) -> Tuple[str]: __snake_case : List[Any] = self._tokenizer.model.save(_A , name=_A) return tuple(_A) def _lowercase (self : int , _A : Optional[int] , _A : Tuple=None , _A : Tuple=None , _A : Dict=False , **_A : Optional[int] , ) -> Optional[Any]: __snake_case : Optional[Any] = BertPreTokenizer() return super().save_pretrained(_A , _A , _A , _A , **_A)
95
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase_ : Any = { """configuration_data2vec_audio""": ["""DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecAudioConfig"""], """configuration_data2vec_text""": [ """DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecTextConfig""", """Data2VecTextOnnxConfig""", ], """configuration_data2vec_vision""": [ """DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecVisionConfig""", """Data2VecVisionOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Optional[int] = [ """DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecAudioForAudioFrameClassification""", """Data2VecAudioForCTC""", """Data2VecAudioForSequenceClassification""", """Data2VecAudioForXVector""", """Data2VecAudioModel""", """Data2VecAudioPreTrainedModel""", ] UpperCAmelCase_ : Union[str, Any] = [ """DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecTextForCausalLM""", """Data2VecTextForMaskedLM""", """Data2VecTextForMultipleChoice""", """Data2VecTextForQuestionAnswering""", """Data2VecTextForSequenceClassification""", """Data2VecTextForTokenClassification""", """Data2VecTextModel""", """Data2VecTextPreTrainedModel""", ] UpperCAmelCase_ : Optional[Any] = [ """DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecVisionForImageClassification""", """Data2VecVisionForMaskedImageModeling""", """Data2VecVisionForSemanticSegmentation""", """Data2VecVisionModel""", """Data2VecVisionPreTrainedModel""", ] if is_tf_available(): UpperCAmelCase_ : Any = [ """TFData2VecVisionForImageClassification""", """TFData2VecVisionForSemanticSegmentation""", """TFData2VecVisionModel""", """TFData2VecVisionPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys UpperCAmelCase_ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
91
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCAmelCase = { 'configuration_m2m_100': ['M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP', 'M2M100Config', 'M2M100OnnxConfig'], 'tokenization_m2m_100': ['M2M100Tokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST', 'M2M100ForConditionalGeneration', 'M2M100Model', 'M2M100PreTrainedModel', ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
341
0
"""simple docstring""" import tempfile import unittest import numpy as np import transformers from transformers import GPTaTokenizer, GPTJConfig, is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax, tooslow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax import jax.numpy as jnp from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) from transformers.models.gptj.modeling_flax_gptj import FlaxGPTJForCausalLM, FlaxGPTJModel if is_torch_available(): import torch class __snake_case : def __init__( self , lowercase , lowercase=14 , lowercase=7 , lowercase=True , lowercase=True , lowercase=False , lowercase=True , lowercase=99 , lowercase=32 , lowercase=4 , lowercase=4 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=5_12 , lowercase=0.02 , ) -> Any: '''simple docstring''' a__: List[str] = parent a__: Any = batch_size a__: str = seq_length a__: str = is_training a__: Optional[int] = use_input_mask a__: Tuple = use_token_type_ids a__: int = use_labels a__: str = vocab_size a__: Optional[int] = hidden_size a__: List[Any] = rotary_dim a__: Any = num_hidden_layers a__: Dict = num_attention_heads a__: Any = intermediate_size a__: Any = hidden_act a__: Any = hidden_dropout_prob a__: int = attention_probs_dropout_prob a__: List[str] = max_position_embeddings a__: int = initializer_range a__: str = None a__: int = vocab_size - 1 a__: Union[str, Any] = vocab_size - 1 a__: List[str] = vocab_size - 1 def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__: Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) a__: Optional[Any] = None if self.use_input_mask: a__: Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length]) a__: Dict = GPTJConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , use_cache=lowercase , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , rotary_dim=self.rotary_dim , ) return (config, input_ids, input_mask) def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: Dict = self.prepare_config_and_inputs() a__ , a__ , a__: Optional[Any] = config_and_inputs a__: Optional[Any] = {'input_ids': input_ids, 'attention_mask': attention_mask} return config, inputs_dict def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' a__: Tuple = 20 a__: List[Any] = model_class_name(lowercase) a__: int = model.init_cache(input_ids.shape[0] , lowercase) a__: List[str] = jnp.ones((input_ids.shape[0], max_decoder_length) , dtype='i4') a__: List[Any] = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1)[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1)) a__: str = model( input_ids[:, :-1] , attention_mask=lowercase , past_key_values=lowercase , position_ids=lowercase , ) a__: Dict = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype='i4') a__: Optional[Any] = model( input_ids[:, -1:] , attention_mask=lowercase , past_key_values=outputs_cache.past_key_values , position_ids=lowercase , ) a__: int = model(lowercase) a__: Optional[int] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]))) self.parent.assertTrue(diff < 1e-3 , msg=f'Max diff is {diff}') def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase) -> List[str]: '''simple docstring''' a__: int = 20 a__: List[Any] = model_class_name(lowercase) a__: str = jnp.concatenate( [attention_mask, jnp.zeros((attention_mask.shape[0], max_decoder_length - attention_mask.shape[1]))] , axis=-1 , ) a__: Dict = model.init_cache(input_ids.shape[0] , lowercase) a__: Optional[int] = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1)[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1)) a__: List[str] = model( input_ids[:, :-1] , attention_mask=lowercase , past_key_values=lowercase , position_ids=lowercase , ) a__: int = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype='i4') a__: Dict = model( input_ids[:, -1:] , past_key_values=outputs_cache.past_key_values , attention_mask=lowercase , position_ids=lowercase , ) a__: Dict = model(lowercase , attention_mask=lowercase) a__: str = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]))) self.parent.assertTrue(diff < 1e-3 , msg=f'Max diff is {diff}') @require_flax class __snake_case ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): a__ = (FlaxGPTJModel, FlaxGPTJForCausalLM) if is_flax_available() else () a__ = (FlaxGPTJForCausalLM,) if is_flax_available() else () def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__: List[str] = FlaxGPTJModelTester(self) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' for model_class_name in self.all_model_classes: a__ , a__ , a__: Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward(lowercase , lowercase , lowercase , lowercase) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' for model_class_name in self.all_model_classes: a__ , a__ , a__: Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward_with_attn_mask( lowercase , lowercase , lowercase , lowercase) @tooslow def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: Union[str, Any] = GPTaTokenizer.from_pretrained('gpt2' , pad_token='<|endoftext|>' , padding_side='left') a__: Optional[int] = tokenizer(['Hello this is a long string', 'Hey'] , return_tensors='np' , padding=lowercase , truncation=lowercase) a__: Dict = FlaxGPTJForCausalLM.from_pretrained('EleutherAI/gpt-j-6B') a__: Dict = False a__: List[Any] = model.config.eos_token_id a__: str = jax.jit(model.generate) a__: Optional[int] = jit_generate( inputs['input_ids'] , attention_mask=inputs['attention_mask'] , pad_token_id=tokenizer.pad_token_id).sequences a__: Union[str, Any] = tokenizer.batch_decode(lowercase , skip_special_tokens=lowercase) a__: int = [ 'Hello this is a long string of text.\n\nI\'m trying to get the text of the', 'Hey, I\'m a little late to the party. I\'m going to', ] self.assertListEqual(lowercase , lowercase) @is_pt_flax_cross_test def lowerCamelCase_ ( self) -> List[str]: '''simple docstring''' a__ , a__: Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): # prepare inputs a__: List[Any] = self._prepare_for_class(lowercase , lowercase) a__: Dict = {k: torch.tensor(v.tolist()) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class a__: Optional[int] = model_class.__name__[4:] # Skip the "Flax" at the beginning a__: Optional[int] = getattr(lowercase , lowercase) a__ , a__: Any = pt_inputs['input_ids'].shape a__: Tuple = np.random.randint(0 , seq_length - 1 , size=(batch_size,)) for batch_idx, start_index in enumerate(lowercase): a__: Optional[int] = 0 a__: List[Any] = 1 a__: Dict = 0 a__: Dict = 1 a__: Optional[Any] = pt_model_class(lowercase).eval() a__: Tuple = model_class(lowercase , dtype=jnp.floataa) a__: int = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , lowercase) a__: Optional[Any] = fx_state with torch.no_grad(): a__: str = pt_model(**lowercase).to_tuple() a__: Dict = fx_model(**lowercase).to_tuple() self.assertEqual(len(lowercase) , len(lowercase) , 'Output lengths differ between Flax and PyTorch') for fx_output, pt_output in zip(lowercase , lowercase): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4e-2) with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(lowercase) a__: int = model_class.from_pretrained(lowercase , from_pt=lowercase) a__: int = fx_model_loaded(**lowercase).to_tuple() self.assertEqual( len(lowercase) , len(lowercase) , 'Output lengths differ between Flax and PyTorch') for fx_output_loaded, pt_output in zip(lowercase , lowercase): self.assert_almost_equals(fx_output_loaded[:, -1] , pt_output[:, -1].numpy() , 4e-2) @is_pt_flax_cross_test def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__ , a__: str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): # prepare inputs a__: List[str] = self._prepare_for_class(lowercase , lowercase) a__: Union[str, Any] = {k: torch.tensor(v.tolist()) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class a__: Optional[Any] = model_class.__name__[4:] # Skip the "Flax" at the beginning a__: str = getattr(lowercase , lowercase) a__: Any = pt_model_class(lowercase).eval() a__: Union[str, Any] = model_class(lowercase , dtype=jnp.floataa) a__: Optional[int] = load_flax_weights_in_pytorch_model(lowercase , fx_model.params) a__ , a__: str = pt_inputs['input_ids'].shape a__: List[str] = np.random.randint(0 , seq_length - 1 , size=(batch_size,)) for batch_idx, start_index in enumerate(lowercase): a__: Optional[Any] = 0 a__: Optional[Any] = 1 a__: str = 0 a__: Dict = 1 # make sure weights are tied in PyTorch pt_model.tie_weights() with torch.no_grad(): a__: Any = pt_model(**lowercase).to_tuple() a__: Union[str, Any] = fx_model(**lowercase).to_tuple() self.assertEqual(len(lowercase) , len(lowercase) , 'Output lengths differ between Flax and PyTorch') for fx_output, pt_output in zip(lowercase , lowercase): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4e-2) with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(lowercase) a__: Dict = pt_model_class.from_pretrained(lowercase , from_flax=lowercase) with torch.no_grad(): a__: Optional[Any] = pt_model_loaded(**lowercase).to_tuple() self.assertEqual( len(lowercase) , len(lowercase) , 'Output lengths differ between Flax and PyTorch') for fx_output, pt_output in zip(lowercase , lowercase): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4e-2) @tooslow def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' for model_class_name in self.all_model_classes: a__: Dict = model_class_name.from_pretrained('EleutherAI/gpt-j-6B') a__: Optional[Any] = model(np.ones((1, 1))) self.assertIsNotNone(lowercase)
203
"""simple docstring""" import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert import BertTokenizer lowercase__ = logging.get_logger(__name__) lowercase__ = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} lowercase__ = { 'vocab_file': { 'facebook/dpr-ctx_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-ctx_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-ctx_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-ctx_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json' ), }, } lowercase__ = { 'vocab_file': { 'facebook/dpr-question_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-question_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-question_encoder-single-nq-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-question_encoder-multiset-base': ( 'https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json' ), }, } lowercase__ = { 'vocab_file': { 'facebook/dpr-reader-single-nq-base': ( 'https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt' ), 'facebook/dpr-reader-multiset-base': ( 'https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'facebook/dpr-reader-single-nq-base': ( 'https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json' ), 'facebook/dpr-reader-multiset-base': ( 'https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json' ), }, } lowercase__ = { 'facebook/dpr-ctx_encoder-single-nq-base': 512, 'facebook/dpr-ctx_encoder-multiset-base': 512, } lowercase__ = { 'facebook/dpr-question_encoder-single-nq-base': 512, 'facebook/dpr-question_encoder-multiset-base': 512, } lowercase__ = { 'facebook/dpr-reader-single-nq-base': 512, 'facebook/dpr-reader-multiset-base': 512, } lowercase__ = { 'facebook/dpr-ctx_encoder-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-ctx_encoder-multiset-base': {'do_lower_case': True}, } lowercase__ = { 'facebook/dpr-question_encoder-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-question_encoder-multiset-base': {'do_lower_case': True}, } lowercase__ = { 'facebook/dpr-reader-single-nq-base': {'do_lower_case': True}, 'facebook/dpr-reader-multiset-base': {'do_lower_case': True}, } class __snake_case ( __lowerCAmelCase ): a__ = VOCAB_FILES_NAMES a__ = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP a__ = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION class __snake_case ( __lowerCAmelCase ): a__ = VOCAB_FILES_NAMES a__ = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP a__ = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION lowercase__ = collections.namedtuple( 'DPRSpanPrediction', ['span_score', 'relevance_score', 'doc_id', 'start_index', 'end_index', 'text'] ) lowercase__ = collections.namedtuple('DPRReaderOutput', ['start_logits', 'end_logits', 'relevance_logits']) lowercase__ = r'\n Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`.\n It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers),\n using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)`\n with the format:\n\n ```\n [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids>\n ```\n\n Args:\n questions (`str` or `List[str]`):\n The questions to be encoded. You can specify one question for many passages. In this case, the question\n will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in\n `titles` or `texts`.\n titles (`str` or `List[str]`):\n The passages titles to be encoded. This can be a string or a list of strings if there are several passages.\n texts (`str` or `List[str]`):\n The passages texts to be encoded. This can be a string or a list of strings if there are several passages.\n padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`):\n Activates and controls padding. Accepts the following values:\n\n - `True` or `\'longest\'`: Pad to the longest sequence in the batch (or no padding if only a single sequence\n if provided).\n - `\'max_length\'`: Pad to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided.\n - `False` or `\'do_not_pad\'` (default): No padding (i.e., can output a batch with sequences of different\n lengths).\n truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`):\n Activates and controls truncation. Accepts the following values:\n\n - `True` or `\'longest_first\'`: Truncate to a maximum length specified with the argument `max_length` or to\n the maximum acceptable input length for the model if that argument is not provided. This will truncate\n token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch\n of pairs) is provided.\n - `\'only_first\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the first\n sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `\'only_second\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum\n acceptable input length for the model if that argument is not provided. This will only truncate the\n second sequence of a pair if a pair of sequences (or a batch of pairs) is provided.\n - `False` or `\'do_not_truncate\'` (default): No truncation (i.e., can output batch with sequence lengths\n greater than the model maximum admissible input size).\n max_length (`int`, *optional*):\n Controls the maximum length to use by one of the truncation/padding parameters.\n\n If left unset or set to `None`, this will use the predefined model maximum length if a maximum length\n is required by one of the truncation/padding parameters. If the model has no specific maximum input\n length (like XLNet) truncation/padding to a maximum length will be deactivated.\n return_tensors (`str` or [`~utils.TensorType`], *optional*):\n If set, will return tensors instead of list of python integers. Acceptable values are:\n\n - `\'tf\'`: Return TensorFlow `tf.constant` objects.\n - `\'pt\'`: Return PyTorch `torch.Tensor` objects.\n - `\'np\'`: Return Numpy `np.ndarray` objects.\n return_attention_mask (`bool`, *optional*):\n Whether or not to return the attention mask. If not set, will return the attention mask according to the\n specific tokenizer\'s default, defined by the `return_outputs` attribute.\n\n [What are attention masks?](../glossary#attention-mask)\n\n Returns:\n `Dict[str, List[List[int]]]`: A dictionary with the following keys:\n\n - `input_ids`: List of token ids to be fed to a model.\n - `attention_mask`: List of indices specifying which tokens should be attended to by the model.\n ' @add_start_docstrings(__lowerCAmelCase ) class __snake_case : def __call__( self , lowercase , lowercase = None , lowercase = None , lowercase = False , lowercase = False , lowercase = None , lowercase = None , lowercase = None , **lowercase , ) -> BatchEncoding: '''simple docstring''' if titles is None and texts is None: return super().__call__( lowercase , padding=lowercase , truncation=lowercase , max_length=lowercase , return_tensors=lowercase , return_attention_mask=lowercase , **lowercase , ) elif titles is None or texts is None: a__: str = titles if texts is None else texts return super().__call__( lowercase , lowercase , padding=lowercase , truncation=lowercase , max_length=lowercase , return_tensors=lowercase , return_attention_mask=lowercase , **lowercase , ) a__: Tuple = titles if not isinstance(lowercase , lowercase) else [titles] a__: Optional[int] = texts if not isinstance(lowercase , lowercase) else [texts] a__: Dict = len(lowercase) a__: Dict = questions if not isinstance(lowercase , lowercase) else [questions] * n_passages if len(lowercase) != len(lowercase): raise ValueError( f'There should be as many titles than texts but got {len(lowercase)} titles and {len(lowercase)} texts.') a__: List[str] = super().__call__(lowercase , lowercase , padding=lowercase , truncation=lowercase)['input_ids'] a__: List[Any] = super().__call__(lowercase , add_special_tokens=lowercase , padding=lowercase , truncation=lowercase)['input_ids'] a__: Optional[Any] = { 'input_ids': [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(lowercase , lowercase) ] } if return_attention_mask is not False: a__: Dict = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id) for input_id in input_ids]) a__: int = attention_mask return self.pad(lowercase , padding=lowercase , max_length=lowercase , return_tensors=lowercase) def lowerCamelCase_ ( self , lowercase , lowercase , lowercase = 16 , lowercase = 64 , lowercase = 4 , ) -> List[DPRSpanPrediction]: '''simple docstring''' a__: Dict = reader_input['input_ids'] a__ , a__ , a__: Tuple = reader_output[:3] a__: Tuple = len(lowercase) a__: Optional[Any] = sorted(range(lowercase) , reverse=lowercase , key=relevance_logits.__getitem__) a__: List[DPRReaderOutput] = [] for doc_id in sorted_docs: a__: Tuple = list(input_ids[doc_id]) # assuming question & title information is at the beginning of the sequence a__: Dict = sequence_ids.index(self.sep_token_id , 2) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: a__: Any = sequence_ids.index(self.pad_token_id) else: a__: Optional[Any] = len(lowercase) a__: Optional[int] = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] , end_logits=end_logits[doc_id][passage_offset:sequence_len] , max_answer_length=lowercase , top_spans=lowercase , ) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] , relevance_score=relevance_logits[doc_id] , doc_id=lowercase , start_index=lowercase , end_index=lowercase , text=self.decode(sequence_ids[start_index : end_index + 1]) , )) if len(lowercase) >= num_spans: break return nbest_spans_predictions[:num_spans] def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , ) -> List[DPRSpanPrediction]: '''simple docstring''' a__: Optional[Any] = [] for start_index, start_score in enumerate(lowercase): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length]): scores.append(((start_index, start_index + answer_length), start_score + end_score)) a__: str = sorted(lowercase , key=lambda lowercase: x[1] , reverse=lowercase) a__: Optional[Any] = [] for (start_index, end_index), score in scores: if start_index > end_index: raise ValueError(f'Wrong span indices: [{start_index}:{end_index}]') a__: str = end_index - start_index + 1 if length > max_answer_length: raise ValueError(f'Span is too long: {length} > {max_answer_length}') if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals): continue chosen_span_intervals.append((start_index, end_index)) if len(lowercase) == top_spans: break return chosen_span_intervals @add_end_docstrings(__lowerCAmelCase ) class __snake_case ( __lowerCAmelCase , __lowerCAmelCase ): a__ = VOCAB_FILES_NAMES a__ = READER_PRETRAINED_VOCAB_FILES_MAP a__ = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ = READER_PRETRAINED_INIT_CONFIGURATION a__ = ["""input_ids""", """attention_mask"""]
203
1
import json import os import unittest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' lowercase_ = CLIPTokenizer lowercase_ = CLIPTokenizerFast lowercase_ = True lowercase_ = {} lowercase_ = False def SCREAMING_SNAKE_CASE_ (self : str) ->Union[str, Any]: '''simple docstring''' super().setUp() # fmt: off lowerCamelCase__: str =["l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "lo", "l</w>", "w</w>", "r</w>", "t</w>", "low</w>", "er</w>", "lowest</w>", "newer</w>", "wider", "<unk>", "<|startoftext|>", "<|endoftext|>"] # fmt: on lowerCamelCase__: Optional[int] =dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_)))) lowerCamelCase__: Dict =["#version: 0.2", "l o", "lo w</w>", "e r</w>"] lowerCamelCase__: List[str] ={"unk_token": "<unk>"} lowerCamelCase__: Dict =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) lowerCamelCase__: Dict =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as fp: fp.write(json.dumps(UpperCAmelCase_) + "\n") with open(self.merges_file , "w" , encoding="utf-8") as fp: fp.write("\n".join(UpperCAmelCase_)) def SCREAMING_SNAKE_CASE_ (self : Optional[int] , **UpperCAmelCase_ : List[str]) ->List[Any]: '''simple docstring''' kwargs.update(self.special_tokens_map) return CLIPTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : Any , **UpperCAmelCase_ : Any) ->Dict: '''simple docstring''' kwargs.update(self.special_tokens_map) return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : int , UpperCAmelCase_ : List[Any]) ->int: '''simple docstring''' lowerCamelCase__: Optional[Any] ="lower newer" lowerCamelCase__: Optional[Any] ="lower newer" return input_text, output_text def SCREAMING_SNAKE_CASE_ (self : Any) ->Optional[Any]: '''simple docstring''' lowerCamelCase__: Any =CLIPTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map) lowerCamelCase__: Any ="lower newer" lowerCamelCase__: List[str] =["lo", "w", "er</w>", "n", "e", "w", "er</w>"] lowerCamelCase__: List[str] =tokenizer.tokenize(UpperCAmelCase_) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_) lowerCamelCase__: Tuple =tokens + [tokenizer.unk_token] lowerCamelCase__: List[str] =[10, 2, 16, 9, 3, 2, 16, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCAmelCase_) , UpperCAmelCase_) @require_ftfy def SCREAMING_SNAKE_CASE_ (self : str) ->List[Any]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})"""): lowerCamelCase__: Tuple =self.tokenizer_class.from_pretrained(UpperCAmelCase_ , **UpperCAmelCase_) lowerCamelCase__: int =self.rust_tokenizer_class.from_pretrained(UpperCAmelCase_ , **UpperCAmelCase_) lowerCamelCase__: Union[str, Any] ="A\n'll 11p223RF☆ho!!to?'d'd''d of a cat to-$''d." lowerCamelCase__: str =tokenizer_s.tokenize(UpperCAmelCase_) lowerCamelCase__: int =tokenizer_r.tokenize(UpperCAmelCase_) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_) # Test that the tokenization is identical on an example containing a character (Latin Small Letter A # with Tilde) encoded in 2 different ways lowerCamelCase__: Optional[int] ="xa\u0303y" + " " + "x\xe3y" lowerCamelCase__: Union[str, Any] =tokenizer_s.tokenize(UpperCAmelCase_) lowerCamelCase__: Tuple =tokenizer_r.tokenize(UpperCAmelCase_) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_) # Test that the tokenization is identical on unicode of space type lowerCamelCase__: Tuple =[ "\u0009", # (horizontal tab, '\t') "\u000B", # (vertical tab) "\u000C", # (form feed) "\u0020", # (space, ' ') "\u200E", # (left-to-right mark):w "\u200F", # (right-to-left mark) ] for unicode_seq in spaces_unicodes: lowerCamelCase__: int =tokenizer_s.tokenize(UpperCAmelCase_) lowerCamelCase__: Optional[Any] =tokenizer_r.tokenize(UpperCAmelCase_) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_) # Test that the tokenization is identical on unicode of line break type lowerCamelCase__: Tuple =[ "\u000A", # (line feed, '\n') "\r\n", # (carriage return and line feed, '\r\n') "\u000D", # (carriage return, '\r') "\r", # (carriage return, '\r') "\u000D", # (carriage return, '\r') "\u2028", # (line separator) "\u2029", # (paragraph separator) # "\u0085", # (next line) ] # The tokenization is not identical for the character "\u0085" (next line). The slow version using ftfy transforms # it into the Horizontal Ellipsis character "…" ("\u2026") while the fast version transforms it into a # space (and thus into an empty list). for unicode_seq in line_break_unicodes: lowerCamelCase__: List[str] =tokenizer_s.tokenize(UpperCAmelCase_) lowerCamelCase__: str =tokenizer_r.tokenize(UpperCAmelCase_) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : Optional[Any]) ->Optional[int]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})"""): lowerCamelCase__: List[str] ="hello" # `hello` is a token in the vocabulary of `pretrained_name` lowerCamelCase__: str =F"""{text_of_1_token} {text_of_1_token}""" lowerCamelCase__: int =self.rust_tokenizer_class.from_pretrained( UpperCAmelCase_ , use_fast=UpperCAmelCase_ , ) lowerCamelCase__: str =tokenizer_r(UpperCAmelCase_ , return_offsets_mapping=UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_) self.assertEqual(encoding.offset_mapping[0] , (0, len(UpperCAmelCase_))) self.assertEqual( encoding.offset_mapping[1] , (len(UpperCAmelCase_) + 1, len(UpperCAmelCase_) + 1 + len(UpperCAmelCase_)) , ) lowerCamelCase__: str =F""" {text}""" lowerCamelCase__: Optional[Any] =self.rust_tokenizer_class.from_pretrained( UpperCAmelCase_ , use_fast=UpperCAmelCase_ , ) lowerCamelCase__: Tuple =tokenizer_r(UpperCAmelCase_ , return_offsets_mapping=UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(UpperCAmelCase_))) self.assertEqual( encoding.offset_mapping[1] , (1 + len(UpperCAmelCase_) + 1, 1 + len(UpperCAmelCase_) + 1 + len(UpperCAmelCase_)) , ) def SCREAMING_SNAKE_CASE_ (self : Any) ->Dict: '''simple docstring''' with self.assertRaises(UpperCAmelCase_) as context: self.rust_tokenizer_class.from_pretrained("robot-test/old-clip-tokenizer") self.assertTrue( context.exception.args[0].startswith( "The `backend_tokenizer` provided does not match the expected format.")) @require_ftfy def SCREAMING_SNAKE_CASE_ (self : Any) ->List[Any]: '''simple docstring''' super().test_tokenization_python_rust_equals() def SCREAMING_SNAKE_CASE_ (self : str) ->Union[str, Any]: '''simple docstring''' pass
10
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase_ = { '''configuration_time_series_transformer''': [ '''TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TimeSeriesTransformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ '''TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TimeSeriesTransformerForPrediction''', '''TimeSeriesTransformerModel''', '''TimeSeriesTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimeSeriesTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimeSeriesTransformerForPrediction, TimeSeriesTransformerModel, TimeSeriesTransformerPreTrainedModel, ) else: import sys lowerCamelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
79
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE : Optional[int] = { "configuration_chinese_clip": [ "CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "ChineseCLIPConfig", "ChineseCLIPOnnxConfig", "ChineseCLIPTextConfig", "ChineseCLIPVisionConfig", ], "processing_chinese_clip": ["ChineseCLIPProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Any = ["ChineseCLIPFeatureExtractor"] SCREAMING_SNAKE_CASE : List[Any] = ["ChineseCLIPImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Tuple = [ "CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "ChineseCLIPModel", "ChineseCLIPPreTrainedModel", "ChineseCLIPTextModel", "ChineseCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys SCREAMING_SNAKE_CASE : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
252
from itertools import count def UpperCamelCase ( _a = 5_0 ) -> int: '''simple docstring''' lowercase_ :Dict = [1] * min_block_length for n in count(_a ): fill_count_functions.append(1 ) for block_length in range(_a , n + 1 ): for block_start in range(n - block_length ): fill_count_functions[n] += fill_count_functions[ n - block_start - block_length - 1 ] fill_count_functions[n] += 1 if fill_count_functions[n] > 1_0_0_0_0_0_0: break return n if __name__ == "__main__": print(f"{solution() = }")
252
1
from __future__ import annotations import unittest from transformers import DebertaVaConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, TFDebertaVaModel, ) class _SCREAMING_SNAKE_CASE : def __init__( self : Optional[Any] , __lowerCamelCase : Dict , __lowerCamelCase : Any=13 , __lowerCamelCase : Tuple=7 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Optional[int]=True , __lowerCamelCase : str=True , __lowerCamelCase : Any=True , __lowerCamelCase : Union[str, Any]=99 , __lowerCamelCase : Tuple=32 , __lowerCamelCase : Any=2 , __lowerCamelCase : int=4 , __lowerCamelCase : List[str]=37 , __lowerCamelCase : Optional[int]="gelu" , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Any=0.1 , __lowerCamelCase : Optional[int]=512 , __lowerCamelCase : Union[str, Any]=16 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Optional[Any]=False , __lowerCamelCase : Tuple=True , __lowerCamelCase : Any="None" , __lowerCamelCase : List[Any]=3 , __lowerCamelCase : Optional[Any]=4 , __lowerCamelCase : Any=None , ): UpperCamelCase :List[Any] = parent UpperCamelCase :Tuple = batch_size UpperCamelCase :Union[str, Any] = seq_length UpperCamelCase :Dict = is_training UpperCamelCase :Dict = use_input_mask UpperCamelCase :Union[str, Any] = use_token_type_ids UpperCamelCase :Any = use_labels UpperCamelCase :int = vocab_size UpperCamelCase :Tuple = hidden_size UpperCamelCase :int = num_hidden_layers UpperCamelCase :List[Any] = num_attention_heads UpperCamelCase :List[str] = intermediate_size UpperCamelCase :List[Any] = hidden_act UpperCamelCase :Any = hidden_dropout_prob UpperCamelCase :str = attention_probs_dropout_prob UpperCamelCase :Union[str, Any] = max_position_embeddings UpperCamelCase :Any = type_vocab_size UpperCamelCase :Tuple = type_sequence_label_size UpperCamelCase :str = initializer_range UpperCamelCase :Tuple = num_labels UpperCamelCase :int = num_choices UpperCamelCase :Optional[int] = relative_attention UpperCamelCase :List[Any] = position_biased_input UpperCamelCase :Optional[Any] = pos_att_type UpperCamelCase :Dict = scope def _A ( self : Dict ): UpperCamelCase :Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase :Optional[int] = None if self.use_input_mask: UpperCamelCase :Optional[int] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase :List[Any] = None if self.use_token_type_ids: UpperCamelCase :Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCamelCase :Dict = None UpperCamelCase :Dict = None UpperCamelCase :Any = None if self.use_labels: UpperCamelCase :str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase :int = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCamelCase :Tuple = DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , initializer_range=self.initializer_range , return_dict=__lowerCamelCase , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _A ( self : Tuple , __lowerCamelCase : str , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] ): UpperCamelCase :Optional[int] = TFDebertaVaModel(config=__lowerCamelCase ) UpperCamelCase :List[str] = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} UpperCamelCase :Optional[int] = [input_ids, input_mask] UpperCamelCase :str = model(__lowerCamelCase ) UpperCamelCase :Optional[Any] = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _A ( self : Optional[int] , __lowerCamelCase : List[str] , __lowerCamelCase : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Any , __lowerCamelCase : Dict , __lowerCamelCase : List[str] , __lowerCamelCase : Dict ): UpperCamelCase :Optional[Any] = TFDebertaVaForMaskedLM(config=__lowerCamelCase ) UpperCamelCase :Optional[int] = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } UpperCamelCase :Union[str, Any] = model(__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _A ( self : List[str] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : str , __lowerCamelCase : Any , __lowerCamelCase : Dict ): UpperCamelCase :Optional[Any] = self.num_labels UpperCamelCase :Optional[int] = TFDebertaVaForSequenceClassification(config=__lowerCamelCase ) UpperCamelCase :Optional[int] = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } UpperCamelCase :Tuple = model(__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _A ( self : Union[str, Any] , __lowerCamelCase : int , __lowerCamelCase : Tuple , __lowerCamelCase : Optional[int] , __lowerCamelCase : Tuple , __lowerCamelCase : Dict , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : List[str] ): UpperCamelCase :int = self.num_labels UpperCamelCase :Optional[Any] = TFDebertaVaForTokenClassification(config=__lowerCamelCase ) UpperCamelCase :Dict = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } UpperCamelCase :int = model(__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _A ( self : Tuple , __lowerCamelCase : List[str] , __lowerCamelCase : int , __lowerCamelCase : Optional[int] , __lowerCamelCase : str , __lowerCamelCase : Optional[int] , __lowerCamelCase : str , __lowerCamelCase : Dict ): UpperCamelCase :Optional[Any] = TFDebertaVaForQuestionAnswering(config=__lowerCamelCase ) UpperCamelCase :Optional[int] = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } UpperCamelCase :Dict = model(__lowerCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _A ( self : Any ): UpperCamelCase :List[Any] = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) :str = config_and_inputs UpperCamelCase :Union[str, Any] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class _SCREAMING_SNAKE_CASE ( _a , _a , unittest.TestCase ): snake_case__ : List[str] = ( ( TFDebertaVaModel, TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, ) if is_tf_available() else () ) snake_case__ : Optional[Any] = ( { """feature-extraction""": TFDebertaVaModel, """fill-mask""": TFDebertaVaForMaskedLM, """question-answering""": TFDebertaVaForQuestionAnswering, """text-classification""": TFDebertaVaForSequenceClassification, """token-classification""": TFDebertaVaForTokenClassification, """zero-shot""": TFDebertaVaForSequenceClassification, } if is_tf_available() else {} ) snake_case__ : Any = False snake_case__ : Union[str, Any] = False def _A ( self : Any ): UpperCamelCase :Tuple = TFDebertaVaModelTester(self ) UpperCamelCase :Optional[int] = ConfigTester(self , config_class=__lowerCamelCase , hidden_size=37 ) def _A ( self : Optional[int] ): self.config_tester.run_common_tests() def _A ( self : str ): UpperCamelCase :List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _A ( self : Optional[Any] ): UpperCamelCase :int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__lowerCamelCase ) def _A ( self : Tuple ): UpperCamelCase :Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__lowerCamelCase ) def _A ( self : Optional[int] ): UpperCamelCase :Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__lowerCamelCase ) def _A ( self : Tuple ): UpperCamelCase :int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__lowerCamelCase ) @slow def _A ( self : Tuple ): UpperCamelCase :Optional[int] = TFDebertaVaModel.from_pretrained("""kamalkraj/deberta-v2-xlarge""" ) self.assertIsNotNone(__lowerCamelCase ) @require_tf class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): @unittest.skip(reason="""Model not available yet""" ) def _A ( self : Union[str, Any] ): pass @slow def _A ( self : Optional[Any] ): UpperCamelCase :Dict = TFDebertaVaModel.from_pretrained("""kamalkraj/deberta-v2-xlarge""" ) UpperCamelCase :Any = tf.constant([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] ) UpperCamelCase :Optional[Any] = tf.constant([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) UpperCamelCase :Optional[Any] = model(__lowerCamelCase , attention_mask=__lowerCamelCase )[0] UpperCamelCase :Tuple = tf.constant( [[[0.2356, 0.1948, 0.0369], [-0.1063, 0.3586, -0.5152], [-0.6399, -0.0259, -0.2525]]] ) tf.debugging.assert_near(output[:, 1:4, 1:4] , __lowerCamelCase , atol=1E-4 )
38
from ....configuration_utils import PretrainedConfig from ....utils import logging UpperCAmelCase_ : List[str] = logging.get_logger(__name__) UpperCAmelCase_ : int = { '''CarlCochet/trajectory-transformer-halfcheetah-medium-v2''': ( '''https://huggingface.co/CarlCochet/trajectory-transformer-halfcheetah-medium-v2/resolve/main/config.json''' ), # See all TrajectoryTransformer models at https://huggingface.co/models?filter=trajectory_transformer } class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : Any = """trajectory_transformer""" snake_case__ : Optional[Any] = ["""past_key_values"""] snake_case__ : Tuple = { """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : Union[str, Any] , __lowerCamelCase : Any=100 , __lowerCamelCase : str=5 , __lowerCamelCase : str=1 , __lowerCamelCase : Optional[int]=1 , __lowerCamelCase : int=249 , __lowerCamelCase : str=6 , __lowerCamelCase : Dict=17 , __lowerCamelCase : Optional[Any]=25 , __lowerCamelCase : List[str]=4 , __lowerCamelCase : str=4 , __lowerCamelCase : Tuple=128 , __lowerCamelCase : Dict=0.1 , __lowerCamelCase : str=0.1 , __lowerCamelCase : Any=0.1 , __lowerCamelCase : int=0.0006 , __lowerCamelCase : List[str]=512 , __lowerCamelCase : str=0.02 , __lowerCamelCase : Any=1E-12 , __lowerCamelCase : int=1 , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : Tuple=1 , __lowerCamelCase : int=50_256 , __lowerCamelCase : Union[str, Any]=50_256 , **__lowerCamelCase : Dict , ): UpperCamelCase :Dict = vocab_size UpperCamelCase :int = action_weight UpperCamelCase :Tuple = reward_weight UpperCamelCase :str = value_weight UpperCamelCase :Tuple = max_position_embeddings UpperCamelCase :Tuple = block_size UpperCamelCase :Optional[int] = action_dim UpperCamelCase :int = observation_dim UpperCamelCase :List[str] = transition_dim UpperCamelCase :List[Any] = learning_rate UpperCamelCase :Optional[Any] = n_layer UpperCamelCase :Any = n_head UpperCamelCase :List[str] = n_embd UpperCamelCase :Any = embd_pdrop UpperCamelCase :str = attn_pdrop UpperCamelCase :Union[str, Any] = resid_pdrop UpperCamelCase :Optional[Any] = initializer_range UpperCamelCase :List[Any] = layer_norm_eps UpperCamelCase :Optional[int] = kaiming_initializer_range UpperCamelCase :Tuple = use_cache super().__init__(pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , **__lowerCamelCase )
38
1
"""simple docstring""" import random class _UpperCAmelCase : '''simple docstring''' @staticmethod def __lowerCAmelCase ( A ) -> tuple[list[int], list[int]]: _UpperCAmelCase : Dict = [ord(A ) for i in text] _UpperCAmelCase : str = [] _UpperCAmelCase : Tuple = [] for i in plain: _UpperCAmelCase : List[str] = random.randint(1 , 3_0_0 ) _UpperCAmelCase : str = (i + k) * k cipher.append(A ) key.append(A ) return cipher, key @staticmethod def __lowerCAmelCase ( A , A ) -> str: _UpperCAmelCase : str = [] for i in range(len(A ) ): _UpperCAmelCase : str = int((cipher[i] - (key[i]) ** 2) / key[i] ) plain.append(chr(A ) ) return "".join(A ) if __name__ == "__main__": _lowerCAmelCase,_lowerCAmelCase :List[str] = Onepad().encrypt('Hello') print(c, k) print(Onepad().decrypt(c, k))
68
"""simple docstring""" import os import time import warnings from dataclasses import dataclass, field from enum import Enum from typing import List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import logging from ..processors.glue import glue_convert_examples_to_features, glue_output_modes, glue_processors from ..processors.utils import InputFeatures _lowerCAmelCase :str = logging.get_logger(__name__) @dataclass class _UpperCAmelCase : '''simple docstring''' a__ =field(metadata={'''help''': '''The name of the task to train on: ''' + ''', '''.join(glue_processors.keys() )} ) a__ =field( metadata={'''help''': '''The input data dir. Should contain the .tsv files (or other data files) for the task.'''} ) a__ =field( default=1_2_8 ,metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } ,) a__ =field( default=a ,metadata={'''help''': '''Overwrite the cached training and evaluation sets'''} ) def __lowerCAmelCase ( self ) -> int: _UpperCAmelCase : str = self.task_name.lower() class _UpperCAmelCase ( a ): '''simple docstring''' a__ ='''train''' a__ ='''dev''' a__ ='''test''' class _UpperCAmelCase ( a ): '''simple docstring''' a__ =42 a__ =42 a__ =42 def __init__( self , A , A , A = None , A = Split.train , A = None , ) -> Dict: warnings.warn( '''This dataset will be removed from the library soon, preprocessing should be handled with the 🤗 Datasets ''' '''library. You can have a look at this example script for pointers: ''' '''https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py''' , A , ) _UpperCAmelCase : Dict = args _UpperCAmelCase : int = glue_processors[args.task_name]() _UpperCAmelCase : Any = glue_output_modes[args.task_name] if isinstance(A , A ): try: _UpperCAmelCase : int = Split[mode] except KeyError: raise KeyError('''mode is not a valid split name''' ) # Load data features from cache or dataset file _UpperCAmelCase : Optional[int] = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f'cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{args.task_name}' , ) _UpperCAmelCase : List[Any] = self.processor.get_labels() if args.task_name in ["mnli", "mnli-mm"] and tokenizer.__class__.__name__ in ( "RobertaTokenizer", "RobertaTokenizerFast", "XLMRobertaTokenizer", "BartTokenizer", "BartTokenizerFast", ): # HACK(label indices are swapped in RoBERTa pretrained model) _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = label_list[2], label_list[1] _UpperCAmelCase : Optional[Any] = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. _UpperCAmelCase : List[Any] = cached_features_file + '''.lock''' with FileLock(A ): if os.path.exists(A ) and not args.overwrite_cache: _UpperCAmelCase : str = time.time() _UpperCAmelCase : Dict = torch.load(A ) logger.info( f'Loading features from cached file {cached_features_file} [took %.3f s]' , time.time() - start ) else: logger.info(f'Creating features from dataset file at {args.data_dir}' ) if mode == Split.dev: _UpperCAmelCase : Union[str, Any] = self.processor.get_dev_examples(args.data_dir ) elif mode == Split.test: _UpperCAmelCase : List[Any] = self.processor.get_test_examples(args.data_dir ) else: _UpperCAmelCase : Optional[Any] = self.processor.get_train_examples(args.data_dir ) if limit_length is not None: _UpperCAmelCase : Tuple = examples[:limit_length] _UpperCAmelCase : str = glue_convert_examples_to_features( A , A , max_length=args.max_seq_length , label_list=A , output_mode=self.output_mode , ) _UpperCAmelCase : Optional[int] = time.time() torch.save(self.features , A ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f'Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]' ) def __len__( self ) -> Union[str, Any]: return len(self.features ) def __getitem__( self , A ) -> InputFeatures: return self.features[i] def __lowerCAmelCase ( self ) -> List[Any]: return self.label_list
68
1
'''simple docstring''' from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time _lowerCAmelCase = Lock() def _SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ): """simple docstring""" global process_lock # we perform n swaps since after n swaps we know we are sorted # we *could* stop early if we are sorted already, but it takes as long to # find out we are sorted as it does to sort the list with this algorithm for i in range(0 , 10 ): if (i + position) % 2 == 0 and r_send is not None: # send your value to your right neighbor process_lock.acquire() r_send[1].send(UpperCamelCase ) process_lock.release() # receive your right neighbor's value process_lock.acquire() lowerCAmelCase__ : Any = rr_cv[0].recv() process_lock.release() # take the lower value since you are on the left lowerCAmelCase__ : str = min(UpperCamelCase , UpperCamelCase ) elif (i + position) % 2 != 0 and l_send is not None: # send your value to your left neighbor process_lock.acquire() l_send[1].send(UpperCamelCase ) process_lock.release() # receive your left neighbor's value process_lock.acquire() lowerCAmelCase__ : Tuple = lr_cv[0].recv() process_lock.release() # take the higher value since you are on the right lowerCAmelCase__ : Tuple = max(UpperCamelCase , UpperCamelCase ) # after all swaps are performed, send the values back to main result_pipe[1].send(UpperCamelCase ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase ): """simple docstring""" lowerCAmelCase__ : List[Any] = [] lowerCAmelCase__ : str = [] # initialize the list of pipes where the values will be retrieved for _ in arr: result_pipe.append(Pipe() ) # creates the processes # the first and last process only have one neighbor so they are made outside # of the loop lowerCAmelCase__ : List[str] = Pipe() lowerCAmelCase__ : Dict = Pipe() process_array_.append( Process( target=UpperCamelCase , args=(0, arr[0], None, temp_rs, None, temp_rr, result_pipe[0]) , ) ) lowerCAmelCase__ : Union[str, Any] = temp_rs lowerCAmelCase__ : Dict = temp_rr for i in range(1 , len(UpperCamelCase ) - 1 ): lowerCAmelCase__ : List[Any] = Pipe() lowerCAmelCase__ : Dict = Pipe() process_array_.append( Process( target=UpperCamelCase , args=(i, arr[i], temp_ls, temp_rs, temp_lr, temp_rr, result_pipe[i]) , ) ) lowerCAmelCase__ : Dict = temp_rs lowerCAmelCase__ : Optional[Any] = temp_rr process_array_.append( Process( target=UpperCamelCase , args=( len(UpperCamelCase ) - 1, arr[len(UpperCamelCase ) - 1], temp_ls, None, temp_lr, None, result_pipe[len(UpperCamelCase ) - 1], ) , ) ) # start the processes for p in process_array_: p.start() # wait for the processes to end and write their values to the list for p in range(0 , len(UpperCamelCase ) ): lowerCAmelCase__ : List[Any] = result_pipe[p][0].recv() process_array_[p].join() return arr def _SCREAMING_SNAKE_CASE ( ): """simple docstring""" lowerCAmelCase__ : List[Any] = list(range(10 , 0 , -1 ) ) print("""Initial List""" ) print(*UpperCamelCase ) lowerCAmelCase__ : str = odd_even_transposition(UpperCamelCase ) print("""Sorted List\n""" ) print(*UpperCamelCase ) if __name__ == "__main__": main()
37
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase : Optional[Any] = logging.get_logger(__name__) UpperCAmelCase : Optional[int] = { """google/canine-s""": """https://huggingface.co/google/canine-s/resolve/main/config.json""", # See all CANINE models at https://huggingface.co/models?filter=canine } class __lowerCAmelCase ( UpperCamelCase__): _lowercase : List[Any] = """canine""" def __init__( self , lowerCAmelCase__=7_6_8 , lowerCAmelCase__=1_2 , lowerCAmelCase__=1_2 , lowerCAmelCase__=3_0_7_2 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=1_6_3_8_4 , lowerCAmelCase__=1_6 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-12 , lowerCAmelCase__=0 , lowerCAmelCase__=0XE0_00 , lowerCAmelCase__=0XE0_01 , lowerCAmelCase__=4 , lowerCAmelCase__=4 , lowerCAmelCase__=8 , lowerCAmelCase__=1_6_3_8_4 , lowerCAmelCase__=1_2_8 , **lowerCAmelCase__ , ) -> Dict: '''simple docstring''' super().__init__(pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) a__ : Optional[int] =max_position_embeddings a__ : str =hidden_size a__ : Optional[Any] =num_hidden_layers a__ : Tuple =num_attention_heads a__ : Optional[Any] =intermediate_size a__ : Optional[int] =hidden_act a__ : List[Any] =hidden_dropout_prob a__ : Union[str, Any] =attention_probs_dropout_prob a__ : Optional[Any] =initializer_range a__ : Union[str, Any] =type_vocab_size a__ : Optional[int] =layer_norm_eps # Character config: a__ : int =downsampling_rate a__ : Optional[Any] =upsampling_kernel_size a__ : Union[str, Any] =num_hash_functions a__ : Any =num_hash_buckets a__ : int =local_transformer_stride
95
0
import time import warnings from abc import ABC from copy import deepcopy from typing import Optional import torch from ..utils import add_start_docstrings, logging a_ = logging.get_logger(__name__) a_ = r'\n Args:\n input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`):\n Indices of input sequence tokens in the vocabulary.\n\n Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and\n [`PreTrainedTokenizer.__call__`] for details.\n\n [What are input IDs?](../glossary#input-ids)\n scores (`torch.FloatTensor` of shape `(batch_size, config.vocab_size)`):\n Prediction scores of a language modeling head. These can be scores for each vocabulary token before SoftMax\n or scores for each vocabulary token after SoftMax.\n kwargs (`Dict[str, Any]`, *optional*):\n Additional stopping criteria specific kwargs.\n\n Return:\n `bool`. `False` indicates we should continue, `True` indicates we should stop.\n\n' class _lowercase ( snake_case_ ): @add_start_docstrings(snake_case ) def __call__( self : str , snake_case : torch.LongTensor , snake_case : torch.FloatTensor , **snake_case : Tuple ) -> bool: """simple docstring""" raise NotImplementedError('StoppingCriteria needs to be subclassed' ) class _lowercase ( snake_case_ ): def __init__( self : str , snake_case : int , snake_case : Optional[int] = None ) -> List[str]: """simple docstring""" UpperCamelCase_ : List[Any] = max_length UpperCamelCase_ : List[Any] = max_position_embeddings @add_start_docstrings(snake_case ) def __call__( self : Dict , snake_case : torch.LongTensor , snake_case : torch.FloatTensor , **snake_case : Optional[Any] ) -> bool: """simple docstring""" UpperCamelCase_ : List[Any] = input_ids.shape[-1] UpperCamelCase_ : str = cur_len >= self.max_length if self.max_position_embeddings is not None and not is_done and cur_len >= self.max_position_embeddings: logger.warning_once( 'This is a friendly reminder - the current text generation call will exceed the model\'s predefined ' f"maximum length ({self.max_position_embeddings}). Depending on the model, you may observe " 'exceptions, performance degradation, or nothing at all.' ) return is_done class _lowercase ( snake_case_ ): def __init__( self : int , snake_case : int , snake_case : int ) -> Any: """simple docstring""" warnings.warn( 'The class `MaxNewTokensCriteria` is deprecated. ' f"Please use `MaxLengthCriteria(max_length={start_length + max_new_tokens})` " 'with `max_length = start_length + max_new_tokens` instead.' , snake_case , ) UpperCamelCase_ : Dict = start_length UpperCamelCase_ : Dict = max_new_tokens UpperCamelCase_ : int = start_length + max_new_tokens @add_start_docstrings(snake_case ) def __call__( self : List[str] , snake_case : torch.LongTensor , snake_case : torch.FloatTensor , **snake_case : List[str] ) -> bool: """simple docstring""" return input_ids.shape[-1] >= self.max_length class _lowercase ( snake_case_ ): def __init__( self : Optional[int] , snake_case : float , snake_case : Optional[float] = None ) -> Dict: """simple docstring""" UpperCamelCase_ : Tuple = max_time UpperCamelCase_ : Any = time.time() if initial_timestamp is None else initial_timestamp @add_start_docstrings(snake_case ) def __call__( self : Optional[Any] , snake_case : torch.LongTensor , snake_case : torch.FloatTensor , **snake_case : Union[str, Any] ) -> bool: """simple docstring""" return time.time() - self.initial_timestamp > self.max_time class _lowercase ( snake_case_ ): @add_start_docstrings(snake_case ) def __call__( self : Union[str, Any] , snake_case : torch.LongTensor , snake_case : torch.FloatTensor , **snake_case : List[str] ) -> bool: """simple docstring""" return any(criteria(snake_case , snake_case ) for criteria in self ) @property def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" for stopping_criterium in self: if isinstance(snake_case , snake_case ): return stopping_criterium.max_length elif isinstance(snake_case , snake_case ): return stopping_criterium.max_length return None def __lowercase ( lowerCamelCase : StoppingCriteriaList , lowerCamelCase : int ): UpperCamelCase_ : Optional[Any] = stopping_criteria.max_length UpperCamelCase_ : Dict = deepcopy(lowerCamelCase ) if stopping_max_length is not None and stopping_max_length != max_length: warnings.warn('You set different `max_length` for stopping criteria and `max_length` parameter' , lowerCamelCase ) elif stopping_max_length is None: new_stopping_criteria.append(MaxLengthCriteria(max_length=lowerCamelCase ) ) return new_stopping_criteria
50
a_ = '\n# Transformers 설치 방법\n! pip install transformers datasets\n# 마지막 릴리스 대신 소스에서 설치하려면, 위 명령을 주석으로 바꾸고 아래 명령을 해제하세요.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' a_ = [{'type': 'code', 'content': INSTALL_CONTENT}] a_ = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
50
1
"""simple docstring""" import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def __lowerCAmelCase ( lowercase : str ) -> Any: """simple docstring""" monkeypatch.setattr("datasets.utils.deprecation_utils._emitted_deprecation_warnings" , set() ) @pytest.fixture def __lowerCAmelCase ( lowercase : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" class _lowerCAmelCase : def __init__( self , UpperCamelCase__ ) -> Any: '''simple docstring''' snake_case : Dict = metric_id class _lowerCAmelCase : __UpperCAmelCase : List[str] = [MetricMock(snake_case_ ) for metric_id in ['''accuracy''', '''mse''', '''precision''', '''codeparrot/apps_metric''']] def lowerCamelCase ( self ) -> Dict: '''simple docstring''' return self._metrics monkeypatch.setattr("datasets.inspect.huggingface_hub" , HfhMock() ) @pytest.mark.parametrize( "func, args" , [(load_metric, ("metrics/mse",)), (list_metrics, ()), (inspect_metric, ("metrics/mse", "tmp_path"))] ) def __lowerCAmelCase ( lowercase : Dict , lowercase : List[str] , lowercase : Tuple , lowercase : str , lowercase : List[str] ) -> str: """simple docstring""" if "tmp_path" in args: snake_case : Any = tuple(arg if arg != "tmp_path" else tmp_path for arg in args ) with pytest.warns(lowercase , match="https://huggingface.co/docs/evaluate" ): func(*lowercase )
203
"""simple docstring""" import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class _lowerCAmelCase ( unittest.TestCase ): __UpperCAmelCase : Union[str, Any] = JukeboxTokenizer __UpperCAmelCase : Union[str, Any] = { '''artist''': '''Zac Brown Band''', '''genres''': '''Country''', '''lyrics''': '''I met a traveller from an antique land, Who said "Two vast and trunkless legs of stone Stand in the desert. . . . Near them, on the sand, Half sunk a shattered visage lies, whose frown, And wrinkled lip, and sneer of cold command, Tell that its sculptor well those passions read Which yet survive, stamped on these lifeless things, The hand that mocked them, and the heart that fed; And on the pedestal, these words appear: My name is Ozymandias, King of Kings; Look on my Works, ye Mighty, and despair! Nothing beside remains. Round the decay Of that colossal Wreck, boundless and bare The lone and level sands stretch far away ''', } @require_torch def lowerCamelCase ( self ) -> int: '''simple docstring''' import torch snake_case : Any = JukeboxTokenizer.from_pretrained("openai/jukebox-1b-lyrics" ) snake_case : Optional[Any] = tokenizer(**self.metas )["input_ids"] # fmt: off snake_case : Optional[int] = [ torch.tensor([[ 0, 0, 0, 7169, 507, 9, 76, 39, 31, 46, 76, 27, 76, 46, 44, 27, 48, 31, 38, 38, 31, 44, 76, 32, 44, 41, 39, 76, 27, 40, 76, 27, 40, 46, 35, 43, 47, 31, 76, 38, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 23, 34, 41, 76, 45, 27, 35, 30, 76, 71, 20, 49, 41, 76, 48, 27, 45, 46, 76, 27, 40, 30, 76, 46, 44, 47, 40, 37, 38, 31, 45, 45, 76, 38, 31, 33, 45, 76, 41, 32, 76, 45, 46, 41, 40, 31, 78, 76, 76, 76, 76, 76, 76, 76, 76, 19, 46, 27, 40, 30, 76, 35, 40, 76, 46, 34, 31, 76, 30, 31, 45, 31, 44, 46, 63, 76, 63, 76, 63, 76, 63, 76, 14, 31, 27, 44, 76, 46, 34, 31, 39, 64, 76, 41, 40, 76, 46, 34, 31, 76, 45, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 8, 27, 38, 32, 76, 45, 47, 40, 37, 76, 27, 76, 45, 34, 27, 46, 46, 31, 44, 31, 30, 76, 48, 35, 45, 27, 33, 31, 76, 38, 35, 31, 45, 64, 76, 49, 34, 41, 45, 31, 76, 32, 44, 41, 49, 40, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 1, 40, 30, 76, 49, 44, 35, 40, 37, 38, 31, 30, 76, 38, 35, 42, 64, 76, 27, 40, 30, 76, 45, 40, 31, 31, 44, 76, 41, 32, 76, 29, 41, 38, 30, 76, 29, 41, 39, 39, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 31, 38, 38, 76, 46, 34, 27, 46, 76, 35, 46, 45, 76, 45, 29, 47, 38, 42, 46, 41, 44, 76, 49, 31, 38, 38, 76, 46, 34, 41, 45, 31, 76, 42, 27, 45, 45, 35, 41, 40, 45, 76, 44, 31, 27, 30, 78, 76, 76, 76, 76, 76, 76, 76, 76, 23, 34, 35, 29, 34, 76, 51, 31, 46, 76, 45, 47, 44, 48, 35, 48, 31, 64, 76, 45, 46, 27, 39, 42, 31, 30, 76, 41, 40, 76, 46, 34, 31, 45, 31, 76, 38, 35, 32, 31, 38, 31, 45, 45, 76, 46, 34, 35, 40, 33, 45, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 34, 31, 76, 34, 27, 40, 30, 76, 46, 34, 27, 46, 76, 39, 41, 29, 37, 31, 30, 76, 46, 34, 31, 39, 64, 76, 27, 40, 30, 76, 46, 34, 31, 76, 34, 31, 27, 44, 46, 76, 46, 34, 27, 46, 76, 32, 31, 30, 66, 78, 76, 76, 76, 76, 76, 76, 76, 76, 1, 40, 30, 76, 41, 40, 76, 46, 34, 31, 76, 42, 31, 30, 31, 45, 46, 27, 38, 64, 76, 46, 34, 31, 45, 31, 76, 49, 41, 44, 30, 45, 76, 27, 42, 42, 31, 27, 44, 65, 78, 76, 76, 76, 76, 76, 76, 76, 76, 13, 51, 76, 40, 27, 39, 31, 76, 35, 45, 76, 15, 52, 51, 39, 27, 40, 30, 35, 27, 45, 64, 76, 11, 35, 40, 33, 76, 41, 32, 76, 11, 35, 40, 33, 45, 66, 78, 76, 76, 76, 76, 76, 76, 76, 76, 12, 41, 41, 37, 76, 41, 40, 76, 39, 51, 76, 23, 41, 44, 37, 45, 64, 76, 51, 31, 76, 13, 35, 33, 34, 46, 51, 64, 76, 27, 40, 30, 76, 30, 31, 45, 42, 27, 35, 44, 67, 78, 76, 76, 76, 76, 76, 76, 76, 76, 14, 41, 46, 34, 35, 40, 33, 76, 28, 31, 45, 35, 30, 31, 76, 44, 31, 39, 27, 35, 40, 45, 63, 76, 18, 41, 47, 40, 30, 76, 46, 34, 31, 76, 30, 31, 29, 27, 51, 78, 76, 76, 76, 76, 76, 76, 76, 76, 15, 32, 76, 46, 34, 27, 46, 76, 29, 41, 38, 41, 45, 45, 27, 38, 76, 23, 44, 31, 29, 37, 64, 76, 28, 41, 47, 40, 30, 38, 31, 45, 45, 76, 27, 40, 30, 76, 28, 27, 44, 31, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 34, 31, 76, 38, 41, 40, 31, 76, 27, 40, 30, 76, 38, 31, 48, 31, 38, 76, 45, 27, 40, 30, 45, 76, 45, 46, 44, 31, 46, 29, 34, 76, 32, 27, 44, 76, 27, 49, 27, 51, 78, 76, 76, 76, 76, 76, 76, 76, 76]] ), torch.tensor([[0, 0, 0, 1069, 11]] ), torch.tensor([[0, 0, 0, 1069, 11]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) ) @require_torch def lowerCamelCase ( self ) -> Any: '''simple docstring''' import torch snake_case : Tuple = JukeboxTokenizer.from_pretrained("openai/jukebox-5b-lyrics" ) snake_case : Optional[Any] = tokenizer(**self.metas )["input_ids"] # fmt: off snake_case : List[Any] = [ torch.tensor([[ 0, 0, 0, 1069, 11, -1, -1, -1, -1, 9, 77, 39, 31, 46, 77, 27, 77, 46, 44, 27, 48, 31, 38, 38, 31, 44, 77, 32, 44, 41, 39, 77, 27, 40, 77, 27, 40, 46, 35, 43, 47, 31, 77, 38, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 23, 34, 41, 77, 45, 27, 35, 30, 77, 72, 20, 49, 41, 77, 48, 27, 45, 46, 77, 27, 40, 30, 77, 46, 44, 47, 40, 37, 38, 31, 45, 45, 77, 38, 31, 33, 45, 77, 41, 32, 77, 45, 46, 41, 40, 31, 79, 77, 77, 77, 77, 77, 77, 77, 77, 19, 46, 27, 40, 30, 77, 35, 40, 77, 46, 34, 31, 77, 30, 31, 45, 31, 44, 46, 63, 77, 63, 77, 63, 77, 63, 77, 14, 31, 27, 44, 77, 46, 34, 31, 39, 64, 77, 41, 40, 77, 46, 34, 31, 77, 45, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 8, 27, 38, 32, 77, 45, 47, 40, 37, 77, 27, 77, 45, 34, 27, 46, 46, 31, 44, 31, 30, 77, 48, 35, 45, 27, 33, 31, 77, 38, 35, 31, 45, 64, 77, 49, 34, 41, 45, 31, 77, 32, 44, 41, 49, 40, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 1, 40, 30, 77, 49, 44, 35, 40, 37, 38, 31, 30, 77, 38, 35, 42, 64, 77, 27, 40, 30, 77, 45, 40, 31, 31, 44, 77, 41, 32, 77, 29, 41, 38, 30, 77, 29, 41, 39, 39, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 31, 38, 38, 77, 46, 34, 27, 46, 77, 35, 46, 45, 77, 45, 29, 47, 38, 42, 46, 41, 44, 77, 49, 31, 38, 38, 77, 46, 34, 41, 45, 31, 77, 42, 27, 45, 45, 35, 41, 40, 45, 77, 44, 31, 27, 30, 79, 77, 77, 77, 77, 77, 77, 77, 77, 23, 34, 35, 29, 34, 77, 51, 31, 46, 77, 45, 47, 44, 48, 35, 48, 31, 64, 77, 45, 46, 27, 39, 42, 31, 30, 77, 41, 40, 77, 46, 34, 31, 45, 31, 77, 38, 35, 32, 31, 38, 31, 45, 45, 77, 46, 34, 35, 40, 33, 45, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 34, 31, 77, 34, 27, 40, 30, 77, 46, 34, 27, 46, 77, 39, 41, 29, 37, 31, 30, 77, 46, 34, 31, 39, 64, 77, 27, 40, 30, 77, 46, 34, 31, 77, 34, 31, 27, 44, 46, 77, 46, 34, 27, 46, 77, 32, 31, 30, 66, 79, 77, 77, 77, 77, 77, 77, 77, 77, 1, 40, 30, 77, 41, 40, 77, 46, 34, 31, 77, 42, 31, 30, 31, 45, 46, 27, 38, 64, 77, 46, 34, 31, 45, 31, 77, 49, 41, 44, 30, 45, 77, 27, 42, 42, 31, 27, 44, 65, 79, 77, 77, 77, 77, 77, 77, 77, 77, 13, 51, 77, 40, 27, 39, 31, 77, 35, 45, 77, 15, 52, 51, 39, 27, 40, 30, 35, 27, 45, 64, 77, 11, 35, 40, 33, 77, 41, 32, 77, 11, 35, 40, 33, 45, 66, 79, 77, 77, 77, 77, 77, 77, 77, 77, 12, 41, 41, 37, 77, 41, 40, 77, 39, 51, 77, 23, 41, 44, 37, 45, 64, 77, 51, 31, 77, 13, 35, 33, 34, 46, 51, 64, 77, 27, 40, 30, 77, 30, 31, 45, 42, 27, 35, 44, 67, 79, 77, 77, 77, 77, 77, 77, 77, 77, 14, 41, 46, 34, 35, 40, 33, 77, 28, 31, 45, 35, 30, 31, 77, 44, 31, 39, 27, 35, 40, 45, 63, 77, 18, 41, 47, 40, 30, 77, 46, 34, 31, 77, 30, 31, 29, 27, 51, 79, 77, 77, 77, 77, 77, 77, 77, 77, 15, 32, 77, 46, 34, 27, 46, 77, 29, 41, 38, 41, 45, 45, 27, 38, 77, 23, 44, 31, 29, 37, 64, 77, 28, 41, 47, 40, 30, 38, 31, 45, 45, 77, 27, 40, 30, 77, 28, 27, 44, 31, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 34, 31, 77, 38, 41, 40, 31, 77, 27, 40, 30, 77, 38, 31, 48, 31, 38, 77, 45, 27, 40, 30, 45, 77, 45, 46, 44, 31, 46, 29, 34, 77, 32, 27, 44, 77, 27, 49, 27, 51, 79, 77, 77, 77, 77, 77, 77, 77, 77]] ), torch.tensor([[0, 0, 0, 1069, 11, -1, -1, -1, -1]] ), torch.tensor([[0, 0, 0, 1069, 11, -1, -1, -1, -1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) )
203
1
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionAttendAndExcitePipeline, UNetaDConditionModel, ) from diffusers.utils import load_numpy, skip_mps, slow from diffusers.utils.testing_utils import require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin a_ : Tuple = False @skip_mps class __UpperCamelCase ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): lowercase : Union[str, Any] =StableDiffusionAttendAndExcitePipeline lowercase : Tuple =False lowercase : Optional[Any] =TEXT_TO_IMAGE_PARAMS lowercase : Dict =TEXT_TO_IMAGE_BATCH_PARAMS.union({'token_indices'} ) lowercase : Optional[Any] =TEXT_TO_IMAGE_IMAGE_PARAMS lowercase : Optional[Any] =TEXT_TO_IMAGE_IMAGE_PARAMS @classmethod def lowercase__ ( cls ): """simple docstring""" super().setUpClass() torch.use_deterministic_algorithms(lowerCAmelCase ) @classmethod def lowercase__ ( cls ): """simple docstring""" super().tearDownClass() torch.use_deterministic_algorithms(lowerCAmelCase ) def lowercase__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowerCamelCase_ =UNetaDConditionModel( block_out_channels=(32, 64), layers_per_block=1, sample_size=32, in_channels=4, out_channels=4, down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D'''), up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D'''), cross_attention_dim=32, attention_head_dim=(2, 4), use_linear_projection=lowerCAmelCase, ) lowerCamelCase_ =DDIMScheduler( beta_start=0.0_0_0_8_5, beta_end=0.0_1_2, beta_schedule='''scaled_linear''', clip_sample=lowerCAmelCase, set_alpha_to_one=lowerCAmelCase, ) torch.manual_seed(0 ) lowerCamelCase_ =AutoencoderKL( block_out_channels=[32, 64], in_channels=3, out_channels=3, down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''], up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''], latent_channels=4, sample_size=128, ) torch.manual_seed(0 ) lowerCamelCase_ =CLIPTextConfig( bos_token_id=0, eos_token_id=2, hidden_size=32, intermediate_size=37, layer_norm_eps=1e-05, num_attention_heads=4, num_hidden_layers=5, pad_token_id=1, vocab_size=1_000, hidden_act='''gelu''', projection_dim=512, ) lowerCamelCase_ =CLIPTextModel(lowerCAmelCase ) lowerCamelCase_ =CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) lowerCamelCase_ ={ '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase=0 ): """simple docstring""" if str(lowerCAmelCase ).startswith('''mps''' ): lowerCamelCase_ =torch.manual_seed(lowerCAmelCase ) else: lowerCamelCase_ =torch.Generator(device=lowerCAmelCase ).manual_seed(lowerCAmelCase ) lowerCamelCase_ =lowerCamelCase_ ={ '''prompt''': '''a cat and a frog''', '''token_indices''': [2, 5], '''generator''': generator, '''num_inference_steps''': 1, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', '''max_iter_to_alter''': 2, '''thresholds''': {0: 0.7}, } return inputs def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ ='''cpu''' lowerCamelCase_ =self.get_dummy_components() lowerCamelCase_ =self.pipeline_class(**lowerCAmelCase ) pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) lowerCamelCase_ =self.get_dummy_inputs(lowerCAmelCase ) lowerCamelCase_ =pipe(**lowerCAmelCase ).images lowerCamelCase_ =image[0, -3:, -3:, -1] self.assertEqual(image.shape, (1, 64, 64, 3) ) lowerCamelCase_ =np.array( [0.6_3_9_0_5_3_6_4, 0.6_2_8_9_7_3_0_7, 0.4_8_5_9_9_0_1_7, 0.5_1_3_3_6_2_4, 0.5_5_5_0_0_4_8, 0.4_5_7_6_9_5_1_6, 0.5_0_3_2_6_9_7_3, 0.5_0_2_3_1_3_9, 0.4_5_3_8_4_4_9_6] ) lowerCamelCase_ =np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCAmelCase, 1e-3 ) def lowercase__ ( self ): """simple docstring""" super().test_cpu_offload_forward_pass(expected_max_diff=5e-4 ) def lowercase__ ( self ): """simple docstring""" self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def lowercase__ ( self ): """simple docstring""" self._test_inference_batch_single_identical(batch_size=2, expected_max_diff=7e-4 ) def lowercase__ ( self ): """simple docstring""" super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def lowercase__ ( self ): """simple docstring""" super().test_pt_np_pil_outputs_equivalent(expected_max_diff=5e-4 ) def lowercase__ ( self ): """simple docstring""" super().test_save_load_local(expected_max_difference=5e-4 ) def lowercase__ ( self ): """simple docstring""" super().test_save_load_optional_components(expected_max_difference=4e-4 ) @require_torch_gpu @slow class __UpperCamelCase ( unittest.TestCase ): @classmethod def lowercase__ ( cls ): """simple docstring""" super().setUpClass() torch.use_deterministic_algorithms(lowerCAmelCase ) @classmethod def lowercase__ ( cls ): """simple docstring""" super().tearDownClass() torch.use_deterministic_algorithms(lowerCAmelCase ) def lowercase__ ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase__ ( self ): """simple docstring""" lowerCamelCase_ =torch.manual_seed(51 ) lowerCamelCase_ =StableDiffusionAttendAndExcitePipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''', safety_checker=lowerCAmelCase, torch_dtype=torch.floataa ) pipe.to('''cuda''' ) lowerCamelCase_ ='''a painting of an elephant with glasses''' lowerCamelCase_ =[5, 7] lowerCamelCase_ =pipe( prompt=lowerCAmelCase, token_indices=lowerCAmelCase, guidance_scale=7.5, generator=lowerCAmelCase, num_inference_steps=5, max_iter_to_alter=5, output_type='''numpy''', ).images[0] lowerCamelCase_ =load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/attend-and-excite/elephant_glasses.npy''' ) assert np.abs((expected_image - image).max() ) < 5e-1
362
'''simple docstring''' import importlib import os import sys # This is required to make the module import works (when the python process is running from the root of the repo) sys.path.append(""".""") def a_ ( __snake_case : Any ) -> Tuple: """simple docstring""" lowerCamelCase_ =test_file.split(os.path.sep ) if components[0:2] != ["tests", "models"]: raise ValueError( '''`test_file` should start with `tests/models/` (with `/` being the OS specific path separator). Got ''' F'''{test_file} instead.''' ) lowerCamelCase_ =components[-1] if not test_fn.endswith('''py''' ): raise ValueError(F'''`test_file` should be a python file. Got {test_fn} instead.''' ) if not test_fn.startswith('''test_modeling_''' ): raise ValueError( F'''`test_file` should point to a file name of the form `test_modeling_*.py`. Got {test_fn} instead.''' ) lowerCamelCase_ =components[:-1] + [test_fn.replace('''.py''' , '''''' )] lowerCamelCase_ ='''.'''.join(__snake_case ) return test_module_path def a_ ( __snake_case : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ =get_module_path(__snake_case ) lowerCamelCase_ =importlib.import_module(__snake_case ) return test_module def a_ ( __snake_case : Dict ) -> Tuple: """simple docstring""" lowerCamelCase_ =[] lowerCamelCase_ =get_test_module(__snake_case ) for attr in dir(__snake_case ): if attr.endswith('''ModelTester''' ): tester_classes.append(getattr(__snake_case , __snake_case ) ) # sort with class names return sorted(__snake_case , key=lambda __snake_case : x.__name__ ) def a_ ( __snake_case : List[Any] ) -> Optional[int]: """simple docstring""" lowerCamelCase_ =[] lowerCamelCase_ =get_test_module(__snake_case ) for attr in dir(__snake_case ): lowerCamelCase_ =getattr(__snake_case , __snake_case ) # (TF/Flax)ModelTesterMixin is also an attribute in specific model test module. Let's exclude them by checking # `all_model_classes` is not empty (which also excludes other special classes). lowerCamelCase_ =getattr(__snake_case , '''all_model_classes''' , [] ) if len(__snake_case ) > 0: test_classes.append(__snake_case ) # sort with class names return sorted(__snake_case , key=lambda __snake_case : x.__name__ ) def a_ ( __snake_case : List[Any] ) -> Optional[int]: """simple docstring""" lowerCamelCase_ =get_test_classes(__snake_case ) lowerCamelCase_ =set() for test_class in test_classes: model_classes.update(test_class.all_model_classes ) # sort with class names return sorted(__snake_case , key=lambda __snake_case : x.__name__ ) def a_ ( __snake_case : str ) -> str: """simple docstring""" lowerCamelCase_ =test_class() if hasattr(__snake_case , '''setUp''' ): test.setUp() lowerCamelCase_ =None if hasattr(__snake_case , '''model_tester''' ): # `(TF/Flax)ModelTesterMixin` has this attribute default to `None`. Let's skip this case. if test.model_tester is not None: lowerCamelCase_ =test.model_tester.__class__ return model_tester def a_ ( __snake_case : Dict , __snake_case : List[str] ) -> List[Any]: """simple docstring""" lowerCamelCase_ =get_test_classes(__snake_case ) lowerCamelCase_ =[] for test_class in test_classes: if model_class in test_class.all_model_classes: target_test_classes.append(__snake_case ) # sort with class names return sorted(__snake_case , key=lambda __snake_case : x.__name__ ) def a_ ( __snake_case : Optional[Any] , __snake_case : List[str] ) -> Any: """simple docstring""" lowerCamelCase_ =get_test_classes_for_model(__snake_case , __snake_case ) lowerCamelCase_ =[] for test_class in test_classes: lowerCamelCase_ =get_model_tester_from_test_class(__snake_case ) if tester_class is not None: tester_classes.append(__snake_case ) # sort with class names return sorted(__snake_case , key=lambda __snake_case : x.__name__ ) def a_ ( __snake_case : Tuple ) -> Tuple: """simple docstring""" lowerCamelCase_ =get_test_classes(__snake_case ) lowerCamelCase_ ={test_class: get_model_tester_from_test_class(__snake_case ) for test_class in test_classes} return test_tester_mapping def a_ ( __snake_case : Dict ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ =get_model_classes(__snake_case ) lowerCamelCase_ ={ model_class: get_test_classes_for_model(__snake_case , __snake_case ) for model_class in model_classes } return model_test_mapping def a_ ( __snake_case : Optional[Any] ) -> Dict: """simple docstring""" lowerCamelCase_ =get_model_classes(__snake_case ) lowerCamelCase_ ={ model_class: get_tester_classes_for_model(__snake_case , __snake_case ) for model_class in model_classes } return model_to_tester_mapping def a_ ( __snake_case : List[str] ) -> List[Any]: """simple docstring""" if isinstance(__snake_case , __snake_case ): return o elif isinstance(__snake_case , __snake_case ): return o.__name__ elif isinstance(__snake_case , (list, tuple) ): return [to_json(__snake_case ) for x in o] elif isinstance(__snake_case , __snake_case ): return {to_json(__snake_case ): to_json(__snake_case ) for k, v in o.items()} else: return o
6
0
import torch from diffusers import EulerDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class __lowercase ( a_ ): """simple docstring""" UpperCamelCase : Tuple = (EulerDiscreteScheduler,) UpperCamelCase : Any = 1_0 def __A ( self , **A ) -> List[Any]: '''simple docstring''' lowerCamelCase = { """num_train_timesteps""": 11_00, """beta_start""": 0.0001, """beta_end""": 0.02, """beta_schedule""": """linear""", } config.update(**A ) return config def __A ( self ) -> int: '''simple docstring''' for timesteps in [10, 50, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=A ) def __A ( self ) -> Tuple: '''simple docstring''' for beta_start, beta_end in zip([0.00001, 0.0001, 0.001] , [0.0002, 0.002, 0.02] ): self.check_over_configs(beta_start=A , beta_end=A ) def __A ( self ) -> str: '''simple docstring''' for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=A ) def __A ( self ) -> List[Any]: '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A ) def __A ( self ) -> str: '''simple docstring''' lowerCamelCase = self.scheduler_classes[0] lowerCamelCase = self.get_scheduler_config() lowerCamelCase = scheduler_class(**A ) scheduler.set_timesteps(self.num_inference_steps ) lowerCamelCase = torch.manual_seed(0 ) lowerCamelCase = self.dummy_model() lowerCamelCase = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCamelCase = sample.to(A ) for i, t in enumerate(scheduler.timesteps ): lowerCamelCase = scheduler.scale_model_input(A , A ) lowerCamelCase = model(A , A ) lowerCamelCase = scheduler.step(A , A , A , generator=A ) lowerCamelCase = output.prev_sample lowerCamelCase = torch.sum(torch.abs(A ) ) lowerCamelCase = torch.mean(torch.abs(A ) ) assert abs(result_sum.item() - 10.0807 ) < 1e-2 assert abs(result_mean.item() - 0.0131 ) < 1e-3 def __A ( self ) -> List[Any]: '''simple docstring''' lowerCamelCase = self.scheduler_classes[0] lowerCamelCase = self.get_scheduler_config(prediction_type="""v_prediction""" ) lowerCamelCase = scheduler_class(**A ) scheduler.set_timesteps(self.num_inference_steps ) lowerCamelCase = torch.manual_seed(0 ) lowerCamelCase = self.dummy_model() lowerCamelCase = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCamelCase = sample.to(A ) for i, t in enumerate(scheduler.timesteps ): lowerCamelCase = scheduler.scale_model_input(A , A ) lowerCamelCase = model(A , A ) lowerCamelCase = scheduler.step(A , A , A , generator=A ) lowerCamelCase = output.prev_sample lowerCamelCase = torch.sum(torch.abs(A ) ) lowerCamelCase = torch.mean(torch.abs(A ) ) assert abs(result_sum.item() - 0.0002 ) < 1e-2 assert abs(result_mean.item() - 2.2_6_7_6e-0_6 ) < 1e-3 def __A ( self ) -> Union[str, Any]: '''simple docstring''' lowerCamelCase = self.scheduler_classes[0] lowerCamelCase = self.get_scheduler_config() lowerCamelCase = scheduler_class(**A ) scheduler.set_timesteps(self.num_inference_steps , device=A ) lowerCamelCase = torch.manual_seed(0 ) lowerCamelCase = self.dummy_model() lowerCamelCase = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() lowerCamelCase = sample.to(A ) for t in scheduler.timesteps: lowerCamelCase = scheduler.scale_model_input(A , A ) lowerCamelCase = model(A , A ) lowerCamelCase = scheduler.step(A , A , A , generator=A ) lowerCamelCase = output.prev_sample lowerCamelCase = torch.sum(torch.abs(A ) ) lowerCamelCase = torch.mean(torch.abs(A ) ) assert abs(result_sum.item() - 10.0807 ) < 1e-2 assert abs(result_mean.item() - 0.0131 ) < 1e-3 def __A ( self ) -> Optional[int]: '''simple docstring''' lowerCamelCase = self.scheduler_classes[0] lowerCamelCase = self.get_scheduler_config() lowerCamelCase = scheduler_class(**A , use_karras_sigmas=A ) scheduler.set_timesteps(self.num_inference_steps , device=A ) lowerCamelCase = torch.manual_seed(0 ) lowerCamelCase = self.dummy_model() lowerCamelCase = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() lowerCamelCase = sample.to(A ) for t in scheduler.timesteps: lowerCamelCase = scheduler.scale_model_input(A , A ) lowerCamelCase = model(A , A ) lowerCamelCase = scheduler.step(A , A , A , generator=A ) lowerCamelCase = output.prev_sample lowerCamelCase = torch.sum(torch.abs(A ) ) lowerCamelCase = torch.mean(torch.abs(A ) ) assert abs(result_sum.item() - 124.52299499511719 ) < 1e-2 assert abs(result_mean.item() - 0.16213932633399963 ) < 1e-3
252
import json import os from typing import Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase : List[str] = logging.get_logger(__name__) UpperCAmelCase : Tuple = { "vocab_file": "vocab.json", "merges_file": "merges.txt", } UpperCAmelCase : Optional[Any] = { "vocab_file": {"ctrl": "https://raw.githubusercontent.com/salesforce/ctrl/master/ctrl-vocab.json"}, "merges_file": {"ctrl": "https://raw.githubusercontent.com/salesforce/ctrl/master/ctrl-merges.txt"}, } UpperCAmelCase : Union[str, Any] = { "ctrl": 2_56, } UpperCAmelCase : List[str] = { "Pregnancy": 16_86_29, "Christianity": 76_75, "Explain": 10_64_23, "Fitness": 6_34_40, "Saving": 6_31_63, "Ask": 2_71_71, "Ass": 9_59_85, "Joke": 16_35_09, "Questions": 4_56_22, "Thoughts": 4_96_05, "Retail": 5_23_42, "Feminism": 16_43_38, "Writing": 1_19_92, "Atheism": 19_22_63, "Netflix": 4_86_16, "Computing": 3_96_39, "Opinion": 4_32_13, "Alone": 4_49_67, "Funny": 5_89_17, "Gaming": 4_03_58, "Human": 40_88, "India": 13_31, "Joker": 7_71_38, "Diet": 3_62_06, "Legal": 1_18_59, "Norman": 49_39, "Tip": 7_26_89, "Weight": 5_23_43, "Movies": 4_62_73, "Running": 2_34_25, "Science": 20_90, "Horror": 3_77_93, "Confession": 6_05_72, "Finance": 1_22_50, "Politics": 1_63_60, "Scary": 19_19_85, "Support": 1_26_54, "Technologies": 3_25_16, "Teenage": 6_61_60, "Event": 3_27_69, "Learned": 6_74_60, "Notion": 18_27_70, "Wikipedia": 3_75_83, "Books": 66_65, "Extract": 7_60_50, "Confessions": 10_27_01, "Conspiracy": 7_59_32, "Links": 6_36_74, "Narcissus": 15_04_25, "Relationship": 5_47_66, "Relationships": 13_47_96, "Reviews": 4_16_71, "News": 42_56, "Translation": 2_68_20, "multilingual": 12_84_06, } def __lowerCamelCase ( lowerCamelCase__ : Any ): '''simple docstring''' lowerCamelCase = set() lowerCamelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowerCamelCase = char lowerCamelCase = set(lowerCamelCase__ ) return pairs class __lowercase ( a_ ): """simple docstring""" UpperCamelCase : Any = VOCAB_FILES_NAMES UpperCamelCase : str = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase : Optional[int] = CONTROL_CODES def __init__( self , A , A , A="<unk>" , **A ) -> int: '''simple docstring''' super().__init__(unk_token=A , **A ) with open(A , encoding="""utf-8""" ) as vocab_handle: lowerCamelCase = json.load(A ) lowerCamelCase = {v: k for k, v in self.encoder.items()} with open(A , encoding="""utf-8""" ) as merges_handle: lowerCamelCase = merges_handle.read().split("""\n""" )[1:-1] lowerCamelCase = [tuple(merge.split() ) for merge in merges] lowerCamelCase = dict(zip(A , range(len(A ) ) ) ) lowerCamelCase = {} @property def __A ( self ) -> Union[str, Any]: '''simple docstring''' return len(self.encoder ) def __A ( self ) -> List[str]: '''simple docstring''' return dict(self.encoder , **self.added_tokens_encoder ) def __A ( self , A ) -> str: '''simple docstring''' if token in self.cache: return self.cache[token] lowerCamelCase = tuple(A ) lowerCamelCase = tuple(list(word[:-1] ) + [word[-1] + """</w>"""] ) lowerCamelCase = get_pairs(A ) if not pairs: return token while True: lowerCamelCase = min(A , key=lambda A : self.bpe_ranks.get(A , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break lowerCamelCase , lowerCamelCase = bigram lowerCamelCase = [] lowerCamelCase = 0 while i < len(A ): try: lowerCamelCase = word.index(A , A ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCamelCase = j if word[i] == first and i < len(A ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCamelCase = tuple(A ) lowerCamelCase = new_word if len(A ) == 1: break else: lowerCamelCase = get_pairs(A ) lowerCamelCase = """@@ """.join(A ) lowerCamelCase = word[:-4] lowerCamelCase = word return word def __A ( self , A ) -> Optional[int]: '''simple docstring''' lowerCamelCase = [] lowerCamelCase = re.findall(r"""\S+\n?""" , A ) for token in words: split_tokens.extend(list(self.bpe(A ).split(""" """ ) ) ) return split_tokens def __A ( self , A ) -> int: '''simple docstring''' return self.encoder.get(A , self.encoder.get(self.unk_token ) ) def __A ( self , A ) -> Any: '''simple docstring''' return self.decoder.get(A , self.unk_token ) def __A ( self , A ) -> str: '''simple docstring''' lowerCamelCase = """ """.join(A ).replace("""@@ """ , """""" ).strip() return out_string def __A ( self , A , A = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(A ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return lowerCamelCase = os.path.join( A , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) lowerCamelCase = os.path.join( A , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(A , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=A , ensure_ascii=A ) + """\n""" ) lowerCamelCase = 0 with open(A , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda A : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' """ Please check that the tokenizer is not corrupted!""" ) lowerCamelCase = token_index writer.write(""" """.join(A ) + """\n""" ) index += 1 return vocab_file, merge_file # def decode(self, token_ids, skip_special_tokens=False, clean_up_tokenization_spaces=True): # filtered_tokens = ' '.join(self.convert_ids_to_tokens(token_ids, skip_special_tokens=skip_special_tokens)) # tokens_generated_so_far = re.sub('(@@ )', '', string=filtered_tokens) # tokens_generated_so_far = re.sub('(@@ ?$)', '', string=tokens_generated_so_far) # return ''.join(tokens_generated_so_far)
252
1
from math import isqrt def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : int ) -> list[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , __UpperCamelCase , __UpperCamelCase ): SCREAMING_SNAKE_CASE__ = False return [i for i in range(2 , __UpperCamelCase ) if is_prime[i]] def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : int = 10**8 ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = calculate_prime_numbers(max_number // 2 ) SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = len(__UpperCamelCase ) - 1 while left <= right: while prime_numbers[left] * prime_numbers[right] >= max_number: right -= 1 semiprimes_count += right - left + 1 left += 1 return semiprimes_count if __name__ == "__main__": print(F"""{solution() = }""")
370
import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class __snake_case ( unittest.TestCase ): def __a ( self : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) SCREAMING_SNAKE_CASE__ = Vector() def __a ( self : int ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(_lowercase ) , """(0,0,0,0,0,1)""" ) def __a ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Vector([1, 2, 3, 4] ) self.assertEqual(len(_lowercase ) , 4 ) def __a ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Vector([1, 2] ) SCREAMING_SNAKE_CASE__ = Vector([1, 2, 3, 4, 5] ) SCREAMING_SNAKE_CASE__ = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) SCREAMING_SNAKE_CASE__ = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.2_36 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.4_16 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.6_16 , 3 ) def __a ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Vector([1, 2, 3] ) SCREAMING_SNAKE_CASE__ = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def __a ( self : Any ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Vector([1, 2, 3] ) SCREAMING_SNAKE_CASE__ = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def __a ( self : str ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Vector([1, 2, 3] ) SCREAMING_SNAKE_CASE__ = Vector([2, -1, 4] ) # for test of dot product SCREAMING_SNAKE_CASE__ = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , """(3.0,6.0,9.0)""" ) self.assertEqual((a * b) , 0 ) def __a ( self : Union[str, Any] ): """simple docstring""" self.assertEqual(str(zero_vector(10 ) ).count("""0""" ) , 10 ) def __a ( self : str ): """simple docstring""" self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , """(0,1,0)""" ) def __a ( self : Any ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Vector([1, 2, 3] ) SCREAMING_SNAKE_CASE__ = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , _lowercase , _lowercase ) ) , """(3,4,7)""" ) def __a ( self : str ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Vector([1, 0, 0, 0, 0, 0] ) SCREAMING_SNAKE_CASE__ = x.copy() self.assertEqual(str(_lowercase ) , str(_lowercase ) ) def __a ( self : str ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(_lowercase ) , """(0,1,0)""" ) def __a ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual("""|1,2,3|\n|2,4,5|\n|6,7,8|\n""" , str(_lowercase ) ) def __a ( self : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) SCREAMING_SNAKE_CASE__ = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(_lowercase , _lowercase ) ) def __a ( self : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) SCREAMING_SNAKE_CASE__ = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(_lowercase , _lowercase ) ) def __a ( self : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def __a ( self : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) SCREAMING_SNAKE_CASE__ = Vector([1, 2, 3] ) self.assertEqual("""(14,32,50)""" , str(a * x ) ) self.assertEqual("""|2,4,6|\n|8,10,12|\n|14,16,18|\n""" , str(a * 2 ) ) def __a ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual("""|1,2,5|\n|2,4,5|\n|6,7,8|\n""" , str(_lowercase ) ) def __a ( self : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def __a ( self : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) SCREAMING_SNAKE_CASE__ = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("""|2,4,10|\n|4,8,10|\n|12,14,18|\n""" , str(a + b ) ) def __a ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) SCREAMING_SNAKE_CASE__ = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("""|0,0,-4|\n|0,0,0|\n|0,0,-2|\n""" , str(a - b ) ) def __a ( self : Any ): """simple docstring""" self.assertEqual( """|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n""" , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
204
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCAmelCase__ = { """configuration_groupvit""": [ """GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GroupViTConfig""", """GroupViTOnnxConfig""", """GroupViTTextConfig""", """GroupViTVisionConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ """GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """GroupViTModel""", """GroupViTPreTrainedModel""", """GroupViTTextModel""", """GroupViTVisionModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ """TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFGroupViTModel""", """TFGroupViTPreTrainedModel""", """TFGroupViTTextModel""", """TFGroupViTVisionModel""", ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
68
import copy import random from transformers import CLIPTokenizer class a__ ( snake_case ): """simple docstring""" def __init__( self , *lowercase , **lowercase ) -> Union[str, Any]: '''simple docstring''' super().__init__(*lowercase , **lowercase ) A__ = {} def UpperCamelCase ( self , lowercase , *lowercase , **lowercase ) -> str: '''simple docstring''' A__ = super().add_tokens(lowercase , *lowercase , **lowercase ) if num_added_tokens == 0: raise ValueError( F'The tokenizer already contains the token {placeholder_token}. Please pass a different' " `placeholder_token` that is not already in the tokenizer." ) def UpperCamelCase ( self , lowercase , *lowercase , lowercase=1 , **lowercase ) -> Any: '''simple docstring''' A__ = [] if num_vec_per_token == 1: self.try_adding_tokens(lowercase , *lowercase , **lowercase ) output.append(lowercase ) else: A__ = [] for i in range(lowercase ): A__ = placeholder_token + F'_{i}' self.try_adding_tokens(lowercase , *lowercase , **lowercase ) output.append(lowercase ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( F'The tokenizer already has placeholder token {token} that can get confused with' F' {placeholder_token}keep placeholder tokens independent' ) A__ = output def UpperCamelCase ( self , lowercase , lowercase=False , lowercase=1.0 ) -> List[Any]: '''simple docstring''' if isinstance(lowercase , lowercase ): A__ = [] for i in range(len(lowercase ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=lowercase ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: A__ = self.token_map[placeholder_token] A__ = tokens[: 1 + int(len(lowercase ) * prop_tokens_to_load )] if vector_shuffle: A__ = copy.copy(lowercase ) random.shuffle(lowercase ) A__ = text.replace(lowercase , " ".join(lowercase ) ) return text def __call__( self , lowercase , *lowercase , lowercase=False , lowercase=1.0 , **lowercase ) -> str: '''simple docstring''' return super().__call__( self.replace_placeholder_tokens_in_text( lowercase , vector_shuffle=lowercase , prop_tokens_to_load=lowercase ) , *lowercase , **lowercase , ) def UpperCamelCase ( self , lowercase , *lowercase , lowercase=False , lowercase=1.0 , **lowercase ) -> List[str]: '''simple docstring''' return super().encode( self.replace_placeholder_tokens_in_text( lowercase , vector_shuffle=lowercase , prop_tokens_to_load=lowercase ) , *lowercase , **lowercase , )
68
1
'''simple docstring''' import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class __UpperCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): '''simple docstring''' __lowercase : str = StableUnCLIPPipeline __lowercase : int = TEXT_TO_IMAGE_PARAMS __lowercase : Dict = TEXT_TO_IMAGE_BATCH_PARAMS __lowercase : Dict = TEXT_TO_IMAGE_IMAGE_PARAMS __lowercase : Tuple = TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false __lowercase : Dict = False def __A ( self ) -> Any: A_ = 32 A_ = embedder_hidden_size # prior components torch.manual_seed(0 ) A_ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) torch.manual_seed(0 ) A_ = CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=_SCREAMING_SNAKE_CASE , projection_dim=_SCREAMING_SNAKE_CASE , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) A_ = PriorTransformer( num_attention_heads=2 , attention_head_dim=12 , embedding_dim=_SCREAMING_SNAKE_CASE , num_layers=1 , ) torch.manual_seed(0 ) A_ = DDPMScheduler( variance_type='''fixed_small_log''' , prediction_type='''sample''' , num_train_timesteps=1000 , clip_sample=_SCREAMING_SNAKE_CASE , clip_sample_range=5.0 , beta_schedule='''squaredcos_cap_v2''' , ) # regular denoising components torch.manual_seed(0 ) A_ = StableUnCLIPImageNormalizer(embedding_dim=_SCREAMING_SNAKE_CASE ) A_ = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''' ) torch.manual_seed(0 ) A_ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) torch.manual_seed(0 ) A_ = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=_SCREAMING_SNAKE_CASE , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) A_ = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=_SCREAMING_SNAKE_CASE , layers_per_block=1 , upcast_attention=_SCREAMING_SNAKE_CASE , use_linear_projection=_SCREAMING_SNAKE_CASE , ) torch.manual_seed(0 ) A_ = DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.00_085 , beta_end=0.012 , prediction_type='''v_prediction''' , set_alpha_to_one=_SCREAMING_SNAKE_CASE , steps_offset=1 , ) torch.manual_seed(0 ) A_ = AutoencoderKL() A_ = { # prior components '''prior_tokenizer''': prior_tokenizer, '''prior_text_encoder''': prior_text_encoder, '''prior''': prior, '''prior_scheduler''': prior_scheduler, # image noising components '''image_normalizer''': image_normalizer, '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder, '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, } return components def __A ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=0 ) -> Tuple: if str(_SCREAMING_SNAKE_CASE ).startswith('''mps''' ): A_ = torch.manual_seed(_SCREAMING_SNAKE_CASE ) else: A_ = torch.Generator(device=_SCREAMING_SNAKE_CASE ).manual_seed(_SCREAMING_SNAKE_CASE ) A_ = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''prior_num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def __A ( self ) -> int: A_ = torch_device == '''cpu''' self._test_attention_slicing_forward_pass(test_max_difference=_SCREAMING_SNAKE_CASE ) def __A ( self ) -> List[str]: A_ = torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=_SCREAMING_SNAKE_CASE ) @slow @require_torch_gpu class __UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def __A ( self ) -> Dict: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __A ( self ) -> List[str]: A_ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy''' ) A_ = StableUnCLIPPipeline.from_pretrained('''fusing/stable-unclip-2-1-l''' , torch_dtype=torch.floataa ) pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() A_ = torch.Generator(device='''cpu''' ).manual_seed(0 ) A_ = pipe('''anime turle''' , generator=_SCREAMING_SNAKE_CASE , output_type='''np''' ) A_ = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __A ( self ) -> Optional[Any]: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() A_ = StableUnCLIPPipeline.from_pretrained('''fusing/stable-unclip-2-1-l''' , torch_dtype=torch.floataa ) A_ = pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() A_ = pipe( '''anime turtle''' , prior_num_inference_steps=2 , num_inference_steps=2 , output_type='''np''' , ) A_ = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
18
'''simple docstring''' # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import platform import numpy as np import psutil import torch from accelerate import __version__ as version from accelerate.commands.config import default_config_file, load_config_from_file from ..utils import is_npu_available, is_xpu_available def _UpperCAmelCase ( _UpperCamelCase : Union[str, Any]=None ) -> List[Any]: if subparsers is not None: A_ = subparsers.add_parser('''env''' ) else: A_ = argparse.ArgumentParser('''Accelerate env command''' ) parser.add_argument( '''--config_file''', default=_UpperCamelCase, help='''The config file to use for the default values in the launching script.''' ) if subparsers is not None: parser.set_defaults(func=_UpperCamelCase ) return parser def _UpperCAmelCase ( _UpperCamelCase : Dict ) -> Dict: A_ = torch.__version__ A_ = torch.cuda.is_available() A_ = is_xpu_available() A_ = is_npu_available() A_ = '''Not found''' # Get the default from the config file. if args.config_file is not None or os.path.isfile(_UpperCamelCase ): A_ = load_config_from_file(args.config_file ).to_dict() A_ = { '''`Accelerate` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''Numpy version''': np.__version__, '''PyTorch version (GPU?)''': F'''{pt_version} ({pt_cuda_available})''', '''PyTorch XPU available''': str(_UpperCamelCase ), '''PyTorch NPU available''': str(_UpperCamelCase ), '''System RAM''': F'''{psutil.virtual_memory().total / 10_24 ** 3:.2f} GB''', } if pt_cuda_available: A_ = torch.cuda.get_device_name() print('''\nCopy-and-paste the text below in your GitHub issue\n''' ) print('''\n'''.join([F'''- {prop}: {val}''' for prop, val in info.items()] ) ) print('''- `Accelerate` default config:''' if args.config_file is None else '''- `Accelerate` config passed:''' ) A_ = ( '''\n'''.join([F'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(_UpperCamelCase, _UpperCamelCase ) else F'''\t{accelerate_config}''' ) print(_UpperCamelCase ) A_ = accelerate_config return info def _UpperCAmelCase ( ) -> int: A_ = env_command_parser() A_ = parser.parse_args() env_command(_UpperCamelCase ) return 0 if __name__ == "__main__": raise SystemExit(main())
18
1
_UpperCAmelCase : Dict = [ """DownloadConfig""", """DownloadManager""", """DownloadMode""", """StreamingDownloadManager""", ] from .download_config import DownloadConfig from .download_manager import DownloadManager, DownloadMode from .streaming_download_manager import StreamingDownloadManager
50
import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=None , _UpperCAmelCase="no" , _UpperCAmelCase="29500" ) -> Tuple: lowerCamelCase__ : Dict = False lowerCamelCase__ : Dict = False if any(key.startswith('KAGGLE' ) for key in os.environ.keys() ): lowerCamelCase__ : Optional[Any] = True elif "IPython" in sys.modules: lowerCamelCase__ : Optional[Any] = 'google.colab' in str(sys.modules['IPython'].get_ipython() ) try: lowerCamelCase__ : List[str] = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( F"""Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.""" ) if (in_colab or in_kaggle) and (os.environ.get('TPU_NAME' , _UpperCAmelCase ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside ' 'your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if num_processes is None: lowerCamelCase__ : Optional[Any] = 8 lowerCamelCase__ : List[str] = PrepareForLaunch(_UpperCAmelCase , distributed_type='TPU' ) print(F"""Launching a training on {num_processes} TPU cores.""" ) xmp.spawn(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on one CPU.' ) function(*_UpperCAmelCase ) else: if num_processes is None: raise ValueError( 'You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call.' ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized ' 'inside your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if torch.cuda.is_initialized(): raise ValueError( 'To launch a multi-GPU training from your notebook, you need to avoid running any instruction ' 'using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA ' 'function.' ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port=_UpperCAmelCase , mixed_precision=_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = PrepareForLaunch(_UpperCAmelCase , distributed_type='MULTI_GPU' ) print(F"""Launching training on {num_processes} GPUs.""" ) try: start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( 'CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. ' 'This likely stems from an outside import causing issues once the `notebook_launcher()` is called. ' 'Please review your imports and test them when running the `notebook_launcher()` to identify ' 'which one is problematic.' ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): lowerCamelCase__ : int = '1' print('Launching training on MPS.' ) elif torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on CPU.' ) function(*_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=2 ) -> Optional[Any]: from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port='29500' , accelerate_mixed_precision='no' , accelerate_debug_rdv_file=tmp_file.name , accelerate_use_cpu='yes' , ): lowerCamelCase__ : Optional[Any] = PrepareForLaunch(_UpperCAmelCase , debug=_UpperCAmelCase ) start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' )
50
1
from __future__ import annotations import inspect import unittest from math import floor import numpy as np from transformers import CvtConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFCvtForImageClassification, TFCvtModel from transformers.models.cvt.modeling_tf_cvt import TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCamelCase( __lowercase ): def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : Tuple = self.config_class(**self.inputs_dict) self.parent.assertTrue(hasattr(_a, 'embed_dim')) self.parent.assertTrue(hasattr(_a, 'num_heads')) class _lowerCamelCase: def __init__( self, lowerCamelCase, lowerCamelCase=13, lowerCamelCase=64, lowerCamelCase=3, lowerCamelCase=[16, 48, 96], lowerCamelCase=[1, 3, 6], lowerCamelCase=[1, 2, 10], lowerCamelCase=[7, 3, 3], lowerCamelCase=[4, 2, 2], lowerCamelCase=[2, 1, 1], lowerCamelCase=[2, 2, 2], lowerCamelCase=[False, False, True], lowerCamelCase=[0.0, 0.0, 0.0], lowerCamelCase=0.0_2, lowerCamelCase=1E-12, lowerCamelCase=True, lowerCamelCase=True, lowerCamelCase=2, ) -> int: """simple docstring""" _lowercase : str = parent _lowercase : int = batch_size _lowercase : Optional[int] = image_size _lowercase : Union[str, Any] = patch_sizes _lowercase : Optional[int] = patch_stride _lowercase : str = patch_padding _lowercase : str = is_training _lowercase : List[Any] = use_labels _lowercase : List[Any] = num_labels _lowercase : int = num_channels _lowercase : Optional[Any] = embed_dim _lowercase : List[Any] = num_heads _lowercase : List[str] = stride_kv _lowercase : List[str] = depth _lowercase : str = cls_token _lowercase : Dict = attention_drop_rate _lowercase : Union[str, Any] = initializer_range _lowercase : List[Any] = layer_norm_eps def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) _lowercase : Tuple = None if self.use_labels: # create a random int32 tensor of given shape _lowercase : int = ids_tensor([self.batch_size], self.num_labels) _lowercase : str = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self) -> List[Any]: """simple docstring""" return CvtConfig( image_size=self.image_size, num_labels=self.num_labels, num_channels=self.num_channels, embed_dim=self.embed_dim, num_heads=self.num_heads, patch_sizes=self.patch_sizes, patch_padding=self.patch_padding, patch_stride=self.patch_stride, stride_kv=self.stride_kv, depth=self.depth, cls_token=self.cls_token, attention_drop_rate=self.attention_drop_rate, initializer_range=self.initializer_range, ) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> List[str]: """simple docstring""" _lowercase : List[Any] = TFCvtModel(config=_a) _lowercase : Dict = model(_a, training=_a) _lowercase : Dict = (self.image_size, self.image_size) _lowercase : Optional[Any] = image_size[0], image_size[1] for i in range(len(self.depth)): _lowercase : Any = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1) _lowercase : List[Any] = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.embed_dim[-1], height, width)) def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase, lowerCamelCase) -> List[Any]: """simple docstring""" _lowercase : List[str] = self.num_labels _lowercase : str = TFCvtForImageClassification(_a) _lowercase : str = model(_a, labels=_a, training=_a) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_labels)) def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Union[str, Any] = self.prepare_config_and_inputs() _lowercase : Optional[Any] = config_and_inputs _lowercase : Union[str, Any] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class _lowerCamelCase( __lowercase, __lowercase, unittest.TestCase ): lowercase_ : int = (TFCvtModel, TFCvtForImageClassification) if is_tf_available() else () lowercase_ : Optional[Any] = ( {"feature-extraction": TFCvtModel, "image-classification": TFCvtForImageClassification} if is_tf_available() else {} ) lowercase_ : List[Any] = False lowercase_ : Union[str, Any] = False lowercase_ : List[str] = False lowercase_ : int = False lowercase_ : int = False def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Dict = TFCvtModelTester(self) _lowercase : Optional[int] = TFCvtConfigTester(self, config_class=_a, has_text_modality=_a, hidden_size=37) def UpperCamelCase ( self) -> Any: """simple docstring""" self.config_tester.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() @unittest.skip(reason='Cvt does not output attentions') def UpperCamelCase ( self) -> List[Any]: """simple docstring""" pass @unittest.skip(reason='Cvt does not use inputs_embeds') def UpperCamelCase ( self) -> str: """simple docstring""" pass @unittest.skip(reason='Cvt does not support input and output embeddings') def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices('GPU')) == 0, reason='TF does not support backprop for grouped convolutions on CPU.', ) def UpperCamelCase ( self) -> Tuple: """simple docstring""" super().test_dataset_conversion() @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices('GPU')) == 0, reason='TF does not support backprop for grouped convolutions on CPU.', ) @slow def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" super().test_keras_fit() @unittest.skip(reason='Get `Failed to determine best cudnn convolution algo.` error after using TF 2.12+cuda 11.8') def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : Dict = tf.keras.mixed_precision.Policy('mixed_float16') tf.keras.mixed_precision.set_global_policy(_a) super().test_keras_fit() tf.keras.mixed_precision.set_global_policy('float32') def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowercase : int = model_class(_a) _lowercase : Tuple = inspect.signature(model.call) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowercase : Union[str, Any] = [*signature.parameters.keys()] _lowercase : List[Any] = ['''pixel_values'''] self.assertListEqual(arg_names[:1], _a) def UpperCamelCase ( self) -> Dict: """simple docstring""" def check_hidden_states_output(lowerCamelCase, lowerCamelCase, lowerCamelCase): _lowercase : str = model_class(_a) _lowercase : List[Any] = model(**self._prepare_for_class(_a, _a)) _lowercase : Any = outputs.hidden_states _lowercase : Tuple = len(self.model_tester.depth) self.assertEqual(len(_a), _a) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:]), [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ], ) _lowercase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowercase : Dict = True check_hidden_states_output(_a, _a, _a) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowercase : Dict = True check_hidden_states_output(_a, _a, _a) def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a) def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_a) @slow def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" for model_name in TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowercase : List[Any] = TFCvtModel.from_pretrained(_a) self.assertIsNotNone(_a) def UpperCamelCase_( ) -> Dict: _lowercase : Optional[int] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class _lowerCamelCase( unittest.TestCase ): @cached_property def UpperCamelCase ( self) -> Tuple: """simple docstring""" return AutoImageProcessor.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0]) @slow def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Dict = TFCvtForImageClassification.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0]) _lowercase : Optional[int] = self.default_image_processor _lowercase : Tuple = prepare_img() _lowercase : Dict = image_processor(images=_a, return_tensors='tf') # forward pass _lowercase : Dict = model(**_a) # verify the logits _lowercase : Optional[int] = tf.TensorShape((1, 10_00)) self.assertEqual(outputs.logits.shape, _a) _lowercase : Tuple = tf.constant([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0]) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy(), _a, atol=1E-4))
365
import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class _lowerCamelCase( unittest.TestCase ): def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Optional[Any] = torch.nn.Linear(10, 10) _lowercase : Optional[int] = torch.optim.SGD(model.parameters(), 0.1) _lowercase : Optional[int] = Accelerator() _lowercase : Optional[int] = accelerator.prepare(lowerCamelCase) try: pickle.loads(pickle.dumps(lowerCamelCase)) except Exception as e: self.fail(F'''Accelerated optimizer pickling failed with {e}''') AcceleratorState._reset_state()
84
0
import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class __A ( lowerCAmelCase ): lowerCAmelCase_ : Tuple = (PNDMScheduler,) lowerCAmelCase_ : Optional[int] = (("num_inference_steps", 50),) def lowercase__ ( self : List[str] , **UpperCAmelCase_ : List[Any] ): lowerCAmelCase : Optional[int] = { 'num_train_timesteps': 1000, 'beta_start': 0.00_01, 'beta_end': 0.02, 'beta_schedule': 'linear', } config.update(**_snake_case ) return config def lowercase__ ( self : Union[str, Any] , UpperCAmelCase_ : Tuple=0 , **UpperCAmelCase_ : int ): lowerCAmelCase : Union[str, Any] = dict(self.forward_default_kwargs ) lowerCAmelCase : Tuple = kwargs.pop('num_inference_steps' , _snake_case ) lowerCAmelCase : int = self.dummy_sample lowerCAmelCase : Tuple = 0.1 * sample lowerCAmelCase : List[Any] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: lowerCAmelCase : str = self.get_scheduler_config(**_snake_case ) lowerCAmelCase : Dict = scheduler_class(**_snake_case ) scheduler.set_timesteps(_snake_case ) # copy over dummy past residuals lowerCAmelCase : str = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_snake_case ) lowerCAmelCase : Optional[Any] = scheduler_class.from_pretrained(_snake_case ) new_scheduler.set_timesteps(_snake_case ) # copy over dummy past residuals lowerCAmelCase : int = dummy_past_residuals[:] lowerCAmelCase : List[Any] = scheduler.step_prk(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample lowerCAmelCase : Dict = new_scheduler.step_prk(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" lowerCAmelCase : Optional[int] = scheduler.step_plms(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample lowerCAmelCase : Optional[Any] = new_scheduler.step_plms(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def lowercase__ ( self : str ): pass def lowercase__ ( self : int , UpperCAmelCase_ : str=0 , **UpperCAmelCase_ : List[str] ): lowerCAmelCase : List[Any] = dict(self.forward_default_kwargs ) lowerCAmelCase : Tuple = kwargs.pop('num_inference_steps' , _snake_case ) lowerCAmelCase : Optional[Any] = self.dummy_sample lowerCAmelCase : Union[str, Any] = 0.1 * sample lowerCAmelCase : List[str] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: lowerCAmelCase : List[str] = self.get_scheduler_config() lowerCAmelCase : Any = scheduler_class(**_snake_case ) scheduler.set_timesteps(_snake_case ) # copy over dummy past residuals (must be after setting timesteps) lowerCAmelCase : List[str] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_snake_case ) lowerCAmelCase : List[Any] = scheduler_class.from_pretrained(_snake_case ) # copy over dummy past residuals new_scheduler.set_timesteps(_snake_case ) # copy over dummy past residual (must be after setting timesteps) lowerCAmelCase : List[str] = dummy_past_residuals[:] lowerCAmelCase : int = scheduler.step_prk(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample lowerCAmelCase : str = new_scheduler.step_prk(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" lowerCAmelCase : Any = scheduler.step_plms(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample lowerCAmelCase : int = new_scheduler.step_plms(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def lowercase__ ( self : List[Any] , **UpperCAmelCase_ : Tuple ): lowerCAmelCase : List[str] = self.scheduler_classes[0] lowerCAmelCase : str = self.get_scheduler_config(**_snake_case ) lowerCAmelCase : Dict = scheduler_class(**_snake_case ) lowerCAmelCase : Tuple = 10 lowerCAmelCase : int = self.dummy_model() lowerCAmelCase : str = self.dummy_sample_deter scheduler.set_timesteps(_snake_case ) for i, t in enumerate(scheduler.prk_timesteps ): lowerCAmelCase : Dict = model(_snake_case , _snake_case ) lowerCAmelCase : Tuple = scheduler.step_prk(_snake_case , _snake_case , _snake_case ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): lowerCAmelCase : Tuple = model(_snake_case , _snake_case ) lowerCAmelCase : List[str] = scheduler.step_plms(_snake_case , _snake_case , _snake_case ).prev_sample return sample def lowercase__ ( self : Dict ): lowerCAmelCase : str = dict(self.forward_default_kwargs ) lowerCAmelCase : List[str] = kwargs.pop('num_inference_steps' , _snake_case ) for scheduler_class in self.scheduler_classes: lowerCAmelCase : Tuple = self.get_scheduler_config() lowerCAmelCase : List[Any] = scheduler_class(**_snake_case ) lowerCAmelCase : List[Any] = self.dummy_sample lowerCAmelCase : Any = 0.1 * sample if num_inference_steps is not None and hasattr(_snake_case , 'set_timesteps' ): scheduler.set_timesteps(_snake_case ) elif num_inference_steps is not None and not hasattr(_snake_case , 'set_timesteps' ): lowerCAmelCase : Any = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) lowerCAmelCase : int = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] lowerCAmelCase : int = dummy_past_residuals[:] lowerCAmelCase : List[str] = scheduler.step_prk(_snake_case , 0 , _snake_case , **_snake_case ).prev_sample lowerCAmelCase : int = scheduler.step_prk(_snake_case , 1 , _snake_case , **_snake_case ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) lowerCAmelCase : str = scheduler.step_plms(_snake_case , 0 , _snake_case , **_snake_case ).prev_sample lowerCAmelCase : Any = scheduler.step_plms(_snake_case , 1 , _snake_case , **_snake_case ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def lowercase__ ( self : Union[str, Any] ): for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=_snake_case ) def lowercase__ ( self : Dict ): for steps_offset in [0, 1]: self.check_over_configs(steps_offset=_snake_case ) lowerCAmelCase : Dict = self.scheduler_classes[0] lowerCAmelCase : List[str] = self.get_scheduler_config(steps_offset=1 ) lowerCAmelCase : Optional[Any] = scheduler_class(**_snake_case ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [901, 851, 851, 801, 801, 751, 751, 701, 701, 651, 651, 601, 601, 501, 401, 301, 201, 101, 1] ) , ) def lowercase__ ( self : Dict ): for beta_start, beta_end in zip([0.00_01, 0.0_01] , [0.0_02, 0.02] ): self.check_over_configs(beta_start=_snake_case , beta_end=_snake_case ) def lowercase__ ( self : int ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=_snake_case ) def lowercase__ ( self : Optional[Any] ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_snake_case ) def lowercase__ ( self : List[Any] ): for t in [1, 5, 10]: self.check_over_forward(time_step=_snake_case ) def lowercase__ ( self : List[Any] ): for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=_snake_case ) def lowercase__ ( self : Any ): lowerCAmelCase : List[Any] = 27 for scheduler_class in self.scheduler_classes: lowerCAmelCase : Dict = self.dummy_sample lowerCAmelCase : int = 0.1 * sample lowerCAmelCase : Optional[Any] = self.get_scheduler_config() lowerCAmelCase : int = scheduler_class(**_snake_case ) scheduler.set_timesteps(_snake_case ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): lowerCAmelCase : Tuple = scheduler.step_prk(_snake_case , _snake_case , _snake_case ).prev_sample def lowercase__ ( self : List[str] ): with self.assertRaises(_snake_case ): lowerCAmelCase : str = self.scheduler_classes[0] lowerCAmelCase : List[Any] = self.get_scheduler_config() lowerCAmelCase : Any = scheduler_class(**_snake_case ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def lowercase__ ( self : List[Any] ): lowerCAmelCase : Optional[int] = self.full_loop() lowerCAmelCase : Dict = torch.sum(torch.abs(_snake_case ) ) lowerCAmelCase : List[Any] = torch.mean(torch.abs(_snake_case ) ) assert abs(result_sum.item() - 1_98.13_18 ) < 1E-2 assert abs(result_mean.item() - 0.25_80 ) < 1E-3 def lowercase__ ( self : List[Any] ): lowerCAmelCase : int = self.full_loop(prediction_type='v_prediction' ) lowerCAmelCase : List[Any] = torch.sum(torch.abs(_snake_case ) ) lowerCAmelCase : Union[str, Any] = torch.mean(torch.abs(_snake_case ) ) assert abs(result_sum.item() - 67.39_86 ) < 1E-2 assert abs(result_mean.item() - 0.08_78 ) < 1E-3 def lowercase__ ( self : List[str] ): lowerCAmelCase : List[Any] = self.full_loop(set_alpha_to_one=_snake_case , beta_start=0.01 ) lowerCAmelCase : Union[str, Any] = torch.sum(torch.abs(_snake_case ) ) lowerCAmelCase : Dict = torch.mean(torch.abs(_snake_case ) ) assert abs(result_sum.item() - 2_30.03_99 ) < 1E-2 assert abs(result_mean.item() - 0.29_95 ) < 1E-3 def lowercase__ ( self : Union[str, Any] ): lowerCAmelCase : Optional[Any] = self.full_loop(set_alpha_to_one=_snake_case , beta_start=0.01 ) lowerCAmelCase : Dict = torch.sum(torch.abs(_snake_case ) ) lowerCAmelCase : int = torch.mean(torch.abs(_snake_case ) ) assert abs(result_sum.item() - 1_86.94_82 ) < 1E-2 assert abs(result_mean.item() - 0.24_34 ) < 1E-3
138
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __A( a ): snake_case_ = ['''image_processor''', '''tokenizer'''] snake_case_ = '''ChineseCLIPImageProcessor''' snake_case_ = ('''BertTokenizer''', '''BertTokenizerFast''') def __init__( self , _snake_case=None , _snake_case=None , **_snake_case ) -> Tuple: '''simple docstring''' __a = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _snake_case , ) __a = kwargs.pop('''feature_extractor''' ) __a = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_snake_case , _snake_case ) __a = self.image_processor def __call__( self , _snake_case=None , _snake_case=None , _snake_case=None , **_snake_case ) -> Optional[Any]: '''simple docstring''' if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: __a = self.tokenizer(_snake_case , return_tensors=_snake_case , **_snake_case ) if images is not None: __a = self.image_processor(_snake_case , return_tensors=_snake_case , **_snake_case ) if text is not None and images is not None: __a = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_snake_case ) , tensor_type=_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , *_snake_case , **_snake_case ) -> str: '''simple docstring''' return self.tokenizer.batch_decode(*_snake_case , **_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , *_snake_case , **_snake_case ) -> Dict: '''simple docstring''' return self.tokenizer.decode(*_snake_case , **_snake_case ) @property def SCREAMING_SNAKE_CASE_ ( self ) -> Any: '''simple docstring''' __a = self.tokenizer.model_input_names __a = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[Any]: '''simple docstring''' warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , _snake_case , ) return self.image_processor_class
6
0
import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class lowercase_ ( lowercase , unittest.TestCase ): __snake_case = VideoToVideoSDPipeline __snake_case = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({'''video'''} ) - {'''image''', '''width''', '''height'''} __snake_case = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'''video'''} ) - {'''image'''} __snake_case = PipelineTesterMixin.required_optional_params - {'''latents'''} __snake_case = False # No `output_type`. __snake_case = frozenset( [ '''num_inference_steps''', '''generator''', '''latents''', '''return_dict''', '''callback''', '''callback_steps''', ] ) def __lowerCAmelCase ( self : Tuple ) ->Dict: """simple docstring""" torch.manual_seed(0 ) a = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''DownBlock3D''') , up_block_types=('''UpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''') , cross_attention_dim=32 , attention_head_dim=4 , ) a = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=__UpperCAmelCase , set_alpha_to_one=__UpperCAmelCase , ) torch.manual_seed(0 ) a = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) a = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act='''gelu''' , projection_dim=512 , ) a = CLIPTextModel(__UpperCAmelCase ) a = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) a = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def __lowerCAmelCase ( self : str , __UpperCAmelCase : int , __UpperCAmelCase : Optional[Any]=0 ) ->Optional[Any]: """simple docstring""" a = floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(__UpperCAmelCase ) ).to(__UpperCAmelCase ) if str(__UpperCAmelCase ).startswith('''mps''' ): a = torch.manual_seed(__UpperCAmelCase ) else: a = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) a = { '''prompt''': '''A painting of a squirrel eating a burger''', '''video''': video, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def __lowerCAmelCase ( self : Optional[int] ) ->Union[str, Any]: """simple docstring""" a = '''cpu''' # ensure determinism for the device-dependent torch.Generator a = self.get_dummy_components() a = VideoToVideoSDPipeline(**__UpperCAmelCase ) a = sd_pipe.to(__UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCAmelCase ) a = self.get_dummy_inputs(__UpperCAmelCase ) a = '''np''' a = sd_pipe(**__UpperCAmelCase ).frames a = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) a = np.array([106, 117, 113, 174, 137, 112, 148, 151, 131] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def __lowerCAmelCase ( self : List[Any] ) ->Dict: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=__UpperCAmelCase , expected_max_diff=5e-3 ) @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''' ) def __lowerCAmelCase ( self : List[str] ) ->Union[str, Any]: """simple docstring""" pass @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''' ) def __lowerCAmelCase ( self : List[str] ) ->Tuple: """simple docstring""" pass @unittest.skip(reason='''`num_images_per_prompt` argument is not supported for this pipeline.''' ) def __lowerCAmelCase ( self : int ) ->List[Any]: """simple docstring""" pass def __lowerCAmelCase ( self : Dict ) ->str: """simple docstring""" return super().test_progress_bar() @slow @skip_mps class lowercase_ ( unittest.TestCase ): def __lowerCAmelCase ( self : Union[str, Any] ) ->str: """simple docstring""" a = VideoToVideoSDPipeline.from_pretrained('''cerspense/zeroscope_v2_XL''' , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames a = torch.Generator(device='''cpu''' ).manual_seed(0 ) a = torch.randn((1, 10, 3, 1_024, 576) , generator=__UpperCAmelCase ) a = video.to('''cuda''' ) a = '''Spiderman is surfing''' a = pipe(__UpperCAmelCase , video=__UpperCAmelCase , generator=__UpperCAmelCase , num_inference_steps=3 , output_type='''pt''' ).frames a = np.array([-1.0458984, -1.1279297, -0.9663086, -0.91503906, -0.75097656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1e-2
351
import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer UpperCAmelCase__ = "bart" UpperCAmelCase__ = True @st.cache(allow_output_mutation=a ) def _a ( ) -> Tuple: if LOAD_DENSE_INDEX: a = AutoTokenizer.from_pretrained('''yjernite/retribert-base-uncased''' ) a = AutoModel.from_pretrained('''yjernite/retribert-base-uncased''' ).to('''cuda:0''' ) a = qar_model.eval() else: a , a = (None, None) if MODEL_TYPE == "bart": a = AutoTokenizer.from_pretrained('''yjernite/bart_eli5''' ) a = AutoModelForSeqaSeqLM.from_pretrained('''yjernite/bart_eli5''' ).to('''cuda:0''' ) a = torch.load('''seq2seq_models/eli5_bart_model_blm_2.pth''' ) sas_model.load_state_dict(save_dict['''model'''] ) a = sas_model.eval() else: a , a = make_qa_sas_model( model_name='''t5-small''' , from_file='''seq2seq_models/eli5_t5_model_1024_4.pth''' , device='''cuda:0''' ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=a ) def _a ( ) -> Dict: if LOAD_DENSE_INDEX: a = faiss.StandardGpuResources() a = datasets.load_dataset(path='''wiki_snippets''' , name='''wiki40b_en_100_0''' )['''train'''] a = np.memmap( '''wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat''' , dtype='''float32''' , mode='''r''' , shape=(wikiaab_passages.num_rows, 128) , ) a = faiss.IndexFlatIP(128 ) a = faiss.index_cpu_to_gpu(a , 1 , a ) wikiaab_gpu_index_flat.add(a ) # TODO fix for larger GPU else: a , a = (None, None) a = Elasticsearch([{'''host''': '''localhost''', '''port''': '''9200'''}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=a ) def _a ( ) -> Optional[int]: a = datasets.load_dataset('''eli5''' , name='''LFQA_reddit''' ) a = elia['''train_eli5'''] a = np.memmap( '''eli5_questions_reps.dat''' , dtype='''float32''' , mode='''r''' , shape=(elia_train.num_rows, 128) ) a = faiss.IndexFlatIP(128 ) eli5_train_q_index.add(a ) return (elia_train, eli5_train_q_index) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = load_indexes() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = load_models() UpperCAmelCase__ , UpperCAmelCase__ = load_train_data() def _a ( a :str , a :Tuple=10 ) -> List[str]: a = embed_questions_for_retrieval([question] , a , a ) a , a = eli5_train_q_index.search(a , a ) a = [elia_train[int(a )] for i in I[0]] return nn_examples def _a ( a :str , a :Any="wiki40b" , a :int="dense" , a :Union[str, Any]=10 ) -> List[str]: if source == "none": a , a = (''' <P> '''.join(['''''' for _ in range(11 )] ).strip(), []) else: if method == "dense": a , a = query_qa_dense_index( a , a , a , a , a , a ) else: a , a = query_es_index( a , a , index_name='''english_wiki40b_snippets_100w''' , n_results=a , ) a = [ (res['''article_title'''], res['''section_title'''].strip(), res['''score'''], res['''passage_text''']) for res in hit_lst ] a = '''question: {} context: {}'''.format(a , a ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda a : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda a : None), } ) def _a ( a :Tuple , a :int , a :int , a :Dict=64 , a :List[Any]=256 , a :List[Any]=False , a :List[Any]=2 , a :Tuple=0.95 , a :Optional[Any]=0.8 ) -> int: with torch.no_grad(): a = qa_sas_generate( a , a , a , num_answers=1 , num_beams=a , min_len=a , max_len=a , do_sample=a , temp=a , top_p=a , top_k=a , max_input_length=1_024 , device='''cuda:0''' , )[0] return (answer, support_list) st.title("Long Form Question Answering with ELI5") # Start sidebar UpperCAmelCase__ = "<img src='https://huggingface.co/front/assets/huggingface_logo.svg'>" UpperCAmelCase__ = "\n<html>\n <head>\n <style>\n .img-container {\n padding-left: 90px;\n padding-right: 90px;\n padding-top: 50px;\n padding-bottom: 50px;\n background-color: #f0f3f9;\n }\n </style>\n </head>\n <body>\n <span class=\"img-container\"> <!-- Inline parent element -->\n %s\n </span>\n </body>\n</html>\n" % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia UpperCAmelCase__ = "\nThis demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html).\nFirst, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset,\na pre-processed fixed snapshot of Wikipedia.\n" st.sidebar.markdown(description, unsafe_allow_html=True) UpperCAmelCase__ = [ "Answer the question", "View the retrieved document only", "View the most similar ELI5 question and answer", "Show me everything, please!", ] UpperCAmelCase__ = st.sidebar.checkbox("Demo options") if demo_options: UpperCAmelCase__ = st.sidebar.selectbox( "", action_list, index=3, ) UpperCAmelCase__ = action_list.index(action_st) UpperCAmelCase__ = st.sidebar.selectbox( "", ["Show full text of passages", "Show passage section titles"], index=0, ) UpperCAmelCase__ = show_type == "Show full text of passages" else: UpperCAmelCase__ = 3 UpperCAmelCase__ = True UpperCAmelCase__ = st.sidebar.checkbox("Retrieval options") if retrieval_options: UpperCAmelCase__ = "\n ### Information retriever options\n\n The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding\n trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs.\n The answer is then generated by sequence to sequence model which takes the question and retrieved document as input.\n " st.sidebar.markdown(retriever_info) UpperCAmelCase__ = st.sidebar.selectbox("Which Wikipedia format should the model use?", ["wiki40b", "none"]) UpperCAmelCase__ = st.sidebar.selectbox("Which Wikipedia indexer should the model use?", ["dense", "sparse", "mixed"]) else: UpperCAmelCase__ = "wiki40b" UpperCAmelCase__ = "dense" UpperCAmelCase__ = "beam" UpperCAmelCase__ = 2 UpperCAmelCase__ = 64 UpperCAmelCase__ = 256 UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = st.sidebar.checkbox("Generation options") if generate_options: UpperCAmelCase__ = "\n ### Answer generation options\n\n The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large)\n weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with\n **beam** search, or **sample** from the decoder's output probabilities.\n " st.sidebar.markdown(generate_info) UpperCAmelCase__ = st.sidebar.selectbox("Would you like to use beam search or sample an answer?", ["beam", "sampled"]) UpperCAmelCase__ = st.sidebar.slider( "Minimum generation length", min_value=8, max_value=256, value=64, step=8, format=None, key=None ) UpperCAmelCase__ = st.sidebar.slider( "Maximum generation length", min_value=64, max_value=512, value=256, step=16, format=None, key=None ) if sampled == "beam": UpperCAmelCase__ = st.sidebar.slider("Beam size", min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: UpperCAmelCase__ = st.sidebar.slider( "Nucleus sampling p", min_value=0.1, max_value=1.0, value=0.95, step=0.01, format=None, key=None ) UpperCAmelCase__ = st.sidebar.slider( "Temperature", min_value=0.1, max_value=1.0, value=0.7, step=0.01, format=None, key=None ) UpperCAmelCase__ = None # start main text UpperCAmelCase__ = [ "<MY QUESTION>", "How do people make chocolate?", "Why do we get a fever when we are sick?", "How can different animals perceive different colors?", "What is natural language processing?", "What's the best way to treat a sunburn?", "What exactly are vitamins ?", "How does nuclear energy provide electricity?", "What's the difference between viruses and bacteria?", "Why are flutes classified as woodwinds when most of them are made out of metal ?", "Why do people like drinking coffee even though it tastes so bad?", "What happens when wine ages? How does it make the wine taste better?", "If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?", "How can we set a date to the beginning or end of an artistic period? Doesn't the change happen gradually?", "How does New Zealand have so many large bird predators?", ] UpperCAmelCase__ = st.selectbox( "What would you like to ask? ---- select <MY QUESTION> to enter a new query", questions_list, index=1, ) if question_s == "<MY QUESTION>": UpperCAmelCase__ = st.text_input("Enter your question here:", "") else: UpperCAmelCase__ = question_s if st.button("Show me!"): if action in [0, 1, 3]: if index_type == "mixed": UpperCAmelCase__ , UpperCAmelCase__ = make_support(question, source=wiki_source, method="dense", n_results=10) UpperCAmelCase__ , UpperCAmelCase__ = make_support(question, source=wiki_source, method="sparse", n_results=10) UpperCAmelCase__ = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] UpperCAmelCase__ = support_list[:10] UpperCAmelCase__ = "<P> " + " <P> ".join([res[-1] for res in support_list]) else: UpperCAmelCase__ , UpperCAmelCase__ = make_support(question, source=wiki_source, method=index_type, n_results=10) if action in [0, 3]: UpperCAmelCase__ , UpperCAmelCase__ = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == "sampled"), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown("### The model generated answer is:") st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown("--- \n ### The model is drawing information from the following Wikipedia passages:") for i, res in enumerate(support_list): UpperCAmelCase__ = "https://en.wikipedia.org/wiki/{}".format(res[0].replace(" ", "_")) UpperCAmelCase__ = res[1].strip() if sec_titles == "": UpperCAmelCase__ = "[{}]({})".format(res[0], wiki_url) else: UpperCAmelCase__ = sec_titles.split(" & ") UpperCAmelCase__ = " & ".join( ["[{}]({}#{})".format(sec.strip(), wiki_url, sec.strip().replace(" ", "_")) for sec in sec_list] ) st.markdown( "{0:02d} - **Article**: {1:<18} <br> _Section_: {2}".format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( "> <span style=\"font-family:arial; font-size:10pt;\">" + res[-1] + "</span>", unsafe_allow_html=True ) if action in [2, 3]: UpperCAmelCase__ = find_nearest_training(question) UpperCAmelCase__ = nn_train_list[0] st.markdown( "--- \n ### The most similar question in the ELI5 training set was: \n\n {}".format(train_exple["title"]) ) UpperCAmelCase__ = [ "{}. {}".format(i + 1, " \n".join([line.strip() for line in ans.split("\n") if line.strip() != ""])) for i, (ans, sc) in enumerate(zip(train_exple["answers"]["text"], train_exple["answers"]["score"])) if i == 0 or sc > 2 ] st.markdown("##### Its answers were: \n\n {}".format("\n".join(answers_st))) UpperCAmelCase__ = "\n---\n\n**Disclaimer**\n\n*The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system.\nEvaluating biases of such a model and ensuring factual generations are still very much open research problems.\nTherefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.*\n" st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
26
0
import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def UpperCamelCase ( __lowerCamelCase : Dict , __lowerCamelCase : Optional[int] ): snake_case : str = checkpoint snake_case : Tuple = {} snake_case : str = vae_state_dict["encoder.conv_in.weight"] snake_case : int = vae_state_dict["encoder.conv_in.bias"] snake_case : Optional[int] = vae_state_dict["encoder.conv_out.weight"] snake_case : Any = vae_state_dict["encoder.conv_out.bias"] snake_case : Tuple = vae_state_dict["encoder.norm_out.weight"] snake_case : Union[str, Any] = vae_state_dict["encoder.norm_out.bias"] snake_case : str = vae_state_dict["decoder.conv_in.weight"] snake_case : Tuple = vae_state_dict["decoder.conv_in.bias"] snake_case : Dict = vae_state_dict["decoder.conv_out.weight"] snake_case : Union[str, Any] = vae_state_dict["decoder.conv_out.bias"] snake_case : Optional[int] = vae_state_dict["decoder.norm_out.weight"] snake_case : Union[str, Any] = vae_state_dict["decoder.norm_out.bias"] snake_case : str = vae_state_dict["quant_conv.weight"] snake_case : Any = vae_state_dict["quant_conv.bias"] snake_case : List[str] = vae_state_dict["post_quant_conv.weight"] snake_case : Tuple = vae_state_dict["post_quant_conv.bias"] # Retrieves the keys for the encoder down blocks only snake_case : Tuple = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "encoder.down" in layer} ) snake_case : List[Any] = { layer_id: [key for key in vae_state_dict if f"""down.{layer_id}""" in key] for layer_id in range(__lowerCamelCase ) } # Retrieves the keys for the decoder up blocks only snake_case : Optional[int] = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "decoder.up" in layer} ) snake_case : Optional[Any] = { layer_id: [key for key in vae_state_dict if f"""up.{layer_id}""" in key] for layer_id in range(__lowerCamelCase ) } for i in range(__lowerCamelCase ): snake_case : Dict = [key for key in down_blocks[i] if f"""down.{i}""" in key and f"""down.{i}.downsample""" not in key] if f"""encoder.down.{i}.downsample.conv.weight""" in vae_state_dict: snake_case : Union[str, Any] = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.weight""" ) snake_case : Tuple = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.bias""" ) snake_case : Any = renew_vae_resnet_paths(__lowerCamelCase ) snake_case : Dict = {"old": f"""down.{i}.block""", "new": f"""down_blocks.{i}.resnets"""} assign_to_checkpoint(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , additional_replacements=[meta_path] , config=__lowerCamelCase ) snake_case : Tuple = [key for key in vae_state_dict if "encoder.mid.block" in key] snake_case : Tuple = 2 for i in range(1 , num_mid_res_blocks + 1 ): snake_case : Dict = [key for key in mid_resnets if f"""encoder.mid.block_{i}""" in key] snake_case : Dict = renew_vae_resnet_paths(__lowerCamelCase ) snake_case : Optional[int] = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , additional_replacements=[meta_path] , config=__lowerCamelCase ) snake_case : Optional[int] = [key for key in vae_state_dict if "encoder.mid.attn" in key] snake_case : Tuple = renew_vae_attention_paths(__lowerCamelCase ) snake_case : Tuple = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , additional_replacements=[meta_path] , config=__lowerCamelCase ) conv_attn_to_linear(__lowerCamelCase ) for i in range(__lowerCamelCase ): snake_case : int = num_up_blocks - 1 - i snake_case : Tuple = [ key for key in up_blocks[block_id] if f"""up.{block_id}""" in key and f"""up.{block_id}.upsample""" not in key ] if f"""decoder.up.{block_id}.upsample.conv.weight""" in vae_state_dict: snake_case : List[Any] = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.weight""" ] snake_case : int = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.bias""" ] snake_case : Tuple = renew_vae_resnet_paths(__lowerCamelCase ) snake_case : Tuple = {"old": f"""up.{block_id}.block""", "new": f"""up_blocks.{i}.resnets"""} assign_to_checkpoint(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , additional_replacements=[meta_path] , config=__lowerCamelCase ) snake_case : List[Any] = [key for key in vae_state_dict if "decoder.mid.block" in key] snake_case : List[Any] = 2 for i in range(1 , num_mid_res_blocks + 1 ): snake_case : Optional[Any] = [key for key in mid_resnets if f"""decoder.mid.block_{i}""" in key] snake_case : Optional[Any] = renew_vae_resnet_paths(__lowerCamelCase ) snake_case : List[str] = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , additional_replacements=[meta_path] , config=__lowerCamelCase ) snake_case : Tuple = [key for key in vae_state_dict if "decoder.mid.attn" in key] snake_case : Optional[int] = renew_vae_attention_paths(__lowerCamelCase ) snake_case : List[str] = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , additional_replacements=[meta_path] , config=__lowerCamelCase ) conv_attn_to_linear(__lowerCamelCase ) return new_checkpoint def UpperCamelCase ( __lowerCamelCase : str , __lowerCamelCase : str , ): snake_case : Optional[int] = requests.get( " https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml" ) snake_case : Union[str, Any] = io.BytesIO(r.content ) snake_case : Optional[int] = OmegaConf.load(__lowerCamelCase ) snake_case : str = 512 snake_case : str = "cuda" if torch.cuda.is_available() else "cpu" if checkpoint_path.endswith("safetensors" ): from safetensors import safe_open snake_case : Any = {} with safe_open(__lowerCamelCase , framework="pt" , device="cpu" ) as f: for key in f.keys(): snake_case : Dict = f.get_tensor(__lowerCamelCase ) else: snake_case : List[Any] = torch.load(__lowerCamelCase , map_location=__lowerCamelCase )["state_dict"] # Convert the VAE model. snake_case : List[str] = create_vae_diffusers_config(__lowerCamelCase , image_size=__lowerCamelCase ) snake_case : Optional[int] = custom_convert_ldm_vae_checkpoint(__lowerCamelCase , __lowerCamelCase ) snake_case : Any = AutoencoderKL(**__lowerCamelCase ) vae.load_state_dict(__lowerCamelCase ) vae.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": __lowerCamelCase = argparse.ArgumentParser() parser.add_argument("""--vae_pt_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") __lowerCamelCase = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
59
from typing import Dict, List, Optional, Tuple, Union import torch from ...models import AutoencoderKL, TransformeraDModel from ...schedulers import KarrasDiffusionSchedulers from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class A( UpperCamelCase ): '''simple docstring''' def __init__( self : str , A_ : TransformeraDModel , A_ : AutoencoderKL , A_ : KarrasDiffusionSchedulers , A_ : Optional[Dict[int, str]] = None , ) -> str: """simple docstring""" super().__init__() self.register_modules(transformer=A_ , vae=A_ , scheduler=A_ ) # create a imagenet -> id dictionary for easier use lowerCamelCase_ = {} if idalabel is not None: for key, value in idalabel.items(): for label in value.split(',' ): lowerCamelCase_ = int(A_ ) lowerCamelCase_ = dict(sorted(self.labels.items() ) ) def a__ ( self : Optional[int] , A_ : Union[str, List[str]] ) -> List[int]: """simple docstring""" if not isinstance(A_ , A_ ): lowerCamelCase_ = list(A_ ) for l in label: if l not in self.labels: raise ValueError( f"""{l} does not exist. Please make sure to select one of the following labels: \n {self.labels}.""" ) return [self.labels[l] for l in label] @torch.no_grad() def __call__( self : Any , A_ : List[int] , A_ : float = 4.0 , A_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , A_ : int = 50 , A_ : Optional[str] = "pil" , A_ : bool = True , ) -> Union[ImagePipelineOutput, Tuple]: """simple docstring""" lowerCamelCase_ = len(A_ ) lowerCamelCase_ = self.transformer.config.sample_size lowerCamelCase_ = self.transformer.config.in_channels lowerCamelCase_ = randn_tensor( shape=(batch_size, latent_channels, latent_size, latent_size) , generator=A_ , device=self.device , dtype=self.transformer.dtype , ) lowerCamelCase_ = torch.cat([latents] * 2 ) if guidance_scale > 1 else latents lowerCamelCase_ = torch.tensor(A_ , device=self.device ).reshape(-1 ) lowerCamelCase_ = torch.tensor([1000] * batch_size , device=self.device ) lowerCamelCase_ = torch.cat([class_labels, class_null] , 0 ) if guidance_scale > 1 else class_labels # set step values self.scheduler.set_timesteps(A_ ) for t in self.progress_bar(self.scheduler.timesteps ): if guidance_scale > 1: lowerCamelCase_ = latent_model_input[: len(A_ ) // 2] lowerCamelCase_ = torch.cat([half, half] , dim=0 ) lowerCamelCase_ = self.scheduler.scale_model_input(A_ , A_ ) lowerCamelCase_ = t if not torch.is_tensor(A_ ): # TODO: this requires sync between CPU and GPU. So try to pass timesteps as tensors if you can # This would be a good case for the `match` statement (Python 3.10+) lowerCamelCase_ = latent_model_input.device.type == 'mps' if isinstance(A_ , A_ ): lowerCamelCase_ = torch.floataa if is_mps else torch.floataa else: lowerCamelCase_ = torch.intaa if is_mps else torch.intaa lowerCamelCase_ = torch.tensor([timesteps] , dtype=A_ , device=latent_model_input.device ) elif len(timesteps.shape ) == 0: lowerCamelCase_ = timesteps[None].to(latent_model_input.device ) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML lowerCamelCase_ = timesteps.expand(latent_model_input.shape[0] ) # predict noise model_output lowerCamelCase_ = self.transformer( A_ , timestep=A_ , class_labels=A_ ).sample # perform guidance if guidance_scale > 1: lowerCamelCase_ , lowerCamelCase_ = noise_pred[:, :latent_channels], noise_pred[:, latent_channels:] lowerCamelCase_ , lowerCamelCase_ = torch.split(A_ , len(A_ ) // 2 , dim=0 ) lowerCamelCase_ = uncond_eps + guidance_scale * (cond_eps - uncond_eps) lowerCamelCase_ = torch.cat([half_eps, half_eps] , dim=0 ) lowerCamelCase_ = torch.cat([eps, rest] , dim=1 ) # learned sigma if self.transformer.config.out_channels // 2 == latent_channels: lowerCamelCase_ , lowerCamelCase_ = torch.split(A_ , A_ , dim=1 ) else: lowerCamelCase_ = noise_pred # compute previous image: x_t -> x_t-1 lowerCamelCase_ = self.scheduler.step(A_ , A_ , A_ ).prev_sample if guidance_scale > 1: lowerCamelCase_ , lowerCamelCase_ = latent_model_input.chunk(2 , dim=0 ) else: lowerCamelCase_ = latent_model_input lowerCamelCase_ = 1 / self.vae.config.scaling_factor * latents lowerCamelCase_ = self.vae.decode(A_ ).sample lowerCamelCase_ = (samples / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 lowerCamelCase_ = samples.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowerCamelCase_ = self.numpy_to_pil(A_ ) if not return_dict: return (samples,) return ImagePipelineOutput(images=A_ )
204
0
'''simple docstring''' # Author: OMKAR PATHAK, Nwachukwu Chidiebere # Use a Python dictionary to construct the graph. from __future__ import annotations from pprint import pformat from typing import Generic, TypeVar lowercase__ = TypeVar("T") class A_ ( Generic[T] ): '''simple docstring''' def __init__( self : List[Any] , lowercase_ : bool = True ) -> None: UpperCAmelCase : dict[T, list[T]] = {} # dictionary of lists UpperCAmelCase : Optional[int] = directed def UpperCAmelCase_ ( self : int , lowercase_ : T , lowercase_ : T ) -> GraphAdjacencyList[T]: if not self.directed: # For undirected graphs # if both source vertex and destination vertex are both present in the # adjacency list, add destination vertex to source vertex list of adjacent # vertices and add source vertex to destination vertex list of adjacent # vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase_ ) self.adj_list[destination_vertex].append(lowercase_ ) # if only source vertex is present in adjacency list, add destination vertex # to source vertex list of adjacent vertices, then create a new vertex with # destination vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase_ ) UpperCAmelCase : Dict = [source_vertex] # if only destination vertex is present in adjacency list, add source vertex # to destination vertex list of adjacent vertices, then create a new vertex # with source vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif destination_vertex in self.adj_list: self.adj_list[destination_vertex].append(lowercase_ ) UpperCAmelCase : str = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and assign a list # containing the destination vertex as it's first adjacent vertex also # create a new vertex with destination vertex as key and assign a list # containing the source vertex as it's first adjacent vertex. else: UpperCAmelCase : List[Any] = [destination_vertex] UpperCAmelCase : int = [source_vertex] else: # For directed graphs # if both source vertex and destination vertex are present in adjacency # list, add destination vertex to source vertex list of adjacent vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase_ ) # if only source vertex is present in adjacency list, add destination # vertex to source vertex list of adjacent vertices and create a new vertex # with destination vertex as key, which has no adjacent vertex elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase_ ) UpperCAmelCase : Dict = [] # if only destination vertex is present in adjacency list, create a new # vertex with source vertex as key and assign a list containing destination # vertex as first adjacent vertex elif destination_vertex in self.adj_list: UpperCAmelCase : Tuple = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and a list containing # destination vertex as it's first adjacent vertex. Then create a new vertex # with destination vertex as key, which has no adjacent vertex else: UpperCAmelCase : int = [destination_vertex] UpperCAmelCase : List[str] = [] return self def __repr__( self : Optional[Any] ) -> str: return pformat(self.adj_list )
280
'''simple docstring''' import copy from typing import Dict, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING from ..detr import DetrConfig from ..swin import SwinConfig lowercase__ = { "facebook/maskformer-swin-base-ade": ( "https://huggingface.co/facebook/maskformer-swin-base-ade/blob/main/config.json" ) # See all MaskFormer models at https://huggingface.co/models?filter=maskformer } lowercase__ = logging.get_logger(__name__) class A_ ( _snake_case ): '''simple docstring''' UpperCAmelCase_ : Any = """maskformer""" UpperCAmelCase_ : Union[str, Any] = {"""hidden_size""": """mask_feature_size"""} UpperCAmelCase_ : Union[str, Any] = ["""resnet""", """swin"""] UpperCAmelCase_ : Optional[int] = ["""detr"""] def __init__( self : Union[str, Any] , lowercase_ : int = 256 , lowercase_ : int = 256 , lowercase_ : float = 0.1 , lowercase_ : bool = False , lowercase_ : Optional[Dict] = None , lowercase_ : Optional[Dict] = None , lowercase_ : float = 0.02 , lowercase_ : float = 1.0 , lowercase_ : float = 1.0 , lowercase_ : float = 1.0 , lowercase_ : float = 20.0 , lowercase_ : Optional[bool] = None , **lowercase_ : int , ) -> Optional[Any]: if backbone_config is None: # fall back to https://huggingface.co/microsoft/swin-base-patch4-window12-384-in22k UpperCAmelCase : List[Any] = SwinConfig( image_size=384 , in_channels=3 , patch_size=4 , embed_dim=128 , depths=[2, 2, 18, 2] , num_heads=[4, 8, 16, 32] , window_size=12 , drop_path_rate=0.3 , out_features=['stage1', 'stage2', 'stage3', 'stage4'] , ) if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase : List[Any] = backbone_config.pop('model_type' ) UpperCAmelCase : Optional[Any] = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase : List[Any] = config_class.from_dict(lowercase_ ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( f"""Backbone {backbone_config.model_type} is not a supported model and may not be compatible with MaskFormer. """ f"""Supported model types: {",".join(self.backbones_supported )}""" ) if decoder_config is None: # fall back to https://huggingface.co/facebook/detr-resnet-50 UpperCAmelCase : Dict = DetrConfig() else: # verify that the decoder is supported UpperCAmelCase : Tuple = ( decoder_config.pop('model_type' ) if isinstance(lowercase_ , lowercase_ ) else decoder_config.model_type ) if decoder_type not in self.decoders_supported: raise ValueError( f"""Transformer Decoder {decoder_type} not supported, please use one of""" f""" {",".join(self.decoders_supported )}""" ) if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase : Union[str, Any] = CONFIG_MAPPING[decoder_type] UpperCAmelCase : str = config_class.from_dict(lowercase_ ) UpperCAmelCase : int = backbone_config UpperCAmelCase : Tuple = decoder_config # main feature dimension for the model UpperCAmelCase : Union[str, Any] = fpn_feature_size UpperCAmelCase : Optional[Any] = mask_feature_size # initializer UpperCAmelCase : Union[str, Any] = init_std UpperCAmelCase : List[str] = init_xavier_std # Hungarian matcher && loss UpperCAmelCase : int = cross_entropy_weight UpperCAmelCase : Tuple = dice_weight UpperCAmelCase : int = mask_weight UpperCAmelCase : Any = use_auxiliary_loss UpperCAmelCase : Tuple = no_object_weight UpperCAmelCase : str = output_auxiliary_logits UpperCAmelCase : Dict = self.decoder_config.encoder_attention_heads UpperCAmelCase : Union[str, Any] = self.decoder_config.num_hidden_layers super().__init__(**lowercase_ ) @classmethod def UpperCAmelCase_ ( cls : Any , lowercase_ : PretrainedConfig , lowercase_ : PretrainedConfig , **lowercase_ : Optional[Any] ) -> str: return cls( backbone_config=lowercase_ , decoder_config=lowercase_ , **lowercase_ , ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Dict[str, any]: UpperCAmelCase : Optional[int] = copy.deepcopy(self.__dict__ ) UpperCAmelCase : List[str] = self.backbone_config.to_dict() UpperCAmelCase : Any = self.decoder_config.to_dict() UpperCAmelCase : Union[str, Any] = self.__class__.model_type return output
280
1
import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging __lowerCamelCase : Optional[Any] = logging.get_logger(__name__) def _snake_case ( lowerCAmelCase : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = R"\w+[.]\d+" SCREAMING_SNAKE_CASE_ : List[str] = re.findall(lowerCAmelCase , lowerCAmelCase ) for pat in pats: SCREAMING_SNAKE_CASE_ : int = key.replace(lowerCAmelCase , "_".join(pat.split("." ) ) ) return key def _snake_case ( lowerCAmelCase : Optional[int] , lowerCAmelCase : Dict , lowerCAmelCase : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = pt_tuple_key[:-1] + ("scale",) if ( any("norm" in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = pt_tuple_key[:-1] + ("scale",) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: SCREAMING_SNAKE_CASE_ : int = pt_tuple_key[:-1] + ("scale",) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: SCREAMING_SNAKE_CASE_ : Tuple = pt_tuple_key[:-1] + ("embedding",) return renamed_pt_tuple_key, pt_tensor # conv layer SCREAMING_SNAKE_CASE_ : int = pt_tuple_key[:-1] + ("kernel",) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: SCREAMING_SNAKE_CASE_ : Optional[Any] = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer SCREAMING_SNAKE_CASE_ : Dict = pt_tuple_key[:-1] + ("kernel",) if pt_tuple_key[-1] == "weight": SCREAMING_SNAKE_CASE_ : str = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight SCREAMING_SNAKE_CASE_ : Tuple = pt_tuple_key[:-1] + ("weight",) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias SCREAMING_SNAKE_CASE_ : Optional[Any] = pt_tuple_key[:-1] + ("bias",) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def _snake_case ( lowerCAmelCase : Tuple , lowerCAmelCase : Optional[Any] , lowerCAmelCase : Optional[Any]=4_2 ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params SCREAMING_SNAKE_CASE_ : int = flax_model.init_weights(PRNGKey(lowerCAmelCase ) ) SCREAMING_SNAKE_CASE_ : int = flatten_dict(lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : List[str] = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): SCREAMING_SNAKE_CASE_ : Optional[Any] = rename_key(lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : int = tuple(renamed_pt_key.split("." ) ) # Correctly rename weight parameters SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : List[Any] = rename_key_and_reshape_tensor(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f'PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ' f'{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) # also add unexpected weight so that warning is thrown SCREAMING_SNAKE_CASE_ : int = jnp.asarray(lowerCAmelCase ) return unflatten_dict(lowerCAmelCase )
18
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class a__ ( A__ ): def __init__( self : Tuple,_A : Optional[int],_A : Any=13,_A : List[str]=7,_A : int=True,_A : Dict=True,_A : Dict=False,_A : List[Any]=True,_A : Any=99,_A : Optional[int]=32,_A : Any=5,_A : List[Any]=4,_A : Dict=64,_A : Optional[Any]="gelu",_A : Tuple=0.1,_A : Any=0.1,_A : List[Any]=512,_A : Dict=16,_A : Optional[Any]=2,_A : Union[str, Any]=0.02,_A : List[str]=3,_A : Optional[Any]=4,_A : Union[str, Any]=None,_A : Tuple=2,_A : List[str]=2,_A : str=2,_A : Dict=2,_A : Optional[Any]=4,_A : Union[str, Any]=1,): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = parent SCREAMING_SNAKE_CASE_ : Optional[int] = batch_size SCREAMING_SNAKE_CASE_ : Dict = seq_length SCREAMING_SNAKE_CASE_ : Dict = is_training SCREAMING_SNAKE_CASE_ : Optional[int] = use_input_mask SCREAMING_SNAKE_CASE_ : int = use_token_type_ids SCREAMING_SNAKE_CASE_ : Optional[int] = use_labels SCREAMING_SNAKE_CASE_ : Tuple = vocab_size SCREAMING_SNAKE_CASE_ : Any = hidden_size SCREAMING_SNAKE_CASE_ : Union[str, Any] = num_hidden_layers SCREAMING_SNAKE_CASE_ : Tuple = num_attention_heads SCREAMING_SNAKE_CASE_ : List[Any] = intermediate_size SCREAMING_SNAKE_CASE_ : List[str] = hidden_act SCREAMING_SNAKE_CASE_ : List[str] = hidden_dropout_prob SCREAMING_SNAKE_CASE_ : Optional[int] = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ : str = max_position_embeddings SCREAMING_SNAKE_CASE_ : str = type_vocab_size SCREAMING_SNAKE_CASE_ : List[str] = type_sequence_label_size SCREAMING_SNAKE_CASE_ : Optional[Any] = initializer_range SCREAMING_SNAKE_CASE_ : Tuple = num_labels SCREAMING_SNAKE_CASE_ : List[Any] = num_choices SCREAMING_SNAKE_CASE_ : Dict = scope SCREAMING_SNAKE_CASE_ : int = q_groups SCREAMING_SNAKE_CASE_ : Tuple = k_groups SCREAMING_SNAKE_CASE_ : List[Any] = v_groups SCREAMING_SNAKE_CASE_ : Tuple = post_attention_groups SCREAMING_SNAKE_CASE_ : int = intermediate_groups SCREAMING_SNAKE_CASE_ : List[Any] = output_groups def __UpperCamelCase ( self : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length],self.vocab_size ) SCREAMING_SNAKE_CASE_ : List[Any] = None if self.use_input_mask: SCREAMING_SNAKE_CASE_ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) SCREAMING_SNAKE_CASE_ : Optional[Any] = None SCREAMING_SNAKE_CASE_ : Any = None SCREAMING_SNAKE_CASE_ : str = None if self.use_labels: SCREAMING_SNAKE_CASE_ : str = ids_tensor([self.batch_size],self.type_sequence_label_size ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length],self.num_labels ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = ids_tensor([self.batch_size],self.num_choices ) SCREAMING_SNAKE_CASE_ : Optional[int] = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCamelCase ( self : str ): """simple docstring""" return SqueezeBertConfig( embedding_size=self.hidden_size,vocab_size=self.vocab_size,hidden_size=self.hidden_size,num_hidden_layers=self.num_hidden_layers,num_attention_heads=self.num_attention_heads,intermediate_size=self.intermediate_size,hidden_act=self.hidden_act,attention_probs_dropout_prob=self.hidden_dropout_prob,attention_dropout=self.attention_probs_dropout_prob,max_position_embeddings=self.max_position_embeddings,initializer_range=self.initializer_range,q_groups=self.q_groups,k_groups=self.k_groups,v_groups=self.v_groups,post_attention_groups=self.post_attention_groups,intermediate_groups=self.intermediate_groups,output_groups=self.output_groups,) def __UpperCamelCase ( self : Tuple,_A : Union[str, Any],_A : Union[str, Any],_A : int,_A : Optional[int],_A : Union[str, Any],_A : Any ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = SqueezeBertModel(config=_A ) model.to(_A ) model.eval() SCREAMING_SNAKE_CASE_ : Any = model(_A,_A ) SCREAMING_SNAKE_CASE_ : List[str] = model(_A ) self.parent.assertEqual(result.last_hidden_state.shape,(self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCamelCase ( self : Dict,_A : Any,_A : Tuple,_A : str,_A : Any,_A : Union[str, Any],_A : Any ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = SqueezeBertForMaskedLM(config=_A ) model.to(_A ) model.eval() SCREAMING_SNAKE_CASE_ : List[str] = model(_A,attention_mask=_A,labels=_A ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCamelCase ( self : Optional[int],_A : Union[str, Any],_A : Union[str, Any],_A : Any,_A : Tuple,_A : int,_A : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = SqueezeBertForQuestionAnswering(config=_A ) model.to(_A ) model.eval() SCREAMING_SNAKE_CASE_ : Dict = model( _A,attention_mask=_A,start_positions=_A,end_positions=_A ) self.parent.assertEqual(result.start_logits.shape,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape,(self.batch_size, self.seq_length) ) def __UpperCamelCase ( self : List[Any],_A : List[str],_A : Tuple,_A : List[Any],_A : List[str],_A : List[str],_A : List[str] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = self.num_labels SCREAMING_SNAKE_CASE_ : List[str] = SqueezeBertForSequenceClassification(_A ) model.to(_A ) model.eval() SCREAMING_SNAKE_CASE_ : Dict = model(_A,attention_mask=_A,labels=_A ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.num_labels) ) def __UpperCamelCase ( self : str,_A : Optional[int],_A : str,_A : List[Any],_A : List[str],_A : str,_A : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = self.num_labels SCREAMING_SNAKE_CASE_ : Optional[int] = SqueezeBertForTokenClassification(config=_A ) model.to(_A ) model.eval() SCREAMING_SNAKE_CASE_ : Optional[int] = model(_A,attention_mask=_A,labels=_A ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.seq_length, self.num_labels) ) def __UpperCamelCase ( self : List[Any],_A : Tuple,_A : str,_A : Optional[Any],_A : int,_A : str,_A : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = self.num_choices SCREAMING_SNAKE_CASE_ : Union[str, Any] = SqueezeBertForMultipleChoice(config=_A ) model.to(_A ) model.eval() SCREAMING_SNAKE_CASE_ : Dict = input_ids.unsqueeze(1 ).expand(-1,self.num_choices,-1 ).contiguous() SCREAMING_SNAKE_CASE_ : str = input_mask.unsqueeze(1 ).expand(-1,self.num_choices,-1 ).contiguous() SCREAMING_SNAKE_CASE_ : Optional[int] = model( _A,attention_mask=_A,labels=_A,) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.num_choices) ) def __UpperCamelCase ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = self.prepare_config_and_inputs() ((SCREAMING_SNAKE_CASE_) , (SCREAMING_SNAKE_CASE_) , (SCREAMING_SNAKE_CASE_) , (SCREAMING_SNAKE_CASE_) , (SCREAMING_SNAKE_CASE_) , (SCREAMING_SNAKE_CASE_)) : Dict = config_and_inputs SCREAMING_SNAKE_CASE_ : Dict = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class a__ ( A__ , A__ , unittest.TestCase ): A = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) A = ( { 'feature-extraction': SqueezeBertModel, 'fill-mask': SqueezeBertForMaskedLM, 'question-answering': SqueezeBertForQuestionAnswering, 'text-classification': SqueezeBertForSequenceClassification, 'token-classification': SqueezeBertForTokenClassification, 'zero-shot': SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) A = False A = True A = False def __UpperCamelCase ( self : int ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = SqueezeBertModelTester(self ) SCREAMING_SNAKE_CASE_ : List[str] = ConfigTester(self,config_class=_A,dim=37 ) def __UpperCamelCase ( self : List[str] ): """simple docstring""" self.config_tester.run_common_tests() def __UpperCamelCase ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*_A ) def __UpperCamelCase ( self : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*_A ) def __UpperCamelCase ( self : int ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*_A ) def __UpperCamelCase ( self : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*_A ) def __UpperCamelCase ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*_A ) def __UpperCamelCase ( self : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*_A ) @slow def __UpperCamelCase ( self : Any ): """simple docstring""" for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE_ : Tuple = SqueezeBertModel.from_pretrained(_A ) self.assertIsNotNone(_A ) @require_sentencepiece @require_tokenizers @require_torch class a__ ( unittest.TestCase ): @slow def __UpperCamelCase ( self : int ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = SqueezeBertForSequenceClassification.from_pretrained("squeezebert/squeezebert-mnli" ) SCREAMING_SNAKE_CASE_ : Optional[int] = torch.tensor([[1, 2_9414, 232, 328, 740, 1140, 1_2695, 69, 13, 1588, 2]] ) SCREAMING_SNAKE_CASE_ : List[Any] = model(_A )[0] SCREAMING_SNAKE_CASE_ : Union[str, Any] = torch.Size((1, 3) ) self.assertEqual(output.shape,_A ) SCREAMING_SNAKE_CASE_ : int = torch.tensor([[0.6401, -0.0349, -0.6041]] ) self.assertTrue(torch.allclose(_A,_A,atol=1E-4 ) )
18
1
'''simple docstring''' import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py _a : Optional[int] = "src/diffusers" _a : int = "." # This is to make sure the diffusers module imported is the one in the repo. _a : List[Any] = importlib.util.spec_from_file_location( """diffusers""", os.path.join(DIFFUSERS_PATH, """__init__.py"""), submodule_search_locations=[DIFFUSERS_PATH], ) _a : Tuple = spec.loader.load_module() def _lowerCAmelCase ( lowercase , lowercase ) -> Union[str, Any]: return line.startswith(lowerCAmelCase__ ) or len(lowerCAmelCase__ ) <= 1 or re.search(R"""^\s*\)(\s*->.*:|:)\s*$""" , lowerCAmelCase__ ) is not None def _lowerCAmelCase ( lowercase ) -> str: __lowerCAmelCase = object_name.split(""".""" ) __lowerCAmelCase = 0 # First let's find the module where our object lives. __lowerCAmelCase = parts[i] while i < len(lowerCAmelCase__ ) and not os.path.isfile(os.path.join(lowerCAmelCase__ , f'{module}.py' ) ): i += 1 if i < len(lowerCAmelCase__ ): __lowerCAmelCase = os.path.join(lowerCAmelCase__ , parts[i] ) if i >= len(lowerCAmelCase__ ): raise ValueError(f'`object_name` should begin with the name of a module of diffusers but got {object_name}.' ) with open(os.path.join(lowerCAmelCase__ , f'{module}.py' ) , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: __lowerCAmelCase = f.readlines() # Now let's find the class / func in the code! __lowerCAmelCase = """""" __lowerCAmelCase = 0 for name in parts[i + 1 :]: while ( line_index < len(lowerCAmelCase__ ) and re.search(Rf'^{indent}(class|def)\s+{name}(\(|\:)' , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(lowerCAmelCase__ ): raise ValueError(f' {object_name} does not match any function or class in {module}.' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). __lowerCAmelCase = line_index while line_index < len(lowerCAmelCase__ ) and _should_continue(lines[line_index] , lowerCAmelCase__ ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 __lowerCAmelCase = lines[start_index:line_index] return "".join(lowerCAmelCase__ ) _a : Union[str, Any] = re.compile(r"""^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)""") _a : Tuple = re.compile(r"""^\s*(\S+)->(\S+)(\s+.*|$)""") _a : Dict = re.compile(r"""<FILL\s+[^>]*>""") def _lowerCAmelCase ( lowercase ) -> int: __lowerCAmelCase = code.split("""\n""" ) __lowerCAmelCase = 0 while idx < len(lowerCAmelCase__ ) and len(lines[idx] ) == 0: idx += 1 if idx < len(lowerCAmelCase__ ): return re.search(R"""^(\s*)\S""" , lines[idx] ).groups()[0] return "" def _lowerCAmelCase ( lowercase ) -> Dict: __lowerCAmelCase = len(get_indent(lowerCAmelCase__ ) ) > 0 if has_indent: __lowerCAmelCase = f'class Bla:\n{code}' __lowerCAmelCase = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=lowerCAmelCase__ ) __lowerCAmelCase = black.format_str(lowerCAmelCase__ , mode=lowerCAmelCase__ ) __lowerCAmelCase , __lowerCAmelCase = style_docstrings_in_code(lowerCAmelCase__ ) return result[len("""class Bla:\n""" ) :] if has_indent else result def _lowerCAmelCase ( lowercase , lowercase=False ) -> Union[str, Any]: with open(lowerCAmelCase__ , """r""" , encoding="""utf-8""" , newline="""\n""" ) as f: __lowerCAmelCase = f.readlines() __lowerCAmelCase = [] __lowerCAmelCase = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(lowerCAmelCase__ ): __lowerCAmelCase = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = search.groups() __lowerCAmelCase = find_code_in_diffusers(lowerCAmelCase__ ) __lowerCAmelCase = get_indent(lowerCAmelCase__ ) __lowerCAmelCase = line_index + 1 if indent == theoretical_indent else line_index + 2 __lowerCAmelCase = theoretical_indent __lowerCAmelCase = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. __lowerCAmelCase = True while line_index < len(lowerCAmelCase__ ) and should_continue: line_index += 1 if line_index >= len(lowerCAmelCase__ ): break __lowerCAmelCase = lines[line_index] __lowerCAmelCase = _should_continue(lowerCAmelCase__ , lowerCAmelCase__ ) and re.search(f'^{indent}# End copy' , lowerCAmelCase__ ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 __lowerCAmelCase = lines[start_index:line_index] __lowerCAmelCase = """""".join(lowerCAmelCase__ ) # Remove any nested `Copied from` comments to avoid circular copies __lowerCAmelCase = [line for line in theoretical_code.split("""\n""" ) if _re_copy_warning.search(lowerCAmelCase__ ) is None] __lowerCAmelCase = """\n""".join(lowerCAmelCase__ ) # Before comparing, use the `replace_pattern` on the original code. if len(lowerCAmelCase__ ) > 0: __lowerCAmelCase = replace_pattern.replace("""with""" , """""" ).split(""",""" ) __lowerCAmelCase = [_re_replace_pattern.search(lowerCAmelCase__ ) for p in patterns] for pattern in patterns: if pattern is None: continue __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = pattern.groups() __lowerCAmelCase = re.sub(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) if option.strip() == "all-casing": __lowerCAmelCase = re.sub(obja.lower() , obja.lower() , lowerCAmelCase__ ) __lowerCAmelCase = re.sub(obja.upper() , obja.upper() , lowerCAmelCase__ ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line __lowerCAmelCase = blackify(lines[start_index - 1] + theoretical_code ) __lowerCAmelCase = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: __lowerCAmelCase = lines[:start_index] + [theoretical_code] + lines[line_index:] __lowerCAmelCase = start_index + 1 if overwrite and len(lowerCAmelCase__ ) > 0: # Warn the user a file has been modified. print(f'Detected changes, rewriting {filename}.' ) with open(lowerCAmelCase__ , """w""" , encoding="""utf-8""" , newline="""\n""" ) as f: f.writelines(lowerCAmelCase__ ) return diffs def _lowerCAmelCase ( lowercase = False ) -> Any: __lowerCAmelCase = glob.glob(os.path.join(lowerCAmelCase__ , """**/*.py""" ) , recursive=lowerCAmelCase__ ) __lowerCAmelCase = [] for filename in all_files: __lowerCAmelCase = is_copy_consistent(lowerCAmelCase__ , lowerCAmelCase__ ) diffs += [f'- {filename}: copy does not match {d[0]} at line {d[1]}' for d in new_diffs] if not overwrite and len(lowerCAmelCase__ ) > 0: __lowerCAmelCase = """\n""".join(lowerCAmelCase__ ) raise Exception( """Found the following copy inconsistencies:\n""" + diff + """\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.""" ) if __name__ == "__main__": _a : List[str] = argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") _a : Optional[Any] = parser.parse_args() check_copies(args.fix_and_overwrite)
365
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a : List[str] = """▁""" _a : Optional[int] = {"""vocab_file""": """spiece.model"""} _a : int = { """vocab_file""": {"""google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"""} } _a : int = { """google/pegasus-xsum""": 5_1_2, } _a : List[Any] = logging.get_logger(__name__) class _UpperCAmelCase ( lowerCAmelCase_ ): a : List[Any] =VOCAB_FILES_NAMES a : Tuple =VOCAB_FILES_NAMES a : Any =PRETRAINED_VOCAB_FILES_MAP a : Dict =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a : List[Any] =["""input_ids""", """attention_mask"""] def __init__( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE="<pad>",__SCREAMING_SNAKE_CASE="</s>",__SCREAMING_SNAKE_CASE="<unk>",__SCREAMING_SNAKE_CASE="<mask_2>",__SCREAMING_SNAKE_CASE="<mask_1>",__SCREAMING_SNAKE_CASE=None,__SCREAMING_SNAKE_CASE=1_03,__SCREAMING_SNAKE_CASE = None,**__SCREAMING_SNAKE_CASE,): '''simple docstring''' __lowerCAmelCase = offset if additional_special_tokens is not None: if not isinstance(__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE ): raise TypeError( f'additional_special_tokens should be of type {type(__SCREAMING_SNAKE_CASE )}, but is' f' {type(__SCREAMING_SNAKE_CASE )}' ) __lowerCAmelCase = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'<unk_{i}>' for i in range(len(__SCREAMING_SNAKE_CASE ),self.offset - 1 ) ] if len(set(__SCREAMING_SNAKE_CASE ) ) != len(__SCREAMING_SNAKE_CASE ): raise ValueError( """Please make sure that the provided additional_special_tokens do not contain an incorrectly""" f' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.' ) __lowerCAmelCase = additional_special_tokens_extended else: __lowerCAmelCase = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'<unk_{i}>' for i in range(2,self.offset )] __lowerCAmelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=__SCREAMING_SNAKE_CASE,unk_token=__SCREAMING_SNAKE_CASE,mask_token=__SCREAMING_SNAKE_CASE,pad_token=__SCREAMING_SNAKE_CASE,mask_token_sent=__SCREAMING_SNAKE_CASE,offset=__SCREAMING_SNAKE_CASE,additional_special_tokens=__SCREAMING_SNAKE_CASE,sp_model_kwargs=self.sp_model_kwargs,**__SCREAMING_SNAKE_CASE,) __lowerCAmelCase = mask_token_sent __lowerCAmelCase = vocab_file __lowerCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__SCREAMING_SNAKE_CASE ) # add special tokens to encoder dict __lowerCAmelCase = { 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1,self.offset - 1 )} ) __lowerCAmelCase = {v: k for k, v in self.encoder.items()} @property def lowerCamelCase__ ( self ): '''simple docstring''' return len(self.sp_model ) + self.offset def lowerCamelCase__ ( self ): '''simple docstring''' __lowerCAmelCase = {self.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): '''simple docstring''' __lowerCAmelCase = self.__dict__.copy() __lowerCAmelCase = None return state def __setstate__( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase = d # for backward compatibility if not hasattr(self,"""sp_model_kwargs""" ): __lowerCAmelCase = {} __lowerCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' return self.sp_model.encode(__SCREAMING_SNAKE_CASE,out_type=__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] __lowerCAmelCase = self.sp_model.piece_to_id(__SCREAMING_SNAKE_CASE ) return sp_id + self.offset def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: __lowerCAmelCase = self.sp_model.IdToPiece(index - self.offset ) return token def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase = [] __lowerCAmelCase = """""" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__SCREAMING_SNAKE_CASE ) + token __lowerCAmelCase = [] else: current_sub_tokens.append(__SCREAMING_SNAKE_CASE ) out_string += self.sp_model.decode(__SCREAMING_SNAKE_CASE ) return out_string.strip() def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE=False ): '''simple docstring''' return 1 def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCAmelCase = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE = None,__SCREAMING_SNAKE_CASE = False ): '''simple docstring''' if already_has_special_tokens: return self._special_token_mask(__SCREAMING_SNAKE_CASE ) elif token_ids_a is None: return self._special_token_mask(__SCREAMING_SNAKE_CASE ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE=None ): '''simple docstring''' if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def lowerCamelCase__ ( self,__SCREAMING_SNAKE_CASE,__SCREAMING_SNAKE_CASE = None ): '''simple docstring''' if not os.path.isdir(__SCREAMING_SNAKE_CASE ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return __lowerCAmelCase = os.path.join( __SCREAMING_SNAKE_CASE,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__SCREAMING_SNAKE_CASE ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file,__SCREAMING_SNAKE_CASE ) elif not os.path.isfile(self.vocab_file ): with open(__SCREAMING_SNAKE_CASE,"""wb""" ) as fi: __lowerCAmelCase = self.sp_model.serialized_model_proto() fi.write(__SCREAMING_SNAKE_CASE ) return (out_vocab_file,)
46
0