code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import os # Precomputes a list of the 100 first triangular numbers _lowerCamelCase = [int(0.5 * n * (n + 1)) for n in range(1, 101)] def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = os.path.dirname(os.path.realpath(UpperCamelCase__ ) ) SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , """words.txt""" ) SCREAMING_SNAKE_CASE__ = """""" with open(UpperCamelCase__ ) as f: SCREAMING_SNAKE_CASE__ = f.readline() SCREAMING_SNAKE_CASE__ = [word.strip("""\"""" ) for word in words.strip("""\r\n""" ).split(""",""" )] SCREAMING_SNAKE_CASE__ = [ word for word in [sum(ord(UpperCamelCase__ ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(UpperCamelCase__ ) if __name__ == "__main__": print(solution())
59
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] ): SCREAMING_SNAKE_CASE__ = [] for data in source_data: for i, el in enumerate(UpperCamelCase__ ): if len(UpperCamelCase__ ) < i + 1: data_lists.append([] ) data_lists[i].append(float(UpperCamelCase__ ) ) return data_lists def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] , UpperCamelCase__: list[int] ): SCREAMING_SNAKE_CASE__ = [] for dlist, weight in zip(UpperCamelCase__ , UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = min(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = max(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = [] # for weight 0 score is 1 - actual score if weight == 0: for item in dlist: try: score.append(1 - ((item - mind) / (maxd - mind)) ) except ZeroDivisionError: score.append(1 ) elif weight == 1: for item in dlist: try: score.append((item - mind) / (maxd - mind) ) except ZeroDivisionError: score.append(0 ) # weight not 0 or 1 else: SCREAMING_SNAKE_CASE__ = f'''Invalid weight of {weight:f} provided''' raise ValueError(UpperCamelCase__ ) score_lists.append(UpperCamelCase__ ) return score_lists def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] ): SCREAMING_SNAKE_CASE__ = [0 for i in range(len(score_lists[0] ) )] for slist in score_lists: for j, ele in enumerate(UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = final_scores[j] + ele return final_scores def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] , UpperCamelCase__: list[int] ): SCREAMING_SNAKE_CASE__ = get_data(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = calculate_each_score(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = generate_final_scores(UpperCamelCase__ ) # append scores to source data for i, ele in enumerate(UpperCamelCase__ ): source_data[i].append(UpperCamelCase__ ) return source_data
59
1
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = generate_pascal_triangle(UpperCamelCase__ ) for row_idx in range(UpperCamelCase__ ): # Print left spaces for _ in range(num_rows - row_idx - 1 ): print(end=""" """ ) # Print row values for col_idx in range(row_idx + 1 ): if col_idx != row_idx: print(triangle[row_idx][col_idx] , end=""" """ ) else: print(triangle[row_idx][col_idx] , end="""""" ) print() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise TypeError("""The input value of 'num_rows' should be 'int'""" ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( """The input value of 'num_rows' should be greater than or equal to 0""" ) SCREAMING_SNAKE_CASE__ = [] for current_row_idx in range(UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = populate_current_row(UpperCamelCase__ , UpperCamelCase__ ) triangle.append(UpperCamelCase__ ) return triangle def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[int]] , UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = [-1] * (current_row_idx + 1) # first and last elements of current row are equal to 1 SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 1, 1 for current_col_idx in range(1 , UpperCamelCase__ ): calculate_current_element( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) return current_row def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[int]] , UpperCamelCase__: list[int] , UpperCamelCase__: int , UpperCamelCase__: int , ): SCREAMING_SNAKE_CASE__ = triangle[current_row_idx - 1][current_col_idx - 1] SCREAMING_SNAKE_CASE__ = triangle[current_row_idx - 1][current_col_idx] SCREAMING_SNAKE_CASE__ = above_to_left_elt + above_to_right_elt def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise TypeError("""The input value of 'num_rows' should be 'int'""" ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( """The input value of 'num_rows' should be greater than or equal to 0""" ) SCREAMING_SNAKE_CASE__ = [[1]] for row_index in range(1 , UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = [0] + result[-1] + [0] SCREAMING_SNAKE_CASE__ = row_index + 1 # Calculate the number of distinct elements in a row SCREAMING_SNAKE_CASE__ = sum(divmod(UpperCamelCase__ , 2 ) ) SCREAMING_SNAKE_CASE__ = [ temp_row[i - 1] + temp_row[i] for i in range(1 , distinct_elements + 1 ) ] SCREAMING_SNAKE_CASE__ = row_first_half[: (row_index + 1) // 2] row_second_half.reverse() SCREAMING_SNAKE_CASE__ = row_first_half + row_second_half result.append(UpperCamelCase__ ) return result def SCREAMING_SNAKE_CASE__ ( ): from collections.abc import Callable from timeit import timeit def benchmark_a_function(UpperCamelCase__: Callable , UpperCamelCase__: int ) -> None: SCREAMING_SNAKE_CASE__ = f'''{func.__name__}({value})''' SCREAMING_SNAKE_CASE__ = timeit(f'''__main__.{call}''' , setup="""import __main__""" ) # print(f"{call:38} = {func(value)} -- {timing:.4f} seconds") print(f'''{call:38} -- {timing:.4f} seconds''' ) for value in range(15 ): # (1, 7, 14): for func in (generate_pascal_triangle, generate_pascal_triangle_optimized): benchmark_a_function(UpperCamelCase__ , UpperCamelCase__ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
59
import warnings from functools import wraps from typing import Callable def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Callable ): @wraps(UpperCamelCase__ ) def _inner_fn(*UpperCamelCase__: Dict , **UpperCamelCase__: Any ): warnings.warn( (f'''\'{fn.__name__}\' is experimental and might be subject to breaking changes in the future.''') , UpperCamelCase__ , ) return fn(*UpperCamelCase__ , **UpperCamelCase__ ) return _inner_fn
59
1
import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartaaTokenizer, MBartaaTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from ...test_tokenization_common import TokenizerTesterMixin _lowerCamelCase = get_tests_dir('fixtures/test_sentencepiece.model') if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right _lowerCamelCase = 250004 _lowerCamelCase = 250020 @require_sentencepiece @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = MBartaaTokenizer lowerCamelCase_ = MBartaaTokenizerFast lowerCamelCase_ = True lowerCamelCase_ = True def _snake_case ( self :str ) -> Any: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing SCREAMING_SNAKE_CASE__ = MBartaaTokenizer(__A , src_lang="""en_XX""" , tgt_lang="""ro_RO""" , keep_accents=__A ) tokenizer.save_pretrained(self.tmpdirname ) def _snake_case ( self :Optional[int] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = """<s>""" SCREAMING_SNAKE_CASE__ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__A ) , __A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__A ) , __A ) def _snake_case ( self :Union[str, Any] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(__A ) , 1054 ) def _snake_case ( self :Optional[Any] ) -> int: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1054 ) def _snake_case ( self :int ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = MBartaaTokenizer(__A , src_lang="""en_XX""" , tgt_lang="""ro_RO""" , keep_accents=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(__A , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__A ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) SCREAMING_SNAKE_CASE__ = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( __A , [SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """."""] , ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_ids(__A ) self.assertListEqual( __A , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_ids_to_tokens(__A ) self.assertListEqual( __A , [SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """."""] , ) @slow def _snake_case ( self :Union[str, Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = {"""input_ids""": [[25_0004, 1_1062, 8_2772, 7, 15, 8_2772, 538, 5_1529, 237, 1_7198, 1290, 206, 9, 21_5175, 1314, 136, 1_7198, 1290, 206, 9, 5_6359, 42, 12_2009, 9, 1_6466, 16, 8_7344, 4537, 9, 4717, 7_8381, 6, 15_9958, 7, 15, 2_4480, 618, 4, 527, 2_2693, 5428, 4, 2777, 2_4480, 9874, 4, 4_3523, 594, 4, 803, 1_8392, 3_3189, 18, 4, 4_3523, 2_4447, 1_2399, 100, 2_4955, 8_3658, 9626, 14_4057, 15, 839, 2_2335, 16, 136, 2_4955, 8_3658, 8_3479, 15, 3_9102, 724, 16, 678, 645, 2789, 1328, 4589, 42, 12_2009, 11_5774, 23, 805, 1328, 4_6876, 7, 136, 5_3894, 1940, 4_2227, 4_1159, 1_7721, 823, 425, 4, 2_7512, 9_8722, 206, 136, 5531, 4970, 919, 1_7336, 5, 2], [25_0004, 2_0080, 618, 83, 8_2775, 47, 479, 9, 1517, 73, 5_3894, 333, 8_0581, 11_0117, 1_8811, 5256, 1295, 51, 15_2526, 297, 7986, 390, 12_4416, 538, 3_5431, 214, 98, 1_5044, 2_5737, 136, 7108, 4_3701, 23, 756, 13_5355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [25_0004, 581, 6_3773, 11_9455, 6, 14_7797, 8_8203, 7, 645, 70, 21, 3285, 1_0269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__A , model_name="""facebook/mbart-large-50""" , revision="""d3913889c59cd5c9e456b269c376325eabad57e2""" , ) def _snake_case ( self :Optional[int] ) -> int: """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return SCREAMING_SNAKE_CASE__ = (self.rust_tokenizer_class, """hf-internal-testing/tiny-random-mbart50""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = tempfile.mkdtemp() SCREAMING_SNAKE_CASE__ = tokenizer_r.save_pretrained(__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.save_pretrained(__A ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) SCREAMING_SNAKE_CASE__ = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f ) self.assertSequenceEqual(__A , __A ) # Checks everything loads correctly in the same way SCREAMING_SNAKE_CASE__ = tokenizer_r.from_pretrained(__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.from_pretrained(__A ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__A , __A ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(__A ) # Save tokenizer rust, legacy_format=True SCREAMING_SNAKE_CASE__ = tempfile.mkdtemp() SCREAMING_SNAKE_CASE__ = tokenizer_r.save_pretrained(__A , legacy_format=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.save_pretrained(__A ) # Checks it save with the same files self.assertSequenceEqual(__A , __A ) # Checks everything loads correctly in the same way SCREAMING_SNAKE_CASE__ = tokenizer_r.from_pretrained(__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.from_pretrained(__A ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__A , __A ) ) shutil.rmtree(__A ) # Save tokenizer rust, legacy_format=False SCREAMING_SNAKE_CASE__ = tempfile.mkdtemp() SCREAMING_SNAKE_CASE__ = tokenizer_r.save_pretrained(__A , legacy_format=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.save_pretrained(__A ) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way SCREAMING_SNAKE_CASE__ = tokenizer_r.from_pretrained(__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.from_pretrained(__A ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__A , __A ) ) shutil.rmtree(__A ) @require_torch @require_sentencepiece @require_tokenizers class UpperCamelCase_ ( unittest.TestCase ): lowerCamelCase_ = "facebook/mbart-large-50-one-to-many-mmt" lowerCamelCase_ = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] lowerCamelCase_ = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] lowerCamelCase_ = [EN_CODE, 82_74, 12_78_73, 2_59_16, 7, 86_22, 20_71, 4_38, 6_74_85, 53, 18_78_95, 23, 5_17_12, 2] @classmethod def _snake_case ( cls :Tuple ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = MBartaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="""en_XX""" , tgt_lang="""ro_RO""" ) SCREAMING_SNAKE_CASE__ = 1 return cls def _snake_case ( self :Any ) -> str: """simple docstring""" self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""ar_AR"""] , 25_0001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""en_EN"""] , 25_0004 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""ro_RO"""] , 25_0020 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""mr_IN"""] , 25_0038 ) def _snake_case ( self :Dict ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , __A ) def _snake_case ( self :int ) -> Tuple: """simple docstring""" self.assertIn(__A , self.tokenizer.all_special_ids ) SCREAMING_SNAKE_CASE__ = [RO_CODE, 884, 9019, 96, 9, 916, 8_6792, 36, 1_8743, 1_5596, 5, 2] SCREAMING_SNAKE_CASE__ = self.tokenizer.decode(__A , skip_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=__A ) self.assertEqual(__A , __A ) self.assertNotIn(self.tokenizer.eos_token , __A ) def _snake_case ( self :Tuple ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""this is gunna be a long sentence """ * 20] assert isinstance(src_text[0] , __A ) SCREAMING_SNAKE_CASE__ = 10 SCREAMING_SNAKE_CASE__ = self.tokenizer(__A , max_length=__A , truncation=__A ).input_ids[0] self.assertEqual(ids[0] , __A ) self.assertEqual(ids[-1] , 2 ) self.assertEqual(len(__A ) , __A ) def _snake_case ( self :List[Any] ) -> List[str]: """simple docstring""" self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["""<mask>""", """ar_AR"""] ) , [25_0053, 25_0001] ) def _snake_case ( self :Optional[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = tempfile.mkdtemp() SCREAMING_SNAKE_CASE__ = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(__A ) SCREAMING_SNAKE_CASE__ = MBartaaTokenizer.from_pretrained(__A ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , __A ) @require_torch def _snake_case ( self :str ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=__A , return_tensors="""pt""" ) SCREAMING_SNAKE_CASE__ = shift_tokens_right(batch["""labels"""] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == RO_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2].tolist() == [2, RO_CODE] @require_torch def _snake_case ( self :Optional[int] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=__A , truncation=__A , max_length=len(self.expected_src_tokens ) , return_tensors="""pt""" , ) SCREAMING_SNAKE_CASE__ = shift_tokens_right(batch["""labels"""] , self.tokenizer.pad_token_id ) self.assertIsInstance(__A , __A ) self.assertEqual((2, 14) , batch.input_ids.shape ) self.assertEqual((2, 14) , batch.attention_mask.shape ) SCREAMING_SNAKE_CASE__ = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , __A ) self.assertEqual(2 , batch.decoder_input_ids[0, 0] ) # decoder_start_token_id # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) def _snake_case ( self :Dict ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer(self.src_text , padding=__A , truncation=__A , max_length=3 , return_tensors="""pt""" ) SCREAMING_SNAKE_CASE__ = self.tokenizer( text_target=self.tgt_text , padding=__A , truncation=__A , max_length=10 , return_tensors="""pt""" ) SCREAMING_SNAKE_CASE__ = targets["""input_ids"""] SCREAMING_SNAKE_CASE__ = shift_tokens_right(__A , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def _snake_case ( self :Dict ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer._build_translation_inputs( """A test""" , return_tensors="""pt""" , src_lang="""en_XX""" , tgt_lang="""ar_AR""" ) self.assertEqual( nested_simplify(__A ) , { # en_XX, A, test, EOS """input_ids""": [[25_0004, 62, 3034, 2]], """attention_mask""": [[1, 1, 1, 1]], # ar_AR """forced_bos_token_id""": 25_0001, } , )
59
import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = RoCBertTokenizer lowerCamelCase_ = None lowerCamelCase_ = False lowerCamelCase_ = True lowerCamelCase_ = filter_non_english def _snake_case ( self :List[Any] ) -> List[Any]: """simple docstring""" super().setUp() SCREAMING_SNAKE_CASE__ = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """你""", """好""", """是""", """谁""", """a""", """b""", """c""", """d"""] SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = {} for i, value in enumerate(__A ): SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""word_shape_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""word_pronunciation_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) with open(self.word_shape_file , """w""" , encoding="""utf-8""" ) as word_shape_writer: json.dump(__A , __A , ensure_ascii=__A ) with open(self.word_pronunciation_file , """w""" , encoding="""utf-8""" ) as word_pronunciation_writer: json.dump(__A , __A , ensure_ascii=__A ) def _snake_case ( self :List[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) SCREAMING_SNAKE_CASE__ = tokenizer.tokenize("""你好[SEP]你是谁""" ) self.assertListEqual(__A , ["""你""", """好""", """[SEP]""", """你""", """是""", """谁"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_shape_ids(__A ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_pronunciation_ids(__A ) , [5, 6, 2, 5, 7, 8] ) def _snake_case ( self :List[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer() self.assertListEqual(tokenizer.tokenize("""ah\u535A\u63A8zz""" ) , ["""ah""", """\u535A""", """\u63A8""", """zz"""] ) def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hällo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""h\u00E9llo"""] ) def _snake_case ( self :Any ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :List[str] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :Union[str, Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :int ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HäLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :Union[str, Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HaLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :List[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , never_split=["""[UNK]"""] ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? [UNK]""" ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?""", """[UNK]"""] ) def _snake_case ( self :Any ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing"""] SCREAMING_SNAKE_CASE__ = {} for i, token in enumerate(__A ): SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = RoCBertWordpieceTokenizer(vocab=__A , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""unwanted running""" ) , ["""un""", """##want""", """##ed""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.tokenize("""unwantedX running""" ) , ["""[UNK]""", """runn""", """##ing"""] ) def _snake_case ( self :Any ) -> str: """simple docstring""" self.assertTrue(_is_whitespace(""" """ ) ) self.assertTrue(_is_whitespace("""\t""" ) ) self.assertTrue(_is_whitespace("""\r""" ) ) self.assertTrue(_is_whitespace("""\n""" ) ) self.assertTrue(_is_whitespace("""\u00A0""" ) ) self.assertFalse(_is_whitespace("""A""" ) ) self.assertFalse(_is_whitespace("""-""" ) ) def _snake_case ( self :int ) -> str: """simple docstring""" self.assertTrue(_is_control("""\u0005""" ) ) self.assertFalse(_is_control("""A""" ) ) self.assertFalse(_is_control(""" """ ) ) self.assertFalse(_is_control("""\t""" ) ) self.assertFalse(_is_control("""\r""" ) ) def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" self.assertTrue(_is_punctuation("""-""" ) ) self.assertTrue(_is_punctuation("""$""" ) ) self.assertTrue(_is_punctuation("""`""" ) ) self.assertTrue(_is_punctuation(""".""" ) ) self.assertFalse(_is_punctuation("""A""" ) ) self.assertFalse(_is_punctuation(""" """ ) ) def _snake_case ( self :str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(__A ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) if self.test_rust_tokenizer: SCREAMING_SNAKE_CASE__ = self.get_rust_tokenizer() self.assertListEqual( [rust_tokenizer.tokenize(__A ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) def _snake_case ( self :int ) -> Any: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = f'''A, naïve {tokenizer_r.mask_token} AllenNLP sentence.''' SCREAMING_SNAKE_CASE__ = tokenizer_r.encode_plus( __A , return_attention_mask=__A , return_token_type_ids=__A , return_offsets_mapping=__A , add_special_tokens=__A , ) SCREAMING_SNAKE_CASE__ = tokenizer_r.do_lower_case if hasattr(__A , """do_lower_case""" ) else False SCREAMING_SNAKE_CASE__ = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """A"""), ((1, 2), ""","""), ((3, 5), """na"""), ((5, 6), """##ï"""), ((6, 8), """##ve"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """Allen"""), ((21, 23), """##NL"""), ((23, 24), """##P"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """a"""), ((1, 2), ""","""), ((3, 8), """naive"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """allen"""), ((21, 23), """##nl"""), ((23, 24), """##p"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["""input_ids"""] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["""offset_mapping"""] ) def _snake_case ( self :Optional[int] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""的""", """人""", """有"""] SCREAMING_SNAKE_CASE__ = """""".join(__A ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.convert_ids_to_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.convert_ids_to_tokens(__A ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(__A , __A ) self.assertListEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.convert_ids_to_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.convert_ids_to_tokens(__A ) # it is expected that only the first Chinese character is not preceded by "##". SCREAMING_SNAKE_CASE__ = [ f'''##{token}''' if idx != 0 else token for idx, token in enumerate(__A ) ] self.assertListEqual(__A , __A ) self.assertListEqual(__A , __A ) @slow def _snake_case ( self :Union[str, Any] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) SCREAMING_SNAKE_CASE__ = tokenizer.encode("""你好""" , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.encode("""你是谁""" , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.build_inputs_with_special_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.build_inputs_with_special_tokens(__A , __A ) assert encoded_sentence == [1] + text + [2] assert encoded_pair == [1] + text + [2] + text_a + [2] def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_tokenizers(do_lower_case=__A ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE__ = """你好,你是谁""" SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_shape_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_pronunciation_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.prepare_for_model( __A , __A , __A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.encode_plus(__A , add_special_tokens=__A ) self.assertEqual(__A , __A )
59
1
import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "Speech2TextFeatureExtractor" lowerCamelCase_ = "Speech2TextTokenizer" def __init__( self :Tuple , __A :Union[str, Any] , __A :Optional[Any] ) -> Tuple: """simple docstring""" super().__init__(__A , __A ) SCREAMING_SNAKE_CASE__ = self.feature_extractor SCREAMING_SNAKE_CASE__ = False def __call__( self :Optional[Any] , *__A :Optional[int] , **__A :Tuple ) -> int: """simple docstring""" if self._in_target_context_manager: return self.current_processor(*__A , **__A ) if "raw_speech" in kwargs: warnings.warn("""Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.""" ) SCREAMING_SNAKE_CASE__ = kwargs.pop("""raw_speech""" ) else: SCREAMING_SNAKE_CASE__ = kwargs.pop("""audio""" , __A ) SCREAMING_SNAKE_CASE__ = kwargs.pop("""sampling_rate""" , __A ) SCREAMING_SNAKE_CASE__ = kwargs.pop("""text""" , __A ) if len(__A ) > 0: SCREAMING_SNAKE_CASE__ = args[0] SCREAMING_SNAKE_CASE__ = args[1:] if audio is None and text is None: raise ValueError("""You need to specify either an `audio` or `text` input to process.""" ) if audio is not None: SCREAMING_SNAKE_CASE__ = self.feature_extractor(__A , *__A , sampling_rate=__A , **__A ) if text is not None: SCREAMING_SNAKE_CASE__ = self.tokenizer(__A , **__A ) if text is None: return inputs elif audio is None: return encodings else: SCREAMING_SNAKE_CASE__ = encodings["""input_ids"""] return inputs def _snake_case ( self :Any , *__A :Tuple , **__A :Any ) -> List[Any]: """simple docstring""" return self.tokenizer.batch_decode(*__A , **__A ) def _snake_case ( self :Tuple , *__A :Optional[Any] , **__A :List[Any] ) -> Dict: """simple docstring""" return self.tokenizer.decode(*__A , **__A ) @contextmanager def _snake_case ( self :str ) -> Optional[int]: """simple docstring""" warnings.warn( """`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your """ """labels by using the argument `text` of the regular `__call__` method (either in the same call as """ """your audio inputs, or in a separate call.""" ) SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = self.tokenizer yield SCREAMING_SNAKE_CASE__ = self.feature_extractor SCREAMING_SNAKE_CASE__ = False
59
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[str, Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = original_name.split(""".""" )[0] SCREAMING_SNAKE_CASE__ = key.split(""".""" ) SCREAMING_SNAKE_CASE__ = int(key_list[key_list.index(UpperCamelCase__ ) - 2] ) SCREAMING_SNAKE_CASE__ = int(key_list[key_list.index(UpperCamelCase__ ) - 1] ) SCREAMING_SNAKE_CASE__ = orig_block_num - offset SCREAMING_SNAKE_CASE__ = key.replace(f'''{orig_block_num}.{layer_num}.{original_name}''' , f'''block.{new_block_num}.{layer_num}.{new_name}''' ) return key def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = OrderedDict() SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 0, 0 for key, value in state_dict.items(): if key.startswith("""network""" ): SCREAMING_SNAKE_CASE__ = key.replace("""network""" , """poolformer.encoder""" ) if "proj" in key: # Works for the first embedding as well as the internal embedding layers if key.endswith("""bias""" ) and "patch_embed" not in key: patch_emb_offset += 1 SCREAMING_SNAKE_CASE__ = key[: key.find("""proj""" )] SCREAMING_SNAKE_CASE__ = key.replace(UpperCamelCase__ , f'''patch_embeddings.{total_embed_found}.''' ) SCREAMING_SNAKE_CASE__ = key.replace("""proj""" , """projection""" ) if key.endswith("""bias""" ): total_embed_found += 1 if "patch_embeddings" in key: SCREAMING_SNAKE_CASE__ = """poolformer.encoder.""" + key if "mlp.fc1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """mlp.fc1""" , """output.conv1""" ) if "mlp.fc2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """mlp.fc2""" , """output.conv2""" ) if "norm1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """norm1""" , """before_norm""" ) if "norm2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """norm2""" , """after_norm""" ) if "layer_scale_1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """layer_scale_1""" , """layer_scale_1""" ) if "layer_scale_2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """layer_scale_2""" , """layer_scale_2""" ) if "head" in key: SCREAMING_SNAKE_CASE__ = key.replace("""head""" , """classifier""" ) SCREAMING_SNAKE_CASE__ = value return new_state_dict def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = """http://images.cocodataset.org/val2017/000000039769.jpg""" SCREAMING_SNAKE_CASE__ = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return image @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = PoolFormerConfig() # set attributes based on model_name SCREAMING_SNAKE_CASE__ = """huggingface/label-files""" SCREAMING_SNAKE_CASE__ = model_name[-3:] SCREAMING_SNAKE_CASE__ = 1_000 SCREAMING_SNAKE_CASE__ = """imagenet-1k-id2label.json""" SCREAMING_SNAKE_CASE__ = (1, 1_000) # set config attributes SCREAMING_SNAKE_CASE__ = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) SCREAMING_SNAKE_CASE__ = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE__ = idalabel SCREAMING_SNAKE_CASE__ = {v: k for k, v in idalabel.items()} if size == "s12": SCREAMING_SNAKE_CASE__ = [2, 2, 6, 2] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "s24": SCREAMING_SNAKE_CASE__ = [4, 4, 12, 4] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "s36": SCREAMING_SNAKE_CASE__ = [6, 6, 18, 6] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "m36": SCREAMING_SNAKE_CASE__ = [6, 6, 18, 6] SCREAMING_SNAKE_CASE__ = [96, 192, 384, 768] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9_5 elif size == "m48": SCREAMING_SNAKE_CASE__ = [8, 8, 24, 8] SCREAMING_SNAKE_CASE__ = [96, 192, 384, 768] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9_5 else: raise ValueError(f'''Size {size} not supported''' ) # load image processor SCREAMING_SNAKE_CASE__ = PoolFormerImageProcessor(crop_pct=UpperCamelCase__ ) # Prepare image SCREAMING_SNAKE_CASE__ = prepare_img() SCREAMING_SNAKE_CASE__ = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ).pixel_values logger.info(f'''Converting model {model_name}...''' ) # load original state dict SCREAMING_SNAKE_CASE__ = torch.load(UpperCamelCase__ , map_location=torch.device("""cpu""" ) ) # rename keys SCREAMING_SNAKE_CASE__ = rename_keys(UpperCamelCase__ ) # create HuggingFace model and load state dict SCREAMING_SNAKE_CASE__ = PoolFormerForImageClassification(UpperCamelCase__ ) model.load_state_dict(UpperCamelCase__ ) model.eval() # Define image processor SCREAMING_SNAKE_CASE__ = PoolFormerImageProcessor(crop_pct=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = image_processor(images=prepare_img() , return_tensors="""pt""" ).pixel_values # forward pass SCREAMING_SNAKE_CASE__ = model(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = outputs.logits # define expected logit slices for different models if size == "s12": SCREAMING_SNAKE_CASE__ = torch.tensor([-0.3_0_4_5, -0.6_7_5_8, -0.4_8_6_9] ) elif size == "s24": SCREAMING_SNAKE_CASE__ = torch.tensor([0.4_4_0_2, -0.1_3_7_4, -0.8_0_4_5] ) elif size == "s36": SCREAMING_SNAKE_CASE__ = torch.tensor([-0.6_0_8_0, -0.5_1_3_3, -0.5_8_9_8] ) elif size == "m36": SCREAMING_SNAKE_CASE__ = torch.tensor([0.3_9_5_2, 0.2_2_6_3, -1.2_6_6_8] ) elif size == "m48": SCREAMING_SNAKE_CASE__ = torch.tensor([0.1_1_6_7, -0.0_6_5_6, -0.3_4_2_3] ) else: raise ValueError(f'''Size {size} not supported''' ) # verify logits assert logits.shape == expected_shape assert torch.allclose(logits[0, :3] , UpperCamelCase__ , atol=1e-2 ) # finally, save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) model.save_pretrained(UpperCamelCase__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '--model_name', default='poolformer_s12', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) _lowerCamelCase = parser.parse_args() convert_poolformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
59
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _lowerCamelCase = { 'configuration_convnext': ['CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ConvNextConfig', 'ConvNextOnnxConfig'] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['ConvNextFeatureExtractor'] _lowerCamelCase = ['ConvNextImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'ConvNextForImageClassification', 'ConvNextModel', 'ConvNextPreTrainedModel', 'ConvNextBackbone', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'TFConvNextForImageClassification', 'TFConvNextModel', 'TFConvNextPreTrainedModel', ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure)
59
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'google/pix2struct-textcaps-base': ( 'https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json' ), } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct_text_model" lowerCamelCase_ = ["past_key_values"] lowerCamelCase_ = { "hidden_size": "hidden_size", "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self :Union[str, Any] , __A :Any=5_0244 , __A :Optional[Any]=768 , __A :Tuple=64 , __A :List[str]=2048 , __A :int=12 , __A :str=12 , __A :Any=32 , __A :Tuple=128 , __A :int=0.1 , __A :str=1E-6 , __A :Optional[Any]=1.0 , __A :Union[str, Any]="gelu_new" , __A :Any=0 , __A :List[str]=False , __A :Optional[Any]=0 , __A :int=1 , __A :Optional[int]=False , __A :Optional[Any]=True , **__A :List[Any] , ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = d_kv SCREAMING_SNAKE_CASE__ = d_ff SCREAMING_SNAKE_CASE__ = num_layers SCREAMING_SNAKE_CASE__ = num_heads SCREAMING_SNAKE_CASE__ = relative_attention_num_buckets SCREAMING_SNAKE_CASE__ = relative_attention_max_distance SCREAMING_SNAKE_CASE__ = dropout_rate SCREAMING_SNAKE_CASE__ = layer_norm_epsilon SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = use_cache SCREAMING_SNAKE_CASE__ = eos_token_id SCREAMING_SNAKE_CASE__ = decoder_start_token_id # for backwards compatibility SCREAMING_SNAKE_CASE__ = dense_act_fn super().__init__( pad_token_id=__A , eos_token_id=__A , decoder_start_token_id=__A , tie_word_embeddings=__A , is_decoder=__A , **__A , ) @classmethod def _snake_case ( cls :Optional[int] , __A :Union[str, os.PathLike] , **__A :Optional[int] ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = cls.get_config_dict(__A , **__A ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get("""model_type""" ) == "pix2struct": SCREAMING_SNAKE_CASE__ = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct_vision_model" def __init__( self :Optional[int] , __A :int=768 , __A :Optional[Any]=768 , __A :Union[str, Any]=2048 , __A :int=64 , __A :Union[str, Any]=12 , __A :str=12 , __A :Any="gelu_new" , __A :List[Any]=1E-6 , __A :Dict=0.0 , __A :int=0.0 , __A :int=1E-10 , __A :Dict=1.0 , __A :int=4096 , __A :int=32 , __A :int=128 , **__A :Tuple , ) -> str: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = patch_embed_hidden_size SCREAMING_SNAKE_CASE__ = d_ff SCREAMING_SNAKE_CASE__ = dropout_rate SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = attention_dropout SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = dense_act_fn SCREAMING_SNAKE_CASE__ = seq_len SCREAMING_SNAKE_CASE__ = relative_attention_num_buckets SCREAMING_SNAKE_CASE__ = relative_attention_max_distance SCREAMING_SNAKE_CASE__ = d_kv @classmethod def _snake_case ( cls :str , __A :Union[str, os.PathLike] , **__A :str ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = cls.get_config_dict(__A , **__A ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get("""model_type""" ) == "pix2struct": SCREAMING_SNAKE_CASE__ = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct" lowerCamelCase_ = True def __init__( self :str , __A :Optional[Any]=None , __A :List[str]=None , __A :Optional[Any]=1.0 , __A :Optional[Any]=0.0_2 , __A :Any=False , __A :Tuple=False , __A :Any=True , **__A :Dict , ) -> Union[str, Any]: """simple docstring""" super().__init__(tie_word_embeddings=__A , is_encoder_decoder=__A , **__A ) if text_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""text_config is None. Initializing the Pix2StructTextConfig with default values.""" ) if vision_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""vision_config is None. Initializing the Pix2StructVisionConfig with default values.""" ) SCREAMING_SNAKE_CASE__ = PixaStructTextConfig(**__A ) SCREAMING_SNAKE_CASE__ = PixaStructVisionConfig(**__A ) SCREAMING_SNAKE_CASE__ = self.text_config.decoder_start_token_id SCREAMING_SNAKE_CASE__ = self.text_config.pad_token_id SCREAMING_SNAKE_CASE__ = self.text_config.eos_token_id SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = self.initializer_range SCREAMING_SNAKE_CASE__ = self.initializer_range SCREAMING_SNAKE_CASE__ = is_vqa @classmethod def _snake_case ( cls :Union[str, Any] , __A :PixaStructTextConfig , __A :PixaStructVisionConfig , **__A :Optional[int] ) -> Optional[Any]: """simple docstring""" return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__A ) def _snake_case ( self :str ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = copy.deepcopy(self.__dict__ ) SCREAMING_SNAKE_CASE__ = self.text_config.to_dict() SCREAMING_SNAKE_CASE__ = self.vision_config.to_dict() SCREAMING_SNAKE_CASE__ = self.__class__.model_type return output
59
1
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments from transformers.hf_argparser import make_choice_type_function, string_to_bool # Since Python 3.10, we can use the builtin `|` operator for Union types # See PEP 604: https://peps.python.org/pep-0604 _lowerCamelCase = sys.version_info >= (3, 10) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any=None , UpperCamelCase__: Tuple=None ): return field(default_factory=lambda: default , metadata=UpperCamelCase__ ) @dataclass class UpperCamelCase_ : lowerCamelCase_ = 42 lowerCamelCase_ = 42 lowerCamelCase_ = 42 lowerCamelCase_ = 42 @dataclass class UpperCamelCase_ : lowerCamelCase_ = 42 lowerCamelCase_ = field(default="toto" , metadata={"help": "help message"} ) @dataclass class UpperCamelCase_ : lowerCamelCase_ = False lowerCamelCase_ = True lowerCamelCase_ = None class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "titi" lowerCamelCase_ = "toto" class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "titi" lowerCamelCase_ = "toto" lowerCamelCase_ = 42 @dataclass class UpperCamelCase_ : lowerCamelCase_ = "toto" def _snake_case ( self :Dict ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = BasicEnum(self.foo ) @dataclass class UpperCamelCase_ : lowerCamelCase_ = "toto" def _snake_case ( self :str ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = MixedTypeEnum(self.foo ) @dataclass class UpperCamelCase_ : lowerCamelCase_ = None lowerCamelCase_ = field(default=UpperCamelCase__ , metadata={"help": "help message"} ) lowerCamelCase_ = None lowerCamelCase_ = list_field(default=[] ) lowerCamelCase_ = list_field(default=[] ) @dataclass class UpperCamelCase_ : lowerCamelCase_ = list_field(default=[] ) lowerCamelCase_ = list_field(default=[1, 2, 3] ) lowerCamelCase_ = list_field(default=["Hallo", "Bonjour", "Hello"] ) lowerCamelCase_ = list_field(default=[0.1, 0.2, 0.3] ) @dataclass class UpperCamelCase_ : lowerCamelCase_ = field() lowerCamelCase_ = field() lowerCamelCase_ = field() def _snake_case ( self :str ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = BasicEnum(self.required_enum ) @dataclass class UpperCamelCase_ : lowerCamelCase_ = 42 lowerCamelCase_ = field() lowerCamelCase_ = None lowerCamelCase_ = field(default="toto" , metadata={"help": "help message"} ) lowerCamelCase_ = list_field(default=["Hallo", "Bonjour", "Hello"] ) if is_python_no_less_than_3_10: @dataclass class UpperCamelCase_ : lowerCamelCase_ = False lowerCamelCase_ = True lowerCamelCase_ = None @dataclass class UpperCamelCase_ : lowerCamelCase_ = None lowerCamelCase_ = field(default=UpperCamelCase__ , metadata={"help": "help message"} ) lowerCamelCase_ = None lowerCamelCase_ = list_field(default=[] ) lowerCamelCase_ = list_field(default=[] ) class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :str , __A :argparse.ArgumentParser , __A :argparse.ArgumentParser ) -> List[str]: """simple docstring""" self.assertEqual(len(a._actions ) , len(b._actions ) ) for x, y in zip(a._actions , b._actions ): SCREAMING_SNAKE_CASE__ = {k: v for k, v in vars(__A ).items() if k != """container"""} SCREAMING_SNAKE_CASE__ = {k: v for k, v in vars(__A ).items() if k != """container"""} # Choices with mixed type have custom function as "type" # So we need to compare results directly for equality if xx.get("""choices""" , __A ) and yy.get("""choices""" , __A ): for expected_choice in yy["choices"] + xx["choices"]: self.assertEqual(xx["""type"""](__A ) , yy["""type"""](__A ) ) del xx["type"], yy["type"] self.assertEqual(__A , __A ) def _snake_case ( self :int ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = HfArgumentParser(__A ) SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() expected.add_argument("""--foo""" , type=__A , required=__A ) expected.add_argument("""--bar""" , type=__A , required=__A ) expected.add_argument("""--baz""" , type=__A , required=__A ) expected.add_argument("""--flag""" , type=__A , default=__A , const=__A , nargs="""?""" ) self.argparsersEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = ["""--foo""", """1""", """--baz""", """quux""", """--bar""", """0.5"""] ((SCREAMING_SNAKE_CASE__) , ) = parser.parse_args_into_dataclasses(__A , look_for_args_file=__A ) self.assertFalse(example.flag ) def _snake_case ( self :Tuple ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = HfArgumentParser(__A ) SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() expected.add_argument("""--foo""" , default=42 , type=__A ) expected.add_argument("""--baz""" , default="""toto""" , type=__A , help="""help message""" ) self.argparsersEqual(__A , __A ) def _snake_case ( self :List[str] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() expected.add_argument("""--foo""" , type=__A , default=__A , const=__A , nargs="""?""" ) expected.add_argument("""--baz""" , type=__A , default=__A , const=__A , nargs="""?""" ) # A boolean no_* argument always has to come after its "default: True" regular counter-part # and its default must be set to False expected.add_argument("""--no_baz""" , action="""store_false""" , default=__A , dest="""baz""" ) expected.add_argument("""--opt""" , type=__A , default=__A ) SCREAMING_SNAKE_CASE__ = [WithDefaultBoolExample] if is_python_no_less_than_3_10: dataclass_types.append(__A ) for dataclass_type in dataclass_types: SCREAMING_SNAKE_CASE__ = HfArgumentParser(__A ) self.argparsersEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = parser.parse_args([] ) self.assertEqual(__A , Namespace(foo=__A , baz=__A , opt=__A ) ) SCREAMING_SNAKE_CASE__ = parser.parse_args(["""--foo""", """--no_baz"""] ) self.assertEqual(__A , Namespace(foo=__A , baz=__A , opt=__A ) ) SCREAMING_SNAKE_CASE__ = parser.parse_args(["""--foo""", """--baz"""] ) self.assertEqual(__A , Namespace(foo=__A , baz=__A , opt=__A ) ) SCREAMING_SNAKE_CASE__ = parser.parse_args(["""--foo""", """True""", """--baz""", """True""", """--opt""", """True"""] ) self.assertEqual(__A , Namespace(foo=__A , baz=__A , opt=__A ) ) SCREAMING_SNAKE_CASE__ = parser.parse_args(["""--foo""", """False""", """--baz""", """False""", """--opt""", """False"""] ) self.assertEqual(__A , Namespace(foo=__A , baz=__A , opt=__A ) ) def _snake_case ( self :Any ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = HfArgumentParser(__A ) SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() expected.add_argument( """--foo""" , default="""toto""" , choices=["""titi""", """toto""", 42] , type=make_choice_type_function(["""titi""", """toto""", 42] ) , ) self.argparsersEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = parser.parse_args([] ) self.assertEqual(args.foo , """toto""" ) SCREAMING_SNAKE_CASE__ = parser.parse_args_into_dataclasses([] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.toto ) SCREAMING_SNAKE_CASE__ = parser.parse_args(["""--foo""", """titi"""] ) self.assertEqual(args.foo , """titi""" ) SCREAMING_SNAKE_CASE__ = parser.parse_args_into_dataclasses(["""--foo""", """titi"""] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.titi ) SCREAMING_SNAKE_CASE__ = parser.parse_args(["""--foo""", """42"""] ) self.assertEqual(args.foo , 42 ) SCREAMING_SNAKE_CASE__ = parser.parse_args_into_dataclasses(["""--foo""", """42"""] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.fourtytwo ) def _snake_case ( self :int ) -> Any: """simple docstring""" @dataclass class UpperCamelCase_ : lowerCamelCase_ = "toto" SCREAMING_SNAKE_CASE__ = HfArgumentParser(__A ) SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() expected.add_argument( """--foo""" , default="""toto""" , choices=("""titi""", """toto""", 42) , type=make_choice_type_function(["""titi""", """toto""", 42] ) , ) self.argparsersEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = parser.parse_args([] ) self.assertEqual(args.foo , """toto""" ) SCREAMING_SNAKE_CASE__ = parser.parse_args(["""--foo""", """titi"""] ) self.assertEqual(args.foo , """titi""" ) SCREAMING_SNAKE_CASE__ = parser.parse_args(["""--foo""", """42"""] ) self.assertEqual(args.foo , 42 ) def _snake_case ( self :Any ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = HfArgumentParser(__A ) SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() expected.add_argument("""--foo_int""" , nargs="""+""" , default=[] , type=__A ) expected.add_argument("""--bar_int""" , nargs="""+""" , default=[1, 2, 3] , type=__A ) expected.add_argument("""--foo_str""" , nargs="""+""" , default=["""Hallo""", """Bonjour""", """Hello"""] , type=__A ) expected.add_argument("""--foo_float""" , nargs="""+""" , default=[0.1, 0.2, 0.3] , type=__A ) self.argparsersEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = parser.parse_args([] ) self.assertEqual( __A , Namespace(foo_int=[] , bar_int=[1, 2, 3] , foo_str=["""Hallo""", """Bonjour""", """Hello"""] , foo_float=[0.1, 0.2, 0.3] ) , ) SCREAMING_SNAKE_CASE__ = parser.parse_args("""--foo_int 1 --bar_int 2 3 --foo_str a b c --foo_float 0.1 0.7""".split() ) self.assertEqual(__A , Namespace(foo_int=[1] , bar_int=[2, 3] , foo_str=["""a""", """b""", """c"""] , foo_float=[0.1, 0.7] ) ) def _snake_case ( self :str ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() expected.add_argument("""--foo""" , default=__A , type=__A ) expected.add_argument("""--bar""" , default=__A , type=__A , help="""help message""" ) expected.add_argument("""--baz""" , default=__A , type=__A ) expected.add_argument("""--ces""" , nargs="""+""" , default=[] , type=__A ) expected.add_argument("""--des""" , nargs="""+""" , default=[] , type=__A ) SCREAMING_SNAKE_CASE__ = [OptionalExample] if is_python_no_less_than_3_10: dataclass_types.append(__A ) for dataclass_type in dataclass_types: SCREAMING_SNAKE_CASE__ = HfArgumentParser(__A ) self.argparsersEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = parser.parse_args([] ) self.assertEqual(__A , Namespace(foo=__A , bar=__A , baz=__A , ces=[] , des=[] ) ) SCREAMING_SNAKE_CASE__ = parser.parse_args("""--foo 12 --bar 3.14 --baz 42 --ces a b c --des 1 2 3""".split() ) self.assertEqual(__A , Namespace(foo=12 , bar=3.1_4 , baz="""42""" , ces=["""a""", """b""", """c"""] , des=[1, 2, 3] ) ) def _snake_case ( self :Dict ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = HfArgumentParser(__A ) SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() expected.add_argument("""--required_list""" , nargs="""+""" , type=__A , required=__A ) expected.add_argument("""--required_str""" , type=__A , required=__A ) expected.add_argument( """--required_enum""" , type=make_choice_type_function(["""titi""", """toto"""] ) , choices=["""titi""", """toto"""] , required=__A , ) self.argparsersEqual(__A , __A ) def _snake_case ( self :Optional[Any] ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = HfArgumentParser(__A ) SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() expected.add_argument("""--foo""" , type=__A , required=__A ) expected.add_argument( """--required_enum""" , type=make_choice_type_function(["""titi""", """toto"""] ) , choices=["""titi""", """toto"""] , required=__A , ) expected.add_argument("""--opt""" , type=__A , default=__A ) expected.add_argument("""--baz""" , default="""toto""" , type=__A , help="""help message""" ) expected.add_argument("""--foo_str""" , nargs="""+""" , default=["""Hallo""", """Bonjour""", """Hello"""] , type=__A ) self.argparsersEqual(__A , __A ) def _snake_case ( self :int ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = HfArgumentParser(__A ) SCREAMING_SNAKE_CASE__ = { """foo""": 12, """bar""": 3.1_4, """baz""": """42""", """flag""": True, } SCREAMING_SNAKE_CASE__ = parser.parse_dict(__A )[0] SCREAMING_SNAKE_CASE__ = BasicExample(**__A ) self.assertEqual(__A , __A ) def _snake_case ( self :Optional[int] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = HfArgumentParser(__A ) SCREAMING_SNAKE_CASE__ = { """foo""": 12, """bar""": 3.1_4, """baz""": """42""", """flag""": True, """extra""": 42, } self.assertRaises(__A , parser.parse_dict , __A , allow_extra_keys=__A ) def _snake_case ( self :str ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = HfArgumentParser(__A ) SCREAMING_SNAKE_CASE__ = { """foo""": 12, """bar""": 3.1_4, """baz""": """42""", """flag""": True, } with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE__ = os.path.join(__A , """temp_json""" ) os.mkdir(__A ) with open(temp_local_path + """.json""" , """w+""" ) as f: json.dump(__A , __A ) SCREAMING_SNAKE_CASE__ = parser.parse_yaml_file(Path(temp_local_path + """.json""" ) )[0] SCREAMING_SNAKE_CASE__ = BasicExample(**__A ) self.assertEqual(__A , __A ) def _snake_case ( self :Optional[int] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = HfArgumentParser(__A ) SCREAMING_SNAKE_CASE__ = { """foo""": 12, """bar""": 3.1_4, """baz""": """42""", """flag""": True, } with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE__ = os.path.join(__A , """temp_yaml""" ) os.mkdir(__A ) with open(temp_local_path + """.yaml""" , """w+""" ) as f: yaml.dump(__A , __A ) SCREAMING_SNAKE_CASE__ = parser.parse_yaml_file(Path(temp_local_path + """.yaml""" ) )[0] SCREAMING_SNAKE_CASE__ = BasicExample(**__A ) self.assertEqual(__A , __A ) def _snake_case ( self :str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = HfArgumentParser(__A ) self.assertIsNotNone(__A )
59
import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = inspect.getfile(accelerate.test_utils ) SCREAMING_SNAKE_CASE__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_script.py"""] ) SCREAMING_SNAKE_CASE__ = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_distributed_data_loop.py"""] ) SCREAMING_SNAKE_CASE__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_ops.py"""] ) @require_multi_gpu def _snake_case ( self :Optional[Any] ) -> Tuple: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Tuple ) -> Optional[Any]: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.operation_file_path] print(f'''Command: {cmd}''' ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Optional[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Optional[int] ) -> str: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices, using 2 devices only''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices="""0,1""" ): execute_subprocess_async(__A , env=os.environ.copy() ) if __name__ == "__main__": _lowerCamelCase = Accelerator() _lowerCamelCase = (accelerator.state.process_index + 2, 10) _lowerCamelCase = torch.randint(0, 10, shape).to(accelerator.device) _lowerCamelCase = '' _lowerCamelCase = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." _lowerCamelCase = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." _lowerCamelCase = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
59
1
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int = 10 ): if not isinstance(UpperCamelCase__ , UpperCamelCase__ ) or n < 0: raise ValueError("""Invalid input""" ) SCREAMING_SNAKE_CASE__ = 10**n SCREAMING_SNAKE_CASE__ = 28_433 * (pow(2 , 7_830_457 , UpperCamelCase__ )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(F'''{solution(10) = }''')
59
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _lowerCamelCase = { 'configuration_layoutlmv3': [ 'LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv3Config', 'LayoutLMv3OnnxConfig', ], 'processing_layoutlmv3': ['LayoutLMv3Processor'], 'tokenization_layoutlmv3': ['LayoutLMv3Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['LayoutLMv3TokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv3ForQuestionAnswering', 'LayoutLMv3ForSequenceClassification', 'LayoutLMv3ForTokenClassification', 'LayoutLMv3Model', 'LayoutLMv3PreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLayoutLMv3ForQuestionAnswering', 'TFLayoutLMv3ForSequenceClassification', 'TFLayoutLMv3ForTokenClassification', 'TFLayoutLMv3Model', 'TFLayoutLMv3PreTrainedModel', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['LayoutLMv3FeatureExtractor'] _lowerCamelCase = ['LayoutLMv3ImageProcessor'] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
1
from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING _lowerCamelCase = logging.get_logger(__name__) @add_end_docstrings(UpperCamelCase__ ) class UpperCamelCase_ ( UpperCamelCase__ ): def __init__( self :List[Any] , **__A :Optional[Any] ) -> int: """simple docstring""" super().__init__(**__A ) requires_backends(self , """vision""" ) requires_backends(self , """torch""" ) if self.framework != "pt": raise ValueError(f'''The {self.__class__} is only available in PyTorch.''' ) self.check_model_type(__A ) def _snake_case ( self :str , **__A :Optional[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = {} # preprocess args if "points_per_batch" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs["""points_per_batch"""] if "points_per_crop" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs["""points_per_crop"""] if "crops_n_layers" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs["""crops_n_layers"""] if "crop_overlap_ratio" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs["""crop_overlap_ratio"""] if "crop_n_points_downscale_factor" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs["""crop_n_points_downscale_factor"""] # postprocess args if "pred_iou_thresh" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs["""pred_iou_thresh"""] if "stability_score_offset" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs["""stability_score_offset"""] if "mask_threshold" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs["""mask_threshold"""] if "stability_score_thresh" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs["""stability_score_thresh"""] if "crops_nms_thresh" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs["""crops_nms_thresh"""] if "output_rle_mask" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs["""output_rle_mask"""] if "output_bboxes_mask" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs["""output_bboxes_mask"""] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__( self :List[Any] , __A :List[str] , *__A :Tuple , __A :Optional[int]=None , __A :str=None , **__A :Union[str, Any] ) -> List[Any]: """simple docstring""" return super().__call__(__A , *__A , num_workers=__A , batch_size=__A , **__A ) def _snake_case ( self :Any , __A :int , __A :Optional[int]=64 , __A :int = 0 , __A :float = 512 / 1500 , __A :Optional[int] = 32 , __A :Optional[int] = 1 , ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = load_image(__A ) SCREAMING_SNAKE_CASE__ = self.image_processor.size["""longest_edge"""] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.image_processor.generate_crop_boxes( __A , __A , __A , __A , __A , __A ) SCREAMING_SNAKE_CASE__ = self.image_processor(images=__A , return_tensors="""pt""" ) with self.device_placement(): if self.framework == "pt": SCREAMING_SNAKE_CASE__ = self.get_inference_context() with inference_context(): SCREAMING_SNAKE_CASE__ = self._ensure_tensor_on_device(__A , device=self.device ) SCREAMING_SNAKE_CASE__ = self.model.get_image_embeddings(model_inputs.pop("""pixel_values""" ) ) SCREAMING_SNAKE_CASE__ = image_embeddings SCREAMING_SNAKE_CASE__ = grid_points.shape[1] SCREAMING_SNAKE_CASE__ = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( """Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. """ """To return all points at once, set points_per_batch to None""" ) for i in range(0 , __A , __A ): SCREAMING_SNAKE_CASE__ = grid_points[:, i : i + points_per_batch, :, :] SCREAMING_SNAKE_CASE__ = input_labels[:, i : i + points_per_batch] SCREAMING_SNAKE_CASE__ = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def _snake_case ( self :Optional[Any] , __A :str , __A :List[Any]=0.8_8 , __A :Tuple=0.9_5 , __A :Optional[Any]=0 , __A :List[Any]=1 , ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = model_inputs.pop("""input_boxes""" ) SCREAMING_SNAKE_CASE__ = model_inputs.pop("""is_last""" ) SCREAMING_SNAKE_CASE__ = model_inputs.pop("""original_sizes""" ).tolist() SCREAMING_SNAKE_CASE__ = model_inputs.pop("""reshaped_input_sizes""" ).tolist() SCREAMING_SNAKE_CASE__ = self.model(**__A ) # post processing happens here in order to avoid CPU GPU copies of ALL the masks SCREAMING_SNAKE_CASE__ = model_outputs["""pred_masks"""] SCREAMING_SNAKE_CASE__ = self.image_processor.post_process_masks( __A , __A , __A , __A , binarize=__A ) SCREAMING_SNAKE_CASE__ = model_outputs["""iou_scores"""] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.image_processor.filter_masks( masks[0] , iou_scores[0] , original_sizes[0] , input_boxes[0] , __A , __A , __A , __A , ) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def _snake_case ( self :Union[str, Any] , __A :Union[str, Any] , __A :Any=False , __A :Any=False , __A :str=0.7 , ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] for model_output in model_outputs: all_scores.append(model_output.pop("""iou_scores""" ) ) all_masks.extend(model_output.pop("""masks""" ) ) all_boxes.append(model_output.pop("""boxes""" ) ) SCREAMING_SNAKE_CASE__ = torch.cat(__A ) SCREAMING_SNAKE_CASE__ = torch.cat(__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.image_processor.post_process_for_mask_generation( __A , __A , __A , __A ) SCREAMING_SNAKE_CASE__ = defaultdict(__A ) for output in model_outputs: for k, v in output.items(): extra[k].append(__A ) SCREAMING_SNAKE_CASE__ = {} if output_rle_mask: SCREAMING_SNAKE_CASE__ = rle_mask if output_bboxes_mask: SCREAMING_SNAKE_CASE__ = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
59
import inspect import unittest class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :str ) -> Union[str, Any]: """simple docstring""" try: import diffusers # noqa: F401 except ImportError: assert False def _snake_case ( self :Any ) -> Any: """simple docstring""" import diffusers from diffusers.dependency_versions_table import deps SCREAMING_SNAKE_CASE__ = inspect.getmembers(__A , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": SCREAMING_SNAKE_CASE__ = """k-diffusion""" elif backend == "invisible_watermark": SCREAMING_SNAKE_CASE__ = """invisible-watermark""" assert backend in deps, f'''{backend} is not in the deps table!'''
59
1
import argparse import random import joblib import numpy as np import torch from igf.igf import ( SecondaryLearner, collect_objective_set, compute_perplexity, generate_datasets, load_gpta, recopy_gpta, set_seed, train_secondary_learner, ) from torch.utils.data import DataLoader, RandomSampler from transformers import GPTaLMHeadModel def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[Any]=32 , UpperCamelCase__: List[Any]=10 , UpperCamelCase__: str=100 , UpperCamelCase__: Any=1_026 , UpperCamelCase__: int=True , UpperCamelCase__: str="data/tokenized_stories_train_wikitext103.jbl" , UpperCamelCase__: str="igf_context_pairs.jbl" , ): set_seed(3 ) # generate train_data and objective_set SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = generate_datasets( UpperCamelCase__ , UpperCamelCase__ , number=UpperCamelCase__ , min_len=1_026 , trim=UpperCamelCase__ ) # keeps model same across runs set_seed(4 ) # model, lm_optimizer, lm_scheduler = recopy_gpt2(model, device, max_steps) # store original model weights # can we train on GPU? SCREAMING_SNAKE_CASE__ = torch.device("""cuda:0""" if torch.cuda.is_available() else """cpu""" ) # load pretrained model SCREAMING_SNAKE_CASE__ = load_gpta("""gpt2""" ).to(UpperCamelCase__ ) print("""computing perplexity on objective set""" ) SCREAMING_SNAKE_CASE__ = compute_perplexity(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ).item() print("""perplexity on objective set:""" , UpperCamelCase__ ) # collect igf pairs and save to file demo.jbl collect_objective_set(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # clean up, delete model and data we don't need anymore del model, train_data, objective_set torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Optional[int]=15 , UpperCamelCase__: Union[str, Any]=128 , UpperCamelCase__: List[Any]=100 , UpperCamelCase__: Optional[int]="igf_model.pt" , ): set_seed(42 ) # Load pre-trained model SCREAMING_SNAKE_CASE__ = GPTaLMHeadModel.from_pretrained("""gpt2""" ) # Initialize secondary learner to use embedding weights of model SCREAMING_SNAKE_CASE__ = SecondaryLearner(UpperCamelCase__ ) # Train secondary learner SCREAMING_SNAKE_CASE__ = train_secondary_learner( UpperCamelCase__ , UpperCamelCase__ , max_epochs=UpperCamelCase__ , batch_size=UpperCamelCase__ , eval_freq=100 , igf_model_path=UpperCamelCase__ , ) del model, secondary_learner_train_data torch.cuda.empty_cache() return secondary_learner def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Tuple , UpperCamelCase__: int , UpperCamelCase__: Optional[int] , UpperCamelCase__: Any=32 , UpperCamelCase__: Any=1_000 , UpperCamelCase__: Union[str, Any]=16 , UpperCamelCase__: Dict=1.0 , UpperCamelCase__: List[str]=recopy_gpta , UpperCamelCase__: List[str]=None , UpperCamelCase__: List[str]=10 , UpperCamelCase__: List[str]="gpt2_finetuned.pt" , ): SCREAMING_SNAKE_CASE__ = torch.device("""cuda:0""" if torch.cuda.is_available() else """cpu""" ) SCREAMING_SNAKE_CASE__ = RandomSampler(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = DataLoader(UpperCamelCase__ , sampler=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = max_steps // (len(UpperCamelCase__ )) + 1 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = torch.zeros((1, context_len) , dtype=torch.long , device=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = recopy_model(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) model.train() if secondary_learner is not None: secondary_learner.to(UpperCamelCase__ ) secondary_learner.eval() SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] # Compute the performance of the transformer model at the beginning SCREAMING_SNAKE_CASE__ = compute_perplexity(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) test_perps.append(UpperCamelCase__ ) print("""Test perplexity, step""" , UpperCamelCase__ , """:""" , UpperCamelCase__ ) for epoch in range(int(UpperCamelCase__ ) ): for step, example in enumerate(UpperCamelCase__ ): torch.cuda.empty_cache() SCREAMING_SNAKE_CASE__ = random.randint(0 , example.size(2 ) - context_len - 1 ) SCREAMING_SNAKE_CASE__ = example[0, 0, start : start + context_len] lm_optimizer.zero_grad() SCREAMING_SNAKE_CASE__ = model(UpperCamelCase__ , labels=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = True if secondary_learner is not None: SCREAMING_SNAKE_CASE__ = secondary_learner.forward( torch.tensor(UpperCamelCase__ , dtype=torch.long , device=UpperCamelCase__ ).unsqueeze(0 ) )[0].item() observed_qs.append(float(UpperCamelCase__ ) ) # Here we implement the simple non-constant threshold for the predicted IG(X) value # We will decay the selectivity of our secondary learner filter from # 1 standard deviation above average to 1 below average after 10 batches. if global_step == 10: SCREAMING_SNAKE_CASE__ = -1 if predicted_q < threshold: SCREAMING_SNAKE_CASE__ = False # If we passed the filter, add the context to the batch! if do_backprop: contexts.append(np.array(context.cpu() ) ) SCREAMING_SNAKE_CASE__ = outputs[0] lm_loss.backward() examples += 1 del outputs # Once the batch is filled with enough contexts, backprop on the batch. if examples == batch_size: torch.cuda.empty_cache() SCREAMING_SNAKE_CASE__ = 0 # Do LM backprop torch.nn.utils.clip_grad_norm_(model.parameters() , 3.0 ) lm_optimizer.step() lm_scheduler.step() # Update learning rate schedule global_step += 1 # Compute the performance of the transformer model at this batch if global_step % eval_interval == 0: SCREAMING_SNAKE_CASE__ = compute_perplexity(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) test_perps.append(UpperCamelCase__ ) print("""Test perplexity, step""" , UpperCamelCase__ , """:""" , UpperCamelCase__ ) # Break out of the loop after 60 batches if max_steps > 0 and global_step > 60: break if max_steps > 0 and global_step > 60: break # save finetuned transformer model torch.save(model.state_dict() , UpperCamelCase__ ) torch.cuda.empty_cache() # Do some cleaning up so we can reinitialize for the next run of this function del lm_optimizer del lm_scheduler return model def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser(description="""Fine-tune a transformer model with IGF on a language modeling task""" ) # Required parameters parser.add_argument( """--data_dir""" , default=UpperCamelCase__ , type=UpperCamelCase__ , required=UpperCamelCase__ , help="""The input data dir. Should contain data files for WikiText.""" , ) parser.add_argument( """--model_name_or_path""" , default=UpperCamelCase__ , type=UpperCamelCase__ , required=UpperCamelCase__ , help="""Path to pretrained model or model identifier from huggingface.co/models""" , ) parser.add_argument( """--data_file""" , type=UpperCamelCase__ , default=UpperCamelCase__ , help=( """A jbl file containing tokenized data which can be split as objective dataset, """ """train_dataset and test_dataset.""" ) , ) parser.add_argument( """--igf_data_file""" , type=UpperCamelCase__ , default=UpperCamelCase__ , help="""A jbl file containing the context and information gain pairs to train secondary learner.""" , ) parser.add_argument( """--output_dir""" , default=UpperCamelCase__ , type=UpperCamelCase__ , required=UpperCamelCase__ , help="""The output directory where the final fine-tuned model is stored.""" , ) parser.add_argument( """--tokenizer_name""" , default=UpperCamelCase__ , type=UpperCamelCase__ , help="""Pretrained tokenizer name or path if not the same as model_name""" , ) parser.add_argument("""--seed""" , type=UpperCamelCase__ , default=UpperCamelCase__ , help="""A seed for reproducible training.""" ) parser.add_argument( """--context_len""" , default=32 , type=UpperCamelCase__ , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument( """--size_objective_set""" , default=100 , type=UpperCamelCase__ , help="""number of articles that are long enough to be used as our objective set""" , ) parser.add_argument( """--eval_freq""" , default=100 , type=UpperCamelCase__ , help="""secondary model evaluation is triggered at eval_freq""" ) parser.add_argument("""--max_steps""" , default=1_000 , type=UpperCamelCase__ , help="""To calculate training epochs""" ) parser.add_argument( """--secondary_learner_batch_size""" , default=128 , type=UpperCamelCase__ , help="""batch size of training data for secondary learner""" , ) parser.add_argument( """--batch_size""" , default=16 , type=UpperCamelCase__ , help="""batch size of training data of language model(gpt2) """ ) parser.add_argument( """--eval_interval""" , default=10 , type=UpperCamelCase__ , help=( """decay the selectivity of our secondary learner filter from""" """1 standard deviation above average to 1 below average after 10 batches""" ) , ) parser.add_argument( """--number""" , default=100 , type=UpperCamelCase__ , help="""The number of examples split to be used as objective_set/test_data""" ) parser.add_argument( """--min_len""" , default=1_026 , type=UpperCamelCase__ , help="""The minimum length of the article to be used as objective set""" ) parser.add_argument( """--secondary_learner_max_epochs""" , default=15 , type=UpperCamelCase__ , help="""number of epochs to train secondary learner""" ) parser.add_argument("""--trim""" , default=UpperCamelCase__ , type=UpperCamelCase__ , help="""truncate the example if it exceeds context length""" ) parser.add_argument( """--threshold""" , default=1.0 , type=UpperCamelCase__ , help=( """The threshold value used by secondary learner to filter the train_data and allow only""" """ informative data as input to the model""" ) , ) parser.add_argument("""--finetuned_model_name""" , default="""gpt2_finetuned.pt""" , type=UpperCamelCase__ , help="""finetuned_model_name""" ) parser.add_argument( """--recopy_model""" , default=UpperCamelCase__ , type=UpperCamelCase__ , help="""Reset the model to the original pretrained GPT-2 weights after each iteration""" , ) # function calls # Collecting *n* pairs of context and information gain(X, IG(X)) for training the secondary learner generate_n_pairs( context_len=32 , max_steps=10 , size_objective_set=100 , min_len=1_026 , trim=UpperCamelCase__ , data_file="""data/tokenized_stories_train_wikitext103.jbl""" , igf_data_file="""igf_context_pairs.jbl""" , ) # Load train data for secondary learner SCREAMING_SNAKE_CASE__ = joblib.load("""data/IGF_values.jbl""" ) # Train secondary learner SCREAMING_SNAKE_CASE__ = training_secondary_learner( UpperCamelCase__ , secondary_learner_max_epochs=15 , secondary_learner_batch_size=128 , eval_freq=100 , igf_model_path="""igf_model.pt""" , ) # load pretrained gpt2 model SCREAMING_SNAKE_CASE__ = GPTaLMHeadModel.from_pretrained("""gpt2""" ) set_seed(42 ) # Generate train and test data to train and evaluate gpt2 model SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = generate_datasets( context_len=32 , file="""data/tokenized_stories_train_wikitext103.jbl""" , number=100 , min_len=1_026 , trim=UpperCamelCase__ ) # fine-tuning of the gpt2 model using igf (Information Gain Filtration) finetune( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , context_len=32 , max_steps=1_000 , batch_size=16 , threshold=1.0 , recopy_model=UpperCamelCase__ , secondary_learner=UpperCamelCase__ , eval_interval=10 , finetuned_model_name="""gpt2_finetuned.pt""" , ) if __name__ == "__main__": main()
59
import warnings from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch from ...models import UNetaDModel from ...schedulers import RePaintScheduler from ...utils import PIL_INTERPOLATION, logging, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput _lowerCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[List, PIL.Image.Image, torch.Tensor] ): warnings.warn( """The preprocess method is deprecated and will be removed in a future version. Please""" """ use VaeImageProcessor.preprocess instead""" , UpperCamelCase__ , ) if isinstance(UpperCamelCase__ , torch.Tensor ): return image elif isinstance(UpperCamelCase__ , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ = [image] if isinstance(image[0] , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = image[0].size SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = (x - x % 8 for x in (w, h)) # resize to integer multiple of 8 SCREAMING_SNAKE_CASE__ = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION["""lanczos"""] ) )[None, :] for i in image] SCREAMING_SNAKE_CASE__ = np.concatenate(UpperCamelCase__ , axis=0 ) SCREAMING_SNAKE_CASE__ = np.array(UpperCamelCase__ ).astype(np.floataa ) / 2_5_5.0 SCREAMING_SNAKE_CASE__ = image.transpose(0 , 3 , 1 , 2 ) SCREAMING_SNAKE_CASE__ = 2.0 * image - 1.0 SCREAMING_SNAKE_CASE__ = torch.from_numpy(UpperCamelCase__ ) elif isinstance(image[0] , torch.Tensor ): SCREAMING_SNAKE_CASE__ = torch.cat(UpperCamelCase__ , dim=0 ) return image def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[List, PIL.Image.Image, torch.Tensor] ): if isinstance(UpperCamelCase__ , torch.Tensor ): return mask elif isinstance(UpperCamelCase__ , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ = [mask] if isinstance(mask[0] , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = mask[0].size SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 SCREAMING_SNAKE_CASE__ = [np.array(m.convert("""L""" ).resize((w, h) , resample=PIL_INTERPOLATION["""nearest"""] ) )[None, :] for m in mask] SCREAMING_SNAKE_CASE__ = np.concatenate(UpperCamelCase__ , axis=0 ) SCREAMING_SNAKE_CASE__ = mask.astype(np.floataa ) / 2_5_5.0 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = torch.from_numpy(UpperCamelCase__ ) elif isinstance(mask[0] , torch.Tensor ): SCREAMING_SNAKE_CASE__ = torch.cat(UpperCamelCase__ , dim=0 ) return mask class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = 42 lowerCamelCase_ = 42 def __init__( self :Any , __A :List[Any] , __A :Optional[Any] ) -> Union[str, Any]: """simple docstring""" super().__init__() self.register_modules(unet=__A , scheduler=__A ) @torch.no_grad() def __call__( self :str , __A :Union[torch.Tensor, PIL.Image.Image] , __A :Union[torch.Tensor, PIL.Image.Image] , __A :int = 250 , __A :float = 0.0 , __A :int = 10 , __A :int = 10 , __A :Optional[Union[torch.Generator, List[torch.Generator]]] = None , __A :Optional[str] = "pil" , __A :bool = True , ) -> Union[ImagePipelineOutput, Tuple]: """simple docstring""" SCREAMING_SNAKE_CASE__ = image SCREAMING_SNAKE_CASE__ = _preprocess_image(__A ) SCREAMING_SNAKE_CASE__ = original_image.to(device=self.device , dtype=self.unet.dtype ) SCREAMING_SNAKE_CASE__ = _preprocess_mask(__A ) SCREAMING_SNAKE_CASE__ = mask_image.to(device=self.device , dtype=self.unet.dtype ) SCREAMING_SNAKE_CASE__ = original_image.shape[0] # sample gaussian noise to begin the loop if isinstance(__A , __A ) and len(__A ) != batch_size: raise ValueError( f'''You have passed a list of generators of length {len(__A )}, but requested an effective batch''' f''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) SCREAMING_SNAKE_CASE__ = original_image.shape SCREAMING_SNAKE_CASE__ = randn_tensor(__A , generator=__A , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(__A , __A , __A , self.device ) SCREAMING_SNAKE_CASE__ = eta SCREAMING_SNAKE_CASE__ = self.scheduler.timesteps[0] + 1 SCREAMING_SNAKE_CASE__ = generator[0] if isinstance(__A , __A ) else generator for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): if t < t_last: # predict the noise residual SCREAMING_SNAKE_CASE__ = self.unet(__A , __A ).sample # compute previous image: x_t -> x_t-1 SCREAMING_SNAKE_CASE__ = self.scheduler.step(__A , __A , __A , __A , __A , __A ).prev_sample else: # compute the reverse: x_t-1 -> x_t SCREAMING_SNAKE_CASE__ = self.scheduler.undo_step(__A , __A , __A ) SCREAMING_SNAKE_CASE__ = t SCREAMING_SNAKE_CASE__ = (image / 2 + 0.5).clamp(0 , 1 ) SCREAMING_SNAKE_CASE__ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": SCREAMING_SNAKE_CASE__ = self.numpy_to_pil(__A ) if not return_dict: return (image,) return ImagePipelineOutput(images=__A )
59
1
import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_MULTIPLE_CHOICE_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, ) from transformers.models.layoutlmva.modeling_layoutlmva import LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class UpperCamelCase_ : def __init__( self :Optional[Any] , __A :Tuple , __A :List[Any]=2 , __A :Tuple=3 , __A :Union[str, Any]=4 , __A :List[Any]=2 , __A :List[str]=7 , __A :str=True , __A :List[Any]=True , __A :str=True , __A :Union[str, Any]=True , __A :Dict=99 , __A :Union[str, Any]=36 , __A :Dict=3 , __A :int=4 , __A :Tuple=37 , __A :Union[str, Any]="gelu" , __A :int=0.1 , __A :Optional[Any]=0.1 , __A :int=512 , __A :List[Any]=16 , __A :Any=2 , __A :Dict=0.0_2 , __A :List[Any]=6 , __A :Union[str, Any]=6 , __A :Tuple=3 , __A :Tuple=4 , __A :Optional[Any]=None , __A :str=1000 , ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = parent SCREAMING_SNAKE_CASE__ = batch_size SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = image_size SCREAMING_SNAKE_CASE__ = patch_size SCREAMING_SNAKE_CASE__ = text_seq_length SCREAMING_SNAKE_CASE__ = is_training SCREAMING_SNAKE_CASE__ = use_input_mask SCREAMING_SNAKE_CASE__ = use_token_type_ids SCREAMING_SNAKE_CASE__ = use_labels SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = type_vocab_size SCREAMING_SNAKE_CASE__ = type_sequence_label_size SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = coordinate_size SCREAMING_SNAKE_CASE__ = shape_size SCREAMING_SNAKE_CASE__ = num_labels SCREAMING_SNAKE_CASE__ = num_choices SCREAMING_SNAKE_CASE__ = scope SCREAMING_SNAKE_CASE__ = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) SCREAMING_SNAKE_CASE__ = text_seq_length SCREAMING_SNAKE_CASE__ = (image_size // patch_size) ** 2 + 1 SCREAMING_SNAKE_CASE__ = self.text_seq_length + self.image_seq_length def _snake_case ( self :str ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: SCREAMING_SNAKE_CASE__ = bbox[i, j, 3] SCREAMING_SNAKE_CASE__ = bbox[i, j, 1] SCREAMING_SNAKE_CASE__ = t if bbox[i, j, 2] < bbox[i, j, 0]: SCREAMING_SNAKE_CASE__ = bbox[i, j, 2] SCREAMING_SNAKE_CASE__ = bbox[i, j, 0] SCREAMING_SNAKE_CASE__ = t SCREAMING_SNAKE_CASE__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE__ = None if self.use_input_mask: SCREAMING_SNAKE_CASE__ = random_attention_mask([self.batch_size, self.text_seq_length] ) SCREAMING_SNAKE_CASE__ = None if self.use_token_type_ids: SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None if self.use_labels: SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) SCREAMING_SNAKE_CASE__ = LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def _snake_case ( self :Any , __A :Optional[Any] , __A :Optional[Any] , __A :List[str] , __A :List[str] , __A :int , __A :Optional[Any] , __A :Tuple , __A :Tuple ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = LayoutLMvaModel(config=__A ) model.to(__A ) model.eval() # text + image SCREAMING_SNAKE_CASE__ = model(__A , pixel_values=__A ) SCREAMING_SNAKE_CASE__ = model( __A , bbox=__A , pixel_values=__A , attention_mask=__A , token_type_ids=__A ) SCREAMING_SNAKE_CASE__ = model(__A , bbox=__A , pixel_values=__A , token_type_ids=__A ) SCREAMING_SNAKE_CASE__ = model(__A , bbox=__A , pixel_values=__A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only SCREAMING_SNAKE_CASE__ = model(__A ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only SCREAMING_SNAKE_CASE__ = model(pixel_values=__A ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def _snake_case ( self :List[Any] , __A :List[Any] , __A :Optional[Any] , __A :Union[str, Any] , __A :int , __A :List[str] , __A :Any , __A :str , __A :Dict ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.num_labels SCREAMING_SNAKE_CASE__ = LayoutLMvaForSequenceClassification(__A ) model.to(__A ) model.eval() SCREAMING_SNAKE_CASE__ = model( __A , bbox=__A , pixel_values=__A , attention_mask=__A , token_type_ids=__A , labels=__A , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _snake_case ( self :List[Any] , __A :int , __A :List[Any] , __A :Optional[int] , __A :Dict , __A :Optional[Any] , __A :Union[str, Any] , __A :List[str] , __A :List[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.num_labels SCREAMING_SNAKE_CASE__ = LayoutLMvaForTokenClassification(config=__A ) model.to(__A ) model.eval() SCREAMING_SNAKE_CASE__ = model( __A , bbox=__A , pixel_values=__A , attention_mask=__A , token_type_ids=__A , labels=__A , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def _snake_case ( self :Union[str, Any] , __A :Tuple , __A :Union[str, Any] , __A :Tuple , __A :Optional[Any] , __A :List[Any] , __A :Union[str, Any] , __A :Optional[int] , __A :int ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = LayoutLMvaForQuestionAnswering(config=__A ) model.to(__A ) model.eval() SCREAMING_SNAKE_CASE__ = model( __A , bbox=__A , pixel_values=__A , attention_mask=__A , token_type_ids=__A , start_positions=__A , end_positions=__A , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _snake_case ( self :Optional[int] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.prepare_config_and_inputs() ( ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ) = config_and_inputs SCREAMING_SNAKE_CASE__ = { """input_ids""": input_ids, """bbox""": bbox, """pixel_values""": pixel_values, """token_type_ids""": token_type_ids, """attention_mask""": input_mask, } return config, inputs_dict @require_torch class UpperCamelCase_ ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = False lowerCamelCase_ = False lowerCamelCase_ = False lowerCamelCase_ = ( ( LayoutLMvaModel, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaForQuestionAnswering, ) if is_torch_available() else () ) lowerCamelCase_ = ( {"document-question-answering": LayoutLMvaForQuestionAnswering, "feature-extraction": LayoutLMvaModel} if is_torch_available() else {} ) def _snake_case ( self :Any , __A :str , __A :Dict , __A :Optional[int] , __A :List[str] , __A :Union[str, Any] ) -> Union[str, Any]: """simple docstring""" return True def _snake_case ( self :int ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = LayoutLMvaModelTester(self ) SCREAMING_SNAKE_CASE__ = ConfigTester(self , config_class=__A , hidden_size=37 ) def _snake_case ( self :str , __A :Any , __A :Union[str, Any] , __A :List[str]=False ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = copy.deepcopy(__A ) if model_class in get_values(__A ): SCREAMING_SNAKE_CASE__ = { k: v.unsqueeze(1 ).expand(-1 , self.model_tester.num_choices , -1 ).contiguous() if isinstance(__A , torch.Tensor ) and v.ndim > 1 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(__A ): SCREAMING_SNAKE_CASE__ = torch.ones(self.model_tester.batch_size , dtype=torch.long , device=__A ) elif model_class in get_values(__A ): SCREAMING_SNAKE_CASE__ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__A ) SCREAMING_SNAKE_CASE__ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__A ) elif model_class in [ *get_values(__A ), ]: SCREAMING_SNAKE_CASE__ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__A ) elif model_class in [ *get_values(__A ), ]: SCREAMING_SNAKE_CASE__ = torch.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=torch.long , device=__A , ) return inputs_dict def _snake_case ( self :str ) -> Optional[int]: """simple docstring""" self.config_tester.run_common_tests() def _snake_case ( self :Optional[int] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__A ) def _snake_case ( self :List[Any] ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: SCREAMING_SNAKE_CASE__ = type self.model_tester.create_and_check_model(*__A ) def _snake_case ( self :Tuple ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__A ) def _snake_case ( self :Any ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__A ) def _snake_case ( self :Optional[Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__A ) @slow def _snake_case ( self :Optional[int] ) -> Optional[int]: """simple docstring""" for model_name in LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE__ = LayoutLMvaModel.from_pretrained(__A ) self.assertIsNotNone(__A ) def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch class UpperCamelCase_ ( unittest.TestCase ): @cached_property def _snake_case ( self :str ) -> int: """simple docstring""" return LayoutLMvaImageProcessor(apply_ocr=__A ) if is_vision_available() else None @slow def _snake_case ( self :str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = LayoutLMvaModel.from_pretrained("""microsoft/layoutlmv3-base""" ).to(__A ) SCREAMING_SNAKE_CASE__ = self.default_image_processor SCREAMING_SNAKE_CASE__ = prepare_img() SCREAMING_SNAKE_CASE__ = image_processor(images=__A , return_tensors="""pt""" ).pixel_values.to(__A ) SCREAMING_SNAKE_CASE__ = torch.tensor([[1, 2]] ) SCREAMING_SNAKE_CASE__ = torch.tensor([[1, 2, 3, 4], [5, 6, 7, 8]] ).unsqueeze(0 ) # forward pass SCREAMING_SNAKE_CASE__ = model( input_ids=input_ids.to(__A ) , bbox=bbox.to(__A ) , pixel_values=pixel_values.to(__A ) , ) # verify the logits SCREAMING_SNAKE_CASE__ = torch.Size((1, 199, 768) ) self.assertEqual(outputs.last_hidden_state.shape , __A ) SCREAMING_SNAKE_CASE__ = torch.tensor( [[-0.0_5_2_9, 0.3_6_1_8, 0.1_6_3_2], [-0.1_5_8_7, -0.1_6_6_7, -0.0_4_0_0], [-0.1_5_5_7, -0.1_6_7_1, -0.0_5_0_5]] ).to(__A ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , __A , atol=1E-4 ) )
59
import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = OpenAIGPTTokenizer lowerCamelCase_ = OpenAIGPTTokenizerFast lowerCamelCase_ = True lowerCamelCase_ = False def _snake_case ( self :Optional[Any] ) -> Dict: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt SCREAMING_SNAKE_CASE__ = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """w</w>""", """r</w>""", """t</w>""", """lo""", """low""", """er</w>""", """low</w>""", """lowest</w>""", """newer</w>""", """wider</w>""", """<unk>""", ] SCREAMING_SNAKE_CASE__ = dict(zip(__A , range(len(__A ) ) ) ) SCREAMING_SNAKE_CASE__ = ["""#version: 0.2""", """l o""", """lo w""", """e r</w>""", """"""] SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" ) as fp: fp.write(json.dumps(__A ) ) with open(self.merges_file , """w""" ) as fp: fp.write("""\n""".join(__A ) ) def _snake_case ( self :Union[str, Any] , __A :str ) -> List[Any]: """simple docstring""" return "lower newer", "lower newer" def _snake_case ( self :Optional[Any] ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = OpenAIGPTTokenizer(self.vocab_file , self.merges_file ) SCREAMING_SNAKE_CASE__ = """lower""" SCREAMING_SNAKE_CASE__ = ["""low""", """er</w>"""] SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(__A ) self.assertListEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = tokens + ["""<unk>"""] SCREAMING_SNAKE_CASE__ = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , __A ) def _snake_case ( self :Optional[Any] , __A :Optional[Any]=15 ) -> Any: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) # Simple input SCREAMING_SNAKE_CASE__ = """This is a simple input""" SCREAMING_SNAKE_CASE__ = ["""This is a simple input 1""", """This is a simple input 2"""] SCREAMING_SNAKE_CASE__ = ("""This is a simple input""", """This is a pair""") SCREAMING_SNAKE_CASE__ = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests self.assertRaises(__A , tokenizer_r.encode , __A , max_length=__A , padding="""max_length""" ) # Simple input self.assertRaises(__A , tokenizer_r.encode_plus , __A , max_length=__A , padding="""max_length""" ) # Simple input self.assertRaises( __A , tokenizer_r.batch_encode_plus , __A , max_length=__A , padding="""max_length""" , ) # Pair input self.assertRaises(__A , tokenizer_r.encode , __A , max_length=__A , padding="""max_length""" ) # Pair input self.assertRaises(__A , tokenizer_r.encode_plus , __A , max_length=__A , padding="""max_length""" ) # Pair input self.assertRaises( __A , tokenizer_r.batch_encode_plus , __A , max_length=__A , padding="""max_length""" , ) def _snake_case ( self :Dict ) -> List[Any]: """simple docstring""" pass @require_ftfy @require_spacy @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ ): pass
59
1
import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int = 3 ): if isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise TypeError("""number of qubits must be a integer.""" ) if number_of_qubits <= 0: raise ValueError("""number of qubits must be > 0.""" ) if math.floor(UpperCamelCase__ ) != number_of_qubits: raise ValueError("""number of qubits must be exact integer.""" ) if number_of_qubits > 10: raise ValueError("""number of qubits too large to simulate(>10).""" ) SCREAMING_SNAKE_CASE__ = QuantumRegister(UpperCamelCase__ , """qr""" ) SCREAMING_SNAKE_CASE__ = ClassicalRegister(UpperCamelCase__ , """cr""" ) SCREAMING_SNAKE_CASE__ = QuantumCircuit(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = number_of_qubits for i in range(UpperCamelCase__ ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(UpperCamelCase__ ): quantum_circuit.cp(np.pi / 2 ** (counter - j) , UpperCamelCase__ , UpperCamelCase__ ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(UpperCamelCase__ , number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(UpperCamelCase__ , UpperCamelCase__ ) # simulate with 10000 shots SCREAMING_SNAKE_CASE__ = Aer.get_backend("""qasm_simulator""" ) SCREAMING_SNAKE_CASE__ = execute(UpperCamelCase__ , UpperCamelCase__ , shots=10_000 ) return job.result().get_counts(UpperCamelCase__ ) if __name__ == "__main__": print( F'''Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}''' )
59
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=UpperCamelCase__ ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = field(default="image-classification" , metadata={"include_in_asdict_even_if_is_default": True} ) lowerCamelCase_ = Features({"image": Image()} ) lowerCamelCase_ = Features({"labels": ClassLabel} ) lowerCamelCase_ = "image" lowerCamelCase_ = "labels" def _snake_case ( self :List[str] , __A :Tuple ) -> Tuple: """simple docstring""" if self.label_column not in features: raise ValueError(f'''Column {self.label_column} is not present in features.''' ) if not isinstance(features[self.label_column] , __A ): raise ValueError(f'''Column {self.label_column} is not a ClassLabel.''' ) SCREAMING_SNAKE_CASE__ = copy.deepcopy(self ) SCREAMING_SNAKE_CASE__ = self.label_schema.copy() SCREAMING_SNAKE_CASE__ = features[self.label_column] SCREAMING_SNAKE_CASE__ = label_schema return task_template @property def _snake_case ( self :Dict ) -> Dict[str, str]: """simple docstring""" return { self.image_column: "image", self.label_column: "labels", }
59
1
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int = 600_851_475_143 ): try: SCREAMING_SNAKE_CASE__ = int(UpperCamelCase__ ) except (TypeError, ValueError): raise TypeError("""Parameter n must be int or castable to int.""" ) if n <= 0: raise ValueError("""Parameter n must be greater than or equal to one.""" ) SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = 2 while i * i <= n: while n % i == 0: SCREAMING_SNAKE_CASE__ = i n //= i i += 1 if n > 1: SCREAMING_SNAKE_CASE__ = n return int(UpperCamelCase__ ) if __name__ == "__main__": print(F'''{solution() = }''')
59
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCamelCase = { 'configuration_xmod': [ 'XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XmodConfig', 'XmodOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'XMOD_PRETRAINED_MODEL_ARCHIVE_LIST', 'XmodForCausalLM', 'XmodForMaskedLM', 'XmodForMultipleChoice', 'XmodForQuestionAnswering', 'XmodForSequenceClassification', 'XmodForTokenClassification', 'XmodModel', 'XmodPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
1
from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker _lowerCamelCase = 'CompVis/stable-diffusion-v1-1' _lowerCamelCase = 'CompVis/stable-diffusion-v1-2' _lowerCamelCase = 'CompVis/stable-diffusion-v1-3' _lowerCamelCase = 'CompVis/stable-diffusion-v1-4' class UpperCamelCase_ ( UpperCamelCase__ ): def __init__( self :Union[str, Any] , __A :AutoencoderKL , __A :CLIPTextModel , __A :CLIPTokenizer , __A :UNetaDConditionModel , __A :Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , __A :StableDiffusionSafetyChecker , __A :CLIPImageProcessor , __A :bool = True , ) -> Dict: """simple docstring""" super()._init_() SCREAMING_SNAKE_CASE__ = StableDiffusionPipeline.from_pretrained(__A ) SCREAMING_SNAKE_CASE__ = StableDiffusionPipeline.from_pretrained(__A ) SCREAMING_SNAKE_CASE__ = StableDiffusionPipeline.from_pretrained(__A ) SCREAMING_SNAKE_CASE__ = StableDiffusionPipeline( vae=__A , text_encoder=__A , tokenizer=__A , unet=__A , scheduler=__A , safety_checker=__A , feature_extractor=__A , requires_safety_checker=__A , ) self.register_modules(pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea ) @property def _snake_case ( self :Optional[int] ) -> Dict[str, Any]: """simple docstring""" return {k: getattr(self , __A ) for k in self.config.keys() if not k.startswith("""_""" )} def _snake_case ( self :str , __A :Optional[Union[str, int]] = "auto" ) -> Dict: """simple docstring""" if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory SCREAMING_SNAKE_CASE__ = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(__A ) def _snake_case ( self :Optional[Any] ) -> Union[str, Any]: """simple docstring""" self.enable_attention_slicing(__A ) @torch.no_grad() def _snake_case ( self :Dict , __A :Union[str, List[str]] , __A :int = 512 , __A :int = 512 , __A :int = 50 , __A :float = 7.5 , __A :Optional[Union[str, List[str]]] = None , __A :Optional[int] = 1 , __A :float = 0.0 , __A :Optional[torch.Generator] = None , __A :Optional[torch.FloatTensor] = None , __A :Optional[str] = "pil" , __A :bool = True , __A :Optional[Callable[[int, int, torch.FloatTensor], None]] = None , __A :int = 1 , **__A :Tuple , ) -> str: """simple docstring""" return self.pipea( prompt=__A , height=__A , width=__A , num_inference_steps=__A , guidance_scale=__A , negative_prompt=__A , num_images_per_prompt=__A , eta=__A , generator=__A , latents=__A , output_type=__A , return_dict=__A , callback=__A , callback_steps=__A , **__A , ) @torch.no_grad() def _snake_case ( self :int , __A :Union[str, List[str]] , __A :int = 512 , __A :int = 512 , __A :int = 50 , __A :float = 7.5 , __A :Optional[Union[str, List[str]]] = None , __A :Optional[int] = 1 , __A :float = 0.0 , __A :Optional[torch.Generator] = None , __A :Optional[torch.FloatTensor] = None , __A :Optional[str] = "pil" , __A :bool = True , __A :Optional[Callable[[int, int, torch.FloatTensor], None]] = None , __A :int = 1 , **__A :Optional[Any] , ) -> Any: """simple docstring""" return self.pipea( prompt=__A , height=__A , width=__A , num_inference_steps=__A , guidance_scale=__A , negative_prompt=__A , num_images_per_prompt=__A , eta=__A , generator=__A , latents=__A , output_type=__A , return_dict=__A , callback=__A , callback_steps=__A , **__A , ) @torch.no_grad() def _snake_case ( self :List[Any] , __A :Union[str, List[str]] , __A :int = 512 , __A :int = 512 , __A :int = 50 , __A :float = 7.5 , __A :Optional[Union[str, List[str]]] = None , __A :Optional[int] = 1 , __A :float = 0.0 , __A :Optional[torch.Generator] = None , __A :Optional[torch.FloatTensor] = None , __A :Optional[str] = "pil" , __A :bool = True , __A :Optional[Callable[[int, int, torch.FloatTensor], None]] = None , __A :int = 1 , **__A :Union[str, Any] , ) -> Any: """simple docstring""" return self.pipea( prompt=__A , height=__A , width=__A , num_inference_steps=__A , guidance_scale=__A , negative_prompt=__A , num_images_per_prompt=__A , eta=__A , generator=__A , latents=__A , output_type=__A , return_dict=__A , callback=__A , callback_steps=__A , **__A , ) @torch.no_grad() def _snake_case ( self :Union[str, Any] , __A :Union[str, List[str]] , __A :int = 512 , __A :int = 512 , __A :int = 50 , __A :float = 7.5 , __A :Optional[Union[str, List[str]]] = None , __A :Optional[int] = 1 , __A :float = 0.0 , __A :Optional[torch.Generator] = None , __A :Optional[torch.FloatTensor] = None , __A :Optional[str] = "pil" , __A :bool = True , __A :Optional[Callable[[int, int, torch.FloatTensor], None]] = None , __A :int = 1 , **__A :List[str] , ) -> Any: """simple docstring""" return self.pipea( prompt=__A , height=__A , width=__A , num_inference_steps=__A , guidance_scale=__A , negative_prompt=__A , num_images_per_prompt=__A , eta=__A , generator=__A , latents=__A , output_type=__A , return_dict=__A , callback=__A , callback_steps=__A , **__A , ) @torch.no_grad() def _snake_case ( self :List[Any] , __A :Union[str, List[str]] , __A :int = 512 , __A :int = 512 , __A :int = 50 , __A :float = 7.5 , __A :Optional[Union[str, List[str]]] = None , __A :Optional[int] = 1 , __A :float = 0.0 , __A :Optional[torch.Generator] = None , __A :Optional[torch.FloatTensor] = None , __A :Optional[str] = "pil" , __A :bool = True , __A :Optional[Callable[[int, int, torch.FloatTensor], None]] = None , __A :int = 1 , **__A :str , ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = """cuda""" if torch.cuda.is_available() else """cpu""" self.to(__A ) # Checks if the height and width are divisible by 8 or not if height % 8 != 0 or width % 8 != 0: raise ValueError(f'''`height` and `width` must be divisible by 8 but are {height} and {width}.''' ) # Get first result from Stable Diffusion Checkpoint v1.1 SCREAMING_SNAKE_CASE__ = self.textaimg_sda_a( prompt=__A , height=__A , width=__A , num_inference_steps=__A , guidance_scale=__A , negative_prompt=__A , num_images_per_prompt=__A , eta=__A , generator=__A , latents=__A , output_type=__A , return_dict=__A , callback=__A , callback_steps=__A , **__A , ) # Get first result from Stable Diffusion Checkpoint v1.2 SCREAMING_SNAKE_CASE__ = self.textaimg_sda_a( prompt=__A , height=__A , width=__A , num_inference_steps=__A , guidance_scale=__A , negative_prompt=__A , num_images_per_prompt=__A , eta=__A , generator=__A , latents=__A , output_type=__A , return_dict=__A , callback=__A , callback_steps=__A , **__A , ) # Get first result from Stable Diffusion Checkpoint v1.3 SCREAMING_SNAKE_CASE__ = self.textaimg_sda_a( prompt=__A , height=__A , width=__A , num_inference_steps=__A , guidance_scale=__A , negative_prompt=__A , num_images_per_prompt=__A , eta=__A , generator=__A , latents=__A , output_type=__A , return_dict=__A , callback=__A , callback_steps=__A , **__A , ) # Get first result from Stable Diffusion Checkpoint v1.4 SCREAMING_SNAKE_CASE__ = self.textaimg_sda_a( prompt=__A , height=__A , width=__A , num_inference_steps=__A , guidance_scale=__A , negative_prompt=__A , num_images_per_prompt=__A , eta=__A , generator=__A , latents=__A , output_type=__A , return_dict=__A , callback=__A , callback_steps=__A , **__A , ) # Get all result images into a single list and pass it via StableDiffusionPipelineOutput for final result return StableDiffusionPipelineOutput([resa[0], resa[0], resa[0], resa[0]] )
59
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "timm_backbone" def __init__( self :Union[str, Any] , __A :str=None , __A :Union[str, Any]=3 , __A :str=True , __A :Any=True , __A :Optional[Any]=None , **__A :List[str] , ) -> Tuple: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = backbone SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = features_only SCREAMING_SNAKE_CASE__ = use_pretrained_backbone SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = out_indices if out_indices is not None else (-1,)
59
1
from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput _lowerCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name class UpperCamelCase_ ( UpperCamelCase__ , UpperCamelCase__ ): @register_to_config def __init__( self :Any , __A :bool , __A :Optional[int] = None , __A :Optional[int] = None ) -> str: """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE__ = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" SCREAMING_SNAKE_CASE__ = torch.zeros(__A , __A ) else: SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = torch.nn.Parameter(__A ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = 42 lowerCamelCase_ = 42 lowerCamelCase_ = 42 lowerCamelCase_ = 42 lowerCamelCase_ = 42 lowerCamelCase_ = 42 def __init__( self :List[Any] , __A :VQModel , __A :CLIPTextModel , __A :CLIPTokenizer , __A :TransformeraDModel , __A :VQDiffusionScheduler , __A :LearnedClassifierFreeSamplingEmbeddings , ) -> Tuple: """simple docstring""" super().__init__() self.register_modules( vqvae=__A , transformer=__A , text_encoder=__A , tokenizer=__A , scheduler=__A , learned_classifier_free_sampling_embeddings=__A , ) def _snake_case ( self :int , __A :int , __A :Dict , __A :Tuple ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = len(__A ) if isinstance(__A , __A ) else 1 # get prompt text embeddings SCREAMING_SNAKE_CASE__ = self.tokenizer( __A , padding="""max_length""" , max_length=self.tokenizer.model_max_length , return_tensors="""pt""" , ) SCREAMING_SNAKE_CASE__ = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: SCREAMING_SNAKE_CASE__ = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( """The following part of your input was truncated because CLIP can only handle sequences up to""" f''' {self.tokenizer.model_max_length} tokens: {removed_text}''' ) SCREAMING_SNAKE_CASE__ = text_input_ids[:, : self.tokenizer.model_max_length] SCREAMING_SNAKE_CASE__ = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 SCREAMING_SNAKE_CASE__ = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=__A ) # duplicate text embeddings for each generation per prompt SCREAMING_SNAKE_CASE__ = prompt_embeds.repeat_interleave(__A , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: SCREAMING_SNAKE_CASE__ = self.learned_classifier_free_sampling_embeddings.embeddings SCREAMING_SNAKE_CASE__ = negative_prompt_embeds.unsqueeze(0 ).repeat(__A , 1 , 1 ) else: SCREAMING_SNAKE_CASE__ = [""""""] * batch_size SCREAMING_SNAKE_CASE__ = text_input_ids.shape[-1] SCREAMING_SNAKE_CASE__ = self.tokenizer( __A , padding="""max_length""" , max_length=__A , truncation=__A , return_tensors="""pt""" , ) SCREAMING_SNAKE_CASE__ = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings SCREAMING_SNAKE_CASE__ = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=__A ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method SCREAMING_SNAKE_CASE__ = negative_prompt_embeds.shape[1] SCREAMING_SNAKE_CASE__ = negative_prompt_embeds.repeat(1 , __A , 1 ) SCREAMING_SNAKE_CASE__ = negative_prompt_embeds.view(batch_size * num_images_per_prompt , __A , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes SCREAMING_SNAKE_CASE__ = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self :Tuple , __A :Union[str, List[str]] , __A :int = 100 , __A :float = 5.0 , __A :float = 1.0 , __A :int = 1 , __A :Optional[Union[torch.Generator, List[torch.Generator]]] = None , __A :Optional[torch.FloatTensor] = None , __A :Optional[str] = "pil" , __A :bool = True , __A :Optional[Callable[[int, int, torch.FloatTensor], None]] = None , __A :int = 1 , ) -> Union[ImagePipelineOutput, Tuple]: """simple docstring""" if isinstance(__A , __A ): SCREAMING_SNAKE_CASE__ = 1 elif isinstance(__A , __A ): SCREAMING_SNAKE_CASE__ = len(__A ) else: raise ValueError(f'''`prompt` has to be of type `str` or `list` but is {type(__A )}''' ) SCREAMING_SNAKE_CASE__ = batch_size * num_images_per_prompt SCREAMING_SNAKE_CASE__ = guidance_scale > 1.0 SCREAMING_SNAKE_CASE__ = self._encode_prompt(__A , __A , __A ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__A , __A ) or callback_steps <= 0) ): raise ValueError( f'''`callback_steps` has to be a positive integer but is {callback_steps} of type''' f''' {type(__A )}.''' ) # get the initial completely masked latents unless the user supplied it SCREAMING_SNAKE_CASE__ = (batch_size, self.transformer.num_latent_pixels) if latents is None: SCREAMING_SNAKE_CASE__ = self.transformer.num_vector_embeds - 1 SCREAMING_SNAKE_CASE__ = torch.full(__A , __A ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {latents_shape}''' ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( """Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0,""" f''' {self.transformer.num_vector_embeds - 1} (inclusive).''' ) SCREAMING_SNAKE_CASE__ = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(__A , device=self.device ) SCREAMING_SNAKE_CASE__ = self.scheduler.timesteps.to(self.device ) SCREAMING_SNAKE_CASE__ = latents for i, t in enumerate(self.progress_bar(__A ) ): # expand the sample if we are doing classifier free guidance SCREAMING_SNAKE_CASE__ = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` SCREAMING_SNAKE_CASE__ = self.transformer(__A , encoder_hidden_states=__A , timestep=__A ).sample if do_classifier_free_guidance: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = model_output.chunk(2 ) SCREAMING_SNAKE_CASE__ = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(__A , dim=1 , keepdim=__A ) SCREAMING_SNAKE_CASE__ = self.truncate(__A , __A ) # remove `log(0)`'s (`-inf`s) SCREAMING_SNAKE_CASE__ = model_output.clamp(-70 ) # compute the previous noisy sample x_t -> x_t-1 SCREAMING_SNAKE_CASE__ = self.scheduler.step(__A , timestep=__A , sample=__A , generator=__A ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__A , __A , __A ) SCREAMING_SNAKE_CASE__ = self.vqvae.config.vq_embed_dim SCREAMING_SNAKE_CASE__ = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) SCREAMING_SNAKE_CASE__ = self.vqvae.quantize.get_codebook_entry(__A , shape=__A ) SCREAMING_SNAKE_CASE__ = self.vqvae.decode(__A , force_not_quantize=__A ).sample SCREAMING_SNAKE_CASE__ = (image / 2 + 0.5).clamp(0 , 1 ) SCREAMING_SNAKE_CASE__ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": SCREAMING_SNAKE_CASE__ = self.numpy_to_pil(__A ) if not return_dict: return (image,) return ImagePipelineOutput(images=__A ) def _snake_case ( self :Any , __A :torch.FloatTensor , __A :float ) -> torch.FloatTensor: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = torch.sort(__A , 1 , descending=__A ) SCREAMING_SNAKE_CASE__ = torch.exp(__A ) SCREAMING_SNAKE_CASE__ = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out SCREAMING_SNAKE_CASE__ = torch.full_like(keep_mask[:, 0:1, :] , __A ) SCREAMING_SNAKE_CASE__ = torch.cat((all_true, keep_mask) , dim=1 ) SCREAMING_SNAKE_CASE__ = keep_mask[:, :-1, :] SCREAMING_SNAKE_CASE__ = keep_mask.gather(1 , indices.argsort(1 ) ) SCREAMING_SNAKE_CASE__ = log_p_x_0.clone() SCREAMING_SNAKE_CASE__ = -torch.inf # -inf = log(0) return rv
59
from math import pow, sqrt def SCREAMING_SNAKE_CASE__ ( *UpperCamelCase__: float ): SCREAMING_SNAKE_CASE__ = len(UpperCamelCase__ ) > 0 and all(value > 0.0 for value in values ) return result def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ ) else ValueError("""Input Error: Molar mass values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) )
59
1
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'google/pix2struct-textcaps-base': ( 'https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json' ), } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct_text_model" lowerCamelCase_ = ["past_key_values"] lowerCamelCase_ = { "hidden_size": "hidden_size", "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self :Union[str, Any] , __A :Any=5_0244 , __A :Optional[Any]=768 , __A :Tuple=64 , __A :List[str]=2048 , __A :int=12 , __A :str=12 , __A :Any=32 , __A :Tuple=128 , __A :int=0.1 , __A :str=1E-6 , __A :Optional[Any]=1.0 , __A :Union[str, Any]="gelu_new" , __A :Any=0 , __A :List[str]=False , __A :Optional[Any]=0 , __A :int=1 , __A :Optional[int]=False , __A :Optional[Any]=True , **__A :List[Any] , ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = d_kv SCREAMING_SNAKE_CASE__ = d_ff SCREAMING_SNAKE_CASE__ = num_layers SCREAMING_SNAKE_CASE__ = num_heads SCREAMING_SNAKE_CASE__ = relative_attention_num_buckets SCREAMING_SNAKE_CASE__ = relative_attention_max_distance SCREAMING_SNAKE_CASE__ = dropout_rate SCREAMING_SNAKE_CASE__ = layer_norm_epsilon SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = use_cache SCREAMING_SNAKE_CASE__ = eos_token_id SCREAMING_SNAKE_CASE__ = decoder_start_token_id # for backwards compatibility SCREAMING_SNAKE_CASE__ = dense_act_fn super().__init__( pad_token_id=__A , eos_token_id=__A , decoder_start_token_id=__A , tie_word_embeddings=__A , is_decoder=__A , **__A , ) @classmethod def _snake_case ( cls :Optional[int] , __A :Union[str, os.PathLike] , **__A :Optional[int] ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = cls.get_config_dict(__A , **__A ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get("""model_type""" ) == "pix2struct": SCREAMING_SNAKE_CASE__ = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct_vision_model" def __init__( self :Optional[int] , __A :int=768 , __A :Optional[Any]=768 , __A :Union[str, Any]=2048 , __A :int=64 , __A :Union[str, Any]=12 , __A :str=12 , __A :Any="gelu_new" , __A :List[Any]=1E-6 , __A :Dict=0.0 , __A :int=0.0 , __A :int=1E-10 , __A :Dict=1.0 , __A :int=4096 , __A :int=32 , __A :int=128 , **__A :Tuple , ) -> str: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = patch_embed_hidden_size SCREAMING_SNAKE_CASE__ = d_ff SCREAMING_SNAKE_CASE__ = dropout_rate SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = attention_dropout SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = dense_act_fn SCREAMING_SNAKE_CASE__ = seq_len SCREAMING_SNAKE_CASE__ = relative_attention_num_buckets SCREAMING_SNAKE_CASE__ = relative_attention_max_distance SCREAMING_SNAKE_CASE__ = d_kv @classmethod def _snake_case ( cls :str , __A :Union[str, os.PathLike] , **__A :str ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = cls.get_config_dict(__A , **__A ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get("""model_type""" ) == "pix2struct": SCREAMING_SNAKE_CASE__ = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct" lowerCamelCase_ = True def __init__( self :str , __A :Optional[Any]=None , __A :List[str]=None , __A :Optional[Any]=1.0 , __A :Optional[Any]=0.0_2 , __A :Any=False , __A :Tuple=False , __A :Any=True , **__A :Dict , ) -> Union[str, Any]: """simple docstring""" super().__init__(tie_word_embeddings=__A , is_encoder_decoder=__A , **__A ) if text_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""text_config is None. Initializing the Pix2StructTextConfig with default values.""" ) if vision_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""vision_config is None. Initializing the Pix2StructVisionConfig with default values.""" ) SCREAMING_SNAKE_CASE__ = PixaStructTextConfig(**__A ) SCREAMING_SNAKE_CASE__ = PixaStructVisionConfig(**__A ) SCREAMING_SNAKE_CASE__ = self.text_config.decoder_start_token_id SCREAMING_SNAKE_CASE__ = self.text_config.pad_token_id SCREAMING_SNAKE_CASE__ = self.text_config.eos_token_id SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = self.initializer_range SCREAMING_SNAKE_CASE__ = self.initializer_range SCREAMING_SNAKE_CASE__ = is_vqa @classmethod def _snake_case ( cls :Union[str, Any] , __A :PixaStructTextConfig , __A :PixaStructVisionConfig , **__A :Optional[int] ) -> Optional[Any]: """simple docstring""" return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__A ) def _snake_case ( self :str ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = copy.deepcopy(self.__dict__ ) SCREAMING_SNAKE_CASE__ = self.text_config.to_dict() SCREAMING_SNAKE_CASE__ = self.vision_config.to_dict() SCREAMING_SNAKE_CASE__ = self.__class__.model_type return output
59
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'shi-labs/nat-mini-in1k-224': 'https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json', # See all Nat models at https://huggingface.co/models?filter=nat } class UpperCamelCase_ ( UpperCamelCase__ , UpperCamelCase__ ): lowerCamelCase_ = "nat" lowerCamelCase_ = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self :List[Any] , __A :Optional[Any]=4 , __A :Any=3 , __A :Optional[int]=64 , __A :Optional[int]=[3, 4, 6, 5] , __A :Union[str, Any]=[2, 4, 8, 16] , __A :Optional[Any]=7 , __A :Optional[Any]=3.0 , __A :List[Any]=True , __A :int=0.0 , __A :Dict=0.0 , __A :Optional[Any]=0.1 , __A :str="gelu" , __A :Optional[Any]=0.0_2 , __A :Optional[int]=1E-5 , __A :Optional[int]=0.0 , __A :Optional[Any]=None , __A :Union[str, Any]=None , **__A :Union[str, Any] , ) -> Optional[int]: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = patch_size SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = embed_dim SCREAMING_SNAKE_CASE__ = depths SCREAMING_SNAKE_CASE__ = len(__A ) SCREAMING_SNAKE_CASE__ = num_heads SCREAMING_SNAKE_CASE__ = kernel_size SCREAMING_SNAKE_CASE__ = mlp_ratio SCREAMING_SNAKE_CASE__ = qkv_bias SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = drop_path_rate SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model SCREAMING_SNAKE_CASE__ = int(embed_dim * 2 ** (len(__A ) - 1) ) SCREAMING_SNAKE_CASE__ = layer_scale_init_value SCREAMING_SNAKE_CASE__ = ["""stem"""] + [f'''stage{idx}''' for idx in range(1 , len(__A ) + 1 )] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = get_aligned_output_features_output_indices( out_features=__A , out_indices=__A , stage_names=self.stage_names )
59
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'YituTech/conv-bert-base': 'https://huggingface.co/YituTech/conv-bert-base/resolve/main/config.json', 'YituTech/conv-bert-medium-small': ( 'https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/config.json' ), 'YituTech/conv-bert-small': 'https://huggingface.co/YituTech/conv-bert-small/resolve/main/config.json', # See all ConvBERT models at https://huggingface.co/models?filter=convbert } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "convbert" def __init__( self :List[Any] , __A :List[Any]=3_0522 , __A :int=768 , __A :Dict=12 , __A :int=12 , __A :int=3072 , __A :Any="gelu" , __A :List[str]=0.1 , __A :Any=0.1 , __A :Union[str, Any]=512 , __A :List[str]=2 , __A :Optional[int]=0.0_2 , __A :Union[str, Any]=1E-12 , __A :Tuple=1 , __A :Tuple=0 , __A :int=2 , __A :Union[str, Any]=768 , __A :int=2 , __A :str=9 , __A :Tuple=1 , __A :Optional[int]=None , **__A :Optional[Any] , ) -> int: """simple docstring""" super().__init__( pad_token_id=__A , bos_token_id=__A , eos_token_id=__A , **__A , ) SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = type_vocab_size SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = embedding_size SCREAMING_SNAKE_CASE__ = head_ratio SCREAMING_SNAKE_CASE__ = conv_kernel_size SCREAMING_SNAKE_CASE__ = num_groups SCREAMING_SNAKE_CASE__ = classifier_dropout class UpperCamelCase_ ( UpperCamelCase__ ): @property def _snake_case ( self :Union[str, Any] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": SCREAMING_SNAKE_CASE__ = {0: """batch""", 1: """choice""", 2: """sequence"""} else: SCREAMING_SNAKE_CASE__ = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
59
from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any ): return {key.lstrip("""-""" ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = ArgumentParser( """HuggingFace Datasets CLI tool""" , usage="""datasets-cli <command> [<args>]""" , allow_abbrev=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = parser.add_subparsers(help="""datasets-cli command helpers""" ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(UpperCamelCase__ ) EnvironmentCommand.register_subcommand(UpperCamelCase__ ) TestCommand.register_subcommand(UpperCamelCase__ ) RunBeamCommand.register_subcommand(UpperCamelCase__ ) DummyDataCommand.register_subcommand(UpperCamelCase__ ) # Parse args SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = parser.parse_known_args() if not hasattr(UpperCamelCase__ , """func""" ): parser.print_help() exit(1 ) SCREAMING_SNAKE_CASE__ = parse_unknown_args(UpperCamelCase__ ) # Run SCREAMING_SNAKE_CASE__ = args.func(UpperCamelCase__ , **UpperCamelCase__ ) service.run() if __name__ == "__main__": main()
59
1
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] ): SCREAMING_SNAKE_CASE__ = [] for data in source_data: for i, el in enumerate(UpperCamelCase__ ): if len(UpperCamelCase__ ) < i + 1: data_lists.append([] ) data_lists[i].append(float(UpperCamelCase__ ) ) return data_lists def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] , UpperCamelCase__: list[int] ): SCREAMING_SNAKE_CASE__ = [] for dlist, weight in zip(UpperCamelCase__ , UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = min(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = max(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = [] # for weight 0 score is 1 - actual score if weight == 0: for item in dlist: try: score.append(1 - ((item - mind) / (maxd - mind)) ) except ZeroDivisionError: score.append(1 ) elif weight == 1: for item in dlist: try: score.append((item - mind) / (maxd - mind) ) except ZeroDivisionError: score.append(0 ) # weight not 0 or 1 else: SCREAMING_SNAKE_CASE__ = f'''Invalid weight of {weight:f} provided''' raise ValueError(UpperCamelCase__ ) score_lists.append(UpperCamelCase__ ) return score_lists def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] ): SCREAMING_SNAKE_CASE__ = [0 for i in range(len(score_lists[0] ) )] for slist in score_lists: for j, ele in enumerate(UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = final_scores[j] + ele return final_scores def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] , UpperCamelCase__: list[int] ): SCREAMING_SNAKE_CASE__ = get_data(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = calculate_each_score(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = generate_final_scores(UpperCamelCase__ ) # append scores to source data for i, ele in enumerate(UpperCamelCase__ ): source_data[i].append(UpperCamelCase__ ) return source_data
59
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase__ ): def __init__( self :List[Any] , *__A :Tuple , **__A :Dict ) -> None: """simple docstring""" warnings.warn( """The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use LayoutLMv2ImageProcessor instead.""" , __A , ) super().__init__(*__A , **__A )
59
1
import math import sys def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): if number != int(UpperCamelCase__ ): raise ValueError("""the value of input must be a natural number""" ) if number < 0: raise ValueError("""the value of input must not be a negative number""" ) if number == 0: return 1 SCREAMING_SNAKE_CASE__ = [-1] * (number + 1) SCREAMING_SNAKE_CASE__ = 0 for i in range(1 , number + 1 ): SCREAMING_SNAKE_CASE__ = sys.maxsize SCREAMING_SNAKE_CASE__ = int(math.sqrt(UpperCamelCase__ ) ) for j in range(1 , root + 1 ): SCREAMING_SNAKE_CASE__ = 1 + answers[i - (j**2)] SCREAMING_SNAKE_CASE__ = min(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = answer return answers[number] if __name__ == "__main__": import doctest doctest.testmod()
59
import unittest from datasets import load_dataset from transformers.pipelines import pipeline from transformers.testing_utils import is_pipeline_test, nested_simplify, require_torch, slow @is_pipeline_test @require_torch class UpperCamelCase_ ( unittest.TestCase ): @require_torch def _snake_case ( self :Dict ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = pipeline( task="""zero-shot-audio-classification""" , model="""hf-internal-testing/tiny-clap-htsat-unfused""" ) SCREAMING_SNAKE_CASE__ = load_dataset("""ashraq/esc50""" ) SCREAMING_SNAKE_CASE__ = dataset["""train"""]["""audio"""][-1]["""array"""] SCREAMING_SNAKE_CASE__ = audio_classifier(__A , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__A ) , [{"""score""": 0.5_0_1, """label""": """Sound of a dog"""}, {"""score""": 0.4_9_9, """label""": """Sound of vaccum cleaner"""}] , ) @unittest.skip("""No models are available in TF""" ) def _snake_case ( self :Dict ) -> List[str]: """simple docstring""" pass @slow @require_torch def _snake_case ( self :Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = pipeline( task="""zero-shot-audio-classification""" , model="""laion/clap-htsat-unfused""" , ) # This is an audio of a dog SCREAMING_SNAKE_CASE__ = load_dataset("""ashraq/esc50""" ) SCREAMING_SNAKE_CASE__ = dataset["""train"""]["""audio"""][-1]["""array"""] SCREAMING_SNAKE_CASE__ = audio_classifier(__A , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__A ) , [ {"""score""": 0.9_9_9, """label""": """Sound of a dog"""}, {"""score""": 0.0_0_1, """label""": """Sound of vaccum cleaner"""}, ] , ) SCREAMING_SNAKE_CASE__ = audio_classifier([audio] * 5 , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__A ) , [ [ {"""score""": 0.9_9_9, """label""": """Sound of a dog"""}, {"""score""": 0.0_0_1, """label""": """Sound of vaccum cleaner"""}, ], ] * 5 , ) SCREAMING_SNAKE_CASE__ = audio_classifier( [audio] * 5 , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] , batch_size=5 ) self.assertEqual( nested_simplify(__A ) , [ [ {"""score""": 0.9_9_9, """label""": """Sound of a dog"""}, {"""score""": 0.0_0_1, """label""": """Sound of vaccum cleaner"""}, ], ] * 5 , ) @unittest.skip("""No models are available in TF""" ) def _snake_case ( self :str ) -> Optional[int]: """simple docstring""" pass
59
1
from math import pow, sqrt def SCREAMING_SNAKE_CASE__ ( *UpperCamelCase__: float ): SCREAMING_SNAKE_CASE__ = len(UpperCamelCase__ ) > 0 and all(value > 0.0 for value in values ) return result def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ ) else ValueError("""Input Error: Molar mass values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) )
59
import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 _lowerCamelCase = data_utils.TransfoXLTokenizer _lowerCamelCase = data_utils.TransfoXLCorpus _lowerCamelCase = data_utils _lowerCamelCase = data_utils def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Dict , UpperCamelCase__: Any , UpperCamelCase__: Any , UpperCamelCase__: Tuple ): if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(UpperCamelCase__ , """rb""" ) as fp: SCREAMING_SNAKE_CASE__ = pickle.load(UpperCamelCase__ , encoding="""latin1""" ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) SCREAMING_SNAKE_CASE__ = pytorch_dump_folder_path + """/""" + VOCAB_FILES_NAMES["""pretrained_vocab_file"""] print(f'''Save vocabulary to {pytorch_vocab_dump_path}''' ) SCREAMING_SNAKE_CASE__ = corpus.vocab.__dict__ torch.save(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = corpus.__dict__ corpus_dict_no_vocab.pop("""vocab""" , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = pytorch_dump_folder_path + """/""" + CORPUS_NAME print(f'''Save dataset to {pytorch_dataset_dump_path}''' ) torch.save(UpperCamelCase__ , UpperCamelCase__ ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model SCREAMING_SNAKE_CASE__ = os.path.abspath(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = os.path.abspath(UpperCamelCase__ ) print(f'''Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.''' ) # Initialise PyTorch model if transfo_xl_config_file == "": SCREAMING_SNAKE_CASE__ = TransfoXLConfig() else: SCREAMING_SNAKE_CASE__ = TransfoXLConfig.from_json_file(UpperCamelCase__ ) print(f'''Building PyTorch model from configuration: {config}''' ) SCREAMING_SNAKE_CASE__ = TransfoXLLMHeadModel(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = load_tf_weights_in_transfo_xl(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # Save pytorch-model SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) print(f'''Save PyTorch model to {os.path.abspath(UpperCamelCase__ )}''' ) torch.save(model.state_dict() , UpperCamelCase__ ) print(f'''Save configuration file to {os.path.abspath(UpperCamelCase__ )}''' ) with open(UpperCamelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--tf_checkpoint_path', default='', type=str, help='An optional path to a TensorFlow checkpoint path to be converted.', ) parser.add_argument( '--transfo_xl_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained BERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--transfo_xl_dataset_file', default='', type=str, help='An optional dataset file to be converted in a vocabulary.', ) _lowerCamelCase = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
59
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'microsoft/cvt-13': 'https://huggingface.co/microsoft/cvt-13/resolve/main/config.json', # See all Cvt models at https://huggingface.co/models?filter=cvt } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "cvt" def __init__( self :Any , __A :List[str]=3 , __A :Any=[7, 3, 3] , __A :Optional[Any]=[4, 2, 2] , __A :Union[str, Any]=[2, 1, 1] , __A :List[str]=[64, 192, 384] , __A :int=[1, 3, 6] , __A :Any=[1, 2, 10] , __A :Optional[int]=[4.0, 4.0, 4.0] , __A :List[Any]=[0.0, 0.0, 0.0] , __A :Tuple=[0.0, 0.0, 0.0] , __A :Any=[0.0, 0.0, 0.1] , __A :Any=[True, True, True] , __A :Any=[False, False, True] , __A :Dict=["dw_bn", "dw_bn", "dw_bn"] , __A :Optional[Any]=[3, 3, 3] , __A :Dict=[1, 1, 1] , __A :List[str]=[2, 2, 2] , __A :Dict=[1, 1, 1] , __A :int=[1, 1, 1] , __A :Any=0.0_2 , __A :Optional[Any]=1E-12 , **__A :Union[str, Any] , ) -> str: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = patch_sizes SCREAMING_SNAKE_CASE__ = patch_stride SCREAMING_SNAKE_CASE__ = patch_padding SCREAMING_SNAKE_CASE__ = embed_dim SCREAMING_SNAKE_CASE__ = num_heads SCREAMING_SNAKE_CASE__ = depth SCREAMING_SNAKE_CASE__ = mlp_ratio SCREAMING_SNAKE_CASE__ = attention_drop_rate SCREAMING_SNAKE_CASE__ = drop_rate SCREAMING_SNAKE_CASE__ = drop_path_rate SCREAMING_SNAKE_CASE__ = qkv_bias SCREAMING_SNAKE_CASE__ = cls_token SCREAMING_SNAKE_CASE__ = qkv_projection_method SCREAMING_SNAKE_CASE__ = kernel_qkv SCREAMING_SNAKE_CASE__ = padding_kv SCREAMING_SNAKE_CASE__ = stride_kv SCREAMING_SNAKE_CASE__ = padding_q SCREAMING_SNAKE_CASE__ = stride_q SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = layer_norm_eps
59
import argparse import tensorflow as tf import torch from transformers import BertConfig, BertForMaskedLM from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertPooler, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging logging.set_verbosity_info() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str , UpperCamelCase__: str , UpperCamelCase__: str ): def get_masked_lm_array(UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = f'''masked_lm/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) def get_encoder_array(UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = f'''encoder/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) def get_encoder_layer_array(UpperCamelCase__: int , UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = f'''encoder/_transformer_layers/{layer_index}/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) def get_encoder_attention_layer_array(UpperCamelCase__: int , UpperCamelCase__: str , UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = f'''encoder/_transformer_layers/{layer_index}/_attention_layer/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = array.reshape(UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) print(f'''Loading model based on config from {config_path}...''' ) SCREAMING_SNAKE_CASE__ = BertConfig.from_json_file(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = BertForMaskedLM(UpperCamelCase__ ) # Layers for layer_index in range(0 , config.num_hidden_layers ): SCREAMING_SNAKE_CASE__ = model.bert.encoder.layer[layer_index] # Self-attention SCREAMING_SNAKE_CASE__ = layer.attention.self SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_query_dense/kernel""" , self_attn.query.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_query_dense/bias""" , self_attn.query.bias.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_key_dense/kernel""" , self_attn.key.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_key_dense/bias""" , self_attn.key.bias.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_value_dense/kernel""" , self_attn.value.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_value_dense/bias""" , self_attn.value.bias.data.shape ) # Self-attention Output SCREAMING_SNAKE_CASE__ = layer.attention.output SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_output_dense/kernel""" , self_output.dense.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_output_dense/bias""" , self_output.dense.bias.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_attention_layer_norm/gamma""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_attention_layer_norm/beta""" ) # Intermediate SCREAMING_SNAKE_CASE__ = layer.intermediate SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_intermediate_dense/kernel""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_intermediate_dense/bias""" ) # Output SCREAMING_SNAKE_CASE__ = layer.output SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_dense/kernel""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_dense/bias""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_layer_norm/gamma""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_layer_norm/beta""" ) # Embeddings SCREAMING_SNAKE_CASE__ = get_encoder_array("""_position_embedding_layer/embeddings""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_type_embedding_layer/embeddings""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_embedding_norm_layer/gamma""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_embedding_norm_layer/beta""" ) # LM Head SCREAMING_SNAKE_CASE__ = model.cls.predictions.transform SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""dense/kernel""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""dense/bias""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""layer_norm/gamma""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""layer_norm/beta""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""embedding_table""" ) # Pooling SCREAMING_SNAKE_CASE__ = BertPooler(config=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_pooler_layer/kernel""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_pooler_layer/bias""" ) # Export final model model.save_pretrained(UpperCamelCase__ ) # Integration test - should load without any errors ;) SCREAMING_SNAKE_CASE__ = BertForMaskedLM.from_pretrained(UpperCamelCase__ ) print(new_model.eval() ) print("""Model conversion was done sucessfully!""" ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '--tf_checkpoint_path', type=str, required=True, help='Path to the TensorFlow Token Dropping checkpoint path.' ) parser.add_argument( '--bert_config_file', type=str, required=True, help='The config json file corresponding to the BERT model. This specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', type=str, required=True, help='Path to the output PyTorch model.', ) _lowerCamelCase = parser.parse_args() convert_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
59
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCamelCase = { 'configuration_git': ['GIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GitConfig', 'GitVisionConfig'], 'processing_git': ['GitProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'GIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'GitForCausalLM', 'GitModel', 'GitPreTrainedModel', 'GitVisionModel', ] if TYPE_CHECKING: from .configuration_git import GIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GitConfig, GitVisionConfig from .processing_git import GitProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_git import ( GIT_PRETRAINED_MODEL_ARCHIVE_LIST, GitForCausalLM, GitModel, GitPreTrainedModel, GitVisionModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _lowerCamelCase = '\\n Text data.\n Second line of data.' _lowerCamelCase = 'file' @pytest.fixture(scope="""session""" ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / (FILE_PATH + """.zstd""") SCREAMING_SNAKE_CASE__ = bytes(UpperCamelCase__ , """utf-8""" ) with zstd.open(UpperCamelCase__ , """wb""" ) as f: f.write(UpperCamelCase__ ) return path @pytest.fixture def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): with open(os.path.join(tmpfs.local_root_dir , UpperCamelCase__ ) , """w""" ) as f: f.write(UpperCamelCase__ ) return FILE_PATH @pytest.mark.parametrize("""compression_format""" , ["""gzip""", """xz""", """zstd"""] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int , UpperCamelCase__: Dict , UpperCamelCase__: int , UpperCamelCase__: str , UpperCamelCase__: Optional[int] , UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = {"""gzip""": gz_file, """xz""": xz_file, """zstd""": zstd_path} SCREAMING_SNAKE_CASE__ = input_paths[compression_format] SCREAMING_SNAKE_CASE__ = tmp_path / """cache""" SCREAMING_SNAKE_CASE__ = DownloadConfig(cache_dir=UpperCamelCase__ , extract_compressed_file=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = cached_path(UpperCamelCase__ , download_config=UpperCamelCase__ ) with open(UpperCamelCase__ ) as f: SCREAMING_SNAKE_CASE__ = f.read() with open(UpperCamelCase__ ) as f: SCREAMING_SNAKE_CASE__ = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("""default_extracted""" , [True, False] ) @pytest.mark.parametrize("""default_cache_dir""" , [True, False] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Tuple , UpperCamelCase__: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Any , UpperCamelCase__: Union[str, Any] ): SCREAMING_SNAKE_CASE__ = """custom_cache""" SCREAMING_SNAKE_CASE__ = """custom_extracted_dir""" SCREAMING_SNAKE_CASE__ = tmp_path / """custom_extracted_path""" if default_extracted: SCREAMING_SNAKE_CASE__ = ("""downloads""" if default_cache_dir else custom_cache_dir, """extracted""") else: monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_DIR""" , UpperCamelCase__ ) monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_PATH""" , str(UpperCamelCase__ ) ) SCREAMING_SNAKE_CASE__ = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) SCREAMING_SNAKE_CASE__ = xz_file SCREAMING_SNAKE_CASE__ = ( DownloadConfig(extract_compressed_file=UpperCamelCase__ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=UpperCamelCase__ ) ) SCREAMING_SNAKE_CASE__ = cached_path(UpperCamelCase__ , download_config=UpperCamelCase__ ) assert Path(UpperCamelCase__ ).parent.parts[-2:] == expected def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[int] ): # absolute path SCREAMING_SNAKE_CASE__ = str(Path(UpperCamelCase__ ).resolve() ) assert cached_path(UpperCamelCase__ ) == text_file # relative path SCREAMING_SNAKE_CASE__ = str(Path(UpperCamelCase__ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(UpperCamelCase__ ) == text_file def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): # absolute path SCREAMING_SNAKE_CASE__ = str(tmp_path.resolve() / """__missing_file__.txt""" ) with pytest.raises(UpperCamelCase__ ): cached_path(UpperCamelCase__ ) # relative path SCREAMING_SNAKE_CASE__ = """./__missing_file__.txt""" with pytest.raises(UpperCamelCase__ ): cached_path(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): SCREAMING_SNAKE_CASE__ = get_from_cache(f'''tmp://{tmpfs_file}''' ) with open(UpperCamelCase__ ) as f: SCREAMING_SNAKE_CASE__ = f.read() assert output_file_content == FILE_CONTENT @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( ): with pytest.raises(UpperCamelCase__ ): cached_path("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(UpperCamelCase__ ): http_get("""https://huggingface.co""" , temp_file=UpperCamelCase__ ) with pytest.raises(UpperCamelCase__ ): http_head("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[Any] ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(UpperCamelCase__ ): ftp_get("""ftp://huggingface.co""" , temp_file=UpperCamelCase__ ) with pytest.raises(UpperCamelCase__ ): ftp_head("""ftp://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(UpperCamelCase__ ): fsspec_get("""s3://huggingface.co""" , temp_file=UpperCamelCase__ ) with pytest.raises(UpperCamelCase__ ): fsspec_head("""s3://huggingface.co""" )
59
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "timm_backbone" def __init__( self :Union[str, Any] , __A :str=None , __A :Union[str, Any]=3 , __A :str=True , __A :Any=True , __A :Optional[Any]=None , **__A :List[str] , ) -> Tuple: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = backbone SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = features_only SCREAMING_SNAKE_CASE__ = use_pretrained_backbone SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = out_indices if out_indices is not None else (-1,)
59
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _lowerCamelCase = logging.getLogger(__name__) def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser( description="""Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset.""" ) parser.add_argument( """--dataset_name""" , type=UpperCamelCase__ , default="""wikitext""" , help="""Name of the training. Explore datasets at: hf.co/datasets.""" , ) parser.add_argument( """--dataset_config""" , type=UpperCamelCase__ , default="""wikitext-103-raw-v1""" , help="""Configuration name of the dataset.""" ) parser.add_argument( """--tokenizer_name_or_path""" , type=UpperCamelCase__ , default="""sayakpaul/unigram-tokenizer-wikitext""" , help="""Tokenizer identifier. Can be a local filepath or a Hub identifier.""" , ) parser.add_argument( """--shard_size""" , type=UpperCamelCase__ , default=1_000 , help="""Number of entries to go in a single shard.""" , ) parser.add_argument("""--split""" , type=UpperCamelCase__ , default="""train""" , choices=["""train""", """test""", """validation"""] ) parser.add_argument( """--limit""" , default=UpperCamelCase__ , type=UpperCamelCase__ , help="""Limit the number of shards (used for debugging).""" , ) parser.add_argument( """--max_length""" , type=UpperCamelCase__ , default=512 , help="""Maximum sequence length. For training on TPUs, it helps to have a maximum""" """ sequence length that is a multiple of 8.""" , ) parser.add_argument( """--output_dir""" , default="""tf-tpu""" , type=UpperCamelCase__ , help="""Output directory where the TFRecord shards will be saved. If the""" """ path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord""" """ shards will be directly saved to a Google Cloud Storage bucket.""" , ) SCREAMING_SNAKE_CASE__ = parser.parse_args() return args def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[Any] ): def fn(UpperCamelCase__: Any ): return tokenizer(examples["""text"""] ) return fn def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = [] for i in range(len(tokenized_data["""input_ids"""] ) ): SCREAMING_SNAKE_CASE__ = { """input_ids""": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["""input_ids"""][i] ) ), """attention_mask""": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["""attention_mask"""][i] ) ), } SCREAMING_SNAKE_CASE__ = tf.train.Features(feature=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = tf.train.Example(features=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = example.SerializeToString() records.append(UpperCamelCase__ ) return records def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): SCREAMING_SNAKE_CASE__ = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: SCREAMING_SNAKE_CASE__ = min(len(UpperCamelCase__ ) , args.limit ) SCREAMING_SNAKE_CASE__ = dataset.select(range(UpperCamelCase__ ) ) print(f'''Limiting the dataset to {args.limit} entries.''' ) SCREAMING_SNAKE_CASE__ = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) SCREAMING_SNAKE_CASE__ = os.path.join(args.output_dir , args.split ) if not os.path.exists(UpperCamelCase__ ): os.makedirs(UpperCamelCase__ ) else: SCREAMING_SNAKE_CASE__ = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. SCREAMING_SNAKE_CASE__ = tokenize_function(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = dataset.map(UpperCamelCase__ , batched=UpperCamelCase__ , num_proc=4 , remove_columns=["""text"""] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(UpperCamelCase__: int ): # Concatenate all texts. SCREAMING_SNAKE_CASE__ = {k: sum(examples[k] , [] ) for k in examples.keys()} SCREAMING_SNAKE_CASE__ = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 SCREAMING_SNAKE_CASE__ = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. SCREAMING_SNAKE_CASE__ = { k: [t[i : i + args.max_length] for i in range(0 , UpperCamelCase__ , args.max_length )] for k, t in concatenated_examples.items() } return result SCREAMING_SNAKE_CASE__ = dataset_tokenized.map(UpperCamelCase__ , batched=UpperCamelCase__ , batch_size=1_000 , num_proc=4 ) SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 0 for shard in range(0 , len(UpperCamelCase__ ) , args.shard_size ): SCREAMING_SNAKE_CASE__ = grouped_dataset[shard : shard + args.shard_size] SCREAMING_SNAKE_CASE__ = len(dataset_snapshot["""input_ids"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , f'''dataset-{shard_count}-{records_containing}.tfrecord''' ) SCREAMING_SNAKE_CASE__ = get_serialized_examples(UpperCamelCase__ ) with tf.io.TFRecordWriter(UpperCamelCase__ ) as out_file: for i in range(len(UpperCamelCase__ ) ): SCREAMING_SNAKE_CASE__ = serialized_examples[i] out_file.write(UpperCamelCase__ ) print("""Wrote file {} containing {} records""".format(UpperCamelCase__ , UpperCamelCase__ ) ) shard_count += 1 total_records += records_containing with open(f'''split-{args.split}-records-count.txt''' , """w""" ) as f: print(f'''Total {args.split} records: {total_records}''' , file=UpperCamelCase__ ) if __name__ == "__main__": _lowerCamelCase = parse_args() main(args)
59
1
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase__ ): def __init__( self :List[Any] , *__A :Tuple , **__A :Dict ) -> None: """simple docstring""" warnings.warn( """The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use LayoutLMv2ImageProcessor instead.""" , __A , ) super().__init__(*__A , **__A )
59
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] ): SCREAMING_SNAKE_CASE__ = [] for data in source_data: for i, el in enumerate(UpperCamelCase__ ): if len(UpperCamelCase__ ) < i + 1: data_lists.append([] ) data_lists[i].append(float(UpperCamelCase__ ) ) return data_lists def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] , UpperCamelCase__: list[int] ): SCREAMING_SNAKE_CASE__ = [] for dlist, weight in zip(UpperCamelCase__ , UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = min(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = max(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = [] # for weight 0 score is 1 - actual score if weight == 0: for item in dlist: try: score.append(1 - ((item - mind) / (maxd - mind)) ) except ZeroDivisionError: score.append(1 ) elif weight == 1: for item in dlist: try: score.append((item - mind) / (maxd - mind) ) except ZeroDivisionError: score.append(0 ) # weight not 0 or 1 else: SCREAMING_SNAKE_CASE__ = f'''Invalid weight of {weight:f} provided''' raise ValueError(UpperCamelCase__ ) score_lists.append(UpperCamelCase__ ) return score_lists def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] ): SCREAMING_SNAKE_CASE__ = [0 for i in range(len(score_lists[0] ) )] for slist in score_lists: for j, ele in enumerate(UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = final_scores[j] + ele return final_scores def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] , UpperCamelCase__: list[int] ): SCREAMING_SNAKE_CASE__ = get_data(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = calculate_each_score(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = generate_final_scores(UpperCamelCase__ ) # append scores to source data for i, ele in enumerate(UpperCamelCase__ ): source_data[i].append(UpperCamelCase__ ) return source_data
59
1
import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _lowerCamelCase = 16 _lowerCamelCase = 32 def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Accelerator , UpperCamelCase__: int = 16 ): SCREAMING_SNAKE_CASE__ = AutoTokenizer.from_pretrained("""bert-base-cased""" ) SCREAMING_SNAKE_CASE__ = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(UpperCamelCase__: Optional[Any] ): # max_length=None => use the model max length (it's actually the default) SCREAMING_SNAKE_CASE__ = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=UpperCamelCase__ , max_length=UpperCamelCase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): SCREAMING_SNAKE_CASE__ = datasets.map( UpperCamelCase__ , batched=UpperCamelCase__ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library SCREAMING_SNAKE_CASE__ = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(UpperCamelCase__: str ): # On TPU it's best to pad everything to the same length or training will be very slow. SCREAMING_SNAKE_CASE__ = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": SCREAMING_SNAKE_CASE__ = 16 elif accelerator.mixed_precision != "no": SCREAMING_SNAKE_CASE__ = 8 else: SCREAMING_SNAKE_CASE__ = None return tokenizer.pad( UpperCamelCase__ , padding="""longest""" , max_length=UpperCamelCase__ , pad_to_multiple_of=UpperCamelCase__ , return_tensors="""pt""" , ) # Instantiate dataloaders. SCREAMING_SNAKE_CASE__ = DataLoader( tokenized_datasets["""train"""] , shuffle=UpperCamelCase__ , collate_fn=UpperCamelCase__ , batch_size=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = DataLoader( tokenized_datasets["""validation"""] , shuffle=UpperCamelCase__ , collate_fn=UpperCamelCase__ , batch_size=UpperCamelCase__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders _lowerCamelCase = mocked_dataloaders # noqa: F811 def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str , UpperCamelCase__: str ): # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""" , UpperCamelCase__ ) == "1": SCREAMING_SNAKE_CASE__ = 2 # Initialize accelerator SCREAMING_SNAKE_CASE__ = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs SCREAMING_SNAKE_CASE__ = config["""lr"""] SCREAMING_SNAKE_CASE__ = int(config["""num_epochs"""] ) SCREAMING_SNAKE_CASE__ = int(config["""seed"""] ) SCREAMING_SNAKE_CASE__ = int(config["""batch_size"""] ) SCREAMING_SNAKE_CASE__ = evaluate.load("""glue""" , """mrpc""" ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=UpperCamelCase__ ) def inner_training_loop(UpperCamelCase__: Any ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(UpperCamelCase__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) SCREAMING_SNAKE_CASE__ = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""" , return_dict=UpperCamelCase__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). SCREAMING_SNAKE_CASE__ = model.to(accelerator.device ) # Instantiate optimizer SCREAMING_SNAKE_CASE__ = AdamW(params=model.parameters() , lr=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = get_dataloaders(UpperCamelCase__ , UpperCamelCase__ ) # Instantiate scheduler SCREAMING_SNAKE_CASE__ = get_linear_schedule_with_warmup( optimizer=UpperCamelCase__ , num_warmup_steps=100 , num_training_steps=(len(UpperCamelCase__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = accelerator.prepare( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # Now we train the model for epoch in range(UpperCamelCase__ ): model.train() for step, batch in enumerate(UpperCamelCase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) SCREAMING_SNAKE_CASE__ = model(**UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = outputs.loss accelerator.backward(UpperCamelCase__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(UpperCamelCase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(**UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = outputs.logits.argmax(dim=-1 ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = accelerator.gather_for_metrics((predictions, batch["""labels"""]) ) metric.add_batch( predictions=UpperCamelCase__ , references=UpperCamelCase__ , ) SCREAMING_SNAKE_CASE__ = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'''epoch {epoch}:''' , UpperCamelCase__ ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""" , type=UpperCamelCase__ , default=UpperCamelCase__ , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) SCREAMING_SNAKE_CASE__ = parser.parse_args() SCREAMING_SNAKE_CASE__ = {"""lr""": 2e-5, """num_epochs""": 3, """seed""": 42, """batch_size""": 16} training_function(UpperCamelCase__ , UpperCamelCase__ ) if __name__ == "__main__": main()
59
import warnings from functools import wraps from typing import Callable def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Callable ): @wraps(UpperCamelCase__ ) def _inner_fn(*UpperCamelCase__: Dict , **UpperCamelCase__: Any ): warnings.warn( (f'''\'{fn.__name__}\' is experimental and might be subject to breaking changes in the future.''') , UpperCamelCase__ , ) return fn(*UpperCamelCase__ , **UpperCamelCase__ ) return _inner_fn
59
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'MIT/ast-finetuned-audioset-10-10-0.4593': ( 'https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json' ), } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "audio-spectrogram-transformer" def __init__( self :List[Any] , __A :Union[str, Any]=768 , __A :Tuple=12 , __A :Any=12 , __A :str=3072 , __A :str="gelu" , __A :int=0.0 , __A :str=0.0 , __A :int=0.0_2 , __A :Union[str, Any]=1E-12 , __A :Union[str, Any]=16 , __A :List[Any]=True , __A :Optional[Any]=10 , __A :List[str]=10 , __A :int=1024 , __A :List[Any]=128 , **__A :Tuple , ) -> Optional[int]: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = patch_size SCREAMING_SNAKE_CASE__ = qkv_bias SCREAMING_SNAKE_CASE__ = frequency_stride SCREAMING_SNAKE_CASE__ = time_stride SCREAMING_SNAKE_CASE__ = max_length SCREAMING_SNAKE_CASE__ = num_mel_bins
59
import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = RoCBertTokenizer lowerCamelCase_ = None lowerCamelCase_ = False lowerCamelCase_ = True lowerCamelCase_ = filter_non_english def _snake_case ( self :List[Any] ) -> List[Any]: """simple docstring""" super().setUp() SCREAMING_SNAKE_CASE__ = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """你""", """好""", """是""", """谁""", """a""", """b""", """c""", """d"""] SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = {} for i, value in enumerate(__A ): SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""word_shape_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""word_pronunciation_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) with open(self.word_shape_file , """w""" , encoding="""utf-8""" ) as word_shape_writer: json.dump(__A , __A , ensure_ascii=__A ) with open(self.word_pronunciation_file , """w""" , encoding="""utf-8""" ) as word_pronunciation_writer: json.dump(__A , __A , ensure_ascii=__A ) def _snake_case ( self :List[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) SCREAMING_SNAKE_CASE__ = tokenizer.tokenize("""你好[SEP]你是谁""" ) self.assertListEqual(__A , ["""你""", """好""", """[SEP]""", """你""", """是""", """谁"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_shape_ids(__A ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_pronunciation_ids(__A ) , [5, 6, 2, 5, 7, 8] ) def _snake_case ( self :List[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer() self.assertListEqual(tokenizer.tokenize("""ah\u535A\u63A8zz""" ) , ["""ah""", """\u535A""", """\u63A8""", """zz"""] ) def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hällo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""h\u00E9llo"""] ) def _snake_case ( self :Any ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :List[str] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :Union[str, Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :int ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HäLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :Union[str, Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HaLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :List[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , never_split=["""[UNK]"""] ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? [UNK]""" ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?""", """[UNK]"""] ) def _snake_case ( self :Any ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing"""] SCREAMING_SNAKE_CASE__ = {} for i, token in enumerate(__A ): SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = RoCBertWordpieceTokenizer(vocab=__A , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""unwanted running""" ) , ["""un""", """##want""", """##ed""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.tokenize("""unwantedX running""" ) , ["""[UNK]""", """runn""", """##ing"""] ) def _snake_case ( self :Any ) -> str: """simple docstring""" self.assertTrue(_is_whitespace(""" """ ) ) self.assertTrue(_is_whitespace("""\t""" ) ) self.assertTrue(_is_whitespace("""\r""" ) ) self.assertTrue(_is_whitespace("""\n""" ) ) self.assertTrue(_is_whitespace("""\u00A0""" ) ) self.assertFalse(_is_whitespace("""A""" ) ) self.assertFalse(_is_whitespace("""-""" ) ) def _snake_case ( self :int ) -> str: """simple docstring""" self.assertTrue(_is_control("""\u0005""" ) ) self.assertFalse(_is_control("""A""" ) ) self.assertFalse(_is_control(""" """ ) ) self.assertFalse(_is_control("""\t""" ) ) self.assertFalse(_is_control("""\r""" ) ) def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" self.assertTrue(_is_punctuation("""-""" ) ) self.assertTrue(_is_punctuation("""$""" ) ) self.assertTrue(_is_punctuation("""`""" ) ) self.assertTrue(_is_punctuation(""".""" ) ) self.assertFalse(_is_punctuation("""A""" ) ) self.assertFalse(_is_punctuation(""" """ ) ) def _snake_case ( self :str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(__A ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) if self.test_rust_tokenizer: SCREAMING_SNAKE_CASE__ = self.get_rust_tokenizer() self.assertListEqual( [rust_tokenizer.tokenize(__A ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) def _snake_case ( self :int ) -> Any: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = f'''A, naïve {tokenizer_r.mask_token} AllenNLP sentence.''' SCREAMING_SNAKE_CASE__ = tokenizer_r.encode_plus( __A , return_attention_mask=__A , return_token_type_ids=__A , return_offsets_mapping=__A , add_special_tokens=__A , ) SCREAMING_SNAKE_CASE__ = tokenizer_r.do_lower_case if hasattr(__A , """do_lower_case""" ) else False SCREAMING_SNAKE_CASE__ = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """A"""), ((1, 2), ""","""), ((3, 5), """na"""), ((5, 6), """##ï"""), ((6, 8), """##ve"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """Allen"""), ((21, 23), """##NL"""), ((23, 24), """##P"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """a"""), ((1, 2), ""","""), ((3, 8), """naive"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """allen"""), ((21, 23), """##nl"""), ((23, 24), """##p"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["""input_ids"""] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["""offset_mapping"""] ) def _snake_case ( self :Optional[int] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""的""", """人""", """有"""] SCREAMING_SNAKE_CASE__ = """""".join(__A ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.convert_ids_to_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.convert_ids_to_tokens(__A ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(__A , __A ) self.assertListEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.convert_ids_to_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.convert_ids_to_tokens(__A ) # it is expected that only the first Chinese character is not preceded by "##". SCREAMING_SNAKE_CASE__ = [ f'''##{token}''' if idx != 0 else token for idx, token in enumerate(__A ) ] self.assertListEqual(__A , __A ) self.assertListEqual(__A , __A ) @slow def _snake_case ( self :Union[str, Any] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) SCREAMING_SNAKE_CASE__ = tokenizer.encode("""你好""" , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.encode("""你是谁""" , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.build_inputs_with_special_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.build_inputs_with_special_tokens(__A , __A ) assert encoded_sentence == [1] + text + [2] assert encoded_pair == [1] + text + [2] + text_a + [2] def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_tokenizers(do_lower_case=__A ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE__ = """你好,你是谁""" SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_shape_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_pronunciation_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.prepare_for_model( __A , __A , __A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.encode_plus(__A , add_special_tokens=__A ) self.assertEqual(__A , __A )
59
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _lowerCamelCase = { 'configuration_layoutlmv3': [ 'LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv3Config', 'LayoutLMv3OnnxConfig', ], 'processing_layoutlmv3': ['LayoutLMv3Processor'], 'tokenization_layoutlmv3': ['LayoutLMv3Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['LayoutLMv3TokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv3ForQuestionAnswering', 'LayoutLMv3ForSequenceClassification', 'LayoutLMv3ForTokenClassification', 'LayoutLMv3Model', 'LayoutLMv3PreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLayoutLMv3ForQuestionAnswering', 'TFLayoutLMv3ForSequenceClassification', 'TFLayoutLMv3ForTokenClassification', 'TFLayoutLMv3Model', 'TFLayoutLMv3PreTrainedModel', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['LayoutLMv3FeatureExtractor'] _lowerCamelCase = ['LayoutLMv3ImageProcessor'] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[str, Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = original_name.split(""".""" )[0] SCREAMING_SNAKE_CASE__ = key.split(""".""" ) SCREAMING_SNAKE_CASE__ = int(key_list[key_list.index(UpperCamelCase__ ) - 2] ) SCREAMING_SNAKE_CASE__ = int(key_list[key_list.index(UpperCamelCase__ ) - 1] ) SCREAMING_SNAKE_CASE__ = orig_block_num - offset SCREAMING_SNAKE_CASE__ = key.replace(f'''{orig_block_num}.{layer_num}.{original_name}''' , f'''block.{new_block_num}.{layer_num}.{new_name}''' ) return key def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = OrderedDict() SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 0, 0 for key, value in state_dict.items(): if key.startswith("""network""" ): SCREAMING_SNAKE_CASE__ = key.replace("""network""" , """poolformer.encoder""" ) if "proj" in key: # Works for the first embedding as well as the internal embedding layers if key.endswith("""bias""" ) and "patch_embed" not in key: patch_emb_offset += 1 SCREAMING_SNAKE_CASE__ = key[: key.find("""proj""" )] SCREAMING_SNAKE_CASE__ = key.replace(UpperCamelCase__ , f'''patch_embeddings.{total_embed_found}.''' ) SCREAMING_SNAKE_CASE__ = key.replace("""proj""" , """projection""" ) if key.endswith("""bias""" ): total_embed_found += 1 if "patch_embeddings" in key: SCREAMING_SNAKE_CASE__ = """poolformer.encoder.""" + key if "mlp.fc1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """mlp.fc1""" , """output.conv1""" ) if "mlp.fc2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """mlp.fc2""" , """output.conv2""" ) if "norm1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """norm1""" , """before_norm""" ) if "norm2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """norm2""" , """after_norm""" ) if "layer_scale_1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """layer_scale_1""" , """layer_scale_1""" ) if "layer_scale_2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """layer_scale_2""" , """layer_scale_2""" ) if "head" in key: SCREAMING_SNAKE_CASE__ = key.replace("""head""" , """classifier""" ) SCREAMING_SNAKE_CASE__ = value return new_state_dict def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = """http://images.cocodataset.org/val2017/000000039769.jpg""" SCREAMING_SNAKE_CASE__ = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return image @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = PoolFormerConfig() # set attributes based on model_name SCREAMING_SNAKE_CASE__ = """huggingface/label-files""" SCREAMING_SNAKE_CASE__ = model_name[-3:] SCREAMING_SNAKE_CASE__ = 1_000 SCREAMING_SNAKE_CASE__ = """imagenet-1k-id2label.json""" SCREAMING_SNAKE_CASE__ = (1, 1_000) # set config attributes SCREAMING_SNAKE_CASE__ = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) SCREAMING_SNAKE_CASE__ = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE__ = idalabel SCREAMING_SNAKE_CASE__ = {v: k for k, v in idalabel.items()} if size == "s12": SCREAMING_SNAKE_CASE__ = [2, 2, 6, 2] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "s24": SCREAMING_SNAKE_CASE__ = [4, 4, 12, 4] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "s36": SCREAMING_SNAKE_CASE__ = [6, 6, 18, 6] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "m36": SCREAMING_SNAKE_CASE__ = [6, 6, 18, 6] SCREAMING_SNAKE_CASE__ = [96, 192, 384, 768] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9_5 elif size == "m48": SCREAMING_SNAKE_CASE__ = [8, 8, 24, 8] SCREAMING_SNAKE_CASE__ = [96, 192, 384, 768] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9_5 else: raise ValueError(f'''Size {size} not supported''' ) # load image processor SCREAMING_SNAKE_CASE__ = PoolFormerImageProcessor(crop_pct=UpperCamelCase__ ) # Prepare image SCREAMING_SNAKE_CASE__ = prepare_img() SCREAMING_SNAKE_CASE__ = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ).pixel_values logger.info(f'''Converting model {model_name}...''' ) # load original state dict SCREAMING_SNAKE_CASE__ = torch.load(UpperCamelCase__ , map_location=torch.device("""cpu""" ) ) # rename keys SCREAMING_SNAKE_CASE__ = rename_keys(UpperCamelCase__ ) # create HuggingFace model and load state dict SCREAMING_SNAKE_CASE__ = PoolFormerForImageClassification(UpperCamelCase__ ) model.load_state_dict(UpperCamelCase__ ) model.eval() # Define image processor SCREAMING_SNAKE_CASE__ = PoolFormerImageProcessor(crop_pct=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = image_processor(images=prepare_img() , return_tensors="""pt""" ).pixel_values # forward pass SCREAMING_SNAKE_CASE__ = model(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = outputs.logits # define expected logit slices for different models if size == "s12": SCREAMING_SNAKE_CASE__ = torch.tensor([-0.3_0_4_5, -0.6_7_5_8, -0.4_8_6_9] ) elif size == "s24": SCREAMING_SNAKE_CASE__ = torch.tensor([0.4_4_0_2, -0.1_3_7_4, -0.8_0_4_5] ) elif size == "s36": SCREAMING_SNAKE_CASE__ = torch.tensor([-0.6_0_8_0, -0.5_1_3_3, -0.5_8_9_8] ) elif size == "m36": SCREAMING_SNAKE_CASE__ = torch.tensor([0.3_9_5_2, 0.2_2_6_3, -1.2_6_6_8] ) elif size == "m48": SCREAMING_SNAKE_CASE__ = torch.tensor([0.1_1_6_7, -0.0_6_5_6, -0.3_4_2_3] ) else: raise ValueError(f'''Size {size} not supported''' ) # verify logits assert logits.shape == expected_shape assert torch.allclose(logits[0, :3] , UpperCamelCase__ , atol=1e-2 ) # finally, save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) model.save_pretrained(UpperCamelCase__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '--model_name', default='poolformer_s12', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) _lowerCamelCase = parser.parse_args() convert_poolformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
59
1
from __future__ import annotations import typing from collections import Counter def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = Counter() for base in range(1 , max_perimeter + 1 ): for perpendicular in range(UpperCamelCase__ , max_perimeter + 1 ): SCREAMING_SNAKE_CASE__ = (base * base + perpendicular * perpendicular) ** 0.5 if hypotenuse == int(UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = int(base + perpendicular + hypotenuse ) if perimeter > max_perimeter: continue triplets[perimeter] += 1 return triplets def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int = 1_000 ): SCREAMING_SNAKE_CASE__ = pythagorean_triple(UpperCamelCase__ ) return triplets.most_common(1 )[0][0] if __name__ == "__main__": print(F'''Perimeter {solution()} has maximum solutions''')
59
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'google/pix2struct-textcaps-base': ( 'https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json' ), } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct_text_model" lowerCamelCase_ = ["past_key_values"] lowerCamelCase_ = { "hidden_size": "hidden_size", "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self :Union[str, Any] , __A :Any=5_0244 , __A :Optional[Any]=768 , __A :Tuple=64 , __A :List[str]=2048 , __A :int=12 , __A :str=12 , __A :Any=32 , __A :Tuple=128 , __A :int=0.1 , __A :str=1E-6 , __A :Optional[Any]=1.0 , __A :Union[str, Any]="gelu_new" , __A :Any=0 , __A :List[str]=False , __A :Optional[Any]=0 , __A :int=1 , __A :Optional[int]=False , __A :Optional[Any]=True , **__A :List[Any] , ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = d_kv SCREAMING_SNAKE_CASE__ = d_ff SCREAMING_SNAKE_CASE__ = num_layers SCREAMING_SNAKE_CASE__ = num_heads SCREAMING_SNAKE_CASE__ = relative_attention_num_buckets SCREAMING_SNAKE_CASE__ = relative_attention_max_distance SCREAMING_SNAKE_CASE__ = dropout_rate SCREAMING_SNAKE_CASE__ = layer_norm_epsilon SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = use_cache SCREAMING_SNAKE_CASE__ = eos_token_id SCREAMING_SNAKE_CASE__ = decoder_start_token_id # for backwards compatibility SCREAMING_SNAKE_CASE__ = dense_act_fn super().__init__( pad_token_id=__A , eos_token_id=__A , decoder_start_token_id=__A , tie_word_embeddings=__A , is_decoder=__A , **__A , ) @classmethod def _snake_case ( cls :Optional[int] , __A :Union[str, os.PathLike] , **__A :Optional[int] ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = cls.get_config_dict(__A , **__A ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get("""model_type""" ) == "pix2struct": SCREAMING_SNAKE_CASE__ = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct_vision_model" def __init__( self :Optional[int] , __A :int=768 , __A :Optional[Any]=768 , __A :Union[str, Any]=2048 , __A :int=64 , __A :Union[str, Any]=12 , __A :str=12 , __A :Any="gelu_new" , __A :List[Any]=1E-6 , __A :Dict=0.0 , __A :int=0.0 , __A :int=1E-10 , __A :Dict=1.0 , __A :int=4096 , __A :int=32 , __A :int=128 , **__A :Tuple , ) -> str: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = patch_embed_hidden_size SCREAMING_SNAKE_CASE__ = d_ff SCREAMING_SNAKE_CASE__ = dropout_rate SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = attention_dropout SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = dense_act_fn SCREAMING_SNAKE_CASE__ = seq_len SCREAMING_SNAKE_CASE__ = relative_attention_num_buckets SCREAMING_SNAKE_CASE__ = relative_attention_max_distance SCREAMING_SNAKE_CASE__ = d_kv @classmethod def _snake_case ( cls :str , __A :Union[str, os.PathLike] , **__A :str ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = cls.get_config_dict(__A , **__A ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get("""model_type""" ) == "pix2struct": SCREAMING_SNAKE_CASE__ = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct" lowerCamelCase_ = True def __init__( self :str , __A :Optional[Any]=None , __A :List[str]=None , __A :Optional[Any]=1.0 , __A :Optional[Any]=0.0_2 , __A :Any=False , __A :Tuple=False , __A :Any=True , **__A :Dict , ) -> Union[str, Any]: """simple docstring""" super().__init__(tie_word_embeddings=__A , is_encoder_decoder=__A , **__A ) if text_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""text_config is None. Initializing the Pix2StructTextConfig with default values.""" ) if vision_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""vision_config is None. Initializing the Pix2StructVisionConfig with default values.""" ) SCREAMING_SNAKE_CASE__ = PixaStructTextConfig(**__A ) SCREAMING_SNAKE_CASE__ = PixaStructVisionConfig(**__A ) SCREAMING_SNAKE_CASE__ = self.text_config.decoder_start_token_id SCREAMING_SNAKE_CASE__ = self.text_config.pad_token_id SCREAMING_SNAKE_CASE__ = self.text_config.eos_token_id SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = self.initializer_range SCREAMING_SNAKE_CASE__ = self.initializer_range SCREAMING_SNAKE_CASE__ = is_vqa @classmethod def _snake_case ( cls :Union[str, Any] , __A :PixaStructTextConfig , __A :PixaStructVisionConfig , **__A :Optional[int] ) -> Optional[Any]: """simple docstring""" return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__A ) def _snake_case ( self :str ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = copy.deepcopy(self.__dict__ ) SCREAMING_SNAKE_CASE__ = self.text_config.to_dict() SCREAMING_SNAKE_CASE__ = self.vision_config.to_dict() SCREAMING_SNAKE_CASE__ = self.__class__.model_type return output
59
1
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'ut/deta': 'https://huggingface.co/ut/deta/resolve/main/config.json', } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "deta" lowerCamelCase_ = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self :Dict , __A :Tuple=None , __A :Union[str, Any]=900 , __A :str=2048 , __A :int=6 , __A :Tuple=2048 , __A :List[Any]=8 , __A :List[Any]=6 , __A :Optional[Any]=1024 , __A :List[Any]=8 , __A :List[str]=0.0 , __A :int=True , __A :Any="relu" , __A :str=256 , __A :int=0.1 , __A :Union[str, Any]=0.0 , __A :Union[str, Any]=0.0 , __A :List[str]=0.0_2 , __A :Dict=1.0 , __A :str=True , __A :Tuple=False , __A :int="sine" , __A :str=5 , __A :List[str]=4 , __A :Optional[int]=4 , __A :List[Any]=True , __A :Dict=300 , __A :Optional[int]=True , __A :List[Any]=True , __A :List[Any]=1 , __A :List[str]=5 , __A :Any=2 , __A :Tuple=1 , __A :List[Any]=1 , __A :Any=5 , __A :List[Any]=2 , __A :Union[str, Any]=0.1 , __A :Optional[Any]=0.2_5 , **__A :int , ) -> Any: """simple docstring""" if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) SCREAMING_SNAKE_CASE__ = CONFIG_MAPPING["""resnet"""](out_features=["""stage2""", """stage3""", """stage4"""] ) else: if isinstance(__A , __A ): SCREAMING_SNAKE_CASE__ = backbone_config.pop("""model_type""" ) SCREAMING_SNAKE_CASE__ = CONFIG_MAPPING[backbone_model_type] SCREAMING_SNAKE_CASE__ = config_class.from_dict(__A ) SCREAMING_SNAKE_CASE__ = backbone_config SCREAMING_SNAKE_CASE__ = num_queries SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = d_model SCREAMING_SNAKE_CASE__ = encoder_ffn_dim SCREAMING_SNAKE_CASE__ = encoder_layers SCREAMING_SNAKE_CASE__ = encoder_attention_heads SCREAMING_SNAKE_CASE__ = decoder_ffn_dim SCREAMING_SNAKE_CASE__ = decoder_layers SCREAMING_SNAKE_CASE__ = decoder_attention_heads SCREAMING_SNAKE_CASE__ = dropout SCREAMING_SNAKE_CASE__ = attention_dropout SCREAMING_SNAKE_CASE__ = activation_dropout SCREAMING_SNAKE_CASE__ = activation_function SCREAMING_SNAKE_CASE__ = init_std SCREAMING_SNAKE_CASE__ = init_xavier_std SCREAMING_SNAKE_CASE__ = encoder_layerdrop SCREAMING_SNAKE_CASE__ = auxiliary_loss SCREAMING_SNAKE_CASE__ = position_embedding_type # deformable attributes SCREAMING_SNAKE_CASE__ = num_feature_levels SCREAMING_SNAKE_CASE__ = encoder_n_points SCREAMING_SNAKE_CASE__ = decoder_n_points SCREAMING_SNAKE_CASE__ = two_stage SCREAMING_SNAKE_CASE__ = two_stage_num_proposals SCREAMING_SNAKE_CASE__ = with_box_refine SCREAMING_SNAKE_CASE__ = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError("""If two_stage is True, with_box_refine must be True.""" ) # Hungarian matcher SCREAMING_SNAKE_CASE__ = class_cost SCREAMING_SNAKE_CASE__ = bbox_cost SCREAMING_SNAKE_CASE__ = giou_cost # Loss coefficients SCREAMING_SNAKE_CASE__ = mask_loss_coefficient SCREAMING_SNAKE_CASE__ = dice_loss_coefficient SCREAMING_SNAKE_CASE__ = bbox_loss_coefficient SCREAMING_SNAKE_CASE__ = giou_loss_coefficient SCREAMING_SNAKE_CASE__ = eos_coefficient SCREAMING_SNAKE_CASE__ = focal_alpha super().__init__(is_encoder_decoder=__A , **__A ) @property def _snake_case ( self :List[str] ) -> int: """simple docstring""" return self.encoder_attention_heads @property def _snake_case ( self :List[Any] ) -> int: """simple docstring""" return self.d_model def _snake_case ( self :Dict ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = copy.deepcopy(self.__dict__ ) SCREAMING_SNAKE_CASE__ = self.backbone_config.to_dict() SCREAMING_SNAKE_CASE__ = self.__class__.model_type return output
59
import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = inspect.getfile(accelerate.test_utils ) SCREAMING_SNAKE_CASE__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_script.py"""] ) SCREAMING_SNAKE_CASE__ = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_distributed_data_loop.py"""] ) SCREAMING_SNAKE_CASE__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_ops.py"""] ) @require_multi_gpu def _snake_case ( self :Optional[Any] ) -> Tuple: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Tuple ) -> Optional[Any]: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.operation_file_path] print(f'''Command: {cmd}''' ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Optional[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Optional[int] ) -> str: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices, using 2 devices only''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices="""0,1""" ): execute_subprocess_async(__A , env=os.environ.copy() ) if __name__ == "__main__": _lowerCamelCase = Accelerator() _lowerCamelCase = (accelerator.state.process_index + 2, 10) _lowerCamelCase = torch.randint(0, 10, shape).to(accelerator.device) _lowerCamelCase = '' _lowerCamelCase = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." _lowerCamelCase = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." _lowerCamelCase = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
59
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'google/bigbird-roberta-base': 'https://huggingface.co/google/bigbird-roberta-base/resolve/main/config.json', 'google/bigbird-roberta-large': 'https://huggingface.co/google/bigbird-roberta-large/resolve/main/config.json', 'google/bigbird-base-trivia-itc': 'https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/config.json', # See all BigBird models at https://huggingface.co/models?filter=big_bird } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "big_bird" def __init__( self :Any , __A :Union[str, Any]=5_0358 , __A :List[str]=768 , __A :Dict=12 , __A :Tuple=12 , __A :Optional[Any]=3072 , __A :Optional[int]="gelu_new" , __A :Union[str, Any]=0.1 , __A :str=0.1 , __A :int=4096 , __A :List[Any]=2 , __A :List[Any]=0.0_2 , __A :int=1E-12 , __A :Optional[Any]=True , __A :List[Any]=0 , __A :Union[str, Any]=1 , __A :List[Any]=2 , __A :List[str]=66 , __A :Any="block_sparse" , __A :Dict=True , __A :List[Any]=False , __A :int=64 , __A :Optional[Any]=3 , __A :List[str]=None , **__A :int , ) -> Optional[Any]: """simple docstring""" super().__init__( pad_token_id=__A , bos_token_id=__A , eos_token_id=__A , sep_token_id=__A , **__A , ) SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = type_vocab_size SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = use_cache SCREAMING_SNAKE_CASE__ = rescale_embeddings SCREAMING_SNAKE_CASE__ = attention_type SCREAMING_SNAKE_CASE__ = use_bias SCREAMING_SNAKE_CASE__ = block_size SCREAMING_SNAKE_CASE__ = num_random_blocks SCREAMING_SNAKE_CASE__ = classifier_dropout class UpperCamelCase_ ( UpperCamelCase__ ): @property def _snake_case ( self :Union[str, Any] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": SCREAMING_SNAKE_CASE__ = {0: """batch""", 1: """choice""", 2: """sequence"""} else: SCREAMING_SNAKE_CASE__ = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
59
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _lowerCamelCase = { 'configuration_layoutlmv3': [ 'LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv3Config', 'LayoutLMv3OnnxConfig', ], 'processing_layoutlmv3': ['LayoutLMv3Processor'], 'tokenization_layoutlmv3': ['LayoutLMv3Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['LayoutLMv3TokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv3ForQuestionAnswering', 'LayoutLMv3ForSequenceClassification', 'LayoutLMv3ForTokenClassification', 'LayoutLMv3Model', 'LayoutLMv3PreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLayoutLMv3ForQuestionAnswering', 'TFLayoutLMv3ForSequenceClassification', 'TFLayoutLMv3ForTokenClassification', 'TFLayoutLMv3Model', 'TFLayoutLMv3PreTrainedModel', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['LayoutLMv3FeatureExtractor'] _lowerCamelCase = ['LayoutLMv3ImageProcessor'] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
1
from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = CustomTokenizer pass
59
import inspect import unittest class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :str ) -> Union[str, Any]: """simple docstring""" try: import diffusers # noqa: F401 except ImportError: assert False def _snake_case ( self :Any ) -> Any: """simple docstring""" import diffusers from diffusers.dependency_versions_table import deps SCREAMING_SNAKE_CASE__ = inspect.getmembers(__A , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": SCREAMING_SNAKE_CASE__ = """k-diffusion""" elif backend == "invisible_watermark": SCREAMING_SNAKE_CASE__ = """invisible-watermark""" assert backend in deps, f'''{backend} is not in the deps table!'''
59
1
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[int]] , UpperCamelCase__: int , UpperCamelCase__: int , UpperCamelCase__: set ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = len(UpperCamelCase__ ), len(grid[0] ) if ( min(UpperCamelCase__ , UpperCamelCase__ ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) SCREAMING_SNAKE_CASE__ = 0 count += depth_first_search(UpperCamelCase__ , row + 1 , UpperCamelCase__ , UpperCamelCase__ ) count += depth_first_search(UpperCamelCase__ , row - 1 , UpperCamelCase__ , UpperCamelCase__ ) count += depth_first_search(UpperCamelCase__ , UpperCamelCase__ , col + 1 , UpperCamelCase__ ) count += depth_first_search(UpperCamelCase__ , UpperCamelCase__ , col - 1 , UpperCamelCase__ ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
59
import warnings from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch from ...models import UNetaDModel from ...schedulers import RePaintScheduler from ...utils import PIL_INTERPOLATION, logging, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput _lowerCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[List, PIL.Image.Image, torch.Tensor] ): warnings.warn( """The preprocess method is deprecated and will be removed in a future version. Please""" """ use VaeImageProcessor.preprocess instead""" , UpperCamelCase__ , ) if isinstance(UpperCamelCase__ , torch.Tensor ): return image elif isinstance(UpperCamelCase__ , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ = [image] if isinstance(image[0] , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = image[0].size SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = (x - x % 8 for x in (w, h)) # resize to integer multiple of 8 SCREAMING_SNAKE_CASE__ = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION["""lanczos"""] ) )[None, :] for i in image] SCREAMING_SNAKE_CASE__ = np.concatenate(UpperCamelCase__ , axis=0 ) SCREAMING_SNAKE_CASE__ = np.array(UpperCamelCase__ ).astype(np.floataa ) / 2_5_5.0 SCREAMING_SNAKE_CASE__ = image.transpose(0 , 3 , 1 , 2 ) SCREAMING_SNAKE_CASE__ = 2.0 * image - 1.0 SCREAMING_SNAKE_CASE__ = torch.from_numpy(UpperCamelCase__ ) elif isinstance(image[0] , torch.Tensor ): SCREAMING_SNAKE_CASE__ = torch.cat(UpperCamelCase__ , dim=0 ) return image def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[List, PIL.Image.Image, torch.Tensor] ): if isinstance(UpperCamelCase__ , torch.Tensor ): return mask elif isinstance(UpperCamelCase__ , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ = [mask] if isinstance(mask[0] , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = mask[0].size SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 SCREAMING_SNAKE_CASE__ = [np.array(m.convert("""L""" ).resize((w, h) , resample=PIL_INTERPOLATION["""nearest"""] ) )[None, :] for m in mask] SCREAMING_SNAKE_CASE__ = np.concatenate(UpperCamelCase__ , axis=0 ) SCREAMING_SNAKE_CASE__ = mask.astype(np.floataa ) / 2_5_5.0 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = torch.from_numpy(UpperCamelCase__ ) elif isinstance(mask[0] , torch.Tensor ): SCREAMING_SNAKE_CASE__ = torch.cat(UpperCamelCase__ , dim=0 ) return mask class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = 42 lowerCamelCase_ = 42 def __init__( self :Any , __A :List[Any] , __A :Optional[Any] ) -> Union[str, Any]: """simple docstring""" super().__init__() self.register_modules(unet=__A , scheduler=__A ) @torch.no_grad() def __call__( self :str , __A :Union[torch.Tensor, PIL.Image.Image] , __A :Union[torch.Tensor, PIL.Image.Image] , __A :int = 250 , __A :float = 0.0 , __A :int = 10 , __A :int = 10 , __A :Optional[Union[torch.Generator, List[torch.Generator]]] = None , __A :Optional[str] = "pil" , __A :bool = True , ) -> Union[ImagePipelineOutput, Tuple]: """simple docstring""" SCREAMING_SNAKE_CASE__ = image SCREAMING_SNAKE_CASE__ = _preprocess_image(__A ) SCREAMING_SNAKE_CASE__ = original_image.to(device=self.device , dtype=self.unet.dtype ) SCREAMING_SNAKE_CASE__ = _preprocess_mask(__A ) SCREAMING_SNAKE_CASE__ = mask_image.to(device=self.device , dtype=self.unet.dtype ) SCREAMING_SNAKE_CASE__ = original_image.shape[0] # sample gaussian noise to begin the loop if isinstance(__A , __A ) and len(__A ) != batch_size: raise ValueError( f'''You have passed a list of generators of length {len(__A )}, but requested an effective batch''' f''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) SCREAMING_SNAKE_CASE__ = original_image.shape SCREAMING_SNAKE_CASE__ = randn_tensor(__A , generator=__A , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(__A , __A , __A , self.device ) SCREAMING_SNAKE_CASE__ = eta SCREAMING_SNAKE_CASE__ = self.scheduler.timesteps[0] + 1 SCREAMING_SNAKE_CASE__ = generator[0] if isinstance(__A , __A ) else generator for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): if t < t_last: # predict the noise residual SCREAMING_SNAKE_CASE__ = self.unet(__A , __A ).sample # compute previous image: x_t -> x_t-1 SCREAMING_SNAKE_CASE__ = self.scheduler.step(__A , __A , __A , __A , __A , __A ).prev_sample else: # compute the reverse: x_t-1 -> x_t SCREAMING_SNAKE_CASE__ = self.scheduler.undo_step(__A , __A , __A ) SCREAMING_SNAKE_CASE__ = t SCREAMING_SNAKE_CASE__ = (image / 2 + 0.5).clamp(0 , 1 ) SCREAMING_SNAKE_CASE__ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": SCREAMING_SNAKE_CASE__ = self.numpy_to_pil(__A ) if not return_dict: return (image,) return ImagePipelineOutput(images=__A )
59
1
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () _lowerCamelCase = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). _lowerCamelCase = [0, 25, 50] _lowerCamelCase = [25, 50, 75] _lowerCamelCase = fuzz.membership.trimf(X, abca) _lowerCamelCase = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. _lowerCamelCase = np.ones(75) _lowerCamelCase = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) _lowerCamelCase = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) _lowerCamelCase = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) _lowerCamelCase = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) _lowerCamelCase = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] _lowerCamelCase = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) _lowerCamelCase = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] _lowerCamelCase = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] _lowerCamelCase = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
59
import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = OpenAIGPTTokenizer lowerCamelCase_ = OpenAIGPTTokenizerFast lowerCamelCase_ = True lowerCamelCase_ = False def _snake_case ( self :Optional[Any] ) -> Dict: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt SCREAMING_SNAKE_CASE__ = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """w</w>""", """r</w>""", """t</w>""", """lo""", """low""", """er</w>""", """low</w>""", """lowest</w>""", """newer</w>""", """wider</w>""", """<unk>""", ] SCREAMING_SNAKE_CASE__ = dict(zip(__A , range(len(__A ) ) ) ) SCREAMING_SNAKE_CASE__ = ["""#version: 0.2""", """l o""", """lo w""", """e r</w>""", """"""] SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" ) as fp: fp.write(json.dumps(__A ) ) with open(self.merges_file , """w""" ) as fp: fp.write("""\n""".join(__A ) ) def _snake_case ( self :Union[str, Any] , __A :str ) -> List[Any]: """simple docstring""" return "lower newer", "lower newer" def _snake_case ( self :Optional[Any] ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = OpenAIGPTTokenizer(self.vocab_file , self.merges_file ) SCREAMING_SNAKE_CASE__ = """lower""" SCREAMING_SNAKE_CASE__ = ["""low""", """er</w>"""] SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(__A ) self.assertListEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = tokens + ["""<unk>"""] SCREAMING_SNAKE_CASE__ = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , __A ) def _snake_case ( self :Optional[Any] , __A :Optional[Any]=15 ) -> Any: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) # Simple input SCREAMING_SNAKE_CASE__ = """This is a simple input""" SCREAMING_SNAKE_CASE__ = ["""This is a simple input 1""", """This is a simple input 2"""] SCREAMING_SNAKE_CASE__ = ("""This is a simple input""", """This is a pair""") SCREAMING_SNAKE_CASE__ = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests self.assertRaises(__A , tokenizer_r.encode , __A , max_length=__A , padding="""max_length""" ) # Simple input self.assertRaises(__A , tokenizer_r.encode_plus , __A , max_length=__A , padding="""max_length""" ) # Simple input self.assertRaises( __A , tokenizer_r.batch_encode_plus , __A , max_length=__A , padding="""max_length""" , ) # Pair input self.assertRaises(__A , tokenizer_r.encode , __A , max_length=__A , padding="""max_length""" ) # Pair input self.assertRaises(__A , tokenizer_r.encode_plus , __A , max_length=__A , padding="""max_length""" ) # Pair input self.assertRaises( __A , tokenizer_r.batch_encode_plus , __A , max_length=__A , padding="""max_length""" , ) def _snake_case ( self :Dict ) -> List[Any]: """simple docstring""" pass @require_ftfy @require_spacy @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ ): pass
59
1
import argparse import gc import json import os import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler _lowerCamelCase = 16 _lowerCamelCase = 32 def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str ): return int(x / 2**20 ) class UpperCamelCase_ : def __enter__( self :int ) -> Tuple: """simple docstring""" gc.collect() torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() # reset the peak gauge to zero SCREAMING_SNAKE_CASE__ = torch.cuda.memory_allocated() return self def __exit__( self :int , *__A :Any ) -> List[str]: """simple docstring""" gc.collect() torch.cuda.empty_cache() SCREAMING_SNAKE_CASE__ = torch.cuda.memory_allocated() SCREAMING_SNAKE_CASE__ = torch.cuda.max_memory_allocated() SCREAMING_SNAKE_CASE__ = bamb(self.end - self.begin ) SCREAMING_SNAKE_CASE__ = bamb(self.peak - self.begin ) # print(f"delta used/peak {self.used:4d}/{self.peaked:4d}") def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Accelerator , UpperCamelCase__: int = 16 , UpperCamelCase__: str = "bert-base-cased" , UpperCamelCase__: int = 320 , UpperCamelCase__: int = 160 , ): SCREAMING_SNAKE_CASE__ = AutoTokenizer.from_pretrained(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = load_dataset( """glue""" , """mrpc""" , split={"""train""": f'''train[:{n_train}]''', """validation""": f'''validation[:{n_val}]'''} ) def tokenize_function(UpperCamelCase__: Optional[Any] ): # max_length=None => use the model max length (it's actually the default) SCREAMING_SNAKE_CASE__ = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=UpperCamelCase__ , max_length=UpperCamelCase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset SCREAMING_SNAKE_CASE__ = datasets.map( UpperCamelCase__ , batched=UpperCamelCase__ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , load_from_cache_file=UpperCamelCase__ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library SCREAMING_SNAKE_CASE__ = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(UpperCamelCase__: str ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(UpperCamelCase__ , padding="""max_length""" , max_length=128 , return_tensors="""pt""" ) return tokenizer.pad(UpperCamelCase__ , padding="""longest""" , return_tensors="""pt""" ) # Instantiate dataloaders. SCREAMING_SNAKE_CASE__ = DataLoader( tokenized_datasets["""train"""] , shuffle=UpperCamelCase__ , collate_fn=UpperCamelCase__ , batch_size=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = DataLoader( tokenized_datasets["""validation"""] , shuffle=UpperCamelCase__ , collate_fn=UpperCamelCase__ , batch_size=UpperCamelCase__ ) return train_dataloader, eval_dataloader def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Tuple , UpperCamelCase__: Tuple ): # Initialize accelerator SCREAMING_SNAKE_CASE__ = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs SCREAMING_SNAKE_CASE__ = config["""lr"""] SCREAMING_SNAKE_CASE__ = int(config["""num_epochs"""] ) SCREAMING_SNAKE_CASE__ = int(config["""seed"""] ) SCREAMING_SNAKE_CASE__ = int(config["""batch_size"""] ) SCREAMING_SNAKE_CASE__ = args.model_name_or_path set_seed(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = get_dataloaders(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , args.n_train , args.n_val ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) SCREAMING_SNAKE_CASE__ = AutoModelForSequenceClassification.from_pretrained(UpperCamelCase__ , return_dict=UpperCamelCase__ ) # Instantiate optimizer SCREAMING_SNAKE_CASE__ = ( AdamW if accelerator.state.deepspeed_plugin is None or """optimizer""" not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) SCREAMING_SNAKE_CASE__ = optimizer_cls(params=model.parameters() , lr=UpperCamelCase__ ) if accelerator.state.deepspeed_plugin is not None: SCREAMING_SNAKE_CASE__ = accelerator.state.deepspeed_plugin.deepspeed_config[ """gradient_accumulation_steps""" ] else: SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = (len(UpperCamelCase__ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): SCREAMING_SNAKE_CASE__ = get_linear_schedule_with_warmup( optimizer=UpperCamelCase__ , num_warmup_steps=0 , num_training_steps=UpperCamelCase__ , ) else: SCREAMING_SNAKE_CASE__ = DummyScheduler(UpperCamelCase__ , total_num_steps=UpperCamelCase__ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = accelerator.prepare( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # We need to keep track of how many total steps we have iterated over SCREAMING_SNAKE_CASE__ = 0 # We also need to keep track of the stating epoch so files are named properly SCREAMING_SNAKE_CASE__ = 0 # Now we train the model SCREAMING_SNAKE_CASE__ = {} for epoch in range(UpperCamelCase__ , UpperCamelCase__ ): with TorchTracemalloc() as tracemalloc: model.train() for step, batch in enumerate(UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = model(**UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = outputs.loss SCREAMING_SNAKE_CASE__ = loss / gradient_accumulation_steps accelerator.backward(UpperCamelCase__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 # Printing the GPU memory usage details such as allocated memory, peak memory, and total memory usage accelerator.print("""Memory before entering the train : {}""".format(bamb(tracemalloc.begin ) ) ) accelerator.print("""Memory consumed at the end of the train (end-begin): {}""".format(tracemalloc.used ) ) accelerator.print("""Peak Memory consumed during the train (max-begin): {}""".format(tracemalloc.peaked ) ) accelerator.print( """Total Peak Memory consumed during the train (max): {}""".format( tracemalloc.peaked + bamb(tracemalloc.begin ) ) ) SCREAMING_SNAKE_CASE__ = tracemalloc.peaked + bamb(tracemalloc.begin ) if args.peak_memory_upper_bound is not None: assert ( train_total_peak_memory[f'''epoch-{epoch}'''] <= args.peak_memory_upper_bound ), "Peak memory usage exceeded the upper bound" accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , """peak_memory_utilization.json""" ) , """w""" ) as f: json.dump(UpperCamelCase__ , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser(description="""Simple example of training script tracking peak GPU memory usage.""" ) parser.add_argument( """--model_name_or_path""" , type=UpperCamelCase__ , default="""bert-base-cased""" , help="""Path to pretrained model or model identifier from huggingface.co/models.""" , required=UpperCamelCase__ , ) parser.add_argument( """--output_dir""" , type=UpperCamelCase__ , default=""".""" , help="""Optional save directory where all checkpoint folders will be stored. Default is the current working directory.""" , ) parser.add_argument( """--peak_memory_upper_bound""" , type=UpperCamelCase__ , default=UpperCamelCase__ , help="""The upper bound of peak memory usage in MB. If set, the training will throw an error if the peak memory usage exceeds this value.""" , ) parser.add_argument( """--n_train""" , type=UpperCamelCase__ , default=320 , help="""Number of training examples to use.""" , ) parser.add_argument( """--n_val""" , type=UpperCamelCase__ , default=160 , help="""Number of validation examples to use.""" , ) parser.add_argument( """--num_epochs""" , type=UpperCamelCase__ , default=1 , help="""Number of train epochs.""" , ) SCREAMING_SNAKE_CASE__ = parser.parse_args() SCREAMING_SNAKE_CASE__ = {"""lr""": 2e-5, """num_epochs""": args.num_epochs, """seed""": 42, """batch_size""": 16} training_function(UpperCamelCase__ , UpperCamelCase__ ) if __name__ == "__main__": main()
59
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=UpperCamelCase__ ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = field(default="image-classification" , metadata={"include_in_asdict_even_if_is_default": True} ) lowerCamelCase_ = Features({"image": Image()} ) lowerCamelCase_ = Features({"labels": ClassLabel} ) lowerCamelCase_ = "image" lowerCamelCase_ = "labels" def _snake_case ( self :List[str] , __A :Tuple ) -> Tuple: """simple docstring""" if self.label_column not in features: raise ValueError(f'''Column {self.label_column} is not present in features.''' ) if not isinstance(features[self.label_column] , __A ): raise ValueError(f'''Column {self.label_column} is not a ClassLabel.''' ) SCREAMING_SNAKE_CASE__ = copy.deepcopy(self ) SCREAMING_SNAKE_CASE__ = self.label_schema.copy() SCREAMING_SNAKE_CASE__ = features[self.label_column] SCREAMING_SNAKE_CASE__ = label_schema return task_template @property def _snake_case ( self :Dict ) -> Dict[str, str]: """simple docstring""" return { self.image_column: "image", self.label_column: "labels", }
59
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _lowerCamelCase = {'configuration_deit': ['DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'DeiTConfig', 'DeiTOnnxConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['DeiTFeatureExtractor'] _lowerCamelCase = ['DeiTImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'DEIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'DeiTForImageClassification', 'DeiTForImageClassificationWithTeacher', 'DeiTForMaskedImageModeling', 'DeiTModel', 'DeiTPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFDeiTForImageClassification', 'TFDeiTForImageClassificationWithTeacher', 'TFDeiTForMaskedImageModeling', 'TFDeiTModel', 'TFDeiTPreTrainedModel', ] if TYPE_CHECKING: from .configuration_deit import DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, DeiTConfig, DeiTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_deit import DeiTFeatureExtractor from .image_processing_deit import DeiTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deit import ( DEIT_PRETRAINED_MODEL_ARCHIVE_LIST, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, DeiTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deit import ( TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, TFDeiTPreTrainedModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCamelCase = { 'configuration_xmod': [ 'XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XmodConfig', 'XmodOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'XMOD_PRETRAINED_MODEL_ARCHIVE_LIST', 'XmodForCausalLM', 'XmodForMaskedLM', 'XmodForMultipleChoice', 'XmodForQuestionAnswering', 'XmodForSequenceClassification', 'XmodForTokenClassification', 'XmodModel', 'XmodPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'edbeeching/decision-transformer-gym-hopper-medium': ( 'https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/config.json' ), # See all DecisionTransformer models at https://huggingface.co/models?filter=decision_transformer } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "decision_transformer" lowerCamelCase_ = ["past_key_values"] lowerCamelCase_ = { "max_position_embeddings": "n_positions", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self :str , __A :List[Any]=17 , __A :Optional[int]=4 , __A :str=128 , __A :int=4096 , __A :Any=True , __A :int=1 , __A :int=1024 , __A :int=3 , __A :Union[str, Any]=1 , __A :int=None , __A :str="relu" , __A :Union[str, Any]=0.1 , __A :Tuple=0.1 , __A :Optional[int]=0.1 , __A :Optional[int]=1E-5 , __A :int=0.0_2 , __A :Any=True , __A :Dict=True , __A :Optional[int]=5_0256 , __A :Dict=5_0256 , __A :str=False , __A :Optional[Any]=False , **__A :Any , ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = state_dim SCREAMING_SNAKE_CASE__ = act_dim SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = max_ep_len SCREAMING_SNAKE_CASE__ = action_tanh SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = n_positions SCREAMING_SNAKE_CASE__ = n_layer SCREAMING_SNAKE_CASE__ = n_head SCREAMING_SNAKE_CASE__ = n_inner SCREAMING_SNAKE_CASE__ = activation_function SCREAMING_SNAKE_CASE__ = resid_pdrop SCREAMING_SNAKE_CASE__ = embd_pdrop SCREAMING_SNAKE_CASE__ = attn_pdrop SCREAMING_SNAKE_CASE__ = layer_norm_epsilon SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = scale_attn_weights SCREAMING_SNAKE_CASE__ = use_cache SCREAMING_SNAKE_CASE__ = scale_attn_by_inverse_layer_idx SCREAMING_SNAKE_CASE__ = reorder_and_upcast_attn SCREAMING_SNAKE_CASE__ = bos_token_id SCREAMING_SNAKE_CASE__ = eos_token_id super().__init__(bos_token_id=__A , eos_token_id=__A , **__A )
59
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "timm_backbone" def __init__( self :Union[str, Any] , __A :str=None , __A :Union[str, Any]=3 , __A :str=True , __A :Any=True , __A :Optional[Any]=None , **__A :List[str] , ) -> Tuple: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = backbone SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = features_only SCREAMING_SNAKE_CASE__ = use_pretrained_backbone SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = out_indices if out_indices is not None else (-1,)
59
1
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int = 3 , UpperCamelCase__: int = 7 , UpperCamelCase__: int = 1_000_000 ): SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 1 for current_denominator in range(1 , limit + 1 ): SCREAMING_SNAKE_CASE__ = current_denominator * numerator // denominator if current_denominator % denominator == 0: current_numerator -= 1 if current_numerator * max_denominator > current_denominator * max_numerator: SCREAMING_SNAKE_CASE__ = current_numerator SCREAMING_SNAKE_CASE__ = current_denominator return max_numerator if __name__ == "__main__": print(solution(numerator=3, denominator=7, limit=1000000))
59
from math import pow, sqrt def SCREAMING_SNAKE_CASE__ ( *UpperCamelCase__: float ): SCREAMING_SNAKE_CASE__ = len(UpperCamelCase__ ) > 0 and all(value > 0.0 for value in values ) return result def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ ) else ValueError("""Input Error: Molar mass values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) )
59
1
from dataclasses import dataclass from typing import Optional import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .modeling_utils import ModelMixin @dataclass class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = 42 class UpperCamelCase_ ( UpperCamelCase__ , UpperCamelCase__ ): @register_to_config def __init__( self :Optional[Any] , __A :int = 16 , __A :int = 88 , __A :Optional[int] = None , __A :Optional[int] = None , __A :int = 1 , __A :float = 0.0 , __A :int = 32 , __A :Optional[int] = None , __A :bool = False , __A :Optional[int] = None , __A :str = "geglu" , __A :bool = True , __A :bool = True , ) -> List[Any]: """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = attention_head_dim SCREAMING_SNAKE_CASE__ = num_attention_heads * attention_head_dim SCREAMING_SNAKE_CASE__ = in_channels SCREAMING_SNAKE_CASE__ = torch.nn.GroupNorm(num_groups=__A , num_channels=__A , eps=1E-6 , affine=__A ) SCREAMING_SNAKE_CASE__ = nn.Linear(__A , __A ) # 3. Define transformers blocks SCREAMING_SNAKE_CASE__ = nn.ModuleList( [ BasicTransformerBlock( __A , __A , __A , dropout=__A , cross_attention_dim=__A , activation_fn=__A , attention_bias=__A , double_self_attention=__A , norm_elementwise_affine=__A , ) for d in range(__A ) ] ) SCREAMING_SNAKE_CASE__ = nn.Linear(__A , __A ) def _snake_case ( self :Any , __A :int , __A :int=None , __A :Tuple=None , __A :str=None , __A :str=1 , __A :str=None , __A :bool = True , ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = hidden_states.shape SCREAMING_SNAKE_CASE__ = batch_frames // num_frames SCREAMING_SNAKE_CASE__ = hidden_states SCREAMING_SNAKE_CASE__ = hidden_states[None, :].reshape(__A , __A , __A , __A , __A ) SCREAMING_SNAKE_CASE__ = hidden_states.permute(0 , 2 , 1 , 3 , 4 ) SCREAMING_SNAKE_CASE__ = self.norm(__A ) SCREAMING_SNAKE_CASE__ = hidden_states.permute(0 , 3 , 4 , 2 , 1 ).reshape(batch_size * height * width , __A , __A ) SCREAMING_SNAKE_CASE__ = self.proj_in(__A ) # 2. Blocks for block in self.transformer_blocks: SCREAMING_SNAKE_CASE__ = block( __A , encoder_hidden_states=__A , timestep=__A , cross_attention_kwargs=__A , class_labels=__A , ) # 3. Output SCREAMING_SNAKE_CASE__ = self.proj_out(__A ) SCREAMING_SNAKE_CASE__ = ( hidden_states[None, None, :] .reshape(__A , __A , __A , __A , __A ) .permute(0 , 3 , 4 , 1 , 2 ) .contiguous() ) SCREAMING_SNAKE_CASE__ = hidden_states.reshape(__A , __A , __A , __A ) SCREAMING_SNAKE_CASE__ = hidden_states + residual if not return_dict: return (output,) return TransformerTemporalModelOutput(sample=__A )
59
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'shi-labs/nat-mini-in1k-224': 'https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json', # See all Nat models at https://huggingface.co/models?filter=nat } class UpperCamelCase_ ( UpperCamelCase__ , UpperCamelCase__ ): lowerCamelCase_ = "nat" lowerCamelCase_ = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self :List[Any] , __A :Optional[Any]=4 , __A :Any=3 , __A :Optional[int]=64 , __A :Optional[int]=[3, 4, 6, 5] , __A :Union[str, Any]=[2, 4, 8, 16] , __A :Optional[Any]=7 , __A :Optional[Any]=3.0 , __A :List[Any]=True , __A :int=0.0 , __A :Dict=0.0 , __A :Optional[Any]=0.1 , __A :str="gelu" , __A :Optional[Any]=0.0_2 , __A :Optional[int]=1E-5 , __A :Optional[int]=0.0 , __A :Optional[Any]=None , __A :Union[str, Any]=None , **__A :Union[str, Any] , ) -> Optional[int]: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = patch_size SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = embed_dim SCREAMING_SNAKE_CASE__ = depths SCREAMING_SNAKE_CASE__ = len(__A ) SCREAMING_SNAKE_CASE__ = num_heads SCREAMING_SNAKE_CASE__ = kernel_size SCREAMING_SNAKE_CASE__ = mlp_ratio SCREAMING_SNAKE_CASE__ = qkv_bias SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = drop_path_rate SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model SCREAMING_SNAKE_CASE__ = int(embed_dim * 2 ** (len(__A ) - 1) ) SCREAMING_SNAKE_CASE__ = layer_scale_init_value SCREAMING_SNAKE_CASE__ = ["""stem"""] + [f'''stage{idx}''' for idx in range(1 , len(__A ) + 1 )] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = get_aligned_output_features_output_indices( out_features=__A , out_indices=__A , stage_names=self.stage_names )
59
1
import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _lowerCamelCase = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece class UpperCamelCase_ ( UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = XLMProphetNetTokenizer lowerCamelCase_ = False lowerCamelCase_ = True def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing SCREAMING_SNAKE_CASE__ = XLMProphetNetTokenizer(__A , keep_accents=__A ) tokenizer.save_pretrained(self.tmpdirname ) def _snake_case ( self :Optional[Any] ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = """[PAD]""" SCREAMING_SNAKE_CASE__ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__A ) , __A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__A ) , __A ) def _snake_case ( self :Any ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """[PAD]""" ) self.assertEqual(vocab_keys[1] , """[CLS]""" ) self.assertEqual(vocab_keys[-1] , """j""" ) self.assertEqual(len(__A ) , 1012 ) def _snake_case ( self :List[Any] ) -> Union[str, Any]: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1012 ) def _snake_case ( self :Union[str, Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = XLMProphetNetTokenizer(__A , keep_accents=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(__A , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__A ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) SCREAMING_SNAKE_CASE__ = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( __A , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_ids(__A ) self.assertListEqual( __A , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] , ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_ids_to_tokens(__A ) self.assertListEqual( __A , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """[UNK]""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """[UNK]""", """.""", ] , ) @cached_property def _snake_case ( self :Tuple ) -> Any: """simple docstring""" return XLMProphetNetTokenizer.from_pretrained("""microsoft/xprophetnet-large-wiki100-cased""" ) @slow def _snake_case ( self :Tuple ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = """Hello World!""" SCREAMING_SNAKE_CASE__ = [3_5389, 6672, 49, 2] self.assertListEqual(__A , self.big_tokenizer.encode(__A ) ) @slow def _snake_case ( self :List[Any] ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = {"""input_ids""": [[1_1073, 8_2783, 18, 26, 8_2783, 549, 5_1540, 248, 1_7209, 1301, 217, 20, 21_5186, 1325, 147, 1_7209, 1301, 217, 20, 5_6370, 53, 12_2020, 20, 1_6477, 27, 8_7355, 4548, 20, 4728, 7_8392, 17, 15_9969, 18, 26, 2_4491, 629, 15, 538, 2_2704, 5439, 15, 2788, 2_4491, 9885, 15, 4_3534, 605, 15, 814, 1_8403, 3_3200, 29, 15, 4_3534, 2_4458, 1_2410, 111, 2_4966, 8_3669, 9637, 14_4068, 26, 850, 2_2346, 27, 147, 2_4966, 8_3669, 8_3490, 26, 3_9113, 735, 27, 689, 656, 2800, 1339, 4600, 53, 12_2020, 11_5785, 34, 816, 1339, 4_6887, 18, 147, 5_3905, 1951, 4_2238, 4_1170, 1_7732, 834, 436, 15, 2_7523, 9_8733, 217, 147, 5542, 4981, 930, 1_7347, 16, 2], [2_0091, 629, 94, 8_2786, 58, 490, 20, 1528, 84, 5_3905, 344, 8_0592, 11_0128, 1_8822, 5267, 1306, 62, 15_2537, 308, 7997, 401, 12_4427, 549, 3_5442, 225, 109, 1_5055, 2_5748, 147, 7119, 4_3712, 34, 767, 13_5366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 6_3784, 11_9466, 17, 14_7808, 8_8214, 18, 656, 81, 32, 3296, 1_0280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__A , model_name="""microsoft/xprophetnet-large-wiki100-cased""" , revision="""1acad1643ddd54a44df6a1b797ada8373685d90e""" , )
59
from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any ): return {key.lstrip("""-""" ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = ArgumentParser( """HuggingFace Datasets CLI tool""" , usage="""datasets-cli <command> [<args>]""" , allow_abbrev=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = parser.add_subparsers(help="""datasets-cli command helpers""" ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(UpperCamelCase__ ) EnvironmentCommand.register_subcommand(UpperCamelCase__ ) TestCommand.register_subcommand(UpperCamelCase__ ) RunBeamCommand.register_subcommand(UpperCamelCase__ ) DummyDataCommand.register_subcommand(UpperCamelCase__ ) # Parse args SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = parser.parse_known_args() if not hasattr(UpperCamelCase__ , """func""" ): parser.print_help() exit(1 ) SCREAMING_SNAKE_CASE__ = parse_unknown_args(UpperCamelCase__ ) # Run SCREAMING_SNAKE_CASE__ = args.func(UpperCamelCase__ , **UpperCamelCase__ ) service.run() if __name__ == "__main__": main()
59
1
import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class UpperCamelCase_ ( UpperCamelCase__ ): def _snake_case ( self :int ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = tempfile.mkdtemp() SCREAMING_SNAKE_CASE__ = 8 # DPR tok SCREAMING_SNAKE_CASE__ = [ """[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , """dpr_tokenizer""" ) os.makedirs(__A , exist_ok=__A ) SCREAMING_SNAKE_CASE__ = os.path.join(__A , DPR_VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) # BART tok SCREAMING_SNAKE_CASE__ = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] SCREAMING_SNAKE_CASE__ = dict(zip(__A , range(len(__A ) ) ) ) SCREAMING_SNAKE_CASE__ = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] SCREAMING_SNAKE_CASE__ = {"""unk_token""": """<unk>"""} SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , """bart_tokenizer""" ) os.makedirs(__A , exist_ok=__A ) SCREAMING_SNAKE_CASE__ = os.path.join(__A , BART_VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(__A , BART_VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__A ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__A ) ) def _snake_case ( self :Union[str, Any] ) -> DPRQuestionEncoderTokenizer: """simple docstring""" return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , """dpr_tokenizer""" ) ) def _snake_case ( self :List[Any] ) -> DPRContextEncoderTokenizer: """simple docstring""" return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , """dpr_tokenizer""" ) ) def _snake_case ( self :Union[str, Any] ) -> BartTokenizer: """simple docstring""" return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , """bart_tokenizer""" ) ) def _snake_case ( self :Union[str, Any] ) -> List[Any]: """simple docstring""" shutil.rmtree(self.tmpdirname ) def _snake_case ( self :Tuple ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = Dataset.from_dict( { """id""": ["""0""", """1"""], """text""": ["""foo""", """bar"""], """title""": ["""Foo""", """Bar"""], """embeddings""": [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index("""embeddings""" , string_factory="""Flat""" , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def _snake_case ( self :str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_dummy_dataset() SCREAMING_SNAKE_CASE__ = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch("""transformers.models.rag.retrieval_rag.load_dataset""" ) as mock_load_dataset: SCREAMING_SNAKE_CASE__ = dataset SCREAMING_SNAKE_CASE__ = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def _snake_case ( self :Union[str, Any] , __A :bool ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_dummy_dataset() SCREAMING_SNAKE_CASE__ = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name="""custom""" , ) if from_disk: SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , """dataset""" ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , """index.faiss""" ) dataset.get_index("""embeddings""" ).save(os.path.join(self.tmpdirname , """index.faiss""" ) ) dataset.drop_index("""embeddings""" ) dataset.save_to_disk(os.path.join(self.tmpdirname , """dataset""" ) ) del dataset SCREAMING_SNAKE_CASE__ = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: SCREAMING_SNAKE_CASE__ = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , __A ) , ) return retriever def _snake_case ( self :Tuple ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = Dataset.from_dict( { """id""": ["""0""", """1"""], """text""": ["""foo""", """bar"""], """title""": ["""Foo""", """Bar"""], """embeddings""": [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index("""embeddings""" , string_factory="""Flat""" , metric_type=faiss.METRIC_INNER_PRODUCT ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , """hf_bert_base.hnswSQ8_correct_phi_128.c_index""" ) dataset.save_faiss_index("""embeddings""" , index_file_name + """.index.dpr""" ) pickle.dump(dataset["""id"""] , open(index_file_name + """.index_meta.dpr""" , """wb""" ) ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , """psgs_w100.tsv.pkl""" ) SCREAMING_SNAKE_CASE__ = {sample["""id"""]: [sample["""text"""], sample["""title"""]] for sample in dataset} pickle.dump(__A , open(__A , """wb""" ) ) SCREAMING_SNAKE_CASE__ = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name="""legacy""" , index_path=self.tmpdirname , ) SCREAMING_SNAKE_CASE__ = RagRetriever( __A , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def _snake_case ( self :str ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = self.get_dummy_canonical_hf_index_retriever() SCREAMING_SNAKE_CASE__ = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["""embeddings""", """id""", """text""", """title"""] ) self.assertEqual(len(doc_dicts[0]["""id"""] ) , __A ) self.assertEqual(doc_dicts[0]["""id"""][0] , """1""" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["""id"""][0] , """0""" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _snake_case ( self :Union[str, Any] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch("""transformers.models.rag.retrieval_rag.load_dataset""" ) as mock_load_dataset: SCREAMING_SNAKE_CASE__ = self.get_dummy_dataset() retriever.save_pretrained(__A ) SCREAMING_SNAKE_CASE__ = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) SCREAMING_SNAKE_CASE__ = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) SCREAMING_SNAKE_CASE__ = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _snake_case ( self :Tuple ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) SCREAMING_SNAKE_CASE__ = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["""embeddings""", """id""", """text""", """title"""] ) self.assertEqual(len(doc_dicts[0]["""id"""] ) , __A ) self.assertEqual(doc_dicts[0]["""id"""][0] , """1""" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["""id"""][0] , """0""" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _snake_case ( self :Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) SCREAMING_SNAKE_CASE__ = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) SCREAMING_SNAKE_CASE__ = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) SCREAMING_SNAKE_CASE__ = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _snake_case ( self :Any ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) SCREAMING_SNAKE_CASE__ = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["""embeddings""", """id""", """text""", """title"""] ) self.assertEqual(len(doc_dicts[0]["""id"""] ) , __A ) self.assertEqual(doc_dicts[0]["""id"""][0] , """1""" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["""id"""][0] , """0""" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _snake_case ( self :Any ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) SCREAMING_SNAKE_CASE__ = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) SCREAMING_SNAKE_CASE__ = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) SCREAMING_SNAKE_CASE__ = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) def _snake_case ( self :Dict ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = self.get_dummy_legacy_index_retriever() SCREAMING_SNAKE_CASE__ = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = retriever.retrieve(__A , n_docs=__A ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(__A ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["""text""", """title"""] ) self.assertEqual(len(doc_dicts[0]["""text"""] ) , __A ) self.assertEqual(doc_dicts[0]["""text"""][0] , """bar""" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["""text"""][0] , """foo""" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _snake_case ( self :int ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(__A ) SCREAMING_SNAKE_CASE__ = RagRetriever.from_pretrained(__A ) self.assertIsInstance(__A , __A ) SCREAMING_SNAKE_CASE__ = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) SCREAMING_SNAKE_CASE__ = retriever.retrieve(__A , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def _snake_case ( self :Dict ) -> Tuple: """simple docstring""" import torch SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = self.get_dummy_canonical_hf_index_retriever() SCREAMING_SNAKE_CASE__ = [[5, 7], [10, 11]] SCREAMING_SNAKE_CASE__ = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) SCREAMING_SNAKE_CASE__ = retriever(__A , __A , prefix=retriever.config.generator.prefix , n_docs=__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = ( out["""context_input_ids"""], out["""context_attention_mask"""], out["""retrieved_doc_embeds"""], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__A , __A ) self.assertIsInstance(__A , __A ) self.assertIsInstance(__A , np.ndarray ) SCREAMING_SNAKE_CASE__ = retriever( __A , __A , prefix=retriever.config.generator.prefix , n_docs=__A , return_tensors="""pt""" , ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = ( # noqa: F841 out["""context_input_ids"""], out["""context_attention_mask"""], out["""retrieved_doc_embeds"""], out["""doc_ids"""], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(__A , torch.Tensor ) self.assertIsInstance(__A , torch.Tensor ) self.assertIsInstance(__A , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def _snake_case ( self :List[str] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_dpr_ctx_encoder_tokenizer() SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = self.get_dummy_custom_hf_index_retriever(from_disk=__A ) retriever.set_ctx_encoder_tokenizer(__A ) SCREAMING_SNAKE_CASE__ = [[5, 7], [10, 11]] SCREAMING_SNAKE_CASE__ = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) SCREAMING_SNAKE_CASE__ = retriever(__A , __A , prefix=retriever.config.generator.prefix , n_docs=__A ) self.assertEqual( len(__A ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ("""tokenized_doc_ids""", """tokenized_doc_attention_mask""") ) , __A ) # check for doc token related keys in dictionary.
59
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase__ ): def __init__( self :List[Any] , *__A :Tuple , **__A :Dict ) -> None: """simple docstring""" warnings.warn( """The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use LayoutLMv2ImageProcessor instead.""" , __A , ) super().__init__(*__A , **__A )
59
1
import os import unittest from transformers.models.cpmant.tokenization_cpmant import VOCAB_FILES_NAMES, CpmAntTokenizer from transformers.testing_utils import require_jieba, tooslow from ...test_tokenization_common import TokenizerTesterMixin @require_jieba class UpperCamelCase_ ( UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = CpmAntTokenizer lowerCamelCase_ = False def _snake_case ( self :Any ) -> List[Any]: """simple docstring""" super().setUp() SCREAMING_SNAKE_CASE__ = [ """<d>""", """</d>""", """<s>""", """</s>""", """</_>""", """<unk>""", """<pad>""", """</n>""", """我""", """是""", """C""", """P""", """M""", """A""", """n""", """t""", ] SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) @tooslow def _snake_case ( self :Dict ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = CpmAntTokenizer.from_pretrained("""openbmb/cpm-ant-10b""" ) SCREAMING_SNAKE_CASE__ = """今天天气真好!""" SCREAMING_SNAKE_CASE__ = ["""今天""", """天气""", """真""", """好""", """!"""] SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(__A ) self.assertListEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = """今天天气真好!""" SCREAMING_SNAKE_CASE__ = [tokenizer.bos_token] + tokens SCREAMING_SNAKE_CASE__ = [6, 9802, 1_4962, 2082, 831, 244] self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , __A ) SCREAMING_SNAKE_CASE__ = tokenizer.decode(__A ) self.assertEqual(__A , __A )
59
import unittest from datasets import load_dataset from transformers.pipelines import pipeline from transformers.testing_utils import is_pipeline_test, nested_simplify, require_torch, slow @is_pipeline_test @require_torch class UpperCamelCase_ ( unittest.TestCase ): @require_torch def _snake_case ( self :Dict ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = pipeline( task="""zero-shot-audio-classification""" , model="""hf-internal-testing/tiny-clap-htsat-unfused""" ) SCREAMING_SNAKE_CASE__ = load_dataset("""ashraq/esc50""" ) SCREAMING_SNAKE_CASE__ = dataset["""train"""]["""audio"""][-1]["""array"""] SCREAMING_SNAKE_CASE__ = audio_classifier(__A , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__A ) , [{"""score""": 0.5_0_1, """label""": """Sound of a dog"""}, {"""score""": 0.4_9_9, """label""": """Sound of vaccum cleaner"""}] , ) @unittest.skip("""No models are available in TF""" ) def _snake_case ( self :Dict ) -> List[str]: """simple docstring""" pass @slow @require_torch def _snake_case ( self :Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = pipeline( task="""zero-shot-audio-classification""" , model="""laion/clap-htsat-unfused""" , ) # This is an audio of a dog SCREAMING_SNAKE_CASE__ = load_dataset("""ashraq/esc50""" ) SCREAMING_SNAKE_CASE__ = dataset["""train"""]["""audio"""][-1]["""array"""] SCREAMING_SNAKE_CASE__ = audio_classifier(__A , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__A ) , [ {"""score""": 0.9_9_9, """label""": """Sound of a dog"""}, {"""score""": 0.0_0_1, """label""": """Sound of vaccum cleaner"""}, ] , ) SCREAMING_SNAKE_CASE__ = audio_classifier([audio] * 5 , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__A ) , [ [ {"""score""": 0.9_9_9, """label""": """Sound of a dog"""}, {"""score""": 0.0_0_1, """label""": """Sound of vaccum cleaner"""}, ], ] * 5 , ) SCREAMING_SNAKE_CASE__ = audio_classifier( [audio] * 5 , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] , batch_size=5 ) self.assertEqual( nested_simplify(__A ) , [ [ {"""score""": 0.9_9_9, """label""": """Sound of a dog"""}, {"""score""": 0.0_0_1, """label""": """Sound of vaccum cleaner"""}, ], ] * 5 , ) @unittest.skip("""No models are available in TF""" ) def _snake_case ( self :str ) -> Optional[int]: """simple docstring""" pass
59
1
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): if upper_limit < 0: raise ValueError("""Limit for the Catalan sequence must be ≥ 0""" ) SCREAMING_SNAKE_CASE__ = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 SCREAMING_SNAKE_CASE__ = 1 if upper_limit > 0: SCREAMING_SNAKE_CASE__ = 1 # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i for i in range(2 , upper_limit + 1 ): for j in range(UpperCamelCase__ ): catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1] return catalan_list if __name__ == "__main__": print('\n********* Catalan Numbers Using Dynamic Programming ************\n') print('\n*** Enter -1 at any time to quit ***') print('\nEnter the upper limit (≥ 0) for the Catalan number sequence: ', end='') try: while True: _lowerCamelCase = int(input().strip()) if N < 0: print('\n********* Goodbye!! ************') break else: print(F'''The Catalan numbers from 0 through {N} are:''') print(catalan_numbers(N)) print('Try another upper limit for the sequence: ', end='') except (NameError, ValueError): print('\n********* Invalid input, goodbye! ************\n') import doctest doctest.testmod()
59
import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 _lowerCamelCase = data_utils.TransfoXLTokenizer _lowerCamelCase = data_utils.TransfoXLCorpus _lowerCamelCase = data_utils _lowerCamelCase = data_utils def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Dict , UpperCamelCase__: Any , UpperCamelCase__: Any , UpperCamelCase__: Tuple ): if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(UpperCamelCase__ , """rb""" ) as fp: SCREAMING_SNAKE_CASE__ = pickle.load(UpperCamelCase__ , encoding="""latin1""" ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) SCREAMING_SNAKE_CASE__ = pytorch_dump_folder_path + """/""" + VOCAB_FILES_NAMES["""pretrained_vocab_file"""] print(f'''Save vocabulary to {pytorch_vocab_dump_path}''' ) SCREAMING_SNAKE_CASE__ = corpus.vocab.__dict__ torch.save(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = corpus.__dict__ corpus_dict_no_vocab.pop("""vocab""" , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = pytorch_dump_folder_path + """/""" + CORPUS_NAME print(f'''Save dataset to {pytorch_dataset_dump_path}''' ) torch.save(UpperCamelCase__ , UpperCamelCase__ ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model SCREAMING_SNAKE_CASE__ = os.path.abspath(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = os.path.abspath(UpperCamelCase__ ) print(f'''Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.''' ) # Initialise PyTorch model if transfo_xl_config_file == "": SCREAMING_SNAKE_CASE__ = TransfoXLConfig() else: SCREAMING_SNAKE_CASE__ = TransfoXLConfig.from_json_file(UpperCamelCase__ ) print(f'''Building PyTorch model from configuration: {config}''' ) SCREAMING_SNAKE_CASE__ = TransfoXLLMHeadModel(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = load_tf_weights_in_transfo_xl(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # Save pytorch-model SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) print(f'''Save PyTorch model to {os.path.abspath(UpperCamelCase__ )}''' ) torch.save(model.state_dict() , UpperCamelCase__ ) print(f'''Save configuration file to {os.path.abspath(UpperCamelCase__ )}''' ) with open(UpperCamelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--tf_checkpoint_path', default='', type=str, help='An optional path to a TensorFlow checkpoint path to be converted.', ) parser.add_argument( '--transfo_xl_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained BERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--transfo_xl_dataset_file', default='', type=str, help='An optional dataset file to be converted in a vocabulary.', ) _lowerCamelCase = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
59
1
import pytest from datasets.splits import SplitDict, SplitInfo from datasets.utils.py_utils import asdict @pytest.mark.parametrize( """split_dict""" , [ SplitDict(), SplitDict({"""train""": SplitInfo(name="""train""" , num_bytes=1_337 , num_examples=42 , dataset_name="""my_dataset""" )} ), SplitDict({"""train""": SplitInfo(name="""train""" , num_bytes=1_337 , num_examples=42 )} ), SplitDict({"""train""": SplitInfo()} ), ] , ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: SplitDict ): SCREAMING_SNAKE_CASE__ = split_dict._to_yaml_list() assert len(UpperCamelCase__ ) == len(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = SplitDict._from_yaml_list(UpperCamelCase__ ) for split_name, split_info in split_dict.items(): # dataset_name field is deprecated, and is therefore not part of the YAML dump SCREAMING_SNAKE_CASE__ = None # the split name of split_dict takes over the name of the split info object SCREAMING_SNAKE_CASE__ = split_name assert split_dict == reloaded @pytest.mark.parametrize( """split_info""" , [SplitInfo(), SplitInfo(dataset_name=UpperCamelCase__ ), SplitInfo(dataset_name="""my_dataset""" )] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[int] ): # For backward compatibility, we need asdict(split_dict) to return split info dictrionaries with the "dataset_name" # field even if it's deprecated. This way old versionso of `datasets` can still reload dataset_infos.json files SCREAMING_SNAKE_CASE__ = asdict(SplitDict({"""train""": split_info} ) ) assert "dataset_name" in split_dict_asdict["train"] assert split_dict_asdict["train"]["dataset_name"] == split_info.dataset_name
59
import argparse import tensorflow as tf import torch from transformers import BertConfig, BertForMaskedLM from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertPooler, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging logging.set_verbosity_info() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str , UpperCamelCase__: str , UpperCamelCase__: str ): def get_masked_lm_array(UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = f'''masked_lm/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) def get_encoder_array(UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = f'''encoder/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) def get_encoder_layer_array(UpperCamelCase__: int , UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = f'''encoder/_transformer_layers/{layer_index}/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) def get_encoder_attention_layer_array(UpperCamelCase__: int , UpperCamelCase__: str , UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = f'''encoder/_transformer_layers/{layer_index}/_attention_layer/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = array.reshape(UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) print(f'''Loading model based on config from {config_path}...''' ) SCREAMING_SNAKE_CASE__ = BertConfig.from_json_file(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = BertForMaskedLM(UpperCamelCase__ ) # Layers for layer_index in range(0 , config.num_hidden_layers ): SCREAMING_SNAKE_CASE__ = model.bert.encoder.layer[layer_index] # Self-attention SCREAMING_SNAKE_CASE__ = layer.attention.self SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_query_dense/kernel""" , self_attn.query.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_query_dense/bias""" , self_attn.query.bias.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_key_dense/kernel""" , self_attn.key.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_key_dense/bias""" , self_attn.key.bias.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_value_dense/kernel""" , self_attn.value.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_value_dense/bias""" , self_attn.value.bias.data.shape ) # Self-attention Output SCREAMING_SNAKE_CASE__ = layer.attention.output SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_output_dense/kernel""" , self_output.dense.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_output_dense/bias""" , self_output.dense.bias.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_attention_layer_norm/gamma""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_attention_layer_norm/beta""" ) # Intermediate SCREAMING_SNAKE_CASE__ = layer.intermediate SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_intermediate_dense/kernel""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_intermediate_dense/bias""" ) # Output SCREAMING_SNAKE_CASE__ = layer.output SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_dense/kernel""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_dense/bias""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_layer_norm/gamma""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_layer_norm/beta""" ) # Embeddings SCREAMING_SNAKE_CASE__ = get_encoder_array("""_position_embedding_layer/embeddings""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_type_embedding_layer/embeddings""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_embedding_norm_layer/gamma""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_embedding_norm_layer/beta""" ) # LM Head SCREAMING_SNAKE_CASE__ = model.cls.predictions.transform SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""dense/kernel""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""dense/bias""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""layer_norm/gamma""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""layer_norm/beta""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""embedding_table""" ) # Pooling SCREAMING_SNAKE_CASE__ = BertPooler(config=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_pooler_layer/kernel""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_pooler_layer/bias""" ) # Export final model model.save_pretrained(UpperCamelCase__ ) # Integration test - should load without any errors ;) SCREAMING_SNAKE_CASE__ = BertForMaskedLM.from_pretrained(UpperCamelCase__ ) print(new_model.eval() ) print("""Model conversion was done sucessfully!""" ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '--tf_checkpoint_path', type=str, required=True, help='Path to the TensorFlow Token Dropping checkpoint path.' ) parser.add_argument( '--bert_config_file', type=str, required=True, help='The config json file corresponding to the BERT model. This specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', type=str, required=True, help='Path to the output PyTorch model.', ) _lowerCamelCase = parser.parse_args() convert_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
59
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _lowerCamelCase = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['MLukeTokenizer'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mluke import MLukeTokenizer else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _lowerCamelCase = '\\n Text data.\n Second line of data.' _lowerCamelCase = 'file' @pytest.fixture(scope="""session""" ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / (FILE_PATH + """.zstd""") SCREAMING_SNAKE_CASE__ = bytes(UpperCamelCase__ , """utf-8""" ) with zstd.open(UpperCamelCase__ , """wb""" ) as f: f.write(UpperCamelCase__ ) return path @pytest.fixture def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): with open(os.path.join(tmpfs.local_root_dir , UpperCamelCase__ ) , """w""" ) as f: f.write(UpperCamelCase__ ) return FILE_PATH @pytest.mark.parametrize("""compression_format""" , ["""gzip""", """xz""", """zstd"""] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int , UpperCamelCase__: Dict , UpperCamelCase__: int , UpperCamelCase__: str , UpperCamelCase__: Optional[int] , UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = {"""gzip""": gz_file, """xz""": xz_file, """zstd""": zstd_path} SCREAMING_SNAKE_CASE__ = input_paths[compression_format] SCREAMING_SNAKE_CASE__ = tmp_path / """cache""" SCREAMING_SNAKE_CASE__ = DownloadConfig(cache_dir=UpperCamelCase__ , extract_compressed_file=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = cached_path(UpperCamelCase__ , download_config=UpperCamelCase__ ) with open(UpperCamelCase__ ) as f: SCREAMING_SNAKE_CASE__ = f.read() with open(UpperCamelCase__ ) as f: SCREAMING_SNAKE_CASE__ = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("""default_extracted""" , [True, False] ) @pytest.mark.parametrize("""default_cache_dir""" , [True, False] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Tuple , UpperCamelCase__: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Any , UpperCamelCase__: Union[str, Any] ): SCREAMING_SNAKE_CASE__ = """custom_cache""" SCREAMING_SNAKE_CASE__ = """custom_extracted_dir""" SCREAMING_SNAKE_CASE__ = tmp_path / """custom_extracted_path""" if default_extracted: SCREAMING_SNAKE_CASE__ = ("""downloads""" if default_cache_dir else custom_cache_dir, """extracted""") else: monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_DIR""" , UpperCamelCase__ ) monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_PATH""" , str(UpperCamelCase__ ) ) SCREAMING_SNAKE_CASE__ = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) SCREAMING_SNAKE_CASE__ = xz_file SCREAMING_SNAKE_CASE__ = ( DownloadConfig(extract_compressed_file=UpperCamelCase__ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=UpperCamelCase__ ) ) SCREAMING_SNAKE_CASE__ = cached_path(UpperCamelCase__ , download_config=UpperCamelCase__ ) assert Path(UpperCamelCase__ ).parent.parts[-2:] == expected def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[int] ): # absolute path SCREAMING_SNAKE_CASE__ = str(Path(UpperCamelCase__ ).resolve() ) assert cached_path(UpperCamelCase__ ) == text_file # relative path SCREAMING_SNAKE_CASE__ = str(Path(UpperCamelCase__ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(UpperCamelCase__ ) == text_file def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): # absolute path SCREAMING_SNAKE_CASE__ = str(tmp_path.resolve() / """__missing_file__.txt""" ) with pytest.raises(UpperCamelCase__ ): cached_path(UpperCamelCase__ ) # relative path SCREAMING_SNAKE_CASE__ = """./__missing_file__.txt""" with pytest.raises(UpperCamelCase__ ): cached_path(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): SCREAMING_SNAKE_CASE__ = get_from_cache(f'''tmp://{tmpfs_file}''' ) with open(UpperCamelCase__ ) as f: SCREAMING_SNAKE_CASE__ = f.read() assert output_file_content == FILE_CONTENT @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( ): with pytest.raises(UpperCamelCase__ ): cached_path("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(UpperCamelCase__ ): http_get("""https://huggingface.co""" , temp_file=UpperCamelCase__ ) with pytest.raises(UpperCamelCase__ ): http_head("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[Any] ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(UpperCamelCase__ ): ftp_get("""ftp://huggingface.co""" , temp_file=UpperCamelCase__ ) with pytest.raises(UpperCamelCase__ ): ftp_head("""ftp://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(UpperCamelCase__ ): fsspec_get("""s3://huggingface.co""" , temp_file=UpperCamelCase__ ) with pytest.raises(UpperCamelCase__ ): fsspec_head("""s3://huggingface.co""" )
59
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _lowerCamelCase = { 'configuration_blip': [ 'BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BlipConfig', 'BlipTextConfig', 'BlipVisionConfig', ], 'processing_blip': ['BlipProcessor'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['BlipImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'BLIP_PRETRAINED_MODEL_ARCHIVE_LIST', 'BlipModel', 'BlipPreTrainedModel', 'BlipForConditionalGeneration', 'BlipForQuestionAnswering', 'BlipVisionModel', 'BlipTextModel', 'BlipForImageTextRetrieval', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFBlipModel', 'TFBlipPreTrainedModel', 'TFBlipForConditionalGeneration', 'TFBlipForQuestionAnswering', 'TFBlipVisionModel', 'TFBlipTextModel', 'TFBlipForImageTextRetrieval', ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _lowerCamelCase = logging.getLogger(__name__) def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser( description="""Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset.""" ) parser.add_argument( """--dataset_name""" , type=UpperCamelCase__ , default="""wikitext""" , help="""Name of the training. Explore datasets at: hf.co/datasets.""" , ) parser.add_argument( """--dataset_config""" , type=UpperCamelCase__ , default="""wikitext-103-raw-v1""" , help="""Configuration name of the dataset.""" ) parser.add_argument( """--tokenizer_name_or_path""" , type=UpperCamelCase__ , default="""sayakpaul/unigram-tokenizer-wikitext""" , help="""Tokenizer identifier. Can be a local filepath or a Hub identifier.""" , ) parser.add_argument( """--shard_size""" , type=UpperCamelCase__ , default=1_000 , help="""Number of entries to go in a single shard.""" , ) parser.add_argument("""--split""" , type=UpperCamelCase__ , default="""train""" , choices=["""train""", """test""", """validation"""] ) parser.add_argument( """--limit""" , default=UpperCamelCase__ , type=UpperCamelCase__ , help="""Limit the number of shards (used for debugging).""" , ) parser.add_argument( """--max_length""" , type=UpperCamelCase__ , default=512 , help="""Maximum sequence length. For training on TPUs, it helps to have a maximum""" """ sequence length that is a multiple of 8.""" , ) parser.add_argument( """--output_dir""" , default="""tf-tpu""" , type=UpperCamelCase__ , help="""Output directory where the TFRecord shards will be saved. If the""" """ path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord""" """ shards will be directly saved to a Google Cloud Storage bucket.""" , ) SCREAMING_SNAKE_CASE__ = parser.parse_args() return args def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[Any] ): def fn(UpperCamelCase__: Any ): return tokenizer(examples["""text"""] ) return fn def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = [] for i in range(len(tokenized_data["""input_ids"""] ) ): SCREAMING_SNAKE_CASE__ = { """input_ids""": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["""input_ids"""][i] ) ), """attention_mask""": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["""attention_mask"""][i] ) ), } SCREAMING_SNAKE_CASE__ = tf.train.Features(feature=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = tf.train.Example(features=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = example.SerializeToString() records.append(UpperCamelCase__ ) return records def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): SCREAMING_SNAKE_CASE__ = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: SCREAMING_SNAKE_CASE__ = min(len(UpperCamelCase__ ) , args.limit ) SCREAMING_SNAKE_CASE__ = dataset.select(range(UpperCamelCase__ ) ) print(f'''Limiting the dataset to {args.limit} entries.''' ) SCREAMING_SNAKE_CASE__ = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) SCREAMING_SNAKE_CASE__ = os.path.join(args.output_dir , args.split ) if not os.path.exists(UpperCamelCase__ ): os.makedirs(UpperCamelCase__ ) else: SCREAMING_SNAKE_CASE__ = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. SCREAMING_SNAKE_CASE__ = tokenize_function(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = dataset.map(UpperCamelCase__ , batched=UpperCamelCase__ , num_proc=4 , remove_columns=["""text"""] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(UpperCamelCase__: int ): # Concatenate all texts. SCREAMING_SNAKE_CASE__ = {k: sum(examples[k] , [] ) for k in examples.keys()} SCREAMING_SNAKE_CASE__ = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 SCREAMING_SNAKE_CASE__ = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. SCREAMING_SNAKE_CASE__ = { k: [t[i : i + args.max_length] for i in range(0 , UpperCamelCase__ , args.max_length )] for k, t in concatenated_examples.items() } return result SCREAMING_SNAKE_CASE__ = dataset_tokenized.map(UpperCamelCase__ , batched=UpperCamelCase__ , batch_size=1_000 , num_proc=4 ) SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 0 for shard in range(0 , len(UpperCamelCase__ ) , args.shard_size ): SCREAMING_SNAKE_CASE__ = grouped_dataset[shard : shard + args.shard_size] SCREAMING_SNAKE_CASE__ = len(dataset_snapshot["""input_ids"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , f'''dataset-{shard_count}-{records_containing}.tfrecord''' ) SCREAMING_SNAKE_CASE__ = get_serialized_examples(UpperCamelCase__ ) with tf.io.TFRecordWriter(UpperCamelCase__ ) as out_file: for i in range(len(UpperCamelCase__ ) ): SCREAMING_SNAKE_CASE__ = serialized_examples[i] out_file.write(UpperCamelCase__ ) print("""Wrote file {} containing {} records""".format(UpperCamelCase__ , UpperCamelCase__ ) ) shard_count += 1 total_records += records_containing with open(f'''split-{args.split}-records-count.txt''' , """w""" ) as f: print(f'''Total {args.split} records: {total_records}''' , file=UpperCamelCase__ ) if __name__ == "__main__": _lowerCamelCase = parse_args() main(args)
59
1
from queue import Queue from typing import TYPE_CHECKING, Optional if TYPE_CHECKING: from ..models.auto import AutoTokenizer class UpperCamelCase_ : def _snake_case ( self :Any , __A :Tuple ) -> Optional[Any]: """simple docstring""" raise NotImplementedError() def _snake_case ( self :Any ) -> Optional[Any]: """simple docstring""" raise NotImplementedError() class UpperCamelCase_ ( UpperCamelCase__ ): def __init__( self :Optional[int] , __A :"AutoTokenizer" , __A :bool = False , **__A :Optional[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = tokenizer SCREAMING_SNAKE_CASE__ = skip_prompt SCREAMING_SNAKE_CASE__ = decode_kwargs # variables used in the streaming process SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = True def _snake_case ( self :Optional[Any] , __A :List[str] ) -> Union[str, Any]: """simple docstring""" if len(value.shape ) > 1 and value.shape[0] > 1: raise ValueError("""TextStreamer only supports batch size 1""" ) elif len(value.shape ) > 1: SCREAMING_SNAKE_CASE__ = value[0] if self.skip_prompt and self.next_tokens_are_prompt: SCREAMING_SNAKE_CASE__ = False return # Add the new token to the cache and decodes the entire thing. self.token_cache.extend(value.tolist() ) SCREAMING_SNAKE_CASE__ = self.tokenizer.decode(self.token_cache , **self.decode_kwargs ) # After the symbol for a new line, we flush the cache. if text.endswith("""\n""" ): SCREAMING_SNAKE_CASE__ = text[self.print_len :] SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = 0 # If the last token is a CJK character, we print the characters. elif len(__A ) > 0 and self._is_chinese_char(ord(text[-1] ) ): SCREAMING_SNAKE_CASE__ = text[self.print_len :] self.print_len += len(__A ) # Otherwise, prints until the last space char (simple heuristic to avoid printing incomplete words, # which may change with the subsequent token -- there are probably smarter ways to do this!) else: SCREAMING_SNAKE_CASE__ = text[self.print_len : text.rfind(""" """ ) + 1] self.print_len += len(__A ) self.on_finalized_text(__A ) def _snake_case ( self :Union[str, Any] ) -> int: """simple docstring""" if len(self.token_cache ) > 0: SCREAMING_SNAKE_CASE__ = self.tokenizer.decode(self.token_cache , **self.decode_kwargs ) SCREAMING_SNAKE_CASE__ = text[self.print_len :] SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = 0 else: SCREAMING_SNAKE_CASE__ = """""" SCREAMING_SNAKE_CASE__ = True self.on_finalized_text(__A , stream_end=__A ) def _snake_case ( self :int , __A :str , __A :bool = False ) -> Union[str, Any]: """simple docstring""" print(__A , flush=__A , end="""""" if not stream_end else None ) def _snake_case ( self :Optional[Any] , __A :Optional[Any] ) -> Union[str, Any]: """simple docstring""" if ( (cp >= 0X4E_00 and cp <= 0X9F_FF) or (cp >= 0X34_00 and cp <= 0X4D_BF) # or (cp >= 0X2_00_00 and cp <= 0X2_A6_DF) # or (cp >= 0X2_A7_00 and cp <= 0X2_B7_3F) # or (cp >= 0X2_B7_40 and cp <= 0X2_B8_1F) # or (cp >= 0X2_B8_20 and cp <= 0X2_CE_AF) # or (cp >= 0XF9_00 and cp <= 0XFA_FF) or (cp >= 0X2_F8_00 and cp <= 0X2_FA_1F) # ): # return True return False class UpperCamelCase_ ( UpperCamelCase__ ): def __init__( self :List[Any] , __A :"AutoTokenizer" , __A :bool = False , __A :Optional[float] = None , **__A :Optional[int] ) -> List[Any]: """simple docstring""" super().__init__(__A , __A , **__A ) SCREAMING_SNAKE_CASE__ = Queue() SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = timeout def _snake_case ( self :str , __A :str , __A :bool = False ) -> str: """simple docstring""" self.text_queue.put(__A , timeout=self.timeout ) if stream_end: self.text_queue.put(self.stop_signal , timeout=self.timeout ) def __iter__( self :List[str] ) -> Tuple: """simple docstring""" return self def _snake_case ( self :Dict ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.text_queue.get(timeout=self.timeout ) if value == self.stop_signal: raise StopIteration() else: return value
59
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] ): SCREAMING_SNAKE_CASE__ = [] for data in source_data: for i, el in enumerate(UpperCamelCase__ ): if len(UpperCamelCase__ ) < i + 1: data_lists.append([] ) data_lists[i].append(float(UpperCamelCase__ ) ) return data_lists def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] , UpperCamelCase__: list[int] ): SCREAMING_SNAKE_CASE__ = [] for dlist, weight in zip(UpperCamelCase__ , UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = min(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = max(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = [] # for weight 0 score is 1 - actual score if weight == 0: for item in dlist: try: score.append(1 - ((item - mind) / (maxd - mind)) ) except ZeroDivisionError: score.append(1 ) elif weight == 1: for item in dlist: try: score.append((item - mind) / (maxd - mind) ) except ZeroDivisionError: score.append(0 ) # weight not 0 or 1 else: SCREAMING_SNAKE_CASE__ = f'''Invalid weight of {weight:f} provided''' raise ValueError(UpperCamelCase__ ) score_lists.append(UpperCamelCase__ ) return score_lists def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] ): SCREAMING_SNAKE_CASE__ = [0 for i in range(len(score_lists[0] ) )] for slist in score_lists: for j, ele in enumerate(UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = final_scores[j] + ele return final_scores def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] , UpperCamelCase__: list[int] ): SCREAMING_SNAKE_CASE__ = get_data(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = calculate_each_score(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = generate_final_scores(UpperCamelCase__ ) # append scores to source data for i, ele in enumerate(UpperCamelCase__ ): source_data[i].append(UpperCamelCase__ ) return source_data
59
1
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = { """en""": """Machine learning is great, isn't it?""", """ru""": """Машинное обучение - это здорово, не так ли?""", """de""": """Maschinelles Lernen ist großartig, oder?""", } # BLUE scores as follows: # "pair": [fairseq, transformers] SCREAMING_SNAKE_CASE__ = { """ru-en""": ["""[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)""", """39.20"""], """en-ru""": ["""[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)""", """33.47"""], """en-de""": ["""[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)""", """42.83"""], """de-en""": ["""[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)""", """41.35"""], } SCREAMING_SNAKE_CASE__ = f'''{src_lang}-{tgt_lang}''' SCREAMING_SNAKE_CASE__ = f''' --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "facebook/wmt19-{src_lang}-{tgt_lang}" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = "{texts[src_lang]}" input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR\'s WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) ''' os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , """README.md""" ) print(f'''Generating {path}''' ) with open(UpperCamelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(UpperCamelCase__ ) # make sure we are under the root of the project _lowerCamelCase = Path(__file__).resolve().parent.parent.parent _lowerCamelCase = repo_dir / 'model_cards' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = model_name.split('-') _lowerCamelCase = model_cards_dir / 'facebook' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
59
import warnings from functools import wraps from typing import Callable def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Callable ): @wraps(UpperCamelCase__ ) def _inner_fn(*UpperCamelCase__: Dict , **UpperCamelCase__: Any ): warnings.warn( (f'''\'{fn.__name__}\' is experimental and might be subject to breaking changes in the future.''') , UpperCamelCase__ , ) return fn(*UpperCamelCase__ , **UpperCamelCase__ ) return _inner_fn
59
1
import sys import turtle def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: tuple[float, float] , UpperCamelCase__: tuple[float, float] ): return (pa[0] + pa[0]) / 2, (pa[1] + pa[1]) / 2 def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: tuple[float, float] , UpperCamelCase__: tuple[float, float] , UpperCamelCase__: tuple[float, float] , UpperCamelCase__: int , ): my_pen.up() my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.down() my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.goto(vertexa[0] , vertexa[1] ) if depth == 0: return triangle(UpperCamelCase__ , get_mid(UpperCamelCase__ , UpperCamelCase__ ) , get_mid(UpperCamelCase__ , UpperCamelCase__ ) , depth - 1 ) triangle(UpperCamelCase__ , get_mid(UpperCamelCase__ , UpperCamelCase__ ) , get_mid(UpperCamelCase__ , UpperCamelCase__ ) , depth - 1 ) triangle(UpperCamelCase__ , get_mid(UpperCamelCase__ , UpperCamelCase__ ) , get_mid(UpperCamelCase__ , UpperCamelCase__ ) , depth - 1 ) if __name__ == "__main__": if len(sys.argv) != 2: raise ValueError( 'Correct format for using this script: ' 'python fractals.py <int:depth_for_fractal>' ) _lowerCamelCase = turtle.Turtle() my_pen.ht() my_pen.speed(5) my_pen.pencolor('red') _lowerCamelCase = [(-175, -125), (0, 175), (175, -125)] # vertices of triangle triangle(vertices[0], vertices[1], vertices[2], int(sys.argv[1]))
59
import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = RoCBertTokenizer lowerCamelCase_ = None lowerCamelCase_ = False lowerCamelCase_ = True lowerCamelCase_ = filter_non_english def _snake_case ( self :List[Any] ) -> List[Any]: """simple docstring""" super().setUp() SCREAMING_SNAKE_CASE__ = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """你""", """好""", """是""", """谁""", """a""", """b""", """c""", """d"""] SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = {} for i, value in enumerate(__A ): SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""word_shape_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""word_pronunciation_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) with open(self.word_shape_file , """w""" , encoding="""utf-8""" ) as word_shape_writer: json.dump(__A , __A , ensure_ascii=__A ) with open(self.word_pronunciation_file , """w""" , encoding="""utf-8""" ) as word_pronunciation_writer: json.dump(__A , __A , ensure_ascii=__A ) def _snake_case ( self :List[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) SCREAMING_SNAKE_CASE__ = tokenizer.tokenize("""你好[SEP]你是谁""" ) self.assertListEqual(__A , ["""你""", """好""", """[SEP]""", """你""", """是""", """谁"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_shape_ids(__A ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_pronunciation_ids(__A ) , [5, 6, 2, 5, 7, 8] ) def _snake_case ( self :List[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer() self.assertListEqual(tokenizer.tokenize("""ah\u535A\u63A8zz""" ) , ["""ah""", """\u535A""", """\u63A8""", """zz"""] ) def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hällo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""h\u00E9llo"""] ) def _snake_case ( self :Any ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :List[str] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :Union[str, Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :int ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HäLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :Union[str, Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HaLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :List[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , never_split=["""[UNK]"""] ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? [UNK]""" ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?""", """[UNK]"""] ) def _snake_case ( self :Any ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing"""] SCREAMING_SNAKE_CASE__ = {} for i, token in enumerate(__A ): SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = RoCBertWordpieceTokenizer(vocab=__A , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""unwanted running""" ) , ["""un""", """##want""", """##ed""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.tokenize("""unwantedX running""" ) , ["""[UNK]""", """runn""", """##ing"""] ) def _snake_case ( self :Any ) -> str: """simple docstring""" self.assertTrue(_is_whitespace(""" """ ) ) self.assertTrue(_is_whitespace("""\t""" ) ) self.assertTrue(_is_whitespace("""\r""" ) ) self.assertTrue(_is_whitespace("""\n""" ) ) self.assertTrue(_is_whitespace("""\u00A0""" ) ) self.assertFalse(_is_whitespace("""A""" ) ) self.assertFalse(_is_whitespace("""-""" ) ) def _snake_case ( self :int ) -> str: """simple docstring""" self.assertTrue(_is_control("""\u0005""" ) ) self.assertFalse(_is_control("""A""" ) ) self.assertFalse(_is_control(""" """ ) ) self.assertFalse(_is_control("""\t""" ) ) self.assertFalse(_is_control("""\r""" ) ) def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" self.assertTrue(_is_punctuation("""-""" ) ) self.assertTrue(_is_punctuation("""$""" ) ) self.assertTrue(_is_punctuation("""`""" ) ) self.assertTrue(_is_punctuation(""".""" ) ) self.assertFalse(_is_punctuation("""A""" ) ) self.assertFalse(_is_punctuation(""" """ ) ) def _snake_case ( self :str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(__A ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) if self.test_rust_tokenizer: SCREAMING_SNAKE_CASE__ = self.get_rust_tokenizer() self.assertListEqual( [rust_tokenizer.tokenize(__A ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) def _snake_case ( self :int ) -> Any: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = f'''A, naïve {tokenizer_r.mask_token} AllenNLP sentence.''' SCREAMING_SNAKE_CASE__ = tokenizer_r.encode_plus( __A , return_attention_mask=__A , return_token_type_ids=__A , return_offsets_mapping=__A , add_special_tokens=__A , ) SCREAMING_SNAKE_CASE__ = tokenizer_r.do_lower_case if hasattr(__A , """do_lower_case""" ) else False SCREAMING_SNAKE_CASE__ = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """A"""), ((1, 2), ""","""), ((3, 5), """na"""), ((5, 6), """##ï"""), ((6, 8), """##ve"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """Allen"""), ((21, 23), """##NL"""), ((23, 24), """##P"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """a"""), ((1, 2), ""","""), ((3, 8), """naive"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """allen"""), ((21, 23), """##nl"""), ((23, 24), """##p"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["""input_ids"""] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["""offset_mapping"""] ) def _snake_case ( self :Optional[int] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""的""", """人""", """有"""] SCREAMING_SNAKE_CASE__ = """""".join(__A ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.convert_ids_to_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.convert_ids_to_tokens(__A ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(__A , __A ) self.assertListEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.convert_ids_to_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.convert_ids_to_tokens(__A ) # it is expected that only the first Chinese character is not preceded by "##". SCREAMING_SNAKE_CASE__ = [ f'''##{token}''' if idx != 0 else token for idx, token in enumerate(__A ) ] self.assertListEqual(__A , __A ) self.assertListEqual(__A , __A ) @slow def _snake_case ( self :Union[str, Any] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) SCREAMING_SNAKE_CASE__ = tokenizer.encode("""你好""" , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.encode("""你是谁""" , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.build_inputs_with_special_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.build_inputs_with_special_tokens(__A , __A ) assert encoded_sentence == [1] + text + [2] assert encoded_pair == [1] + text + [2] + text_a + [2] def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_tokenizers(do_lower_case=__A ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE__ = """你好,你是谁""" SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_shape_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_pronunciation_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.prepare_for_model( __A , __A , __A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.encode_plus(__A , add_special_tokens=__A ) self.assertEqual(__A , __A )
59
1
_lowerCamelCase = [ 'Audio', 'Array2D', 'Array3D', 'Array4D', 'Array5D', 'ClassLabel', 'Features', 'Sequence', 'Value', 'Image', 'Translation', 'TranslationVariableLanguages', ] from .audio import Audio from .features import ArrayaD, ArrayaD, ArrayaD, ArrayaD, ClassLabel, Features, Sequence, Value from .image import Image from .translation import Translation, TranslationVariableLanguages
59
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[str, Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = original_name.split(""".""" )[0] SCREAMING_SNAKE_CASE__ = key.split(""".""" ) SCREAMING_SNAKE_CASE__ = int(key_list[key_list.index(UpperCamelCase__ ) - 2] ) SCREAMING_SNAKE_CASE__ = int(key_list[key_list.index(UpperCamelCase__ ) - 1] ) SCREAMING_SNAKE_CASE__ = orig_block_num - offset SCREAMING_SNAKE_CASE__ = key.replace(f'''{orig_block_num}.{layer_num}.{original_name}''' , f'''block.{new_block_num}.{layer_num}.{new_name}''' ) return key def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = OrderedDict() SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 0, 0 for key, value in state_dict.items(): if key.startswith("""network""" ): SCREAMING_SNAKE_CASE__ = key.replace("""network""" , """poolformer.encoder""" ) if "proj" in key: # Works for the first embedding as well as the internal embedding layers if key.endswith("""bias""" ) and "patch_embed" not in key: patch_emb_offset += 1 SCREAMING_SNAKE_CASE__ = key[: key.find("""proj""" )] SCREAMING_SNAKE_CASE__ = key.replace(UpperCamelCase__ , f'''patch_embeddings.{total_embed_found}.''' ) SCREAMING_SNAKE_CASE__ = key.replace("""proj""" , """projection""" ) if key.endswith("""bias""" ): total_embed_found += 1 if "patch_embeddings" in key: SCREAMING_SNAKE_CASE__ = """poolformer.encoder.""" + key if "mlp.fc1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """mlp.fc1""" , """output.conv1""" ) if "mlp.fc2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """mlp.fc2""" , """output.conv2""" ) if "norm1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """norm1""" , """before_norm""" ) if "norm2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """norm2""" , """after_norm""" ) if "layer_scale_1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """layer_scale_1""" , """layer_scale_1""" ) if "layer_scale_2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """layer_scale_2""" , """layer_scale_2""" ) if "head" in key: SCREAMING_SNAKE_CASE__ = key.replace("""head""" , """classifier""" ) SCREAMING_SNAKE_CASE__ = value return new_state_dict def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = """http://images.cocodataset.org/val2017/000000039769.jpg""" SCREAMING_SNAKE_CASE__ = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return image @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = PoolFormerConfig() # set attributes based on model_name SCREAMING_SNAKE_CASE__ = """huggingface/label-files""" SCREAMING_SNAKE_CASE__ = model_name[-3:] SCREAMING_SNAKE_CASE__ = 1_000 SCREAMING_SNAKE_CASE__ = """imagenet-1k-id2label.json""" SCREAMING_SNAKE_CASE__ = (1, 1_000) # set config attributes SCREAMING_SNAKE_CASE__ = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) SCREAMING_SNAKE_CASE__ = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE__ = idalabel SCREAMING_SNAKE_CASE__ = {v: k for k, v in idalabel.items()} if size == "s12": SCREAMING_SNAKE_CASE__ = [2, 2, 6, 2] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "s24": SCREAMING_SNAKE_CASE__ = [4, 4, 12, 4] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "s36": SCREAMING_SNAKE_CASE__ = [6, 6, 18, 6] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "m36": SCREAMING_SNAKE_CASE__ = [6, 6, 18, 6] SCREAMING_SNAKE_CASE__ = [96, 192, 384, 768] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9_5 elif size == "m48": SCREAMING_SNAKE_CASE__ = [8, 8, 24, 8] SCREAMING_SNAKE_CASE__ = [96, 192, 384, 768] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9_5 else: raise ValueError(f'''Size {size} not supported''' ) # load image processor SCREAMING_SNAKE_CASE__ = PoolFormerImageProcessor(crop_pct=UpperCamelCase__ ) # Prepare image SCREAMING_SNAKE_CASE__ = prepare_img() SCREAMING_SNAKE_CASE__ = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ).pixel_values logger.info(f'''Converting model {model_name}...''' ) # load original state dict SCREAMING_SNAKE_CASE__ = torch.load(UpperCamelCase__ , map_location=torch.device("""cpu""" ) ) # rename keys SCREAMING_SNAKE_CASE__ = rename_keys(UpperCamelCase__ ) # create HuggingFace model and load state dict SCREAMING_SNAKE_CASE__ = PoolFormerForImageClassification(UpperCamelCase__ ) model.load_state_dict(UpperCamelCase__ ) model.eval() # Define image processor SCREAMING_SNAKE_CASE__ = PoolFormerImageProcessor(crop_pct=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = image_processor(images=prepare_img() , return_tensors="""pt""" ).pixel_values # forward pass SCREAMING_SNAKE_CASE__ = model(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = outputs.logits # define expected logit slices for different models if size == "s12": SCREAMING_SNAKE_CASE__ = torch.tensor([-0.3_0_4_5, -0.6_7_5_8, -0.4_8_6_9] ) elif size == "s24": SCREAMING_SNAKE_CASE__ = torch.tensor([0.4_4_0_2, -0.1_3_7_4, -0.8_0_4_5] ) elif size == "s36": SCREAMING_SNAKE_CASE__ = torch.tensor([-0.6_0_8_0, -0.5_1_3_3, -0.5_8_9_8] ) elif size == "m36": SCREAMING_SNAKE_CASE__ = torch.tensor([0.3_9_5_2, 0.2_2_6_3, -1.2_6_6_8] ) elif size == "m48": SCREAMING_SNAKE_CASE__ = torch.tensor([0.1_1_6_7, -0.0_6_5_6, -0.3_4_2_3] ) else: raise ValueError(f'''Size {size} not supported''' ) # verify logits assert logits.shape == expected_shape assert torch.allclose(logits[0, :3] , UpperCamelCase__ , atol=1e-2 ) # finally, save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) model.save_pretrained(UpperCamelCase__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '--model_name', default='poolformer_s12', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) _lowerCamelCase = parser.parse_args() convert_poolformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
59
1
import itertools import string from collections.abc import Generator, Iterable def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Iterable[str] , UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = iter(UpperCamelCase__ ) while True: SCREAMING_SNAKE_CASE__ = tuple(itertools.islice(UpperCamelCase__ , UpperCamelCase__ ) ) if not chunk: return yield chunk def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = """""".join([c.upper() for c in dirty if c in string.ascii_letters] ) SCREAMING_SNAKE_CASE__ = """""" if len(UpperCamelCase__ ) < 2: return dirty for i in range(len(UpperCamelCase__ ) - 1 ): clean += dirty[i] if dirty[i] == dirty[i + 1]: clean += "X" clean += dirty[-1] if len(UpperCamelCase__ ) & 1: clean += "X" return clean def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str ): # I and J are used interchangeably to allow # us to use a 5x5 table (25 letters) SCREAMING_SNAKE_CASE__ = """ABCDEFGHIKLMNOPQRSTUVWXYZ""" # we're using a list instead of a '2d' array because it makes the math # for setting up the table and doing the actual encoding/decoding simpler SCREAMING_SNAKE_CASE__ = [] # copy key chars into the table if they are in `alphabet` ignoring duplicates for char in key.upper(): if char not in table and char in alphabet: table.append(UpperCamelCase__ ) # fill the rest of the table in with the remaining alphabet chars for char in alphabet: if char not in table: table.append(UpperCamelCase__ ) return table def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str , UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = generate_table(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = prepare_input(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = """""" # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(UpperCamelCase__ , 2 ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = divmod(table.index(UpperCamelCase__ ) , 5 ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = divmod(table.index(UpperCamelCase__ ) , 5 ) if rowa == rowa: ciphertext += table[rowa * 5 + (cola + 1) % 5] ciphertext += table[rowa * 5 + (cola + 1) % 5] elif cola == cola: ciphertext += table[((rowa + 1) % 5) * 5 + cola] ciphertext += table[((rowa + 1) % 5) * 5 + cola] else: # rectangle ciphertext += table[rowa * 5 + cola] ciphertext += table[rowa * 5 + cola] return ciphertext def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str , UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = generate_table(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = """""" # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(UpperCamelCase__ , 2 ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = divmod(table.index(UpperCamelCase__ ) , 5 ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = divmod(table.index(UpperCamelCase__ ) , 5 ) if rowa == rowa: plaintext += table[rowa * 5 + (cola - 1) % 5] plaintext += table[rowa * 5 + (cola - 1) % 5] elif cola == cola: plaintext += table[((rowa - 1) % 5) * 5 + cola] plaintext += table[((rowa - 1) % 5) * 5 + cola] else: # rectangle plaintext += table[rowa * 5 + cola] plaintext += table[rowa * 5 + cola] return plaintext
59
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'google/pix2struct-textcaps-base': ( 'https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json' ), } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct_text_model" lowerCamelCase_ = ["past_key_values"] lowerCamelCase_ = { "hidden_size": "hidden_size", "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self :Union[str, Any] , __A :Any=5_0244 , __A :Optional[Any]=768 , __A :Tuple=64 , __A :List[str]=2048 , __A :int=12 , __A :str=12 , __A :Any=32 , __A :Tuple=128 , __A :int=0.1 , __A :str=1E-6 , __A :Optional[Any]=1.0 , __A :Union[str, Any]="gelu_new" , __A :Any=0 , __A :List[str]=False , __A :Optional[Any]=0 , __A :int=1 , __A :Optional[int]=False , __A :Optional[Any]=True , **__A :List[Any] , ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = d_kv SCREAMING_SNAKE_CASE__ = d_ff SCREAMING_SNAKE_CASE__ = num_layers SCREAMING_SNAKE_CASE__ = num_heads SCREAMING_SNAKE_CASE__ = relative_attention_num_buckets SCREAMING_SNAKE_CASE__ = relative_attention_max_distance SCREAMING_SNAKE_CASE__ = dropout_rate SCREAMING_SNAKE_CASE__ = layer_norm_epsilon SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = use_cache SCREAMING_SNAKE_CASE__ = eos_token_id SCREAMING_SNAKE_CASE__ = decoder_start_token_id # for backwards compatibility SCREAMING_SNAKE_CASE__ = dense_act_fn super().__init__( pad_token_id=__A , eos_token_id=__A , decoder_start_token_id=__A , tie_word_embeddings=__A , is_decoder=__A , **__A , ) @classmethod def _snake_case ( cls :Optional[int] , __A :Union[str, os.PathLike] , **__A :Optional[int] ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = cls.get_config_dict(__A , **__A ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get("""model_type""" ) == "pix2struct": SCREAMING_SNAKE_CASE__ = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct_vision_model" def __init__( self :Optional[int] , __A :int=768 , __A :Optional[Any]=768 , __A :Union[str, Any]=2048 , __A :int=64 , __A :Union[str, Any]=12 , __A :str=12 , __A :Any="gelu_new" , __A :List[Any]=1E-6 , __A :Dict=0.0 , __A :int=0.0 , __A :int=1E-10 , __A :Dict=1.0 , __A :int=4096 , __A :int=32 , __A :int=128 , **__A :Tuple , ) -> str: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = patch_embed_hidden_size SCREAMING_SNAKE_CASE__ = d_ff SCREAMING_SNAKE_CASE__ = dropout_rate SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = attention_dropout SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = dense_act_fn SCREAMING_SNAKE_CASE__ = seq_len SCREAMING_SNAKE_CASE__ = relative_attention_num_buckets SCREAMING_SNAKE_CASE__ = relative_attention_max_distance SCREAMING_SNAKE_CASE__ = d_kv @classmethod def _snake_case ( cls :str , __A :Union[str, os.PathLike] , **__A :str ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = cls.get_config_dict(__A , **__A ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get("""model_type""" ) == "pix2struct": SCREAMING_SNAKE_CASE__ = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct" lowerCamelCase_ = True def __init__( self :str , __A :Optional[Any]=None , __A :List[str]=None , __A :Optional[Any]=1.0 , __A :Optional[Any]=0.0_2 , __A :Any=False , __A :Tuple=False , __A :Any=True , **__A :Dict , ) -> Union[str, Any]: """simple docstring""" super().__init__(tie_word_embeddings=__A , is_encoder_decoder=__A , **__A ) if text_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""text_config is None. Initializing the Pix2StructTextConfig with default values.""" ) if vision_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""vision_config is None. Initializing the Pix2StructVisionConfig with default values.""" ) SCREAMING_SNAKE_CASE__ = PixaStructTextConfig(**__A ) SCREAMING_SNAKE_CASE__ = PixaStructVisionConfig(**__A ) SCREAMING_SNAKE_CASE__ = self.text_config.decoder_start_token_id SCREAMING_SNAKE_CASE__ = self.text_config.pad_token_id SCREAMING_SNAKE_CASE__ = self.text_config.eos_token_id SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = self.initializer_range SCREAMING_SNAKE_CASE__ = self.initializer_range SCREAMING_SNAKE_CASE__ = is_vqa @classmethod def _snake_case ( cls :Union[str, Any] , __A :PixaStructTextConfig , __A :PixaStructVisionConfig , **__A :Optional[int] ) -> Optional[Any]: """simple docstring""" return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__A ) def _snake_case ( self :str ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = copy.deepcopy(self.__dict__ ) SCREAMING_SNAKE_CASE__ = self.text_config.to_dict() SCREAMING_SNAKE_CASE__ = self.vision_config.to_dict() SCREAMING_SNAKE_CASE__ = self.__class__.model_type return output
59
1
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = ["image_processor", "tokenizer"] lowerCamelCase_ = "CLIPImageProcessor" lowerCamelCase_ = ("XLMRobertaTokenizer", "XLMRobertaTokenizerFast") def __init__( self :Optional[int] , __A :Dict=None , __A :str=None , **__A :str ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , __A , ) SCREAMING_SNAKE_CASE__ = kwargs.pop("""feature_extractor""" ) SCREAMING_SNAKE_CASE__ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(__A , __A ) def __call__( self :Any , __A :Dict=None , __A :List[str]=None , __A :int=None , **__A :int ) -> Any: """simple docstring""" if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: SCREAMING_SNAKE_CASE__ = self.tokenizer(__A , return_tensors=__A , **__A ) if images is not None: SCREAMING_SNAKE_CASE__ = self.image_processor(__A , return_tensors=__A , **__A ) if text is not None and images is not None: SCREAMING_SNAKE_CASE__ = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__A ) , tensor_type=__A ) def _snake_case ( self :Union[str, Any] , *__A :List[Any] , **__A :Optional[int] ) -> Optional[int]: """simple docstring""" return self.tokenizer.batch_decode(*__A , **__A ) def _snake_case ( self :Union[str, Any] , *__A :Dict , **__A :Tuple ) -> Optional[int]: """simple docstring""" return self.tokenizer.decode(*__A , **__A ) @property def _snake_case ( self :int ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer.model_input_names SCREAMING_SNAKE_CASE__ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
59
import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = inspect.getfile(accelerate.test_utils ) SCREAMING_SNAKE_CASE__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_script.py"""] ) SCREAMING_SNAKE_CASE__ = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_distributed_data_loop.py"""] ) SCREAMING_SNAKE_CASE__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_ops.py"""] ) @require_multi_gpu def _snake_case ( self :Optional[Any] ) -> Tuple: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Tuple ) -> Optional[Any]: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.operation_file_path] print(f'''Command: {cmd}''' ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Optional[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Optional[int] ) -> str: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices, using 2 devices only''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices="""0,1""" ): execute_subprocess_async(__A , env=os.environ.copy() ) if __name__ == "__main__": _lowerCamelCase = Accelerator() _lowerCamelCase = (accelerator.state.process_index + 2, 10) _lowerCamelCase = torch.randint(0, 10, shape).to(accelerator.device) _lowerCamelCase = '' _lowerCamelCase = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." _lowerCamelCase = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." _lowerCamelCase = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
59
1
from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class UpperCamelCase_ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): lowerCamelCase_ = [R"h\.\d+\.attn\.bias", R"h\.\d+\.attn\.masked_bias"] @register_to_config def __init__( self :Any , __A :int , __A :int , __A :Optional[int] = None , __A :int = 5_0257 , __A :int = 1024 , __A :int = 768 , __A :int = 12 , __A :int = 12 , __A :Optional[int] = None , __A :str = "gelu_new" , __A :float = 0.1 , __A :float = 0.1 , __A :float = 0.1 , __A :float = 1E-5 , __A :float = 0.0_2 , __A :bool = True , __A :bool = True , __A :bool = False , __A :bool = False , ) -> List[Any]: """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE__ = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( f'''`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and''' f''' `n_embd`: {n_embd} are not equal.''' ) SCREAMING_SNAKE_CASE__ = prefix_inner_dim SCREAMING_SNAKE_CASE__ = prefix_hidden_dim SCREAMING_SNAKE_CASE__ = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) SCREAMING_SNAKE_CASE__ = ( nn.Linear(self.prefix_hidden_dim , __A ) if self.prefix_hidden_dim is not None else nn.Identity() ) SCREAMING_SNAKE_CASE__ = GPTaConfig( vocab_size=__A , n_positions=__A , n_embd=__A , n_layer=__A , n_head=__A , n_inner=__A , activation_function=__A , resid_pdrop=__A , embd_pdrop=__A , attn_pdrop=__A , layer_norm_epsilon=__A , initializer_range=__A , scale_attn_weights=__A , use_cache=__A , scale_attn_by_inverse_layer_idx=__A , reorder_and_upcast_attn=__A , ) SCREAMING_SNAKE_CASE__ = GPTaLMHeadModel(__A ) def _snake_case ( self :Optional[Any] , __A :torch.Tensor , __A :torch.Tensor , __A :Optional[torch.Tensor] = None , __A :Optional[torch.Tensor] = None , ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.transformer.transformer.wte(__A ) SCREAMING_SNAKE_CASE__ = self.encode_prefix(__A ) SCREAMING_SNAKE_CASE__ = self.decode_prefix(__A ) SCREAMING_SNAKE_CASE__ = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: SCREAMING_SNAKE_CASE__ = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) SCREAMING_SNAKE_CASE__ = torch.cat((dummy_token, input_ids) , dim=1 ) SCREAMING_SNAKE_CASE__ = self.transformer(inputs_embeds=__A , labels=__A , attention_mask=__A ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def _snake_case ( self :Optional[int] , __A :int , __A :torch.device ) -> torch.Tensor: """simple docstring""" return torch.zeros(__A , self.prefix_length , dtype=torch.intaa , device=__A ) def _snake_case ( self :Optional[Any] , __A :Dict ) -> Optional[int]: """simple docstring""" return self.encode_prefix(__A ) @torch.no_grad() def _snake_case ( self :str , __A :List[str] , __A :List[Any] , __A :List[Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = torch.split(__A , 1 , dim=0 ) SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] for feature in features: SCREAMING_SNAKE_CASE__ = self.decode_prefix(feature.to(__A ) ) # back to the clip feature # Only support beam search for now SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.generate_beam( input_embeds=__A , device=__A , eos_token_id=__A ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) SCREAMING_SNAKE_CASE__ = torch.stack(__A ) SCREAMING_SNAKE_CASE__ = torch.stack(__A ) return generated_tokens, generated_seq_lengths @torch.no_grad() def _snake_case ( self :List[str] , __A :Union[str, Any]=None , __A :Optional[int]=None , __A :List[Any]=None , __A :int = 5 , __A :int = 67 , __A :float = 1.0 , __A :Optional[int] = None , ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = eos_token_id SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = torch.ones(__A , device=__A , dtype=torch.int ) SCREAMING_SNAKE_CASE__ = torch.zeros(__A , device=__A , dtype=torch.bool ) if input_embeds is not None: SCREAMING_SNAKE_CASE__ = input_embeds else: SCREAMING_SNAKE_CASE__ = self.transformer.transformer.wte(__A ) for i in range(__A ): SCREAMING_SNAKE_CASE__ = self.transformer(inputs_embeds=__A ) SCREAMING_SNAKE_CASE__ = outputs.logits SCREAMING_SNAKE_CASE__ = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) SCREAMING_SNAKE_CASE__ = logits.softmax(-1 ).log() if scores is None: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = logits.topk(__A , -1 ) SCREAMING_SNAKE_CASE__ = generated.expand(__A , *generated.shape[1:] ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: SCREAMING_SNAKE_CASE__ = next_tokens else: SCREAMING_SNAKE_CASE__ = tokens.expand(__A , *tokens.shape[1:] ) SCREAMING_SNAKE_CASE__ = torch.cat((tokens, next_tokens) , dim=1 ) else: SCREAMING_SNAKE_CASE__ = -float(np.inf ) SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = scores[:, None] + logits seq_lengths[~is_stopped] += 1 SCREAMING_SNAKE_CASE__ = scores_sum / seq_lengths[:, None] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = scores_sum_average.view(-1 ).topk(__A , -1 ) SCREAMING_SNAKE_CASE__ = next_tokens // scores_sum.shape[1] SCREAMING_SNAKE_CASE__ = seq_lengths[next_tokens_source] SCREAMING_SNAKE_CASE__ = next_tokens % scores_sum.shape[1] SCREAMING_SNAKE_CASE__ = next_tokens.unsqueeze(1 ) SCREAMING_SNAKE_CASE__ = tokens[next_tokens_source] SCREAMING_SNAKE_CASE__ = torch.cat((tokens, next_tokens) , dim=1 ) SCREAMING_SNAKE_CASE__ = generated[next_tokens_source] SCREAMING_SNAKE_CASE__ = scores_sum_average * seq_lengths SCREAMING_SNAKE_CASE__ = is_stopped[next_tokens_source] SCREAMING_SNAKE_CASE__ = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) SCREAMING_SNAKE_CASE__ = torch.cat((generated, next_token_embed) , dim=1 ) SCREAMING_SNAKE_CASE__ = is_stopped + next_tokens.eq(__A ).squeeze() if is_stopped.all(): break SCREAMING_SNAKE_CASE__ = scores / seq_lengths SCREAMING_SNAKE_CASE__ = scores.argsort(descending=__A ) # tokens tensors are already padded to max_seq_length SCREAMING_SNAKE_CASE__ = [tokens[i] for i in order] SCREAMING_SNAKE_CASE__ = torch.stack(__A , dim=0 ) SCREAMING_SNAKE_CASE__ = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
59
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _lowerCamelCase = { 'configuration_layoutlmv3': [ 'LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv3Config', 'LayoutLMv3OnnxConfig', ], 'processing_layoutlmv3': ['LayoutLMv3Processor'], 'tokenization_layoutlmv3': ['LayoutLMv3Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['LayoutLMv3TokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv3ForQuestionAnswering', 'LayoutLMv3ForSequenceClassification', 'LayoutLMv3ForTokenClassification', 'LayoutLMv3Model', 'LayoutLMv3PreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLayoutLMv3ForQuestionAnswering', 'TFLayoutLMv3ForSequenceClassification', 'TFLayoutLMv3ForTokenClassification', 'TFLayoutLMv3Model', 'TFLayoutLMv3PreTrainedModel', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['LayoutLMv3FeatureExtractor'] _lowerCamelCase = ['LayoutLMv3ImageProcessor'] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
1
from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str , UpperCamelCase__: float | Decimal , UpperCamelCase__: float = 10**-10 ): SCREAMING_SNAKE_CASE__ = a while True: SCREAMING_SNAKE_CASE__ = Decimal(UpperCamelCase__ ) - ( Decimal(eval(UpperCamelCase__ ) ) / Decimal(eval(str(diff(UpperCamelCase__ ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(UpperCamelCase__ ) ) < precision: # noqa: S307 return float(UpperCamelCase__ ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F'''The root of sin(x) = 0 is {newton_raphson('sin(x)', 2)}''') # Find root of polynomial print(F'''The root of x**2 - 5*x + 2 = 0 is {newton_raphson('x**2 - 5*x + 2', 0.4)}''') # Find Square Root of 5 print(F'''The root of log(x) - 1 = 0 is {newton_raphson('log(x) - 1', 2)}''') # Exponential Roots print(F'''The root of exp(x) - 1 = 0 is {newton_raphson('exp(x) - 1', 0)}''')
59
import inspect import unittest class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :str ) -> Union[str, Any]: """simple docstring""" try: import diffusers # noqa: F401 except ImportError: assert False def _snake_case ( self :Any ) -> Any: """simple docstring""" import diffusers from diffusers.dependency_versions_table import deps SCREAMING_SNAKE_CASE__ = inspect.getmembers(__A , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": SCREAMING_SNAKE_CASE__ = """k-diffusion""" elif backend == "invisible_watermark": SCREAMING_SNAKE_CASE__ = """invisible-watermark""" assert backend in deps, f'''{backend} is not in the deps table!'''
59
1
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int = 100 ): SCREAMING_SNAKE_CASE__ = (n * (n + 1) // 2) ** 2 SCREAMING_SNAKE_CASE__ = n * (n + 1) * (2 * n + 1) // 6 return sum_cubes - sum_squares if __name__ == "__main__": print(F'''{solution() = }''')
59
import warnings from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch from ...models import UNetaDModel from ...schedulers import RePaintScheduler from ...utils import PIL_INTERPOLATION, logging, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput _lowerCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[List, PIL.Image.Image, torch.Tensor] ): warnings.warn( """The preprocess method is deprecated and will be removed in a future version. Please""" """ use VaeImageProcessor.preprocess instead""" , UpperCamelCase__ , ) if isinstance(UpperCamelCase__ , torch.Tensor ): return image elif isinstance(UpperCamelCase__ , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ = [image] if isinstance(image[0] , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = image[0].size SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = (x - x % 8 for x in (w, h)) # resize to integer multiple of 8 SCREAMING_SNAKE_CASE__ = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION["""lanczos"""] ) )[None, :] for i in image] SCREAMING_SNAKE_CASE__ = np.concatenate(UpperCamelCase__ , axis=0 ) SCREAMING_SNAKE_CASE__ = np.array(UpperCamelCase__ ).astype(np.floataa ) / 2_5_5.0 SCREAMING_SNAKE_CASE__ = image.transpose(0 , 3 , 1 , 2 ) SCREAMING_SNAKE_CASE__ = 2.0 * image - 1.0 SCREAMING_SNAKE_CASE__ = torch.from_numpy(UpperCamelCase__ ) elif isinstance(image[0] , torch.Tensor ): SCREAMING_SNAKE_CASE__ = torch.cat(UpperCamelCase__ , dim=0 ) return image def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[List, PIL.Image.Image, torch.Tensor] ): if isinstance(UpperCamelCase__ , torch.Tensor ): return mask elif isinstance(UpperCamelCase__ , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ = [mask] if isinstance(mask[0] , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = mask[0].size SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 SCREAMING_SNAKE_CASE__ = [np.array(m.convert("""L""" ).resize((w, h) , resample=PIL_INTERPOLATION["""nearest"""] ) )[None, :] for m in mask] SCREAMING_SNAKE_CASE__ = np.concatenate(UpperCamelCase__ , axis=0 ) SCREAMING_SNAKE_CASE__ = mask.astype(np.floataa ) / 2_5_5.0 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = torch.from_numpy(UpperCamelCase__ ) elif isinstance(mask[0] , torch.Tensor ): SCREAMING_SNAKE_CASE__ = torch.cat(UpperCamelCase__ , dim=0 ) return mask class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = 42 lowerCamelCase_ = 42 def __init__( self :Any , __A :List[Any] , __A :Optional[Any] ) -> Union[str, Any]: """simple docstring""" super().__init__() self.register_modules(unet=__A , scheduler=__A ) @torch.no_grad() def __call__( self :str , __A :Union[torch.Tensor, PIL.Image.Image] , __A :Union[torch.Tensor, PIL.Image.Image] , __A :int = 250 , __A :float = 0.0 , __A :int = 10 , __A :int = 10 , __A :Optional[Union[torch.Generator, List[torch.Generator]]] = None , __A :Optional[str] = "pil" , __A :bool = True , ) -> Union[ImagePipelineOutput, Tuple]: """simple docstring""" SCREAMING_SNAKE_CASE__ = image SCREAMING_SNAKE_CASE__ = _preprocess_image(__A ) SCREAMING_SNAKE_CASE__ = original_image.to(device=self.device , dtype=self.unet.dtype ) SCREAMING_SNAKE_CASE__ = _preprocess_mask(__A ) SCREAMING_SNAKE_CASE__ = mask_image.to(device=self.device , dtype=self.unet.dtype ) SCREAMING_SNAKE_CASE__ = original_image.shape[0] # sample gaussian noise to begin the loop if isinstance(__A , __A ) and len(__A ) != batch_size: raise ValueError( f'''You have passed a list of generators of length {len(__A )}, but requested an effective batch''' f''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) SCREAMING_SNAKE_CASE__ = original_image.shape SCREAMING_SNAKE_CASE__ = randn_tensor(__A , generator=__A , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(__A , __A , __A , self.device ) SCREAMING_SNAKE_CASE__ = eta SCREAMING_SNAKE_CASE__ = self.scheduler.timesteps[0] + 1 SCREAMING_SNAKE_CASE__ = generator[0] if isinstance(__A , __A ) else generator for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): if t < t_last: # predict the noise residual SCREAMING_SNAKE_CASE__ = self.unet(__A , __A ).sample # compute previous image: x_t -> x_t-1 SCREAMING_SNAKE_CASE__ = self.scheduler.step(__A , __A , __A , __A , __A , __A ).prev_sample else: # compute the reverse: x_t-1 -> x_t SCREAMING_SNAKE_CASE__ = self.scheduler.undo_step(__A , __A , __A ) SCREAMING_SNAKE_CASE__ = t SCREAMING_SNAKE_CASE__ = (image / 2 + 0.5).clamp(0 , 1 ) SCREAMING_SNAKE_CASE__ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": SCREAMING_SNAKE_CASE__ = self.numpy_to_pil(__A ) if not return_dict: return (image,) return ImagePipelineOutput(images=__A )
59
1
import pyarrow.parquet as pq import pytest from datasets import Audio, Dataset, DatasetDict, Features, NamedSplit, Sequence, Value, config from datasets.features.image import Image from datasets.io.parquet import ParquetDatasetReader, ParquetDatasetWriter, get_writer_batch_size from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Tuple , UpperCamelCase__: Optional[Any] ): assert isinstance(UpperCamelCase__ , UpperCamelCase__ ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("""keep_in_memory""" , [False, True] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Tuple ): SCREAMING_SNAKE_CASE__ = tmp_path / """cache""" SCREAMING_SNAKE_CASE__ = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): SCREAMING_SNAKE_CASE__ = ParquetDatasetReader(UpperCamelCase__ , cache_dir=UpperCamelCase__ , keep_in_memory=UpperCamelCase__ ).read() _check_parquet_dataset(UpperCamelCase__ , UpperCamelCase__ ) @pytest.mark.parametrize( """features""" , [ None, {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""}, {"""col_1""": """string""", """col_2""": """string""", """col_3""": """string"""}, {"""col_1""": """int32""", """col_2""": """int32""", """col_3""": """int32"""}, {"""col_1""": """float32""", """col_2""": """float32""", """col_3""": """float32"""}, ] , ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Tuple , UpperCamelCase__: int , UpperCamelCase__: Tuple ): SCREAMING_SNAKE_CASE__ = tmp_path / """cache""" SCREAMING_SNAKE_CASE__ = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} SCREAMING_SNAKE_CASE__ = features.copy() if features else default_expected_features SCREAMING_SNAKE_CASE__ = ( Features({feature: Value(UpperCamelCase__ ) for feature, dtype in features.items()} ) if features is not None else None ) SCREAMING_SNAKE_CASE__ = ParquetDatasetReader(UpperCamelCase__ , features=UpperCamelCase__ , cache_dir=UpperCamelCase__ ).read() _check_parquet_dataset(UpperCamelCase__ , UpperCamelCase__ ) @pytest.mark.parametrize("""split""" , [None, NamedSplit("""train""" ), """train""", """test"""] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = tmp_path / """cache""" SCREAMING_SNAKE_CASE__ = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} SCREAMING_SNAKE_CASE__ = ParquetDatasetReader(UpperCamelCase__ , cache_dir=UpperCamelCase__ , split=UpperCamelCase__ ).read() _check_parquet_dataset(UpperCamelCase__ , UpperCamelCase__ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize("""path_type""" , [str, list] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any , UpperCamelCase__: Tuple , UpperCamelCase__: Union[str, Any] ): if issubclass(UpperCamelCase__ , UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = parquet_path elif issubclass(UpperCamelCase__ , UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = [parquet_path] SCREAMING_SNAKE_CASE__ = tmp_path / """cache""" SCREAMING_SNAKE_CASE__ = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} SCREAMING_SNAKE_CASE__ = ParquetDatasetReader(UpperCamelCase__ , cache_dir=UpperCamelCase__ ).read() _check_parquet_dataset(UpperCamelCase__ , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[Any] , UpperCamelCase__: Dict , UpperCamelCase__: Any=("train",) ): assert isinstance(UpperCamelCase__ , UpperCamelCase__ ) for split in splits: SCREAMING_SNAKE_CASE__ = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("""keep_in_memory""" , [False, True] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[Any] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Dict ): SCREAMING_SNAKE_CASE__ = tmp_path / """cache""" SCREAMING_SNAKE_CASE__ = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): SCREAMING_SNAKE_CASE__ = ParquetDatasetReader( {"""train""": parquet_path} , cache_dir=UpperCamelCase__ , keep_in_memory=UpperCamelCase__ ).read() _check_parquet_datasetdict(UpperCamelCase__ , UpperCamelCase__ ) @pytest.mark.parametrize( """features""" , [ None, {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""}, {"""col_1""": """string""", """col_2""": """string""", """col_3""": """string"""}, {"""col_1""": """int32""", """col_2""": """int32""", """col_3""": """int32"""}, {"""col_1""": """float32""", """col_2""": """float32""", """col_3""": """float32"""}, ] , ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int , UpperCamelCase__: List[str] , UpperCamelCase__: Tuple ): SCREAMING_SNAKE_CASE__ = tmp_path / """cache""" SCREAMING_SNAKE_CASE__ = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} SCREAMING_SNAKE_CASE__ = features.copy() if features else default_expected_features SCREAMING_SNAKE_CASE__ = ( Features({feature: Value(UpperCamelCase__ ) for feature, dtype in features.items()} ) if features is not None else None ) SCREAMING_SNAKE_CASE__ = ParquetDatasetReader({"""train""": parquet_path} , features=UpperCamelCase__ , cache_dir=UpperCamelCase__ ).read() _check_parquet_datasetdict(UpperCamelCase__ , UpperCamelCase__ ) @pytest.mark.parametrize("""split""" , [None, NamedSplit("""train""" ), """train""", """test"""] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: str ): if split: SCREAMING_SNAKE_CASE__ = {split: parquet_path} else: SCREAMING_SNAKE_CASE__ = """train""" SCREAMING_SNAKE_CASE__ = {"""train""": parquet_path, """test""": parquet_path} SCREAMING_SNAKE_CASE__ = tmp_path / """cache""" SCREAMING_SNAKE_CASE__ = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} SCREAMING_SNAKE_CASE__ = ParquetDatasetReader(UpperCamelCase__ , cache_dir=UpperCamelCase__ ).read() _check_parquet_datasetdict(UpperCamelCase__ , UpperCamelCase__ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str , UpperCamelCase__: Optional[int] ): SCREAMING_SNAKE_CASE__ = ParquetDatasetWriter(UpperCamelCase__ , tmp_path / """foo.parquet""" ) assert writer.write() > 0 SCREAMING_SNAKE_CASE__ = pq.ParquetFile(tmp_path / """foo.parquet""" ) SCREAMING_SNAKE_CASE__ = pf.read() assert dataset.data.table == output_table def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str , UpperCamelCase__: List[Any] ): SCREAMING_SNAKE_CASE__ = str(shared_datadir / """test_image_rgb.jpg""" ) SCREAMING_SNAKE_CASE__ = {"""image""": [image_path]} SCREAMING_SNAKE_CASE__ = Features({"""image""": Image()} ) SCREAMING_SNAKE_CASE__ = Dataset.from_dict(UpperCamelCase__ , features=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = ParquetDatasetWriter(UpperCamelCase__ , tmp_path / """foo.parquet""" ) assert writer.write() > 0 SCREAMING_SNAKE_CASE__ = Dataset.from_parquet(str(tmp_path / """foo.parquet""" ) ) assert dataset.features == reloaded_dataset.features SCREAMING_SNAKE_CASE__ = ParquetDatasetReader(str(tmp_path / """foo.parquet""" ) , streaming=UpperCamelCase__ ).read() assert dataset.features == reloaded_iterable_dataset.features @pytest.mark.parametrize( """feature, expected""" , [ (Features({"""foo""": Value("""int32""" )} ), None), (Features({"""image""": Image(), """foo""": Value("""int32""" )} ), config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS), (Features({"""nested""": Sequence(Audio() )} ), config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS), ] , ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] , UpperCamelCase__: Any ): assert get_writer_batch_size(UpperCamelCase__ ) == expected
59
import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = OpenAIGPTTokenizer lowerCamelCase_ = OpenAIGPTTokenizerFast lowerCamelCase_ = True lowerCamelCase_ = False def _snake_case ( self :Optional[Any] ) -> Dict: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt SCREAMING_SNAKE_CASE__ = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """w</w>""", """r</w>""", """t</w>""", """lo""", """low""", """er</w>""", """low</w>""", """lowest</w>""", """newer</w>""", """wider</w>""", """<unk>""", ] SCREAMING_SNAKE_CASE__ = dict(zip(__A , range(len(__A ) ) ) ) SCREAMING_SNAKE_CASE__ = ["""#version: 0.2""", """l o""", """lo w""", """e r</w>""", """"""] SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" ) as fp: fp.write(json.dumps(__A ) ) with open(self.merges_file , """w""" ) as fp: fp.write("""\n""".join(__A ) ) def _snake_case ( self :Union[str, Any] , __A :str ) -> List[Any]: """simple docstring""" return "lower newer", "lower newer" def _snake_case ( self :Optional[Any] ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = OpenAIGPTTokenizer(self.vocab_file , self.merges_file ) SCREAMING_SNAKE_CASE__ = """lower""" SCREAMING_SNAKE_CASE__ = ["""low""", """er</w>"""] SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(__A ) self.assertListEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = tokens + ["""<unk>"""] SCREAMING_SNAKE_CASE__ = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , __A ) def _snake_case ( self :Optional[Any] , __A :Optional[Any]=15 ) -> Any: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) # Simple input SCREAMING_SNAKE_CASE__ = """This is a simple input""" SCREAMING_SNAKE_CASE__ = ["""This is a simple input 1""", """This is a simple input 2"""] SCREAMING_SNAKE_CASE__ = ("""This is a simple input""", """This is a pair""") SCREAMING_SNAKE_CASE__ = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests self.assertRaises(__A , tokenizer_r.encode , __A , max_length=__A , padding="""max_length""" ) # Simple input self.assertRaises(__A , tokenizer_r.encode_plus , __A , max_length=__A , padding="""max_length""" ) # Simple input self.assertRaises( __A , tokenizer_r.batch_encode_plus , __A , max_length=__A , padding="""max_length""" , ) # Pair input self.assertRaises(__A , tokenizer_r.encode , __A , max_length=__A , padding="""max_length""" ) # Pair input self.assertRaises(__A , tokenizer_r.encode_plus , __A , max_length=__A , padding="""max_length""" ) # Pair input self.assertRaises( __A , tokenizer_r.batch_encode_plus , __A , max_length=__A , padding="""max_length""" , ) def _snake_case ( self :Dict ) -> List[Any]: """simple docstring""" pass @require_ftfy @require_spacy @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ ): pass
59
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _lowerCamelCase = '▁' _lowerCamelCase = {'vocab_file': 'spiece.model'} _lowerCamelCase = { 'vocab_file': {'google/pegasus-xsum': 'https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'} } _lowerCamelCase = { 'google/pegasus-xsum': 512, } _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = VOCAB_FILES_NAMES lowerCamelCase_ = VOCAB_FILES_NAMES lowerCamelCase_ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase_ = ["input_ids", "attention_mask"] def __init__( self :int , __A :List[Any] , __A :str="<pad>" , __A :List[Any]="</s>" , __A :Optional[int]="<unk>" , __A :int="<mask_2>" , __A :Dict="<mask_1>" , __A :Tuple=None , __A :Optional[int]=103 , __A :Optional[Dict[str, Any]] = None , **__A :List[str] , ) -> None: """simple docstring""" SCREAMING_SNAKE_CASE__ = offset if additional_special_tokens is not None: if not isinstance(__A , __A ): raise TypeError( f'''additional_special_tokens should be of type {type(__A )}, but is''' f''' {type(__A )}''' ) SCREAMING_SNAKE_CASE__ = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'''<unk_{i}>''' for i in range(len(__A ) , self.offset - 1 ) ] if len(set(__A ) ) != len(__A ): raise ValueError( """Please make sure that the provided additional_special_tokens do not contain an incorrectly""" f''' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.''' ) SCREAMING_SNAKE_CASE__ = additional_special_tokens_extended else: SCREAMING_SNAKE_CASE__ = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'''<unk_{i}>''' for i in range(2 , self.offset )] SCREAMING_SNAKE_CASE__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=__A , unk_token=__A , mask_token=__A , pad_token=__A , mask_token_sent=__A , offset=__A , additional_special_tokens=__A , sp_model_kwargs=self.sp_model_kwargs , **__A , ) SCREAMING_SNAKE_CASE__ = mask_token_sent SCREAMING_SNAKE_CASE__ = vocab_file SCREAMING_SNAKE_CASE__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__A ) # add special tokens to encoder dict SCREAMING_SNAKE_CASE__ = { 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 , self.offset - 1 )} ) SCREAMING_SNAKE_CASE__ = {v: k for k, v in self.encoder.items()} @property def _snake_case ( self :Tuple ) -> int: """simple docstring""" return len(self.sp_model ) + self.offset def _snake_case ( self :str ) -> Dict[str, int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = {self.convert_ids_to_tokens(__A ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self :Optional[int] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.__dict__.copy() SCREAMING_SNAKE_CASE__ = None return state def __setstate__( self :int , __A :int ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _snake_case ( self :Union[str, Any] , __A :str ) -> List[str]: """simple docstring""" return self.sp_model.encode(__A , out_type=__A ) def _snake_case ( self :str , __A :str ) -> int: """simple docstring""" if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] SCREAMING_SNAKE_CASE__ = self.sp_model.piece_to_id(__A ) return sp_id + self.offset def _snake_case ( self :Tuple , __A :int ) -> str: """simple docstring""" if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: SCREAMING_SNAKE_CASE__ = self.sp_model.IdToPiece(index - self.offset ) return token def _snake_case ( self :Optional[int] , __A :List[Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = """""" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__A ) + token SCREAMING_SNAKE_CASE__ = [] else: current_sub_tokens.append(__A ) out_string += self.sp_model.decode(__A ) return out_string.strip() def _snake_case ( self :List[str] , __A :Union[str, Any]=False ) -> Any: """simple docstring""" return 1 def _snake_case ( self :Dict , __A :int ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def _snake_case ( self :Optional[int] , __A :List , __A :Optional[List] = None , __A :bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return self._special_token_mask(__A ) elif token_ids_a is None: return self._special_token_mask(__A ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def _snake_case ( self :List[str] , __A :Union[str, Any] , __A :List[Any]=None ) -> List[int]: """simple docstring""" if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _snake_case ( self :str , __A :str , __A :Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__A ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return SCREAMING_SNAKE_CASE__ = os.path.join( __A , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__A ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __A ) elif not os.path.isfile(self.vocab_file ): with open(__A , """wb""" ) as fi: SCREAMING_SNAKE_CASE__ = self.sp_model.serialized_model_proto() fi.write(__A ) return (out_vocab_file,)
59
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=UpperCamelCase__ ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = field(default="image-classification" , metadata={"include_in_asdict_even_if_is_default": True} ) lowerCamelCase_ = Features({"image": Image()} ) lowerCamelCase_ = Features({"labels": ClassLabel} ) lowerCamelCase_ = "image" lowerCamelCase_ = "labels" def _snake_case ( self :List[str] , __A :Tuple ) -> Tuple: """simple docstring""" if self.label_column not in features: raise ValueError(f'''Column {self.label_column} is not present in features.''' ) if not isinstance(features[self.label_column] , __A ): raise ValueError(f'''Column {self.label_column} is not a ClassLabel.''' ) SCREAMING_SNAKE_CASE__ = copy.deepcopy(self ) SCREAMING_SNAKE_CASE__ = self.label_schema.copy() SCREAMING_SNAKE_CASE__ = features[self.label_column] SCREAMING_SNAKE_CASE__ = label_schema return task_template @property def _snake_case ( self :Dict ) -> Dict[str, str]: """simple docstring""" return { self.image_column: "image", self.label_column: "labels", }
59
1
import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --host <host> --key_path <key_path> <example> <args>, fill them in as BYO cluster # If user passes --instance <instance> --provider <provider> <example> <args>, fill them in as on-demand cluster # Throw an error if user passes both BYO and on-demand cluster args # Otherwise, use default values _lowerCamelCase = argparse.ArgumentParser() parser.add_argument('--user', type=str, default='ubuntu') parser.add_argument('--host', type=str, default='localhost') parser.add_argument('--key_path', type=str, default=None) parser.add_argument('--instance', type=str, default='V100:1') parser.add_argument('--provider', type=str, default='cheapest') parser.add_argument('--use_spot', type=bool, default=False) parser.add_argument('--example', type=str, default='pytorch/text-generation/run_generation.py') _lowerCamelCase , _lowerCamelCase = parser.parse_known_args() if args.host != "localhost": if args.instance != "V100:1" or args.provider != "cheapest": raise ValueError('Cannot specify both BYO and on-demand cluster args') _lowerCamelCase = rh.cluster( name='rh-cluster', ips=[args.host], ssh_creds={'ssh_user': args.user, 'ssh_private_key': args.key_path} ) else: _lowerCamelCase = rh.cluster( name='rh-cluster', instance_type=args.instance, provider=args.provider, use_spot=args.use_spot ) _lowerCamelCase = args.example.rsplit('/', 1)[0] # Set up remote environment cluster.install_packages(['pip:./']) # Installs transformers from local source # Note transformers is copied into the home directory on the remote machine, so we can install from there cluster.run([F'''pip install -r transformers/examples/{example_dir}/requirements.txt''']) cluster.run(['pip install torch --upgrade --extra-index-url https://download.pytorch.org/whl/cu117']) # Run example. You can bypass the CLI wrapper and paste your own code here. cluster.run([F'''python transformers/examples/{args.example} {' '.join(shlex.quote(arg) for arg in unknown)}''']) # Alternatively, we can just import and run a training function (especially if there's no wrapper CLI): # from my_script... import train # reqs = ['pip:./', 'torch', 'datasets', 'accelerate', 'evaluate', 'tqdm', 'scipy', 'scikit-learn', 'tensorboard'] # launch_train_gpu = rh.function(fn=train, # system=gpu, # reqs=reqs, # name='train_bert_glue') # # We can pass in arguments just like we would to a function: # launch_train_gpu(num_epochs = 3, lr = 2e-5, seed = 42, batch_size = 16 # stream_logs=True)
59
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCamelCase = { 'configuration_xmod': [ 'XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XmodConfig', 'XmodOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'XMOD_PRETRAINED_MODEL_ARCHIVE_LIST', 'XmodForCausalLM', 'XmodForMaskedLM', 'XmodForMultipleChoice', 'XmodForQuestionAnswering', 'XmodForSequenceClassification', 'XmodForTokenClassification', 'XmodModel', 'XmodPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
1
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list , UpperCamelCase__: int , UpperCamelCase__: int = 0 , UpperCamelCase__: int = 0 ): SCREAMING_SNAKE_CASE__ = right or len(UpperCamelCase__ ) - 1 if left > right: return -1 elif list_data[left] == key: return left elif list_data[right] == key: return right else: return search(UpperCamelCase__ , UpperCamelCase__ , left + 1 , right - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
59
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "timm_backbone" def __init__( self :Union[str, Any] , __A :str=None , __A :Union[str, Any]=3 , __A :str=True , __A :Any=True , __A :Optional[Any]=None , **__A :List[str] , ) -> Tuple: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = backbone SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = features_only SCREAMING_SNAKE_CASE__ = use_pretrained_backbone SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = out_indices if out_indices is not None else (-1,)
59
1
import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = 42 lowerCamelCase_ = jnp.floataa lowerCamelCase_ = True def _snake_case ( self :Optional[int] ) -> List[Any]: """simple docstring""" super().setup() SCREAMING_SNAKE_CASE__ = nn.Dense(5 , dtype=self.dtype ) def __call__( self :Union[str, Any] , *__A :Optional[Any] , **__A :Optional[Any] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = super().__call__(*__A , **__A ) SCREAMING_SNAKE_CASE__ = self.cls(outputs[2] ) return outputs[:2] + (cls_out,) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = FlaxBigBirdForNaturalQuestionsModule def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] , UpperCamelCase__: List[str] , UpperCamelCase__: Tuple , UpperCamelCase__: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Optional[Any] ): def cross_entropy(UpperCamelCase__: List[Any] , UpperCamelCase__: Tuple , UpperCamelCase__: Union[str, Any]=None ): SCREAMING_SNAKE_CASE__ = logits.shape[-1] SCREAMING_SNAKE_CASE__ = (labels[..., None] == jnp.arange(UpperCamelCase__ )[None]).astype("""f4""" ) SCREAMING_SNAKE_CASE__ = jax.nn.log_softmax(UpperCamelCase__ , axis=-1 ) SCREAMING_SNAKE_CASE__ = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: SCREAMING_SNAKE_CASE__ = reduction(UpperCamelCase__ ) return loss SCREAMING_SNAKE_CASE__ = partial(UpperCamelCase__ , reduction=jnp.mean ) SCREAMING_SNAKE_CASE__ = cross_entropy(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = cross_entropy(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = cross_entropy(UpperCamelCase__ , UpperCamelCase__ ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class UpperCamelCase_ : lowerCamelCase_ = "google/bigbird-roberta-base" lowerCamelCase_ = 30_00 lowerCamelCase_ = 1_05_00 lowerCamelCase_ = 1_28 lowerCamelCase_ = 3 lowerCamelCase_ = 1 lowerCamelCase_ = 5 # tx_args lowerCamelCase_ = 3e-5 lowerCamelCase_ = 0.0 lowerCamelCase_ = 2_00_00 lowerCamelCase_ = 0.0095 lowerCamelCase_ = "bigbird-roberta-natural-questions" lowerCamelCase_ = "training-expt" lowerCamelCase_ = "data/nq-training.jsonl" lowerCamelCase_ = "data/nq-validation.jsonl" def _snake_case ( self :int ) -> List[str]: """simple docstring""" os.makedirs(self.base_dir , exist_ok=__A ) SCREAMING_SNAKE_CASE__ = os.path.join(self.base_dir , self.save_dir ) SCREAMING_SNAKE_CASE__ = self.batch_size_per_device * jax.device_count() @dataclass class UpperCamelCase_ : lowerCamelCase_ = 42 lowerCamelCase_ = 40_96 # no dynamic padding on TPUs def __call__( self :str , __A :Union[str, Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.collate_fn(__A ) SCREAMING_SNAKE_CASE__ = jax.tree_util.tree_map(__A , __A ) return batch def _snake_case ( self :List[str] , __A :Union[str, Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.fetch_inputs(features["""input_ids"""] ) SCREAMING_SNAKE_CASE__ = { """input_ids""": jnp.array(__A , dtype=jnp.intaa ), """attention_mask""": jnp.array(__A , dtype=jnp.intaa ), """start_labels""": jnp.array(features["""start_token"""] , dtype=jnp.intaa ), """end_labels""": jnp.array(features["""end_token"""] , dtype=jnp.intaa ), """pooled_labels""": jnp.array(features["""category"""] , dtype=jnp.intaa ), } return batch def _snake_case ( self :List[Any] , __A :list ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = [self._fetch_inputs(__A ) for ids in input_ids] return zip(*__A ) def _snake_case ( self :Union[str, Any] , __A :list ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = [1 for _ in range(len(__A ) )] while len(__A ) < self.max_length: input_ids.append(self.pad_id ) attention_mask.append(0 ) return input_ids, attention_mask def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[Any] , UpperCamelCase__: Any , UpperCamelCase__: int=None ): if seed is not None: SCREAMING_SNAKE_CASE__ = dataset.shuffle(seed=UpperCamelCase__ ) for i in range(len(UpperCamelCase__ ) // batch_size ): SCREAMING_SNAKE_CASE__ = dataset[i * batch_size : (i + 1) * batch_size] yield dict(UpperCamelCase__ ) @partial(jax.pmap , axis_name="""batch""" ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[Any] , UpperCamelCase__: int , **UpperCamelCase__: Dict ): def loss_fn(UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = model_inputs.pop("""start_labels""" ) SCREAMING_SNAKE_CASE__ = model_inputs.pop("""end_labels""" ) SCREAMING_SNAKE_CASE__ = model_inputs.pop("""pooled_labels""" ) SCREAMING_SNAKE_CASE__ = state.apply_fn(**UpperCamelCase__ , params=UpperCamelCase__ , dropout_rng=UpperCamelCase__ , train=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = outputs return state.loss_fn( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = jax.random.split(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = jax.value_and_grad(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = grad_fn(state.params ) SCREAMING_SNAKE_CASE__ = jax.lax.pmean({"""loss""": loss} , axis_name="""batch""" ) SCREAMING_SNAKE_CASE__ = jax.lax.pmean(UpperCamelCase__ , """batch""" ) SCREAMING_SNAKE_CASE__ = state.apply_gradients(grads=UpperCamelCase__ ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name="""batch""" ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[Any] , **UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = model_inputs.pop("""start_labels""" ) SCREAMING_SNAKE_CASE__ = model_inputs.pop("""end_labels""" ) SCREAMING_SNAKE_CASE__ = model_inputs.pop("""pooled_labels""" ) SCREAMING_SNAKE_CASE__ = state.apply_fn(**UpperCamelCase__ , params=state.params , train=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = outputs SCREAMING_SNAKE_CASE__ = state.loss_fn(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = jax.lax.pmean({"""loss""": loss} , axis_name="""batch""" ) return metrics class UpperCamelCase_ ( train_state.TrainState ): lowerCamelCase_ = struct.field(pytree_node=UpperCamelCase__ ) @dataclass class UpperCamelCase_ : lowerCamelCase_ = 42 lowerCamelCase_ = 42 lowerCamelCase_ = 42 lowerCamelCase_ = 42 lowerCamelCase_ = 42 lowerCamelCase_ = 42 lowerCamelCase_ = None def _snake_case ( self :Dict , __A :int , __A :str , __A :Optional[Any] , __A :Union[str, Any]=None ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = model.params SCREAMING_SNAKE_CASE__ = TrainState.create( apply_fn=model.__call__ , params=__A , tx=__A , loss_fn=__A , ) if ckpt_dir is not None: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = restore_checkpoint(__A , __A ) SCREAMING_SNAKE_CASE__ = { """lr""": args.lr, """init_lr""": args.init_lr, """warmup_steps""": args.warmup_steps, """num_train_steps""": num_train_steps, """weight_decay""": args.weight_decay, } SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = build_tx(**__A ) SCREAMING_SNAKE_CASE__ = train_state.TrainState( step=__A , apply_fn=model.__call__ , params=__A , tx=__A , opt_state=__A , ) SCREAMING_SNAKE_CASE__ = args SCREAMING_SNAKE_CASE__ = data_collator SCREAMING_SNAKE_CASE__ = lr SCREAMING_SNAKE_CASE__ = params SCREAMING_SNAKE_CASE__ = jax_utils.replicate(__A ) return state def _snake_case ( self :Union[str, Any] , __A :Any , __A :Optional[Any] , __A :Optional[Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.args SCREAMING_SNAKE_CASE__ = len(__A ) // args.batch_size SCREAMING_SNAKE_CASE__ = jax.random.PRNGKey(0 ) SCREAMING_SNAKE_CASE__ = jax.random.split(__A , jax.device_count() ) for epoch in range(args.max_epochs ): SCREAMING_SNAKE_CASE__ = jnp.array(0 , dtype=jnp.floataa ) SCREAMING_SNAKE_CASE__ = get_batched_dataset(__A , args.batch_size , seed=__A ) SCREAMING_SNAKE_CASE__ = 0 for batch in tqdm(__A , total=__A , desc=f'''Running EPOCH-{epoch}''' ): SCREAMING_SNAKE_CASE__ = self.data_collator(__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.train_step_fn(__A , __A , **__A ) running_loss += jax_utils.unreplicate(metrics["""loss"""] ) i += 1 if i % args.logging_steps == 0: SCREAMING_SNAKE_CASE__ = jax_utils.unreplicate(state.step ) SCREAMING_SNAKE_CASE__ = running_loss.item() / i SCREAMING_SNAKE_CASE__ = self.scheduler_fn(state_step - 1 ) SCREAMING_SNAKE_CASE__ = self.evaluate(__A , __A ) SCREAMING_SNAKE_CASE__ = { """step""": state_step.item(), """eval_loss""": eval_loss.item(), """tr_loss""": tr_loss, """lr""": lr.item(), } tqdm.write(str(__A ) ) self.logger.log(__A , commit=__A ) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f'''-e{epoch}-s{i}''' , state=__A ) def _snake_case ( self :str , __A :Union[str, Any] , __A :Optional[int] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = get_batched_dataset(__A , self.args.batch_size ) SCREAMING_SNAKE_CASE__ = len(__A ) // self.args.batch_size SCREAMING_SNAKE_CASE__ = jnp.array(0 , dtype=jnp.floataa ) SCREAMING_SNAKE_CASE__ = 0 for batch in tqdm(__A , total=__A , desc="""Evaluating ... """ ): SCREAMING_SNAKE_CASE__ = self.data_collator(__A ) SCREAMING_SNAKE_CASE__ = self.val_step_fn(__A , **__A ) running_loss += jax_utils.unreplicate(metrics["""loss"""] ) i += 1 return running_loss / i def _snake_case ( self :int , __A :int , __A :Tuple ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = jax_utils.unreplicate(__A ) print(f'''SAVING CHECKPOINT IN {save_dir}''' , end=""" ... """ ) self.model_save_fn(__A , params=state.params ) with open(os.path.join(__A , """opt_state.msgpack""" ) , """wb""" ) as f: f.write(to_bytes(state.opt_state ) ) joblib.dump(self.args , os.path.join(__A , """args.joblib""" ) ) joblib.dump(self.data_collator , os.path.join(__A , """data_collator.joblib""" ) ) with open(os.path.join(__A , """training_state.json""" ) , """w""" ) as f: json.dump({"""step""": state.step.item()} , __A ) print("""DONE""" ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] , UpperCamelCase__: str ): print(f'''RESTORING CHECKPOINT FROM {save_dir}''' , end=""" ... """ ) with open(os.path.join(UpperCamelCase__ , """flax_model.msgpack""" ) , """rb""" ) as f: SCREAMING_SNAKE_CASE__ = from_bytes(state.params , f.read() ) with open(os.path.join(UpperCamelCase__ , """opt_state.msgpack""" ) , """rb""" ) as f: SCREAMING_SNAKE_CASE__ = from_bytes(state.opt_state , f.read() ) SCREAMING_SNAKE_CASE__ = joblib.load(os.path.join(UpperCamelCase__ , """args.joblib""" ) ) SCREAMING_SNAKE_CASE__ = joblib.load(os.path.join(UpperCamelCase__ , """data_collator.joblib""" ) ) with open(os.path.join(UpperCamelCase__ , """training_state.json""" ) , """r""" ) as f: SCREAMING_SNAKE_CASE__ = json.load(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = training_state["""step"""] print("""DONE""" ) return params, opt_state, step, args, data_collator def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Optional[int] ): SCREAMING_SNAKE_CASE__ = num_train_steps - warmup_steps SCREAMING_SNAKE_CASE__ = optax.linear_schedule(init_value=UpperCamelCase__ , end_value=UpperCamelCase__ , transition_steps=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = optax.linear_schedule(init_value=UpperCamelCase__ , end_value=1e-7 , transition_steps=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[Any] , UpperCamelCase__: Optional[int] , UpperCamelCase__: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Dict ): def weight_decay_mask(UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = traverse_util.flatten_dict(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = {k: (v[-1] != """bias""" and v[-2:] != ("""LayerNorm""", """scale""")) for k, v in params.items()} return traverse_util.unflatten_dict(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = scheduler_fn(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = optax.adamw(learning_rate=UpperCamelCase__ , weight_decay=UpperCamelCase__ , mask=UpperCamelCase__ ) return tx, lr
59
from math import pow, sqrt def SCREAMING_SNAKE_CASE__ ( *UpperCamelCase__: float ): SCREAMING_SNAKE_CASE__ = len(UpperCamelCase__ ) > 0 and all(value > 0.0 for value in values ) return result def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ ) else ValueError("""Input Error: Molar mass values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) )
59
1
from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax _lowerCamelCase = logging.get_logger(__name__) @add_end_docstrings(UpperCamelCase__ ) class UpperCamelCase_ ( UpperCamelCase__ ): def __init__( self :Union[str, Any] , **__A :Dict ) -> Tuple: """simple docstring""" super().__init__(**__A ) requires_backends(self , """vision""" ) self.check_model_type( TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if self.framework == """tf""" else MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING ) def __call__( self :Optional[int] , __A :Union[str, List[str], "Image", List["Image"]] , **__A :Any ) -> List[str]: """simple docstring""" return super().__call__(__A , **__A ) def _snake_case ( self :Optional[int] , **__A :List[Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = {} if "candidate_labels" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs["""candidate_labels"""] if "hypothesis_template" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs["""hypothesis_template"""] return preprocess_params, {}, {} def _snake_case ( self :int , __A :Any , __A :List[str]=None , __A :Any="This is a photo of {}." ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = load_image(__A ) SCREAMING_SNAKE_CASE__ = self.image_processor(images=[image] , return_tensors=self.framework ) SCREAMING_SNAKE_CASE__ = candidate_labels SCREAMING_SNAKE_CASE__ = [hypothesis_template.format(__A ) for x in candidate_labels] SCREAMING_SNAKE_CASE__ = self.tokenizer(__A , return_tensors=self.framework , padding=__A ) SCREAMING_SNAKE_CASE__ = [text_inputs] return inputs def _snake_case ( self :str , __A :Optional[int] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = model_inputs.pop("""candidate_labels""" ) SCREAMING_SNAKE_CASE__ = model_inputs.pop("""text_inputs""" ) if isinstance(text_inputs[0] , __A ): SCREAMING_SNAKE_CASE__ = text_inputs[0] else: # Batching case. SCREAMING_SNAKE_CASE__ = text_inputs[0][0] SCREAMING_SNAKE_CASE__ = self.model(**__A , **__A ) SCREAMING_SNAKE_CASE__ = { """candidate_labels""": candidate_labels, """logits""": outputs.logits_per_image, } return model_outputs def _snake_case ( self :List[Any] , __A :str ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = model_outputs.pop("""candidate_labels""" ) SCREAMING_SNAKE_CASE__ = model_outputs["""logits"""][0] if self.framework == "pt": SCREAMING_SNAKE_CASE__ = logits.softmax(dim=-1 ).squeeze(-1 ) SCREAMING_SNAKE_CASE__ = probs.tolist() if not isinstance(__A , __A ): SCREAMING_SNAKE_CASE__ = [scores] elif self.framework == "tf": SCREAMING_SNAKE_CASE__ = stable_softmax(__A , axis=-1 ) SCREAMING_SNAKE_CASE__ = probs.numpy().tolist() else: raise ValueError(f'''Unsupported framework: {self.framework}''' ) SCREAMING_SNAKE_CASE__ = [ {"""score""": score, """label""": candidate_label} for score, candidate_label in sorted(zip(__A , __A ) , key=lambda __A : -x[0] ) ] return result
59
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'shi-labs/nat-mini-in1k-224': 'https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json', # See all Nat models at https://huggingface.co/models?filter=nat } class UpperCamelCase_ ( UpperCamelCase__ , UpperCamelCase__ ): lowerCamelCase_ = "nat" lowerCamelCase_ = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self :List[Any] , __A :Optional[Any]=4 , __A :Any=3 , __A :Optional[int]=64 , __A :Optional[int]=[3, 4, 6, 5] , __A :Union[str, Any]=[2, 4, 8, 16] , __A :Optional[Any]=7 , __A :Optional[Any]=3.0 , __A :List[Any]=True , __A :int=0.0 , __A :Dict=0.0 , __A :Optional[Any]=0.1 , __A :str="gelu" , __A :Optional[Any]=0.0_2 , __A :Optional[int]=1E-5 , __A :Optional[int]=0.0 , __A :Optional[Any]=None , __A :Union[str, Any]=None , **__A :Union[str, Any] , ) -> Optional[int]: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = patch_size SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = embed_dim SCREAMING_SNAKE_CASE__ = depths SCREAMING_SNAKE_CASE__ = len(__A ) SCREAMING_SNAKE_CASE__ = num_heads SCREAMING_SNAKE_CASE__ = kernel_size SCREAMING_SNAKE_CASE__ = mlp_ratio SCREAMING_SNAKE_CASE__ = qkv_bias SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = drop_path_rate SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model SCREAMING_SNAKE_CASE__ = int(embed_dim * 2 ** (len(__A ) - 1) ) SCREAMING_SNAKE_CASE__ = layer_scale_init_value SCREAMING_SNAKE_CASE__ = ["""stem"""] + [f'''stage{idx}''' for idx in range(1 , len(__A ) + 1 )] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = get_aligned_output_features_output_indices( out_features=__A , out_indices=__A , stage_names=self.stage_names )
59
1
import tempfile import unittest import numpy as np import transformers from transformers import GPTaTokenizer, GPTJConfig, is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax, tooslow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax import jax.numpy as jnp from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) from transformers.models.gptj.modeling_flax_gptj import FlaxGPTJForCausalLM, FlaxGPTJModel if is_torch_available(): import torch class UpperCamelCase_ : def __init__( self :str , __A :List[Any] , __A :Any=14 , __A :Optional[Any]=7 , __A :Optional[int]=True , __A :List[Any]=True , __A :int=False , __A :str=True , __A :Optional[int]=99 , __A :Union[str, Any]=32 , __A :Union[str, Any]=4 , __A :List[Any]=4 , __A :int=4 , __A :List[str]=37 , __A :int="gelu" , __A :List[Any]=0.1 , __A :List[str]=0.1 , __A :Optional[Any]=512 , __A :int=0.0_2 , ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = parent SCREAMING_SNAKE_CASE__ = batch_size SCREAMING_SNAKE_CASE__ = seq_length SCREAMING_SNAKE_CASE__ = is_training SCREAMING_SNAKE_CASE__ = use_input_mask SCREAMING_SNAKE_CASE__ = use_token_type_ids SCREAMING_SNAKE_CASE__ = use_labels SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = rotary_dim SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = vocab_size - 1 SCREAMING_SNAKE_CASE__ = vocab_size - 1 SCREAMING_SNAKE_CASE__ = vocab_size - 1 def _snake_case ( self :Optional[Any] ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE__ = None if self.use_input_mask: SCREAMING_SNAKE_CASE__ = random_attention_mask([self.batch_size, self.seq_length] ) SCREAMING_SNAKE_CASE__ = GPTJConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , use_cache=__A , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , rotary_dim=self.rotary_dim , ) return (config, input_ids, input_mask) def _snake_case ( self :Optional[int] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = config_and_inputs SCREAMING_SNAKE_CASE__ = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict def _snake_case ( self :int , __A :int , __A :Any , __A :Any , __A :Optional[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = 20 SCREAMING_SNAKE_CASE__ = model_class_name(__A ) SCREAMING_SNAKE_CASE__ = model.init_cache(input_ids.shape[0] , __A ) SCREAMING_SNAKE_CASE__ = jnp.ones((input_ids.shape[0], max_decoder_length) , dtype="""i4""" ) SCREAMING_SNAKE_CASE__ = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) SCREAMING_SNAKE_CASE__ = model( input_ids[:, :-1] , attention_mask=__A , past_key_values=__A , position_ids=__A , ) SCREAMING_SNAKE_CASE__ = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype="""i4""" ) SCREAMING_SNAKE_CASE__ = model( input_ids[:, -1:] , attention_mask=__A , past_key_values=outputs_cache.past_key_values , position_ids=__A , ) SCREAMING_SNAKE_CASE__ = model(__A ) SCREAMING_SNAKE_CASE__ = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f'''Max diff is {diff}''' ) def _snake_case ( self :Dict , __A :List[Any] , __A :List[str] , __A :Optional[Any] , __A :Union[str, Any] ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = 20 SCREAMING_SNAKE_CASE__ = model_class_name(__A ) SCREAMING_SNAKE_CASE__ = jnp.concatenate( [attention_mask, jnp.zeros((attention_mask.shape[0], max_decoder_length - attention_mask.shape[1]) )] , axis=-1 , ) SCREAMING_SNAKE_CASE__ = model.init_cache(input_ids.shape[0] , __A ) SCREAMING_SNAKE_CASE__ = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) SCREAMING_SNAKE_CASE__ = model( input_ids[:, :-1] , attention_mask=__A , past_key_values=__A , position_ids=__A , ) SCREAMING_SNAKE_CASE__ = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype="""i4""" ) SCREAMING_SNAKE_CASE__ = model( input_ids[:, -1:] , past_key_values=outputs_cache.past_key_values , attention_mask=__A , position_ids=__A , ) SCREAMING_SNAKE_CASE__ = model(__A , attention_mask=__A ) SCREAMING_SNAKE_CASE__ = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f'''Max diff is {diff}''' ) @require_flax class UpperCamelCase_ ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = (FlaxGPTJModel, FlaxGPTJForCausalLM) if is_flax_available() else () lowerCamelCase_ = (FlaxGPTJForCausalLM,) if is_flax_available() else () def _snake_case ( self :Tuple ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = FlaxGPTJModelTester(self ) def _snake_case ( self :Union[str, Any] ) -> int: """simple docstring""" for model_class_name in self.all_model_classes: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward(__A , __A , __A , __A ) def _snake_case ( self :Dict ) -> Dict: """simple docstring""" for model_class_name in self.all_model_classes: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward_with_attn_mask( __A , __A , __A , __A ) @tooslow def _snake_case ( self :List[str] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = GPTaTokenizer.from_pretrained("""gpt2""" , pad_token="""<|endoftext|>""" , padding_side="""left""" ) SCREAMING_SNAKE_CASE__ = tokenizer(["""Hello this is a long string""", """Hey"""] , return_tensors="""np""" , padding=__A , truncation=__A ) SCREAMING_SNAKE_CASE__ = FlaxGPTJForCausalLM.from_pretrained("""EleutherAI/gpt-j-6B""" ) SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = model.config.eos_token_id SCREAMING_SNAKE_CASE__ = jax.jit(model.generate ) SCREAMING_SNAKE_CASE__ = jit_generate( inputs["""input_ids"""] , attention_mask=inputs["""attention_mask"""] , pad_token_id=tokenizer.pad_token_id ).sequences SCREAMING_SNAKE_CASE__ = tokenizer.batch_decode(__A , skip_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = [ """Hello this is a long string of text.\n\nI'm trying to get the text of the""", """Hey, I'm a little late to the party. I'm going to""", ] self.assertListEqual(__A , __A ) @is_pt_flax_cross_test def _snake_case ( self :Any ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs SCREAMING_SNAKE_CASE__ = self._prepare_for_class(__A , __A ) SCREAMING_SNAKE_CASE__ = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class SCREAMING_SNAKE_CASE__ = model_class.__name__[4:] # Skip the "Flax" at the beginning SCREAMING_SNAKE_CASE__ = getattr(__A , __A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = pt_inputs["""input_ids"""].shape SCREAMING_SNAKE_CASE__ = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(__A ): SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = pt_model_class(__A ).eval() SCREAMING_SNAKE_CASE__ = model_class(__A , dtype=jnp.floataa ) SCREAMING_SNAKE_CASE__ = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , __A ) SCREAMING_SNAKE_CASE__ = fx_state with torch.no_grad(): SCREAMING_SNAKE_CASE__ = pt_model(**__A ).to_tuple() SCREAMING_SNAKE_CASE__ = fx_model(**__A ).to_tuple() self.assertEqual(len(__A ) , len(__A ) , """Output lengths differ between Flax and PyTorch""" ) for fx_output, pt_output in zip(__A , __A ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(__A ) SCREAMING_SNAKE_CASE__ = model_class.from_pretrained(__A , from_pt=__A ) SCREAMING_SNAKE_CASE__ = fx_model_loaded(**__A ).to_tuple() self.assertEqual( len(__A ) , len(__A ) , """Output lengths differ between Flax and PyTorch""" ) for fx_output_loaded, pt_output in zip(__A , __A ): self.assert_almost_equals(fx_output_loaded[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @is_pt_flax_cross_test def _snake_case ( self :Any ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs SCREAMING_SNAKE_CASE__ = self._prepare_for_class(__A , __A ) SCREAMING_SNAKE_CASE__ = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class SCREAMING_SNAKE_CASE__ = model_class.__name__[4:] # Skip the "Flax" at the beginning SCREAMING_SNAKE_CASE__ = getattr(__A , __A ) SCREAMING_SNAKE_CASE__ = pt_model_class(__A ).eval() SCREAMING_SNAKE_CASE__ = model_class(__A , dtype=jnp.floataa ) SCREAMING_SNAKE_CASE__ = load_flax_weights_in_pytorch_model(__A , fx_model.params ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = pt_inputs["""input_ids"""].shape SCREAMING_SNAKE_CASE__ = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(__A ): SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 1 # make sure weights are tied in PyTorch pt_model.tie_weights() with torch.no_grad(): SCREAMING_SNAKE_CASE__ = pt_model(**__A ).to_tuple() SCREAMING_SNAKE_CASE__ = fx_model(**__A ).to_tuple() self.assertEqual(len(__A ) , len(__A ) , """Output lengths differ between Flax and PyTorch""" ) for fx_output, pt_output in zip(__A , __A ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(__A ) SCREAMING_SNAKE_CASE__ = pt_model_class.from_pretrained(__A , from_flax=__A ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ = pt_model_loaded(**__A ).to_tuple() self.assertEqual( len(__A ) , len(__A ) , """Output lengths differ between Flax and PyTorch""" ) for fx_output, pt_output in zip(__A , __A ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @tooslow def _snake_case ( self :Union[str, Any] ) -> Tuple: """simple docstring""" for model_class_name in self.all_model_classes: SCREAMING_SNAKE_CASE__ = model_class_name.from_pretrained("""EleutherAI/gpt-j-6B""" ) SCREAMING_SNAKE_CASE__ = model(np.ones((1, 1) ) ) self.assertIsNotNone(__A )
59
from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any ): return {key.lstrip("""-""" ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = ArgumentParser( """HuggingFace Datasets CLI tool""" , usage="""datasets-cli <command> [<args>]""" , allow_abbrev=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = parser.add_subparsers(help="""datasets-cli command helpers""" ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(UpperCamelCase__ ) EnvironmentCommand.register_subcommand(UpperCamelCase__ ) TestCommand.register_subcommand(UpperCamelCase__ ) RunBeamCommand.register_subcommand(UpperCamelCase__ ) DummyDataCommand.register_subcommand(UpperCamelCase__ ) # Parse args SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = parser.parse_known_args() if not hasattr(UpperCamelCase__ , """func""" ): parser.print_help() exit(1 ) SCREAMING_SNAKE_CASE__ = parse_unknown_args(UpperCamelCase__ ) # Run SCREAMING_SNAKE_CASE__ = args.func(UpperCamelCase__ , **UpperCamelCase__ ) service.run() if __name__ == "__main__": main()
59
1
def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = [31, 28, 31, 30, 31, 30, 31, 31, 30, 31, 30, 31] SCREAMING_SNAKE_CASE__ = 6 SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = 1_901 SCREAMING_SNAKE_CASE__ = 0 while year < 2_001: day += 7 if (year % 4 == 0 and year % 100 != 0) or (year % 400 == 0): if day > days_per_month[month - 1] and month != 2: month += 1 SCREAMING_SNAKE_CASE__ = day - days_per_month[month - 2] elif day > 29 and month == 2: month += 1 SCREAMING_SNAKE_CASE__ = day - 29 else: if day > days_per_month[month - 1]: month += 1 SCREAMING_SNAKE_CASE__ = day - days_per_month[month - 2] if month > 12: year += 1 SCREAMING_SNAKE_CASE__ = 1 if year < 2_001 and day == 1: sundays += 1 return sundays if __name__ == "__main__": print(solution())
59
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase__ ): def __init__( self :List[Any] , *__A :Tuple , **__A :Dict ) -> None: """simple docstring""" warnings.warn( """The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use LayoutLMv2ImageProcessor instead.""" , __A , ) super().__init__(*__A , **__A )
59
1
import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = inspect.getfile(accelerate.test_utils ) SCREAMING_SNAKE_CASE__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_script.py"""] ) SCREAMING_SNAKE_CASE__ = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_distributed_data_loop.py"""] ) SCREAMING_SNAKE_CASE__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_ops.py"""] ) @require_multi_gpu def _snake_case ( self :Optional[Any] ) -> Tuple: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Tuple ) -> Optional[Any]: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.operation_file_path] print(f'''Command: {cmd}''' ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Optional[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Optional[int] ) -> str: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices, using 2 devices only''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices="""0,1""" ): execute_subprocess_async(__A , env=os.environ.copy() ) if __name__ == "__main__": _lowerCamelCase = Accelerator() _lowerCamelCase = (accelerator.state.process_index + 2, 10) _lowerCamelCase = torch.randint(0, 10, shape).to(accelerator.device) _lowerCamelCase = '' _lowerCamelCase = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." _lowerCamelCase = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." _lowerCamelCase = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
59
import unittest from datasets import load_dataset from transformers.pipelines import pipeline from transformers.testing_utils import is_pipeline_test, nested_simplify, require_torch, slow @is_pipeline_test @require_torch class UpperCamelCase_ ( unittest.TestCase ): @require_torch def _snake_case ( self :Dict ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = pipeline( task="""zero-shot-audio-classification""" , model="""hf-internal-testing/tiny-clap-htsat-unfused""" ) SCREAMING_SNAKE_CASE__ = load_dataset("""ashraq/esc50""" ) SCREAMING_SNAKE_CASE__ = dataset["""train"""]["""audio"""][-1]["""array"""] SCREAMING_SNAKE_CASE__ = audio_classifier(__A , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__A ) , [{"""score""": 0.5_0_1, """label""": """Sound of a dog"""}, {"""score""": 0.4_9_9, """label""": """Sound of vaccum cleaner"""}] , ) @unittest.skip("""No models are available in TF""" ) def _snake_case ( self :Dict ) -> List[str]: """simple docstring""" pass @slow @require_torch def _snake_case ( self :Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = pipeline( task="""zero-shot-audio-classification""" , model="""laion/clap-htsat-unfused""" , ) # This is an audio of a dog SCREAMING_SNAKE_CASE__ = load_dataset("""ashraq/esc50""" ) SCREAMING_SNAKE_CASE__ = dataset["""train"""]["""audio"""][-1]["""array"""] SCREAMING_SNAKE_CASE__ = audio_classifier(__A , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__A ) , [ {"""score""": 0.9_9_9, """label""": """Sound of a dog"""}, {"""score""": 0.0_0_1, """label""": """Sound of vaccum cleaner"""}, ] , ) SCREAMING_SNAKE_CASE__ = audio_classifier([audio] * 5 , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__A ) , [ [ {"""score""": 0.9_9_9, """label""": """Sound of a dog"""}, {"""score""": 0.0_0_1, """label""": """Sound of vaccum cleaner"""}, ], ] * 5 , ) SCREAMING_SNAKE_CASE__ = audio_classifier( [audio] * 5 , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] , batch_size=5 ) self.assertEqual( nested_simplify(__A ) , [ [ {"""score""": 0.9_9_9, """label""": """Sound of a dog"""}, {"""score""": 0.0_0_1, """label""": """Sound of vaccum cleaner"""}, ], ] * 5 , ) @unittest.skip("""No models are available in TF""" ) def _snake_case ( self :str ) -> Optional[int]: """simple docstring""" pass
59
1
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'nvidia/segformer-b0-finetuned-ade-512-512': ( 'https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json' ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "segformer" def __init__( self :Optional[int] , __A :List[Any]=3 , __A :Optional[Any]=4 , __A :int=[2, 2, 2, 2] , __A :Union[str, Any]=[8, 4, 2, 1] , __A :Optional[int]=[32, 64, 160, 256] , __A :int=[7, 3, 3, 3] , __A :Tuple=[4, 2, 2, 2] , __A :str=[1, 2, 5, 8] , __A :Union[str, Any]=[4, 4, 4, 4] , __A :Any="gelu" , __A :List[str]=0.0 , __A :List[Any]=0.0 , __A :Any=0.1 , __A :Optional[int]=0.0_2 , __A :List[Any]=0.1 , __A :int=1E-6 , __A :int=256 , __A :Dict=255 , **__A :Any , ) -> List[Any]: """simple docstring""" super().__init__(**__A ) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( """Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be""" """ removed, as the behaviour will default to that of reshape_last_stage = True.""" , __A , ) SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = num_encoder_blocks SCREAMING_SNAKE_CASE__ = depths SCREAMING_SNAKE_CASE__ = sr_ratios SCREAMING_SNAKE_CASE__ = hidden_sizes SCREAMING_SNAKE_CASE__ = patch_sizes SCREAMING_SNAKE_CASE__ = strides SCREAMING_SNAKE_CASE__ = mlp_ratios SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = classifier_dropout_prob SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = drop_path_rate SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = decoder_hidden_size SCREAMING_SNAKE_CASE__ = kwargs.get("""reshape_last_stage""" , __A ) SCREAMING_SNAKE_CASE__ = semantic_loss_ignore_index class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = version.parse("1.11" ) @property def _snake_case ( self :Union[str, Any] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def _snake_case ( self :Optional[Any] ) -> float: """simple docstring""" return 1E-4 @property def _snake_case ( self :str ) -> int: """simple docstring""" return 12
59
import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 _lowerCamelCase = data_utils.TransfoXLTokenizer _lowerCamelCase = data_utils.TransfoXLCorpus _lowerCamelCase = data_utils _lowerCamelCase = data_utils def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Dict , UpperCamelCase__: Any , UpperCamelCase__: Any , UpperCamelCase__: Tuple ): if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(UpperCamelCase__ , """rb""" ) as fp: SCREAMING_SNAKE_CASE__ = pickle.load(UpperCamelCase__ , encoding="""latin1""" ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) SCREAMING_SNAKE_CASE__ = pytorch_dump_folder_path + """/""" + VOCAB_FILES_NAMES["""pretrained_vocab_file"""] print(f'''Save vocabulary to {pytorch_vocab_dump_path}''' ) SCREAMING_SNAKE_CASE__ = corpus.vocab.__dict__ torch.save(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = corpus.__dict__ corpus_dict_no_vocab.pop("""vocab""" , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = pytorch_dump_folder_path + """/""" + CORPUS_NAME print(f'''Save dataset to {pytorch_dataset_dump_path}''' ) torch.save(UpperCamelCase__ , UpperCamelCase__ ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model SCREAMING_SNAKE_CASE__ = os.path.abspath(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = os.path.abspath(UpperCamelCase__ ) print(f'''Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.''' ) # Initialise PyTorch model if transfo_xl_config_file == "": SCREAMING_SNAKE_CASE__ = TransfoXLConfig() else: SCREAMING_SNAKE_CASE__ = TransfoXLConfig.from_json_file(UpperCamelCase__ ) print(f'''Building PyTorch model from configuration: {config}''' ) SCREAMING_SNAKE_CASE__ = TransfoXLLMHeadModel(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = load_tf_weights_in_transfo_xl(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # Save pytorch-model SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) print(f'''Save PyTorch model to {os.path.abspath(UpperCamelCase__ )}''' ) torch.save(model.state_dict() , UpperCamelCase__ ) print(f'''Save configuration file to {os.path.abspath(UpperCamelCase__ )}''' ) with open(UpperCamelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--tf_checkpoint_path', default='', type=str, help='An optional path to a TensorFlow checkpoint path to be converted.', ) parser.add_argument( '--transfo_xl_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained BERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--transfo_xl_dataset_file', default='', type=str, help='An optional dataset file to be converted in a vocabulary.', ) _lowerCamelCase = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
59
1
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 _lowerCamelCase = get_tests_dir('fixtures') _lowerCamelCase = get_tests_dir('fixtures/dummy_feature_extractor_config.json') _lowerCamelCase = get_tests_dir('fixtures/dummy-config.json') class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :Optional[Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = 0 def _snake_case ( self :str ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained("""facebook/wav2vec2-base-960h""" ) self.assertIsInstance(__A , __A ) def _snake_case ( self :int ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained(__A ) self.assertIsInstance(__A , __A ) def _snake_case ( self :Optional[int] ) -> List[Any]: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: SCREAMING_SNAKE_CASE__ = WavaVecaConfig() # remove feature_extractor_type to make sure config.json alone is enough to load feature processor locally SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained(__A ).to_dict() config_dict.pop("""feature_extractor_type""" ) SCREAMING_SNAKE_CASE__ = WavaVecaFeatureExtractor(**__A ) # save in new folder model_config.save_pretrained(__A ) config.save_pretrained(__A ) SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained(__A ) # make sure private variable is not incorrectly saved SCREAMING_SNAKE_CASE__ = json.loads(config.to_json_string() ) self.assertTrue("""_processor_class""" not in dict_as_saved ) self.assertIsInstance(__A , __A ) def _snake_case ( self :List[str] ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained(__A ) self.assertIsInstance(__A , __A ) def _snake_case ( self :Dict ) -> Union[str, Any]: """simple docstring""" with self.assertRaisesRegex( __A , """bert-base is not a local folder and is not a valid model identifier""" ): SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained("""bert-base""" ) def _snake_case ( self :Any ) -> List[Any]: """simple docstring""" with self.assertRaisesRegex( __A , r"""aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)""" ): SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained(__A , revision="""aaaaaa""" ) def _snake_case ( self :List[str] ) -> Optional[Any]: """simple docstring""" with self.assertRaisesRegex( __A , """hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.""" , ): SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained("""hf-internal-testing/config-no-model""" ) def _snake_case ( self :List[Any] ) -> List[str]: """simple docstring""" with self.assertRaises(__A ): SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" ) # If remote code is disabled, we can't load this config. with self.assertRaises(__A ): SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=__A ) SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=__A ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) # Test feature extractor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(__A ) SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained(__A , trust_remote_code=__A ) self.assertEqual(reloaded_feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) def _snake_case ( self :List[Any] ) -> str: """simple docstring""" try: AutoConfig.register("""custom""" , __A ) AutoFeatureExtractor.register(__A , __A ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__A ): AutoFeatureExtractor.register(__A , __A ) # Now that the config is registered, it can be used as any other config with the auto-API SCREAMING_SNAKE_CASE__ = CustomFeatureExtractor.from_pretrained(__A ) with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(__A ) SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained(__A ) self.assertIsInstance(__A , __A ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] def _snake_case ( self :Any ) -> Optional[int]: """simple docstring""" class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = True try: AutoConfig.register("""custom""" , __A ) AutoFeatureExtractor.register(__A , __A ) # If remote code is not set, the default is to use local SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) self.assertTrue(feature_extractor.is_local ) # If remote code is disabled, we load the local one. SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=__A ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) self.assertTrue(feature_extractor.is_local ) # If remote is enabled, we load from the Hub SCREAMING_SNAKE_CASE__ = AutoFeatureExtractor.from_pretrained( """hf-internal-testing/test_dynamic_feature_extractor""" , trust_remote_code=__A ) self.assertEqual(feature_extractor.__class__.__name__ , """NewFeatureExtractor""" ) self.assertTrue(not hasattr(__A , """is_local""" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig]
59
import argparse import tensorflow as tf import torch from transformers import BertConfig, BertForMaskedLM from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertPooler, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging logging.set_verbosity_info() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str , UpperCamelCase__: str , UpperCamelCase__: str ): def get_masked_lm_array(UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = f'''masked_lm/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) def get_encoder_array(UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = f'''encoder/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) def get_encoder_layer_array(UpperCamelCase__: int , UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = f'''encoder/_transformer_layers/{layer_index}/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) def get_encoder_attention_layer_array(UpperCamelCase__: int , UpperCamelCase__: str , UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = f'''encoder/_transformer_layers/{layer_index}/_attention_layer/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = array.reshape(UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) print(f'''Loading model based on config from {config_path}...''' ) SCREAMING_SNAKE_CASE__ = BertConfig.from_json_file(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = BertForMaskedLM(UpperCamelCase__ ) # Layers for layer_index in range(0 , config.num_hidden_layers ): SCREAMING_SNAKE_CASE__ = model.bert.encoder.layer[layer_index] # Self-attention SCREAMING_SNAKE_CASE__ = layer.attention.self SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_query_dense/kernel""" , self_attn.query.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_query_dense/bias""" , self_attn.query.bias.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_key_dense/kernel""" , self_attn.key.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_key_dense/bias""" , self_attn.key.bias.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_value_dense/kernel""" , self_attn.value.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_value_dense/bias""" , self_attn.value.bias.data.shape ) # Self-attention Output SCREAMING_SNAKE_CASE__ = layer.attention.output SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_output_dense/kernel""" , self_output.dense.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_output_dense/bias""" , self_output.dense.bias.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_attention_layer_norm/gamma""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_attention_layer_norm/beta""" ) # Intermediate SCREAMING_SNAKE_CASE__ = layer.intermediate SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_intermediate_dense/kernel""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_intermediate_dense/bias""" ) # Output SCREAMING_SNAKE_CASE__ = layer.output SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_dense/kernel""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_dense/bias""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_layer_norm/gamma""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_layer_norm/beta""" ) # Embeddings SCREAMING_SNAKE_CASE__ = get_encoder_array("""_position_embedding_layer/embeddings""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_type_embedding_layer/embeddings""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_embedding_norm_layer/gamma""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_embedding_norm_layer/beta""" ) # LM Head SCREAMING_SNAKE_CASE__ = model.cls.predictions.transform SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""dense/kernel""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""dense/bias""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""layer_norm/gamma""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""layer_norm/beta""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""embedding_table""" ) # Pooling SCREAMING_SNAKE_CASE__ = BertPooler(config=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_pooler_layer/kernel""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_pooler_layer/bias""" ) # Export final model model.save_pretrained(UpperCamelCase__ ) # Integration test - should load without any errors ;) SCREAMING_SNAKE_CASE__ = BertForMaskedLM.from_pretrained(UpperCamelCase__ ) print(new_model.eval() ) print("""Model conversion was done sucessfully!""" ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '--tf_checkpoint_path', type=str, required=True, help='Path to the TensorFlow Token Dropping checkpoint path.' ) parser.add_argument( '--bert_config_file', type=str, required=True, help='The config json file corresponding to the BERT model. This specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', type=str, required=True, help='Path to the output PyTorch model.', ) _lowerCamelCase = parser.parse_args() convert_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
59
1
from collections import defaultdict from pathlib import Path import pandas as pd from rouge_cli import calculate_rouge_path from utils import calculate_rouge _lowerCamelCase = [ 'Prosecutor: "No videos were used in the crash investigation" German papers say they saw a cell phone video of the' ' final seconds on board Flight 9525. The Germanwings co-pilot says he had a "previous episode of severe' ' depression\" German airline confirms it knew of Andreas Lubitz\'s depression years before he took control.', 'The Palestinian Authority officially becomes the 123rd member of the International Criminal Court. The formal' ' accession was marked with a ceremony at The Hague, in the Netherlands. The Palestinians signed the ICC\'s' ' founding Rome Statute in January. Israel and the United States opposed the Palestinians\' efforts to join the' ' body.', 'Amnesty International releases its annual report on the death penalty. The report catalogs the use of' ' state-sanctioned killing as a punitive measure across the globe. At least 607 people were executed around the' ' world in 2014, compared to 778 in 2013. The U.S. remains one of the worst offenders for imposing capital' ' punishment.', ] _lowerCamelCase = [ 'Marseille prosecutor says "so far no videos were used in the crash investigation" despite media reports .' ' Journalists at Bild and Paris Match are "very confident" the video clip is real, an editor says . Andreas Lubitz' ' had informed his Lufthansa training school of an episode of severe depression, airline says .', 'Membership gives the ICC jurisdiction over alleged crimes committed in Palestinian territories since last June .' ' Israel and the United States opposed the move, which could open the door to war crimes investigations against' ' Israelis .', 'Amnesty\'s annual death penalty report catalogs encouraging signs, but setbacks in numbers of those sentenced to' ' death . Organization claims that governments around the world are using the threat of terrorism to advance' ' executions . The number of executions worldwide has gone down by almost 22% compared with 2013, but death' ' sentences up by 28% .', ] def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , bootstrap_aggregation=UpperCamelCase__ , rouge_keys=["""rouge2""", """rougeL"""] ) assert isinstance(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , bootstrap_aggregation=UpperCamelCase__ , rouge_keys=["""rouge2"""] ) assert ( pd.DataFrame(no_aggregation["""rouge2"""] ).fmeasure.mean() == pd.DataFrame(no_aggregation_just_ra["""rouge2"""] ).fmeasure.mean() ) def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = """rougeLsum""" SCREAMING_SNAKE_CASE__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , newline_sep=UpperCamelCase__ , rouge_keys=[k] )[k] SCREAMING_SNAKE_CASE__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , newline_sep=UpperCamelCase__ , rouge_keys=[k] )[k] assert score > score_no_sep def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = ["""rouge1""", """rouge2""", """rougeL"""] SCREAMING_SNAKE_CASE__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , newline_sep=UpperCamelCase__ , rouge_keys=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , newline_sep=UpperCamelCase__ , rouge_keys=UpperCamelCase__ ) assert score_sep == score_no_sep def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = [ """Her older sister, Margot Frank, died in 1945, a month earlier than previously thought.""", """Marseille prosecutor says \"so far no videos were used in the crash investigation\" despite media reports .""", ] SCREAMING_SNAKE_CASE__ = [ """Margot Frank, died in 1945, a month earlier than previously thought.""", """Prosecutor: \"No videos were used in the crash investigation\" German papers say they saw a cell phone video of""" """ the final seconds on board Flight 9525.""", ] assert calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , newline_sep=UpperCamelCase__ ) == calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , newline_sep=UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = [ """\" \"a person who has such a video needs to immediately give it to the investigators,\" prosecutor says .<n> \"it is a very disturbing scene,\" editor-in-chief of bild online tells \"erin burnett: outfront\" """ ] SCREAMING_SNAKE_CASE__ = [ """ Marseille prosecutor says \"so far no videos were used in the crash investigation\" despite media reports . Journalists at Bild and Paris Match are \"very confident\" the video clip is real, an editor says . Andreas Lubitz had informed his Lufthansa training school of an episode of severe depression, airline says .""" ] SCREAMING_SNAKE_CASE__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , rouge_keys=["""rougeLsum"""] , newline_sep=UpperCamelCase__ )["""rougeLsum"""] SCREAMING_SNAKE_CASE__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , rouge_keys=["""rougeLsum"""] )["""rougeLsum"""] assert new_score > prev_score def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = Path("""examples/seq2seq/test_data/wmt_en_ro""" ) SCREAMING_SNAKE_CASE__ = calculate_rouge_path(data_dir.joinpath("""test.source""" ) , data_dir.joinpath("""test.target""" ) ) assert isinstance(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = calculate_rouge_path( data_dir.joinpath("""test.source""" ) , data_dir.joinpath("""test.target""" ) , bootstrap_aggregation=UpperCamelCase__ ) assert isinstance(UpperCamelCase__ , UpperCamelCase__ )
59
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _lowerCamelCase = '\\n Text data.\n Second line of data.' _lowerCamelCase = 'file' @pytest.fixture(scope="""session""" ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / (FILE_PATH + """.zstd""") SCREAMING_SNAKE_CASE__ = bytes(UpperCamelCase__ , """utf-8""" ) with zstd.open(UpperCamelCase__ , """wb""" ) as f: f.write(UpperCamelCase__ ) return path @pytest.fixture def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): with open(os.path.join(tmpfs.local_root_dir , UpperCamelCase__ ) , """w""" ) as f: f.write(UpperCamelCase__ ) return FILE_PATH @pytest.mark.parametrize("""compression_format""" , ["""gzip""", """xz""", """zstd"""] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int , UpperCamelCase__: Dict , UpperCamelCase__: int , UpperCamelCase__: str , UpperCamelCase__: Optional[int] , UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = {"""gzip""": gz_file, """xz""": xz_file, """zstd""": zstd_path} SCREAMING_SNAKE_CASE__ = input_paths[compression_format] SCREAMING_SNAKE_CASE__ = tmp_path / """cache""" SCREAMING_SNAKE_CASE__ = DownloadConfig(cache_dir=UpperCamelCase__ , extract_compressed_file=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = cached_path(UpperCamelCase__ , download_config=UpperCamelCase__ ) with open(UpperCamelCase__ ) as f: SCREAMING_SNAKE_CASE__ = f.read() with open(UpperCamelCase__ ) as f: SCREAMING_SNAKE_CASE__ = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("""default_extracted""" , [True, False] ) @pytest.mark.parametrize("""default_cache_dir""" , [True, False] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Tuple , UpperCamelCase__: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Any , UpperCamelCase__: Union[str, Any] ): SCREAMING_SNAKE_CASE__ = """custom_cache""" SCREAMING_SNAKE_CASE__ = """custom_extracted_dir""" SCREAMING_SNAKE_CASE__ = tmp_path / """custom_extracted_path""" if default_extracted: SCREAMING_SNAKE_CASE__ = ("""downloads""" if default_cache_dir else custom_cache_dir, """extracted""") else: monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_DIR""" , UpperCamelCase__ ) monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_PATH""" , str(UpperCamelCase__ ) ) SCREAMING_SNAKE_CASE__ = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) SCREAMING_SNAKE_CASE__ = xz_file SCREAMING_SNAKE_CASE__ = ( DownloadConfig(extract_compressed_file=UpperCamelCase__ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=UpperCamelCase__ ) ) SCREAMING_SNAKE_CASE__ = cached_path(UpperCamelCase__ , download_config=UpperCamelCase__ ) assert Path(UpperCamelCase__ ).parent.parts[-2:] == expected def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[int] ): # absolute path SCREAMING_SNAKE_CASE__ = str(Path(UpperCamelCase__ ).resolve() ) assert cached_path(UpperCamelCase__ ) == text_file # relative path SCREAMING_SNAKE_CASE__ = str(Path(UpperCamelCase__ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(UpperCamelCase__ ) == text_file def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): # absolute path SCREAMING_SNAKE_CASE__ = str(tmp_path.resolve() / """__missing_file__.txt""" ) with pytest.raises(UpperCamelCase__ ): cached_path(UpperCamelCase__ ) # relative path SCREAMING_SNAKE_CASE__ = """./__missing_file__.txt""" with pytest.raises(UpperCamelCase__ ): cached_path(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): SCREAMING_SNAKE_CASE__ = get_from_cache(f'''tmp://{tmpfs_file}''' ) with open(UpperCamelCase__ ) as f: SCREAMING_SNAKE_CASE__ = f.read() assert output_file_content == FILE_CONTENT @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( ): with pytest.raises(UpperCamelCase__ ): cached_path("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(UpperCamelCase__ ): http_get("""https://huggingface.co""" , temp_file=UpperCamelCase__ ) with pytest.raises(UpperCamelCase__ ): http_head("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[Any] ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(UpperCamelCase__ ): ftp_get("""ftp://huggingface.co""" , temp_file=UpperCamelCase__ ) with pytest.raises(UpperCamelCase__ ): ftp_head("""ftp://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(UpperCamelCase__ ): fsspec_get("""s3://huggingface.co""" , temp_file=UpperCamelCase__ ) with pytest.raises(UpperCamelCase__ ): fsspec_head("""s3://huggingface.co""" )
59
1
from typing import Optional from .. import Features, NamedSplit from ..packaged_modules.text.text import Text from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class UpperCamelCase_ ( UpperCamelCase__ ): def __init__( self :Optional[Any] , __A :NestedDataStructureLike[PathLike] , __A :Optional[NamedSplit] = None , __A :Optional[Features] = None , __A :str = None , __A :bool = False , __A :bool = False , __A :Optional[int] = None , **__A :List[str] , ) -> int: """simple docstring""" super().__init__( __A , split=__A , features=__A , cache_dir=__A , keep_in_memory=__A , streaming=__A , num_proc=__A , **__A , ) SCREAMING_SNAKE_CASE__ = path_or_paths if isinstance(__A , __A ) else {self.split: path_or_paths} SCREAMING_SNAKE_CASE__ = Text( cache_dir=__A , data_files=__A , features=__A , **__A , ) def _snake_case ( self :List[str] ) -> int: """simple docstring""" if self.streaming: SCREAMING_SNAKE_CASE__ = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None self.builder.download_and_prepare( download_config=__A , download_mode=__A , verification_mode=__A , base_path=__A , num_proc=self.num_proc , ) SCREAMING_SNAKE_CASE__ = self.builder.as_dataset( split=self.split , verification_mode=__A , in_memory=self.keep_in_memory ) return dataset
59
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _lowerCamelCase = logging.getLogger(__name__) def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser( description="""Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset.""" ) parser.add_argument( """--dataset_name""" , type=UpperCamelCase__ , default="""wikitext""" , help="""Name of the training. Explore datasets at: hf.co/datasets.""" , ) parser.add_argument( """--dataset_config""" , type=UpperCamelCase__ , default="""wikitext-103-raw-v1""" , help="""Configuration name of the dataset.""" ) parser.add_argument( """--tokenizer_name_or_path""" , type=UpperCamelCase__ , default="""sayakpaul/unigram-tokenizer-wikitext""" , help="""Tokenizer identifier. Can be a local filepath or a Hub identifier.""" , ) parser.add_argument( """--shard_size""" , type=UpperCamelCase__ , default=1_000 , help="""Number of entries to go in a single shard.""" , ) parser.add_argument("""--split""" , type=UpperCamelCase__ , default="""train""" , choices=["""train""", """test""", """validation"""] ) parser.add_argument( """--limit""" , default=UpperCamelCase__ , type=UpperCamelCase__ , help="""Limit the number of shards (used for debugging).""" , ) parser.add_argument( """--max_length""" , type=UpperCamelCase__ , default=512 , help="""Maximum sequence length. For training on TPUs, it helps to have a maximum""" """ sequence length that is a multiple of 8.""" , ) parser.add_argument( """--output_dir""" , default="""tf-tpu""" , type=UpperCamelCase__ , help="""Output directory where the TFRecord shards will be saved. If the""" """ path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord""" """ shards will be directly saved to a Google Cloud Storage bucket.""" , ) SCREAMING_SNAKE_CASE__ = parser.parse_args() return args def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[Any] ): def fn(UpperCamelCase__: Any ): return tokenizer(examples["""text"""] ) return fn def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = [] for i in range(len(tokenized_data["""input_ids"""] ) ): SCREAMING_SNAKE_CASE__ = { """input_ids""": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["""input_ids"""][i] ) ), """attention_mask""": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["""attention_mask"""][i] ) ), } SCREAMING_SNAKE_CASE__ = tf.train.Features(feature=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = tf.train.Example(features=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = example.SerializeToString() records.append(UpperCamelCase__ ) return records def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): SCREAMING_SNAKE_CASE__ = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: SCREAMING_SNAKE_CASE__ = min(len(UpperCamelCase__ ) , args.limit ) SCREAMING_SNAKE_CASE__ = dataset.select(range(UpperCamelCase__ ) ) print(f'''Limiting the dataset to {args.limit} entries.''' ) SCREAMING_SNAKE_CASE__ = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) SCREAMING_SNAKE_CASE__ = os.path.join(args.output_dir , args.split ) if not os.path.exists(UpperCamelCase__ ): os.makedirs(UpperCamelCase__ ) else: SCREAMING_SNAKE_CASE__ = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. SCREAMING_SNAKE_CASE__ = tokenize_function(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = dataset.map(UpperCamelCase__ , batched=UpperCamelCase__ , num_proc=4 , remove_columns=["""text"""] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(UpperCamelCase__: int ): # Concatenate all texts. SCREAMING_SNAKE_CASE__ = {k: sum(examples[k] , [] ) for k in examples.keys()} SCREAMING_SNAKE_CASE__ = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 SCREAMING_SNAKE_CASE__ = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. SCREAMING_SNAKE_CASE__ = { k: [t[i : i + args.max_length] for i in range(0 , UpperCamelCase__ , args.max_length )] for k, t in concatenated_examples.items() } return result SCREAMING_SNAKE_CASE__ = dataset_tokenized.map(UpperCamelCase__ , batched=UpperCamelCase__ , batch_size=1_000 , num_proc=4 ) SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 0 for shard in range(0 , len(UpperCamelCase__ ) , args.shard_size ): SCREAMING_SNAKE_CASE__ = grouped_dataset[shard : shard + args.shard_size] SCREAMING_SNAKE_CASE__ = len(dataset_snapshot["""input_ids"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , f'''dataset-{shard_count}-{records_containing}.tfrecord''' ) SCREAMING_SNAKE_CASE__ = get_serialized_examples(UpperCamelCase__ ) with tf.io.TFRecordWriter(UpperCamelCase__ ) as out_file: for i in range(len(UpperCamelCase__ ) ): SCREAMING_SNAKE_CASE__ = serialized_examples[i] out_file.write(UpperCamelCase__ ) print("""Wrote file {} containing {} records""".format(UpperCamelCase__ , UpperCamelCase__ ) ) shard_count += 1 total_records += records_containing with open(f'''split-{args.split}-records-count.txt''' , """w""" ) as f: print(f'''Total {args.split} records: {total_records}''' , file=UpperCamelCase__ ) if __name__ == "__main__": _lowerCamelCase = parse_args() main(args)
59
1
from __future__ import absolute_import, division, print_function, unicode_literals from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers import RobertaConfig from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.roberta.modeling_roberta import ( ROBERTA_INPUTS_DOCSTRING, ROBERTA_START_DOCSTRING, RobertaEmbeddings, ) from .modeling_highway_bert import BertPreTrainedModel, DeeBertModel, HighwayException, entropy @add_start_docstrings( "The RoBERTa Model transformer with early exiting (DeeRoBERTa). " , UpperCamelCase__ , ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = RobertaConfig lowerCamelCase_ = "roberta" def __init__( self :Optional[Any] , __A :List[Any] ) -> Tuple: """simple docstring""" super().__init__(__A ) SCREAMING_SNAKE_CASE__ = RobertaEmbeddings(__A ) self.init_weights() @add_start_docstrings( "RoBERTa Model (with early exiting - DeeRoBERTa) with a classifier on top,\n also takes care of multi-layer training. " , UpperCamelCase__ , ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = RobertaConfig lowerCamelCase_ = "roberta" def __init__( self :Optional[Any] , __A :Union[str, Any] ) -> Any: """simple docstring""" super().__init__(__A ) SCREAMING_SNAKE_CASE__ = config.num_labels SCREAMING_SNAKE_CASE__ = config.num_hidden_layers SCREAMING_SNAKE_CASE__ = DeeRobertaModel(__A ) SCREAMING_SNAKE_CASE__ = nn.Dropout(config.hidden_dropout_prob ) SCREAMING_SNAKE_CASE__ = nn.Linear(config.hidden_size , self.config.num_labels ) @add_start_docstrings_to_model_forward(__A ) def _snake_case ( self :Union[str, Any] , __A :int=None , __A :str=None , __A :List[Any]=None , __A :Any=None , __A :List[Any]=None , __A :Optional[int]=None , __A :List[str]=None , __A :Union[str, Any]=-1 , __A :int=False , ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.num_layers try: SCREAMING_SNAKE_CASE__ = self.roberta( __A , attention_mask=__A , token_type_ids=__A , position_ids=__A , head_mask=__A , inputs_embeds=__A , ) SCREAMING_SNAKE_CASE__ = outputs[1] SCREAMING_SNAKE_CASE__ = self.dropout(__A ) SCREAMING_SNAKE_CASE__ = self.classifier(__A ) SCREAMING_SNAKE_CASE__ = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: SCREAMING_SNAKE_CASE__ = e.message SCREAMING_SNAKE_CASE__ = e.exit_layer SCREAMING_SNAKE_CASE__ = outputs[0] if not self.training: SCREAMING_SNAKE_CASE__ = entropy(__A ) SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = [] if labels is not None: if self.num_labels == 1: # We are doing regression SCREAMING_SNAKE_CASE__ = MSELoss() SCREAMING_SNAKE_CASE__ = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: SCREAMING_SNAKE_CASE__ = CrossEntropyLoss() SCREAMING_SNAKE_CASE__ = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits SCREAMING_SNAKE_CASE__ = [] for highway_exit in outputs[-1]: SCREAMING_SNAKE_CASE__ = highway_exit[0] if not self.training: highway_logits_all.append(__A ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression SCREAMING_SNAKE_CASE__ = MSELoss() SCREAMING_SNAKE_CASE__ = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: SCREAMING_SNAKE_CASE__ = CrossEntropyLoss() SCREAMING_SNAKE_CASE__ = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(__A ) if train_highway: SCREAMING_SNAKE_CASE__ = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: SCREAMING_SNAKE_CASE__ = (loss,) + outputs if not self.training: SCREAMING_SNAKE_CASE__ = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: SCREAMING_SNAKE_CASE__ = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), entropy
59
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] ): SCREAMING_SNAKE_CASE__ = [] for data in source_data: for i, el in enumerate(UpperCamelCase__ ): if len(UpperCamelCase__ ) < i + 1: data_lists.append([] ) data_lists[i].append(float(UpperCamelCase__ ) ) return data_lists def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] , UpperCamelCase__: list[int] ): SCREAMING_SNAKE_CASE__ = [] for dlist, weight in zip(UpperCamelCase__ , UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = min(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = max(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = [] # for weight 0 score is 1 - actual score if weight == 0: for item in dlist: try: score.append(1 - ((item - mind) / (maxd - mind)) ) except ZeroDivisionError: score.append(1 ) elif weight == 1: for item in dlist: try: score.append((item - mind) / (maxd - mind) ) except ZeroDivisionError: score.append(0 ) # weight not 0 or 1 else: SCREAMING_SNAKE_CASE__ = f'''Invalid weight of {weight:f} provided''' raise ValueError(UpperCamelCase__ ) score_lists.append(UpperCamelCase__ ) return score_lists def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] ): SCREAMING_SNAKE_CASE__ = [0 for i in range(len(score_lists[0] ) )] for slist in score_lists: for j, ele in enumerate(UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = final_scores[j] + ele return final_scores def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] , UpperCamelCase__: list[int] ): SCREAMING_SNAKE_CASE__ = get_data(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = calculate_each_score(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = generate_final_scores(UpperCamelCase__ ) # append scores to source data for i, ele in enumerate(UpperCamelCase__ ): source_data[i].append(UpperCamelCase__ ) return source_data
59
1
import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[str, Any] ): SCREAMING_SNAKE_CASE__ = [ """encoder.version""", """decoder.version""", """model.encoder.version""", """model.decoder.version""", """decoder.output_projection.weight""", """_float_tensor""", """encoder.embed_positions._float_tensor""", """decoder.embed_positions._float_tensor""", ] for k in ignore_keys: state_dict.pop(UpperCamelCase__ , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Tuple ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = emb.weight.shape SCREAMING_SNAKE_CASE__ = nn.Linear(UpperCamelCase__ , UpperCamelCase__ , bias=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = emb.weight.data return lin_layer def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): SCREAMING_SNAKE_CASE__ = torch.load(UpperCamelCase__ , map_location="""cpu""" ) SCREAMING_SNAKE_CASE__ = mam_aaa["""args"""] or mam_aaa["""cfg"""]["""model"""] SCREAMING_SNAKE_CASE__ = mam_aaa["""model"""] remove_ignore_keys_(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = state_dict["""encoder.embed_tokens.weight"""].shape[0] SCREAMING_SNAKE_CASE__ = MaMaaaConfig( vocab_size=UpperCamelCase__ , max_position_embeddings=1_024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function="""relu""" , ) SCREAMING_SNAKE_CASE__ = state_dict["""decoder.embed_tokens.weight"""] SCREAMING_SNAKE_CASE__ = MaMaaaForConditionalGeneration(UpperCamelCase__ ) model.model.load_state_dict(UpperCamelCase__ , strict=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument('fairseq_path', type=str, help='path to a model.pt on local filesystem.') parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') _lowerCamelCase = parser.parse_args() _lowerCamelCase = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
59
import warnings from functools import wraps from typing import Callable def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Callable ): @wraps(UpperCamelCase__ ) def _inner_fn(*UpperCamelCase__: Dict , **UpperCamelCase__: Any ): warnings.warn( (f'''\'{fn.__name__}\' is experimental and might be subject to breaking changes in the future.''') , UpperCamelCase__ , ) return fn(*UpperCamelCase__ , **UpperCamelCase__ ) return _inner_fn
59
1
import pytest _lowerCamelCase = '__dummy_dataset1__' _lowerCamelCase = '\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = "https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/"\nURLS = {"train": REPO_URL + "wikiann-bn-train.jsonl", "validation": REPO_URL + "wikiann-bn-validation.jsonl"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n "tokens": datasets.Sequence(datasets.Value("string")),\n "ner_tags": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n "O",\n "B-PER",\n "I-PER",\n "B-ORG",\n "I-ORG",\n "B-LOC",\n "I-LOC",\n ]\n )\n ),\n "langs": datasets.Sequence(datasets.Value("string")),\n "spans": datasets.Sequence(datasets.Value("string")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={"filepath": dl_path["train"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={"filepath": dl_path["validation"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, "r", encoding="utf-8") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n' @pytest.fixture def SCREAMING_SNAKE_CASE__ ( ): return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def SCREAMING_SNAKE_CASE__ ( ): return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] , UpperCamelCase__: Tuple , UpperCamelCase__: Dict ): SCREAMING_SNAKE_CASE__ = dataset_loading_script_name SCREAMING_SNAKE_CASE__ = tmp_path / """datasets""" / script_name script_dir.mkdir(parents=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = script_dir / f'''{script_name}.py''' with open(UpperCamelCase__ , """w""" ) as f: f.write(UpperCamelCase__ ) return str(UpperCamelCase__ )
59
import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = RoCBertTokenizer lowerCamelCase_ = None lowerCamelCase_ = False lowerCamelCase_ = True lowerCamelCase_ = filter_non_english def _snake_case ( self :List[Any] ) -> List[Any]: """simple docstring""" super().setUp() SCREAMING_SNAKE_CASE__ = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """你""", """好""", """是""", """谁""", """a""", """b""", """c""", """d"""] SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = {} for i, value in enumerate(__A ): SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""word_shape_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""word_pronunciation_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) with open(self.word_shape_file , """w""" , encoding="""utf-8""" ) as word_shape_writer: json.dump(__A , __A , ensure_ascii=__A ) with open(self.word_pronunciation_file , """w""" , encoding="""utf-8""" ) as word_pronunciation_writer: json.dump(__A , __A , ensure_ascii=__A ) def _snake_case ( self :List[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) SCREAMING_SNAKE_CASE__ = tokenizer.tokenize("""你好[SEP]你是谁""" ) self.assertListEqual(__A , ["""你""", """好""", """[SEP]""", """你""", """是""", """谁"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_shape_ids(__A ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_pronunciation_ids(__A ) , [5, 6, 2, 5, 7, 8] ) def _snake_case ( self :List[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer() self.assertListEqual(tokenizer.tokenize("""ah\u535A\u63A8zz""" ) , ["""ah""", """\u535A""", """\u63A8""", """zz"""] ) def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hällo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""h\u00E9llo"""] ) def _snake_case ( self :Any ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :List[str] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :Union[str, Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :int ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HäLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :Union[str, Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HaLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :List[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , never_split=["""[UNK]"""] ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? [UNK]""" ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?""", """[UNK]"""] ) def _snake_case ( self :Any ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing"""] SCREAMING_SNAKE_CASE__ = {} for i, token in enumerate(__A ): SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = RoCBertWordpieceTokenizer(vocab=__A , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""unwanted running""" ) , ["""un""", """##want""", """##ed""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.tokenize("""unwantedX running""" ) , ["""[UNK]""", """runn""", """##ing"""] ) def _snake_case ( self :Any ) -> str: """simple docstring""" self.assertTrue(_is_whitespace(""" """ ) ) self.assertTrue(_is_whitespace("""\t""" ) ) self.assertTrue(_is_whitespace("""\r""" ) ) self.assertTrue(_is_whitespace("""\n""" ) ) self.assertTrue(_is_whitespace("""\u00A0""" ) ) self.assertFalse(_is_whitespace("""A""" ) ) self.assertFalse(_is_whitespace("""-""" ) ) def _snake_case ( self :int ) -> str: """simple docstring""" self.assertTrue(_is_control("""\u0005""" ) ) self.assertFalse(_is_control("""A""" ) ) self.assertFalse(_is_control(""" """ ) ) self.assertFalse(_is_control("""\t""" ) ) self.assertFalse(_is_control("""\r""" ) ) def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" self.assertTrue(_is_punctuation("""-""" ) ) self.assertTrue(_is_punctuation("""$""" ) ) self.assertTrue(_is_punctuation("""`""" ) ) self.assertTrue(_is_punctuation(""".""" ) ) self.assertFalse(_is_punctuation("""A""" ) ) self.assertFalse(_is_punctuation(""" """ ) ) def _snake_case ( self :str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(__A ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) if self.test_rust_tokenizer: SCREAMING_SNAKE_CASE__ = self.get_rust_tokenizer() self.assertListEqual( [rust_tokenizer.tokenize(__A ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) def _snake_case ( self :int ) -> Any: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = f'''A, naïve {tokenizer_r.mask_token} AllenNLP sentence.''' SCREAMING_SNAKE_CASE__ = tokenizer_r.encode_plus( __A , return_attention_mask=__A , return_token_type_ids=__A , return_offsets_mapping=__A , add_special_tokens=__A , ) SCREAMING_SNAKE_CASE__ = tokenizer_r.do_lower_case if hasattr(__A , """do_lower_case""" ) else False SCREAMING_SNAKE_CASE__ = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """A"""), ((1, 2), ""","""), ((3, 5), """na"""), ((5, 6), """##ï"""), ((6, 8), """##ve"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """Allen"""), ((21, 23), """##NL"""), ((23, 24), """##P"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """a"""), ((1, 2), ""","""), ((3, 8), """naive"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """allen"""), ((21, 23), """##nl"""), ((23, 24), """##p"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["""input_ids"""] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["""offset_mapping"""] ) def _snake_case ( self :Optional[int] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""的""", """人""", """有"""] SCREAMING_SNAKE_CASE__ = """""".join(__A ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.convert_ids_to_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.convert_ids_to_tokens(__A ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(__A , __A ) self.assertListEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.convert_ids_to_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.convert_ids_to_tokens(__A ) # it is expected that only the first Chinese character is not preceded by "##". SCREAMING_SNAKE_CASE__ = [ f'''##{token}''' if idx != 0 else token for idx, token in enumerate(__A ) ] self.assertListEqual(__A , __A ) self.assertListEqual(__A , __A ) @slow def _snake_case ( self :Union[str, Any] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) SCREAMING_SNAKE_CASE__ = tokenizer.encode("""你好""" , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.encode("""你是谁""" , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.build_inputs_with_special_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.build_inputs_with_special_tokens(__A , __A ) assert encoded_sentence == [1] + text + [2] assert encoded_pair == [1] + text + [2] + text_a + [2] def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_tokenizers(do_lower_case=__A ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE__ = """你好,你是谁""" SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_shape_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_pronunciation_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.prepare_for_model( __A , __A , __A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.encode_plus(__A , add_special_tokens=__A ) self.assertEqual(__A , __A )
59
1
from __future__ import annotations _lowerCamelCase = [] def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[int]] , UpperCamelCase__: int , UpperCamelCase__: int ): for i in range(len(UpperCamelCase__ ) ): if board[row][i] == 1: return False for i in range(len(UpperCamelCase__ ) ): if board[i][column] == 1: return False for i, j in zip(range(UpperCamelCase__ , -1 , -1 ) , range(UpperCamelCase__ , -1 , -1 ) ): if board[i][j] == 1: return False for i, j in zip(range(UpperCamelCase__ , -1 , -1 ) , range(UpperCamelCase__ , len(UpperCamelCase__ ) ) ): if board[i][j] == 1: return False return True def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[int]] , UpperCamelCase__: int ): if row >= len(UpperCamelCase__ ): solution.append(UpperCamelCase__ ) printboard(UpperCamelCase__ ) print() return True for i in range(len(UpperCamelCase__ ) ): if is_safe(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = 1 solve(UpperCamelCase__ , row + 1 ) SCREAMING_SNAKE_CASE__ = 0 return False def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[int]] ): for i in range(len(UpperCamelCase__ ) ): for j in range(len(UpperCamelCase__ ) ): if board[i][j] == 1: print("""Q""" , end=""" """ ) else: print(""".""" , end=""" """ ) print() # n=int(input("The no. of queens")) _lowerCamelCase = 8 _lowerCamelCase = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print('The total no. of solutions are :', len(solution))
59
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[str, Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = original_name.split(""".""" )[0] SCREAMING_SNAKE_CASE__ = key.split(""".""" ) SCREAMING_SNAKE_CASE__ = int(key_list[key_list.index(UpperCamelCase__ ) - 2] ) SCREAMING_SNAKE_CASE__ = int(key_list[key_list.index(UpperCamelCase__ ) - 1] ) SCREAMING_SNAKE_CASE__ = orig_block_num - offset SCREAMING_SNAKE_CASE__ = key.replace(f'''{orig_block_num}.{layer_num}.{original_name}''' , f'''block.{new_block_num}.{layer_num}.{new_name}''' ) return key def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = OrderedDict() SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 0, 0 for key, value in state_dict.items(): if key.startswith("""network""" ): SCREAMING_SNAKE_CASE__ = key.replace("""network""" , """poolformer.encoder""" ) if "proj" in key: # Works for the first embedding as well as the internal embedding layers if key.endswith("""bias""" ) and "patch_embed" not in key: patch_emb_offset += 1 SCREAMING_SNAKE_CASE__ = key[: key.find("""proj""" )] SCREAMING_SNAKE_CASE__ = key.replace(UpperCamelCase__ , f'''patch_embeddings.{total_embed_found}.''' ) SCREAMING_SNAKE_CASE__ = key.replace("""proj""" , """projection""" ) if key.endswith("""bias""" ): total_embed_found += 1 if "patch_embeddings" in key: SCREAMING_SNAKE_CASE__ = """poolformer.encoder.""" + key if "mlp.fc1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """mlp.fc1""" , """output.conv1""" ) if "mlp.fc2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """mlp.fc2""" , """output.conv2""" ) if "norm1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """norm1""" , """before_norm""" ) if "norm2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """norm2""" , """after_norm""" ) if "layer_scale_1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """layer_scale_1""" , """layer_scale_1""" ) if "layer_scale_2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """layer_scale_2""" , """layer_scale_2""" ) if "head" in key: SCREAMING_SNAKE_CASE__ = key.replace("""head""" , """classifier""" ) SCREAMING_SNAKE_CASE__ = value return new_state_dict def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = """http://images.cocodataset.org/val2017/000000039769.jpg""" SCREAMING_SNAKE_CASE__ = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return image @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = PoolFormerConfig() # set attributes based on model_name SCREAMING_SNAKE_CASE__ = """huggingface/label-files""" SCREAMING_SNAKE_CASE__ = model_name[-3:] SCREAMING_SNAKE_CASE__ = 1_000 SCREAMING_SNAKE_CASE__ = """imagenet-1k-id2label.json""" SCREAMING_SNAKE_CASE__ = (1, 1_000) # set config attributes SCREAMING_SNAKE_CASE__ = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) SCREAMING_SNAKE_CASE__ = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE__ = idalabel SCREAMING_SNAKE_CASE__ = {v: k for k, v in idalabel.items()} if size == "s12": SCREAMING_SNAKE_CASE__ = [2, 2, 6, 2] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "s24": SCREAMING_SNAKE_CASE__ = [4, 4, 12, 4] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "s36": SCREAMING_SNAKE_CASE__ = [6, 6, 18, 6] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "m36": SCREAMING_SNAKE_CASE__ = [6, 6, 18, 6] SCREAMING_SNAKE_CASE__ = [96, 192, 384, 768] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9_5 elif size == "m48": SCREAMING_SNAKE_CASE__ = [8, 8, 24, 8] SCREAMING_SNAKE_CASE__ = [96, 192, 384, 768] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9_5 else: raise ValueError(f'''Size {size} not supported''' ) # load image processor SCREAMING_SNAKE_CASE__ = PoolFormerImageProcessor(crop_pct=UpperCamelCase__ ) # Prepare image SCREAMING_SNAKE_CASE__ = prepare_img() SCREAMING_SNAKE_CASE__ = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ).pixel_values logger.info(f'''Converting model {model_name}...''' ) # load original state dict SCREAMING_SNAKE_CASE__ = torch.load(UpperCamelCase__ , map_location=torch.device("""cpu""" ) ) # rename keys SCREAMING_SNAKE_CASE__ = rename_keys(UpperCamelCase__ ) # create HuggingFace model and load state dict SCREAMING_SNAKE_CASE__ = PoolFormerForImageClassification(UpperCamelCase__ ) model.load_state_dict(UpperCamelCase__ ) model.eval() # Define image processor SCREAMING_SNAKE_CASE__ = PoolFormerImageProcessor(crop_pct=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = image_processor(images=prepare_img() , return_tensors="""pt""" ).pixel_values # forward pass SCREAMING_SNAKE_CASE__ = model(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = outputs.logits # define expected logit slices for different models if size == "s12": SCREAMING_SNAKE_CASE__ = torch.tensor([-0.3_0_4_5, -0.6_7_5_8, -0.4_8_6_9] ) elif size == "s24": SCREAMING_SNAKE_CASE__ = torch.tensor([0.4_4_0_2, -0.1_3_7_4, -0.8_0_4_5] ) elif size == "s36": SCREAMING_SNAKE_CASE__ = torch.tensor([-0.6_0_8_0, -0.5_1_3_3, -0.5_8_9_8] ) elif size == "m36": SCREAMING_SNAKE_CASE__ = torch.tensor([0.3_9_5_2, 0.2_2_6_3, -1.2_6_6_8] ) elif size == "m48": SCREAMING_SNAKE_CASE__ = torch.tensor([0.1_1_6_7, -0.0_6_5_6, -0.3_4_2_3] ) else: raise ValueError(f'''Size {size} not supported''' ) # verify logits assert logits.shape == expected_shape assert torch.allclose(logits[0, :3] , UpperCamelCase__ , atol=1e-2 ) # finally, save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) model.save_pretrained(UpperCamelCase__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '--model_name', default='poolformer_s12', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) _lowerCamelCase = parser.parse_args() convert_poolformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
59
1
import argparse import collections import json from pathlib import Path import requests import torch import yaml from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTImageProcessor, MobileViTVaConfig, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, ) from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[Any] ): print("""Loading config file...""" ) def flatten_yaml_as_dict(UpperCamelCase__: Optional[Any] , UpperCamelCase__: List[Any]="" , UpperCamelCase__: Union[str, Any]="." ): SCREAMING_SNAKE_CASE__ = [] for k, v in d.items(): SCREAMING_SNAKE_CASE__ = parent_key + sep + k if parent_key else k if isinstance(UpperCamelCase__ , collections.abc.MutableMapping ): items.extend(flatten_yaml_as_dict(UpperCamelCase__ , UpperCamelCase__ , sep=UpperCamelCase__ ).items() ) else: items.append((new_key, v) ) return dict(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = argparse.Namespace() with open(UpperCamelCase__ , """r""" ) as yaml_file: try: SCREAMING_SNAKE_CASE__ = yaml.load(UpperCamelCase__ , Loader=yaml.FullLoader ) SCREAMING_SNAKE_CASE__ = flatten_yaml_as_dict(UpperCamelCase__ ) for k, v in flat_cfg.items(): setattr(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) except yaml.YAMLError as exc: logger.error("""Error while loading config file: {}. Error message: {}""".format(UpperCamelCase__ , str(UpperCamelCase__ ) ) ) return config def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str , UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = MobileViTVaConfig() SCREAMING_SNAKE_CASE__ = False # dataset if task_name.startswith("""imagenet1k_""" ): SCREAMING_SNAKE_CASE__ = 1_000 if int(task_name.strip().split("""_""" )[-1] ) == 384: SCREAMING_SNAKE_CASE__ = 384 else: SCREAMING_SNAKE_CASE__ = 256 SCREAMING_SNAKE_CASE__ = """imagenet-1k-id2label.json""" elif task_name.startswith("""imagenet21k_to_1k_""" ): SCREAMING_SNAKE_CASE__ = 21_000 if int(task_name.strip().split("""_""" )[-1] ) == 384: SCREAMING_SNAKE_CASE__ = 384 else: SCREAMING_SNAKE_CASE__ = 256 SCREAMING_SNAKE_CASE__ = """imagenet-22k-id2label.json""" elif task_name.startswith("""ade20k_""" ): SCREAMING_SNAKE_CASE__ = 151 SCREAMING_SNAKE_CASE__ = 512 SCREAMING_SNAKE_CASE__ = """ade20k-id2label.json""" SCREAMING_SNAKE_CASE__ = True elif task_name.startswith("""voc_""" ): SCREAMING_SNAKE_CASE__ = 21 SCREAMING_SNAKE_CASE__ = 512 SCREAMING_SNAKE_CASE__ = """pascal-voc-id2label.json""" SCREAMING_SNAKE_CASE__ = True # orig_config SCREAMING_SNAKE_CASE__ = load_orig_config_file(UpperCamelCase__ ) assert getattr(UpperCamelCase__ , """model.classification.name""" , -1 ) == "mobilevit_v2", "Invalid model" SCREAMING_SNAKE_CASE__ = getattr(UpperCamelCase__ , """model.classification.mitv2.width_multiplier""" , 1.0 ) assert ( getattr(UpperCamelCase__ , """model.classification.mitv2.attn_norm_layer""" , -1 ) == "layer_norm_2d" ), "Norm layers other than layer_norm_2d is not supported" SCREAMING_SNAKE_CASE__ = getattr(UpperCamelCase__ , """model.classification.activation.name""" , """swish""" ) # config.image_size == getattr(orig_config, 'sampler.bs.crop_size_width', 256) if is_segmentation_model: SCREAMING_SNAKE_CASE__ = getattr(UpperCamelCase__ , """model.segmentation.output_stride""" , 16 ) if "_deeplabv3" in task_name: SCREAMING_SNAKE_CASE__ = getattr(UpperCamelCase__ , """model.segmentation.deeplabv3.aspp_rates""" , [12, 24, 36] ) SCREAMING_SNAKE_CASE__ = getattr(UpperCamelCase__ , """model.segmentation.deeplabv3.aspp_out_channels""" , 512 ) SCREAMING_SNAKE_CASE__ = getattr(UpperCamelCase__ , """model.segmentation.deeplabv3.aspp_dropout""" , 0.1 ) # id2label SCREAMING_SNAKE_CASE__ = """huggingface/label-files""" SCREAMING_SNAKE_CASE__ = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) SCREAMING_SNAKE_CASE__ = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE__ = idalabel SCREAMING_SNAKE_CASE__ = {v: k for k, v in idalabel.items()} return config def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Dict , UpperCamelCase__: Dict , UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = dct.pop(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = val def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[Any] , UpperCamelCase__: Tuple=False ): if base_model: SCREAMING_SNAKE_CASE__ = """""" else: SCREAMING_SNAKE_CASE__ = """mobilevitv2.""" SCREAMING_SNAKE_CASE__ = [] for k in state_dict.keys(): if k[:8] == "encoder.": SCREAMING_SNAKE_CASE__ = k[8:] else: SCREAMING_SNAKE_CASE__ = k if ".block." in k: SCREAMING_SNAKE_CASE__ = k_new.replace(""".block.""" , """.""" ) if ".conv." in k: SCREAMING_SNAKE_CASE__ = k_new.replace(""".conv.""" , """.convolution.""" ) if ".norm." in k: SCREAMING_SNAKE_CASE__ = k_new.replace(""".norm.""" , """.normalization.""" ) if "conv_1." in k: SCREAMING_SNAKE_CASE__ = k_new.replace("""conv_1.""" , f'''{model_prefix}conv_stem.''' ) for i in [1, 2]: if f'''layer_{i}.''' in k: SCREAMING_SNAKE_CASE__ = k_new.replace(f'''layer_{i}.''' , f'''{model_prefix}encoder.layer.{i-1}.layer.''' ) if ".exp_1x1." in k: SCREAMING_SNAKE_CASE__ = k_new.replace(""".exp_1x1.""" , """.expand_1x1.""" ) if ".red_1x1." in k: SCREAMING_SNAKE_CASE__ = k_new.replace(""".red_1x1.""" , """.reduce_1x1.""" ) for i in [3, 4, 5]: if f'''layer_{i}.0.''' in k: SCREAMING_SNAKE_CASE__ = k_new.replace(f'''layer_{i}.0.''' , f'''{model_prefix}encoder.layer.{i-1}.downsampling_layer.''' ) if f'''layer_{i}.1.local_rep.0.''' in k: SCREAMING_SNAKE_CASE__ = k_new.replace(f'''layer_{i}.1.local_rep.0.''' , f'''{model_prefix}encoder.layer.{i-1}.conv_kxk.''' ) if f'''layer_{i}.1.local_rep.1.''' in k: SCREAMING_SNAKE_CASE__ = k_new.replace(f'''layer_{i}.1.local_rep.1.''' , f'''{model_prefix}encoder.layer.{i-1}.conv_1x1.''' ) for i in [3, 4, 5]: if i == 3: SCREAMING_SNAKE_CASE__ = [0, 1] elif i == 4: SCREAMING_SNAKE_CASE__ = [0, 1, 2, 3] elif i == 5: SCREAMING_SNAKE_CASE__ = [0, 1, 2] for j in j_in: if f'''layer_{i}.1.global_rep.{j}.''' in k: SCREAMING_SNAKE_CASE__ = k_new.replace( f'''layer_{i}.1.global_rep.{j}.''' , f'''{model_prefix}encoder.layer.{i-1}.transformer.layer.{j}.''' ) if f'''layer_{i}.1.global_rep.{j+1}.''' in k: SCREAMING_SNAKE_CASE__ = k_new.replace( f'''layer_{i}.1.global_rep.{j+1}.''' , f'''{model_prefix}encoder.layer.{i-1}.layernorm.''' ) if f'''layer_{i}.1.conv_proj.''' in k: SCREAMING_SNAKE_CASE__ = k_new.replace(f'''layer_{i}.1.conv_proj.''' , f'''{model_prefix}encoder.layer.{i-1}.conv_projection.''' ) if "pre_norm_attn.0." in k: SCREAMING_SNAKE_CASE__ = k_new.replace("""pre_norm_attn.0.""" , """layernorm_before.""" ) if "pre_norm_attn.1." in k: SCREAMING_SNAKE_CASE__ = k_new.replace("""pre_norm_attn.1.""" , """attention.""" ) if "pre_norm_ffn.0." in k: SCREAMING_SNAKE_CASE__ = k_new.replace("""pre_norm_ffn.0.""" , """layernorm_after.""" ) if "pre_norm_ffn.1." in k: SCREAMING_SNAKE_CASE__ = k_new.replace("""pre_norm_ffn.1.""" , """ffn.conv1.""" ) if "pre_norm_ffn.3." in k: SCREAMING_SNAKE_CASE__ = k_new.replace("""pre_norm_ffn.3.""" , """ffn.conv2.""" ) if "classifier.1." in k: SCREAMING_SNAKE_CASE__ = k_new.replace("""classifier.1.""" , """classifier.""" ) if "seg_head." in k: SCREAMING_SNAKE_CASE__ = k_new.replace("""seg_head.""" , """segmentation_head.""" ) if ".aspp_layer." in k: SCREAMING_SNAKE_CASE__ = k_new.replace(""".aspp_layer.""" , """.""" ) if ".aspp_pool." in k: SCREAMING_SNAKE_CASE__ = k_new.replace(""".aspp_pool.""" , """.""" ) rename_keys.append((k, k_new) ) return rename_keys def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = [] for k in state_dict.keys(): if k.startswith("""seg_head.aux_head.""" ): keys_to_ignore.append(UpperCamelCase__ ) for k in keys_to_ignore: state_dict.pop(UpperCamelCase__ , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = """http://images.cocodataset.org/val2017/000000039769.jpg""" # url = "https://cdn.britannica.com/86/141086-050-9D7C75EE/Gulfstream-G450-business-jet-passengers.jpg" SCREAMING_SNAKE_CASE__ = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Dict , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Union[str, Any] ): SCREAMING_SNAKE_CASE__ = get_mobilevitva_config(UpperCamelCase__ , UpperCamelCase__ ) # load original state_dict SCREAMING_SNAKE_CASE__ = torch.load(UpperCamelCase__ , map_location="""cpu""" ) # load huggingface model if task_name.startswith("""ade20k_""" ) or task_name.startswith("""voc_""" ): SCREAMING_SNAKE_CASE__ = MobileViTVaForSemanticSegmentation(UpperCamelCase__ ).eval() SCREAMING_SNAKE_CASE__ = False else: SCREAMING_SNAKE_CASE__ = MobileViTVaForImageClassification(UpperCamelCase__ ).eval() SCREAMING_SNAKE_CASE__ = False # remove and rename some keys of load the original model SCREAMING_SNAKE_CASE__ = checkpoint remove_unused_keys(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = create_rename_keys(UpperCamelCase__ , base_model=UpperCamelCase__ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # load modified state_dict model.load_state_dict(UpperCamelCase__ ) # Check outputs on an image, prepared by MobileViTImageProcessor SCREAMING_SNAKE_CASE__ = MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 32 ) SCREAMING_SNAKE_CASE__ = image_processor(images=prepare_img() , return_tensors="""pt""" ) SCREAMING_SNAKE_CASE__ = model(**UpperCamelCase__ ) # verify classification model if task_name.startswith("""imagenet""" ): SCREAMING_SNAKE_CASE__ = outputs.logits SCREAMING_SNAKE_CASE__ = logits.argmax(-1 ).item() print("""Predicted class:""" , model.config.idalabel[predicted_class_idx] ) if task_name.startswith("""imagenet1k_256""" ) and config.width_multiplier == 1.0: # expected_logits for base variant SCREAMING_SNAKE_CASE__ = torch.tensor([-1.6_3_3_6e0_0, -7.3_2_0_4e-0_2, -5.1_8_8_3e-0_1] ) assert torch.allclose(logits[0, :3] , UpperCamelCase__ , atol=1e-4 ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) print(f'''Saving model {task_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--task', default='imagenet1k_256', type=str, help=( 'Name of the task for which the MobileViTV2 model you\'d like to convert is trained on . ' '\n Classification (ImageNet-1k)\n - MobileViTV2 (256x256) : imagenet1k_256\n - MobileViTV2 (Trained on 256x256 and Finetuned on 384x384) : imagenet1k_384\n - MobileViTV2 (Trained on ImageNet-21k and Finetuned on ImageNet-1k 256x256) :\n imagenet21k_to_1k_256\n - MobileViTV2 (Trained on ImageNet-21k, Finetuned on ImageNet-1k 256x256, and Finetuned on\n ImageNet-1k 384x384) : imagenet21k_to_1k_384\n Segmentation\n - ADE20K Dataset : ade20k_deeplabv3\n - Pascal VOC 2012 Dataset: voc_deeplabv3\n ' ), choices=[ 'imagenet1k_256', 'imagenet1k_384', 'imagenet21k_to_1k_256', 'imagenet21k_to_1k_384', 'ade20k_deeplabv3', 'voc_deeplabv3', ], ) parser.add_argument( '--orig_checkpoint_path', required=True, type=str, help='Path to the original state dict (.pt file).' ) parser.add_argument('--orig_config_path', required=True, type=str, help='Path to the original config file.') parser.add_argument( '--pytorch_dump_folder_path', required=True, type=str, help='Path to the output PyTorch model directory.' ) _lowerCamelCase = parser.parse_args() convert_mobilevitva_checkpoint( args.task, args.orig_checkpoint_path, args.orig_config_path, args.pytorch_dump_folder_path )
59
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'google/pix2struct-textcaps-base': ( 'https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json' ), } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct_text_model" lowerCamelCase_ = ["past_key_values"] lowerCamelCase_ = { "hidden_size": "hidden_size", "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self :Union[str, Any] , __A :Any=5_0244 , __A :Optional[Any]=768 , __A :Tuple=64 , __A :List[str]=2048 , __A :int=12 , __A :str=12 , __A :Any=32 , __A :Tuple=128 , __A :int=0.1 , __A :str=1E-6 , __A :Optional[Any]=1.0 , __A :Union[str, Any]="gelu_new" , __A :Any=0 , __A :List[str]=False , __A :Optional[Any]=0 , __A :int=1 , __A :Optional[int]=False , __A :Optional[Any]=True , **__A :List[Any] , ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = d_kv SCREAMING_SNAKE_CASE__ = d_ff SCREAMING_SNAKE_CASE__ = num_layers SCREAMING_SNAKE_CASE__ = num_heads SCREAMING_SNAKE_CASE__ = relative_attention_num_buckets SCREAMING_SNAKE_CASE__ = relative_attention_max_distance SCREAMING_SNAKE_CASE__ = dropout_rate SCREAMING_SNAKE_CASE__ = layer_norm_epsilon SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = use_cache SCREAMING_SNAKE_CASE__ = eos_token_id SCREAMING_SNAKE_CASE__ = decoder_start_token_id # for backwards compatibility SCREAMING_SNAKE_CASE__ = dense_act_fn super().__init__( pad_token_id=__A , eos_token_id=__A , decoder_start_token_id=__A , tie_word_embeddings=__A , is_decoder=__A , **__A , ) @classmethod def _snake_case ( cls :Optional[int] , __A :Union[str, os.PathLike] , **__A :Optional[int] ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = cls.get_config_dict(__A , **__A ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get("""model_type""" ) == "pix2struct": SCREAMING_SNAKE_CASE__ = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct_vision_model" def __init__( self :Optional[int] , __A :int=768 , __A :Optional[Any]=768 , __A :Union[str, Any]=2048 , __A :int=64 , __A :Union[str, Any]=12 , __A :str=12 , __A :Any="gelu_new" , __A :List[Any]=1E-6 , __A :Dict=0.0 , __A :int=0.0 , __A :int=1E-10 , __A :Dict=1.0 , __A :int=4096 , __A :int=32 , __A :int=128 , **__A :Tuple , ) -> str: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = patch_embed_hidden_size SCREAMING_SNAKE_CASE__ = d_ff SCREAMING_SNAKE_CASE__ = dropout_rate SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = attention_dropout SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = dense_act_fn SCREAMING_SNAKE_CASE__ = seq_len SCREAMING_SNAKE_CASE__ = relative_attention_num_buckets SCREAMING_SNAKE_CASE__ = relative_attention_max_distance SCREAMING_SNAKE_CASE__ = d_kv @classmethod def _snake_case ( cls :str , __A :Union[str, os.PathLike] , **__A :str ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = cls.get_config_dict(__A , **__A ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get("""model_type""" ) == "pix2struct": SCREAMING_SNAKE_CASE__ = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct" lowerCamelCase_ = True def __init__( self :str , __A :Optional[Any]=None , __A :List[str]=None , __A :Optional[Any]=1.0 , __A :Optional[Any]=0.0_2 , __A :Any=False , __A :Tuple=False , __A :Any=True , **__A :Dict , ) -> Union[str, Any]: """simple docstring""" super().__init__(tie_word_embeddings=__A , is_encoder_decoder=__A , **__A ) if text_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""text_config is None. Initializing the Pix2StructTextConfig with default values.""" ) if vision_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""vision_config is None. Initializing the Pix2StructVisionConfig with default values.""" ) SCREAMING_SNAKE_CASE__ = PixaStructTextConfig(**__A ) SCREAMING_SNAKE_CASE__ = PixaStructVisionConfig(**__A ) SCREAMING_SNAKE_CASE__ = self.text_config.decoder_start_token_id SCREAMING_SNAKE_CASE__ = self.text_config.pad_token_id SCREAMING_SNAKE_CASE__ = self.text_config.eos_token_id SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = self.initializer_range SCREAMING_SNAKE_CASE__ = self.initializer_range SCREAMING_SNAKE_CASE__ = is_vqa @classmethod def _snake_case ( cls :Union[str, Any] , __A :PixaStructTextConfig , __A :PixaStructVisionConfig , **__A :Optional[int] ) -> Optional[Any]: """simple docstring""" return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__A ) def _snake_case ( self :str ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = copy.deepcopy(self.__dict__ ) SCREAMING_SNAKE_CASE__ = self.text_config.to_dict() SCREAMING_SNAKE_CASE__ = self.vision_config.to_dict() SCREAMING_SNAKE_CASE__ = self.__class__.model_type return output
59
1
import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 _lowerCamelCase = data_utils.TransfoXLTokenizer _lowerCamelCase = data_utils.TransfoXLCorpus _lowerCamelCase = data_utils _lowerCamelCase = data_utils def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Dict , UpperCamelCase__: Any , UpperCamelCase__: Any , UpperCamelCase__: Tuple ): if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(UpperCamelCase__ , """rb""" ) as fp: SCREAMING_SNAKE_CASE__ = pickle.load(UpperCamelCase__ , encoding="""latin1""" ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) SCREAMING_SNAKE_CASE__ = pytorch_dump_folder_path + """/""" + VOCAB_FILES_NAMES["""pretrained_vocab_file"""] print(f'''Save vocabulary to {pytorch_vocab_dump_path}''' ) SCREAMING_SNAKE_CASE__ = corpus.vocab.__dict__ torch.save(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = corpus.__dict__ corpus_dict_no_vocab.pop("""vocab""" , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = pytorch_dump_folder_path + """/""" + CORPUS_NAME print(f'''Save dataset to {pytorch_dataset_dump_path}''' ) torch.save(UpperCamelCase__ , UpperCamelCase__ ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model SCREAMING_SNAKE_CASE__ = os.path.abspath(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = os.path.abspath(UpperCamelCase__ ) print(f'''Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.''' ) # Initialise PyTorch model if transfo_xl_config_file == "": SCREAMING_SNAKE_CASE__ = TransfoXLConfig() else: SCREAMING_SNAKE_CASE__ = TransfoXLConfig.from_json_file(UpperCamelCase__ ) print(f'''Building PyTorch model from configuration: {config}''' ) SCREAMING_SNAKE_CASE__ = TransfoXLLMHeadModel(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = load_tf_weights_in_transfo_xl(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # Save pytorch-model SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) print(f'''Save PyTorch model to {os.path.abspath(UpperCamelCase__ )}''' ) torch.save(model.state_dict() , UpperCamelCase__ ) print(f'''Save configuration file to {os.path.abspath(UpperCamelCase__ )}''' ) with open(UpperCamelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--tf_checkpoint_path', default='', type=str, help='An optional path to a TensorFlow checkpoint path to be converted.', ) parser.add_argument( '--transfo_xl_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained BERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--transfo_xl_dataset_file', default='', type=str, help='An optional dataset file to be converted in a vocabulary.', ) _lowerCamelCase = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
59
import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = inspect.getfile(accelerate.test_utils ) SCREAMING_SNAKE_CASE__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_script.py"""] ) SCREAMING_SNAKE_CASE__ = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_distributed_data_loop.py"""] ) SCREAMING_SNAKE_CASE__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_ops.py"""] ) @require_multi_gpu def _snake_case ( self :Optional[Any] ) -> Tuple: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Tuple ) -> Optional[Any]: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.operation_file_path] print(f'''Command: {cmd}''' ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Optional[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Optional[int] ) -> str: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices, using 2 devices only''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices="""0,1""" ): execute_subprocess_async(__A , env=os.environ.copy() ) if __name__ == "__main__": _lowerCamelCase = Accelerator() _lowerCamelCase = (accelerator.state.process_index + 2, 10) _lowerCamelCase = torch.randint(0, 10, shape).to(accelerator.device) _lowerCamelCase = '' _lowerCamelCase = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." _lowerCamelCase = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." _lowerCamelCase = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
59
1
import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class UpperCamelCase_ ( UpperCamelCase__ ): def __init__( self :Optional[int] , __A :Any , __A :Union[str, Any]=13 , __A :List[Any]=7 , __A :List[str]=True , __A :List[Any]=True , __A :str=False , __A :Union[str, Any]=True , __A :Any=99 , __A :List[Any]=32 , __A :int=5 , __A :Union[str, Any]=4 , __A :Tuple=64 , __A :Tuple="gelu" , __A :List[Any]=0.1 , __A :Optional[Any]=0.1 , __A :int=512 , __A :int=16 , __A :Optional[int]=2 , __A :Tuple=0.0_2 , __A :Union[str, Any]=3 , __A :Optional[int]=4 , __A :List[str]=None , __A :Any=2 , __A :Optional[Any]=2 , __A :Tuple=2 , __A :Optional[Any]=2 , __A :Optional[Any]=4 , __A :Dict=1 , ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = parent SCREAMING_SNAKE_CASE__ = batch_size SCREAMING_SNAKE_CASE__ = seq_length SCREAMING_SNAKE_CASE__ = is_training SCREAMING_SNAKE_CASE__ = use_input_mask SCREAMING_SNAKE_CASE__ = use_token_type_ids SCREAMING_SNAKE_CASE__ = use_labels SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = type_vocab_size SCREAMING_SNAKE_CASE__ = type_sequence_label_size SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = num_labels SCREAMING_SNAKE_CASE__ = num_choices SCREAMING_SNAKE_CASE__ = scope SCREAMING_SNAKE_CASE__ = q_groups SCREAMING_SNAKE_CASE__ = k_groups SCREAMING_SNAKE_CASE__ = v_groups SCREAMING_SNAKE_CASE__ = post_attention_groups SCREAMING_SNAKE_CASE__ = intermediate_groups SCREAMING_SNAKE_CASE__ = output_groups def _snake_case ( self :List[Any] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE__ = None if self.use_input_mask: SCREAMING_SNAKE_CASE__ = random_attention_mask([self.batch_size, self.seq_length] ) SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None if self.use_labels: SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size] , self.num_choices ) SCREAMING_SNAKE_CASE__ = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _snake_case ( self :Any ) -> Optional[Any]: """simple docstring""" return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def _snake_case ( self :Tuple , __A :Dict , __A :Union[str, Any] , __A :Any , __A :Union[str, Any] , __A :Optional[int] , __A :Optional[Any] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = SqueezeBertModel(config=__A ) model.to(__A ) model.eval() SCREAMING_SNAKE_CASE__ = model(__A , __A ) SCREAMING_SNAKE_CASE__ = model(__A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _snake_case ( self :int , __A :Dict , __A :Optional[Any] , __A :List[Any] , __A :str , __A :str , __A :int ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = SqueezeBertForMaskedLM(config=__A ) model.to(__A ) model.eval() SCREAMING_SNAKE_CASE__ = model(__A , attention_mask=__A , labels=__A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _snake_case ( self :int , __A :str , __A :Optional[Any] , __A :Tuple , __A :Optional[Any] , __A :Optional[Any] , __A :str ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = SqueezeBertForQuestionAnswering(config=__A ) model.to(__A ) model.eval() SCREAMING_SNAKE_CASE__ = model( __A , attention_mask=__A , start_positions=__A , end_positions=__A ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _snake_case ( self :int , __A :Optional[Any] , __A :int , __A :str , __A :int , __A :List[Any] , __A :str ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.num_labels SCREAMING_SNAKE_CASE__ = SqueezeBertForSequenceClassification(__A ) model.to(__A ) model.eval() SCREAMING_SNAKE_CASE__ = model(__A , attention_mask=__A , labels=__A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _snake_case ( self :Union[str, Any] , __A :Tuple , __A :List[Any] , __A :int , __A :List[str] , __A :Any , __A :Optional[Any] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.num_labels SCREAMING_SNAKE_CASE__ = SqueezeBertForTokenClassification(config=__A ) model.to(__A ) model.eval() SCREAMING_SNAKE_CASE__ = model(__A , attention_mask=__A , labels=__A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _snake_case ( self :Any , __A :Any , __A :List[str] , __A :List[Any] , __A :int , __A :str , __A :Optional[int] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.num_choices SCREAMING_SNAKE_CASE__ = SqueezeBertForMultipleChoice(config=__A ) model.to(__A ) model.eval() SCREAMING_SNAKE_CASE__ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() SCREAMING_SNAKE_CASE__ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() SCREAMING_SNAKE_CASE__ = model( __A , attention_mask=__A , labels=__A , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _snake_case ( self :Optional[int] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.prepare_config_and_inputs() ((SCREAMING_SNAKE_CASE__) , (SCREAMING_SNAKE_CASE__) , (SCREAMING_SNAKE_CASE__) , (SCREAMING_SNAKE_CASE__) , (SCREAMING_SNAKE_CASE__) , (SCREAMING_SNAKE_CASE__)) = config_and_inputs SCREAMING_SNAKE_CASE__ = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class UpperCamelCase_ ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) lowerCamelCase_ = ( { "feature-extraction": SqueezeBertModel, "fill-mask": SqueezeBertForMaskedLM, "question-answering": SqueezeBertForQuestionAnswering, "text-classification": SqueezeBertForSequenceClassification, "token-classification": SqueezeBertForTokenClassification, "zero-shot": SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) lowerCamelCase_ = False lowerCamelCase_ = True lowerCamelCase_ = False def _snake_case ( self :List[Any] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = SqueezeBertModelTester(self ) SCREAMING_SNAKE_CASE__ = ConfigTester(self , config_class=__A , dim=37 ) def _snake_case ( self :List[Any] ) -> int: """simple docstring""" self.config_tester.run_common_tests() def _snake_case ( self :Dict ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*__A ) def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*__A ) def _snake_case ( self :List[str] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*__A ) def _snake_case ( self :List[Any] ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*__A ) def _snake_case ( self :Union[str, Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*__A ) def _snake_case ( self :Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*__A ) @slow def _snake_case ( self :Dict ) -> Union[str, Any]: """simple docstring""" for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE__ = SqueezeBertModel.from_pretrained(__A ) self.assertIsNotNone(__A ) @require_sentencepiece @require_tokenizers @require_torch class UpperCamelCase_ ( unittest.TestCase ): @slow def _snake_case ( self :str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = SqueezeBertForSequenceClassification.from_pretrained("""squeezebert/squeezebert-mnli""" ) SCREAMING_SNAKE_CASE__ = torch.tensor([[1, 2_9414, 232, 328, 740, 1140, 1_2695, 69, 13, 1588, 2]] ) SCREAMING_SNAKE_CASE__ = model(__A )[0] SCREAMING_SNAKE_CASE__ = torch.Size((1, 3) ) self.assertEqual(output.shape , __A ) SCREAMING_SNAKE_CASE__ = torch.tensor([[0.6_4_0_1, -0.0_3_4_9, -0.6_0_4_1]] ) self.assertTrue(torch.allclose(__A , __A , atol=1E-4 ) )
59
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _lowerCamelCase = { 'configuration_layoutlmv3': [ 'LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv3Config', 'LayoutLMv3OnnxConfig', ], 'processing_layoutlmv3': ['LayoutLMv3Processor'], 'tokenization_layoutlmv3': ['LayoutLMv3Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['LayoutLMv3TokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv3ForQuestionAnswering', 'LayoutLMv3ForSequenceClassification', 'LayoutLMv3ForTokenClassification', 'LayoutLMv3Model', 'LayoutLMv3PreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLayoutLMv3ForQuestionAnswering', 'TFLayoutLMv3ForSequenceClassification', 'TFLayoutLMv3ForTokenClassification', 'TFLayoutLMv3Model', 'TFLayoutLMv3PreTrainedModel', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['LayoutLMv3FeatureExtractor'] _lowerCamelCase = ['LayoutLMv3ImageProcessor'] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'weiweishi/roc-bert-base-zh': 'https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json', } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "roc_bert" def __init__( self :List[str] , __A :Union[str, Any]=3_0522 , __A :Dict=768 , __A :Dict=12 , __A :Union[str, Any]=12 , __A :Union[str, Any]=3072 , __A :List[str]="gelu" , __A :Dict=0.1 , __A :Any=0.1 , __A :Optional[Any]=512 , __A :List[str]=2 , __A :Dict=0.0_2 , __A :List[str]=1E-12 , __A :Optional[Any]=True , __A :Any=0 , __A :str="absolute" , __A :Union[str, Any]=None , __A :Dict=True , __A :List[Any]=True , __A :Any=768 , __A :str=910 , __A :int=512 , __A :Optional[Any]=2_4858 , __A :Optional[Any]=True , **__A :Dict , ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = type_vocab_size SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = use_cache SCREAMING_SNAKE_CASE__ = enable_pronunciation SCREAMING_SNAKE_CASE__ = enable_shape SCREAMING_SNAKE_CASE__ = pronunciation_embed_dim SCREAMING_SNAKE_CASE__ = pronunciation_vocab_size SCREAMING_SNAKE_CASE__ = shape_embed_dim SCREAMING_SNAKE_CASE__ = shape_vocab_size SCREAMING_SNAKE_CASE__ = concat_input SCREAMING_SNAKE_CASE__ = position_embedding_type SCREAMING_SNAKE_CASE__ = classifier_dropout super().__init__(pad_token_id=__A , **__A )
59
import inspect import unittest class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :str ) -> Union[str, Any]: """simple docstring""" try: import diffusers # noqa: F401 except ImportError: assert False def _snake_case ( self :Any ) -> Any: """simple docstring""" import diffusers from diffusers.dependency_versions_table import deps SCREAMING_SNAKE_CASE__ = inspect.getmembers(__A , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": SCREAMING_SNAKE_CASE__ = """k-diffusion""" elif backend == "invisible_watermark": SCREAMING_SNAKE_CASE__ = """invisible-watermark""" assert backend in deps, f'''{backend} is not in the deps table!'''
59
1
from __future__ import annotations from typing import Dict from ...configuration_utils import PretrainedConfig _lowerCamelCase = { 'susnato/ernie-m-base_pytorch': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/config.json', 'susnato/ernie-m-large_pytorch': 'https://huggingface.co/susnato/ernie-m-large_pytorch/blob/main/config.json', } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "ernie_m" lowerCamelCase_ = {"dropout": "classifier_dropout", "num_classes": "num_labels"} def __init__( self :str , __A :int = 25_0002 , __A :int = 768 , __A :int = 12 , __A :int = 12 , __A :int = 3072 , __A :str = "gelu" , __A :float = 0.1 , __A :float = 0.1 , __A :int = 514 , __A :float = 0.0_2 , __A :int = 1 , __A :float = 1E-05 , __A :Any=None , __A :Dict=False , __A :Dict=0.0 , **__A :Dict , ) -> Dict: """simple docstring""" super().__init__(pad_token_id=__A , **__A ) SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = classifier_dropout SCREAMING_SNAKE_CASE__ = is_decoder SCREAMING_SNAKE_CASE__ = act_dropout
59
import warnings from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch from ...models import UNetaDModel from ...schedulers import RePaintScheduler from ...utils import PIL_INTERPOLATION, logging, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput _lowerCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[List, PIL.Image.Image, torch.Tensor] ): warnings.warn( """The preprocess method is deprecated and will be removed in a future version. Please""" """ use VaeImageProcessor.preprocess instead""" , UpperCamelCase__ , ) if isinstance(UpperCamelCase__ , torch.Tensor ): return image elif isinstance(UpperCamelCase__ , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ = [image] if isinstance(image[0] , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = image[0].size SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = (x - x % 8 for x in (w, h)) # resize to integer multiple of 8 SCREAMING_SNAKE_CASE__ = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION["""lanczos"""] ) )[None, :] for i in image] SCREAMING_SNAKE_CASE__ = np.concatenate(UpperCamelCase__ , axis=0 ) SCREAMING_SNAKE_CASE__ = np.array(UpperCamelCase__ ).astype(np.floataa ) / 2_5_5.0 SCREAMING_SNAKE_CASE__ = image.transpose(0 , 3 , 1 , 2 ) SCREAMING_SNAKE_CASE__ = 2.0 * image - 1.0 SCREAMING_SNAKE_CASE__ = torch.from_numpy(UpperCamelCase__ ) elif isinstance(image[0] , torch.Tensor ): SCREAMING_SNAKE_CASE__ = torch.cat(UpperCamelCase__ , dim=0 ) return image def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[List, PIL.Image.Image, torch.Tensor] ): if isinstance(UpperCamelCase__ , torch.Tensor ): return mask elif isinstance(UpperCamelCase__ , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ = [mask] if isinstance(mask[0] , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = mask[0].size SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 SCREAMING_SNAKE_CASE__ = [np.array(m.convert("""L""" ).resize((w, h) , resample=PIL_INTERPOLATION["""nearest"""] ) )[None, :] for m in mask] SCREAMING_SNAKE_CASE__ = np.concatenate(UpperCamelCase__ , axis=0 ) SCREAMING_SNAKE_CASE__ = mask.astype(np.floataa ) / 2_5_5.0 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = torch.from_numpy(UpperCamelCase__ ) elif isinstance(mask[0] , torch.Tensor ): SCREAMING_SNAKE_CASE__ = torch.cat(UpperCamelCase__ , dim=0 ) return mask class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = 42 lowerCamelCase_ = 42 def __init__( self :Any , __A :List[Any] , __A :Optional[Any] ) -> Union[str, Any]: """simple docstring""" super().__init__() self.register_modules(unet=__A , scheduler=__A ) @torch.no_grad() def __call__( self :str , __A :Union[torch.Tensor, PIL.Image.Image] , __A :Union[torch.Tensor, PIL.Image.Image] , __A :int = 250 , __A :float = 0.0 , __A :int = 10 , __A :int = 10 , __A :Optional[Union[torch.Generator, List[torch.Generator]]] = None , __A :Optional[str] = "pil" , __A :bool = True , ) -> Union[ImagePipelineOutput, Tuple]: """simple docstring""" SCREAMING_SNAKE_CASE__ = image SCREAMING_SNAKE_CASE__ = _preprocess_image(__A ) SCREAMING_SNAKE_CASE__ = original_image.to(device=self.device , dtype=self.unet.dtype ) SCREAMING_SNAKE_CASE__ = _preprocess_mask(__A ) SCREAMING_SNAKE_CASE__ = mask_image.to(device=self.device , dtype=self.unet.dtype ) SCREAMING_SNAKE_CASE__ = original_image.shape[0] # sample gaussian noise to begin the loop if isinstance(__A , __A ) and len(__A ) != batch_size: raise ValueError( f'''You have passed a list of generators of length {len(__A )}, but requested an effective batch''' f''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) SCREAMING_SNAKE_CASE__ = original_image.shape SCREAMING_SNAKE_CASE__ = randn_tensor(__A , generator=__A , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(__A , __A , __A , self.device ) SCREAMING_SNAKE_CASE__ = eta SCREAMING_SNAKE_CASE__ = self.scheduler.timesteps[0] + 1 SCREAMING_SNAKE_CASE__ = generator[0] if isinstance(__A , __A ) else generator for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): if t < t_last: # predict the noise residual SCREAMING_SNAKE_CASE__ = self.unet(__A , __A ).sample # compute previous image: x_t -> x_t-1 SCREAMING_SNAKE_CASE__ = self.scheduler.step(__A , __A , __A , __A , __A , __A ).prev_sample else: # compute the reverse: x_t-1 -> x_t SCREAMING_SNAKE_CASE__ = self.scheduler.undo_step(__A , __A , __A ) SCREAMING_SNAKE_CASE__ = t SCREAMING_SNAKE_CASE__ = (image / 2 + 0.5).clamp(0 , 1 ) SCREAMING_SNAKE_CASE__ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": SCREAMING_SNAKE_CASE__ = self.numpy_to_pil(__A ) if not return_dict: return (image,) return ImagePipelineOutput(images=__A )
59
1
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list ): for i in range(len(UpperCamelCase__ ) - 1 , 0 , -1 ): SCREAMING_SNAKE_CASE__ = False for j in range(UpperCamelCase__ , 0 , -1 ): if unsorted[j] < unsorted[j - 1]: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = unsorted[j - 1], unsorted[j] SCREAMING_SNAKE_CASE__ = True for j in range(UpperCamelCase__ ): if unsorted[j] > unsorted[j + 1]: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = unsorted[j + 1], unsorted[j] SCREAMING_SNAKE_CASE__ = True if not swapped: break return unsorted if __name__ == "__main__": import doctest doctest.testmod() _lowerCamelCase = input('Enter numbers separated by a comma:\n').strip() _lowerCamelCase = [int(item) for item in user_input.split(',')] print(F'''{cocktail_shaker_sort(unsorted) = }''')
59
import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = OpenAIGPTTokenizer lowerCamelCase_ = OpenAIGPTTokenizerFast lowerCamelCase_ = True lowerCamelCase_ = False def _snake_case ( self :Optional[Any] ) -> Dict: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt SCREAMING_SNAKE_CASE__ = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """w</w>""", """r</w>""", """t</w>""", """lo""", """low""", """er</w>""", """low</w>""", """lowest</w>""", """newer</w>""", """wider</w>""", """<unk>""", ] SCREAMING_SNAKE_CASE__ = dict(zip(__A , range(len(__A ) ) ) ) SCREAMING_SNAKE_CASE__ = ["""#version: 0.2""", """l o""", """lo w""", """e r</w>""", """"""] SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" ) as fp: fp.write(json.dumps(__A ) ) with open(self.merges_file , """w""" ) as fp: fp.write("""\n""".join(__A ) ) def _snake_case ( self :Union[str, Any] , __A :str ) -> List[Any]: """simple docstring""" return "lower newer", "lower newer" def _snake_case ( self :Optional[Any] ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = OpenAIGPTTokenizer(self.vocab_file , self.merges_file ) SCREAMING_SNAKE_CASE__ = """lower""" SCREAMING_SNAKE_CASE__ = ["""low""", """er</w>"""] SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(__A ) self.assertListEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = tokens + ["""<unk>"""] SCREAMING_SNAKE_CASE__ = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , __A ) def _snake_case ( self :Optional[Any] , __A :Optional[Any]=15 ) -> Any: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) # Simple input SCREAMING_SNAKE_CASE__ = """This is a simple input""" SCREAMING_SNAKE_CASE__ = ["""This is a simple input 1""", """This is a simple input 2"""] SCREAMING_SNAKE_CASE__ = ("""This is a simple input""", """This is a pair""") SCREAMING_SNAKE_CASE__ = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests self.assertRaises(__A , tokenizer_r.encode , __A , max_length=__A , padding="""max_length""" ) # Simple input self.assertRaises(__A , tokenizer_r.encode_plus , __A , max_length=__A , padding="""max_length""" ) # Simple input self.assertRaises( __A , tokenizer_r.batch_encode_plus , __A , max_length=__A , padding="""max_length""" , ) # Pair input self.assertRaises(__A , tokenizer_r.encode , __A , max_length=__A , padding="""max_length""" ) # Pair input self.assertRaises(__A , tokenizer_r.encode_plus , __A , max_length=__A , padding="""max_length""" ) # Pair input self.assertRaises( __A , tokenizer_r.batch_encode_plus , __A , max_length=__A , padding="""max_length""" , ) def _snake_case ( self :Dict ) -> List[Any]: """simple docstring""" pass @require_ftfy @require_spacy @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ ): pass
59
1
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class UpperCamelCase_ ( UpperCamelCase__ ): def _snake_case ( self :Dict ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__A , """hidden_sizes""" ) ) self.parent.assertTrue(hasattr(__A , """neck_hidden_sizes""" ) ) self.parent.assertTrue(hasattr(__A , """num_attention_heads""" ) ) class UpperCamelCase_ : def __init__( self :List[str] , __A :int , __A :List[Any]=13 , __A :Tuple=32 , __A :Optional[Any]=2 , __A :Optional[Any]=3 , __A :Tuple=640 , __A :Optional[int]=4 , __A :Optional[int]="silu" , __A :Any=3 , __A :List[str]=32 , __A :List[Any]=0.1 , __A :Any=0.1 , __A :Union[str, Any]=0.1 , __A :str=0.0_2 , __A :List[Any]=True , __A :Tuple=True , __A :Optional[int]=10 , __A :Dict=None , ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = parent SCREAMING_SNAKE_CASE__ = batch_size SCREAMING_SNAKE_CASE__ = image_size SCREAMING_SNAKE_CASE__ = patch_size SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = last_hidden_size SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = conv_kernel_size SCREAMING_SNAKE_CASE__ = output_stride SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = classifier_dropout_prob SCREAMING_SNAKE_CASE__ = use_labels SCREAMING_SNAKE_CASE__ = is_training SCREAMING_SNAKE_CASE__ = num_labels SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = scope def _snake_case ( self :str ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None if self.use_labels: SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size] , self.num_labels ) SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) SCREAMING_SNAKE_CASE__ = self.get_config() return config, pixel_values, labels, pixel_labels def _snake_case ( self :Tuple ) -> int: """simple docstring""" return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def _snake_case ( self :int , __A :int , __A :List[Any] , __A :Optional[Any] , __A :Dict ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = MobileViTModel(config=__A ) model.to(__A ) model.eval() SCREAMING_SNAKE_CASE__ = model(__A ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def _snake_case ( self :int , __A :str , __A :str , __A :List[str] , __A :Any ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.num_labels SCREAMING_SNAKE_CASE__ = MobileViTForImageClassification(__A ) model.to(__A ) model.eval() SCREAMING_SNAKE_CASE__ = model(__A , labels=__A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _snake_case ( self :Any , __A :Any , __A :Union[str, Any] , __A :List[Any] , __A :List[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.num_labels SCREAMING_SNAKE_CASE__ = MobileViTForSemanticSegmentation(__A ) model.to(__A ) model.eval() SCREAMING_SNAKE_CASE__ = model(__A ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) SCREAMING_SNAKE_CASE__ = model(__A , labels=__A ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def _snake_case ( self :Dict ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = config_and_inputs SCREAMING_SNAKE_CASE__ = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class UpperCamelCase_ ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) lowerCamelCase_ = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) lowerCamelCase_ = False lowerCamelCase_ = False lowerCamelCase_ = False lowerCamelCase_ = False def _snake_case ( self :str ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = MobileViTModelTester(self ) SCREAMING_SNAKE_CASE__ = MobileViTConfigTester(self , config_class=__A , has_text_modality=__A ) def _snake_case ( self :List[str] ) -> Union[str, Any]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""MobileViT does not use inputs_embeds""" ) def _snake_case ( self :Tuple ) -> str: """simple docstring""" pass @unittest.skip(reason="""MobileViT does not support input and output embeddings""" ) def _snake_case ( self :Dict ) -> Dict: """simple docstring""" pass @unittest.skip(reason="""MobileViT does not output attentions""" ) def _snake_case ( self :Dict ) -> Optional[int]: """simple docstring""" pass def _snake_case ( self :Optional[Any] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ = model_class(__A ) SCREAMING_SNAKE_CASE__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE__ = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE__ = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __A ) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def _snake_case ( self :str ) -> Any: """simple docstring""" pass def _snake_case ( self :Optional[Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__A ) def _snake_case ( self :Any ) -> str: """simple docstring""" def check_hidden_states_output(__A :Union[str, Any] , __A :Any , __A :Any ): SCREAMING_SNAKE_CASE__ = model_class(__A ) model.to(__A ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(**self._prepare_for_class(__A , __A ) ) SCREAMING_SNAKE_CASE__ = outputs.hidden_states SCREAMING_SNAKE_CASE__ = 5 self.assertEqual(len(__A ) , __A ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. SCREAMING_SNAKE_CASE__ = 2 for i in range(len(__A ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ = True check_hidden_states_output(__A , __A , __A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] SCREAMING_SNAKE_CASE__ = True check_hidden_states_output(__A , __A , __A ) def _snake_case ( self :List[str] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__A ) def _snake_case ( self :Any ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*__A ) @slow def _snake_case ( self :List[Any] ) -> Union[str, Any]: """simple docstring""" for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE__ = MobileViTModel.from_pretrained(__A ) self.assertIsNotNone(__A ) def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class UpperCamelCase_ ( unittest.TestCase ): @cached_property def _snake_case ( self :Optional[int] ) -> Union[str, Any]: """simple docstring""" return MobileViTImageProcessor.from_pretrained("""apple/mobilevit-xx-small""" ) if is_vision_available() else None @slow def _snake_case ( self :int ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = MobileViTForImageClassification.from_pretrained("""apple/mobilevit-xx-small""" ).to(__A ) SCREAMING_SNAKE_CASE__ = self.default_image_processor SCREAMING_SNAKE_CASE__ = prepare_img() SCREAMING_SNAKE_CASE__ = image_processor(images=__A , return_tensors="""pt""" ).to(__A ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(**__A ) # verify the logits SCREAMING_SNAKE_CASE__ = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , __A ) SCREAMING_SNAKE_CASE__ = torch.tensor([-1.9_3_6_4, -1.2_3_2_7, -0.4_6_5_3] ).to(__A ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __A , atol=1E-4 ) ) @slow def _snake_case ( self :int ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = MobileViTForSemanticSegmentation.from_pretrained("""apple/deeplabv3-mobilevit-xx-small""" ) SCREAMING_SNAKE_CASE__ = model.to(__A ) SCREAMING_SNAKE_CASE__ = MobileViTImageProcessor.from_pretrained("""apple/deeplabv3-mobilevit-xx-small""" ) SCREAMING_SNAKE_CASE__ = prepare_img() SCREAMING_SNAKE_CASE__ = image_processor(images=__A , return_tensors="""pt""" ).to(__A ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(**__A ) SCREAMING_SNAKE_CASE__ = outputs.logits # verify the logits SCREAMING_SNAKE_CASE__ = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , __A ) SCREAMING_SNAKE_CASE__ = torch.tensor( [ [[6.9_7_1_3, 6.9_7_8_6, 7.2_4_2_2], [7.2_8_9_3, 7.2_8_2_5, 7.4_4_4_6], [7.6_5_8_0, 7.8_7_9_7, 7.9_4_2_0]], [[-1_0.6_8_6_9, -1_0.3_2_5_0, -1_0.3_4_7_1], [-1_0.4_2_2_8, -9.9_8_6_8, -9.7_1_3_2], [-1_1.0_4_0_5, -1_1.0_2_2_1, -1_0.7_3_1_8]], [[-3.3_0_8_9, -2.8_5_3_9, -2.6_7_4_0], [-3.2_7_0_6, -2.5_6_2_1, -2.5_1_0_8], [-3.2_5_3_4, -2.6_6_1_5, -2.6_6_5_1]], ] , device=__A , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , __A , atol=1E-4 ) ) @slow def _snake_case ( self :List[Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = MobileViTForSemanticSegmentation.from_pretrained("""apple/deeplabv3-mobilevit-xx-small""" ) SCREAMING_SNAKE_CASE__ = model.to(__A ) SCREAMING_SNAKE_CASE__ = MobileViTImageProcessor.from_pretrained("""apple/deeplabv3-mobilevit-xx-small""" ) SCREAMING_SNAKE_CASE__ = prepare_img() SCREAMING_SNAKE_CASE__ = image_processor(images=__A , return_tensors="""pt""" ).to(__A ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(**__A ) SCREAMING_SNAKE_CASE__ = outputs.logits.detach().cpu() SCREAMING_SNAKE_CASE__ = image_processor.post_process_semantic_segmentation(outputs=__A , target_sizes=[(50, 60)] ) SCREAMING_SNAKE_CASE__ = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , __A ) SCREAMING_SNAKE_CASE__ = image_processor.post_process_semantic_segmentation(outputs=__A ) SCREAMING_SNAKE_CASE__ = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , __A )
59
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=UpperCamelCase__ ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = field(default="image-classification" , metadata={"include_in_asdict_even_if_is_default": True} ) lowerCamelCase_ = Features({"image": Image()} ) lowerCamelCase_ = Features({"labels": ClassLabel} ) lowerCamelCase_ = "image" lowerCamelCase_ = "labels" def _snake_case ( self :List[str] , __A :Tuple ) -> Tuple: """simple docstring""" if self.label_column not in features: raise ValueError(f'''Column {self.label_column} is not present in features.''' ) if not isinstance(features[self.label_column] , __A ): raise ValueError(f'''Column {self.label_column} is not a ClassLabel.''' ) SCREAMING_SNAKE_CASE__ = copy.deepcopy(self ) SCREAMING_SNAKE_CASE__ = self.label_schema.copy() SCREAMING_SNAKE_CASE__ = features[self.label_column] SCREAMING_SNAKE_CASE__ = label_schema return task_template @property def _snake_case ( self :Dict ) -> Dict[str, str]: """simple docstring""" return { self.image_column: "image", self.label_column: "labels", }
59
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowerCamelCase = { 'configuration_vivit': ['VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'VivitConfig'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['VivitImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'VivitModel', 'VivitPreTrainedModel', 'VivitForVideoClassification', ] if TYPE_CHECKING: from .configuration_vivit import VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, VivitConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_vivit import VivitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vivit import ( VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST, VivitForVideoClassification, VivitModel, VivitPreTrainedModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCamelCase = { 'configuration_xmod': [ 'XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XmodConfig', 'XmodOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'XMOD_PRETRAINED_MODEL_ARCHIVE_LIST', 'XmodForCausalLM', 'XmodForMaskedLM', 'XmodForMultipleChoice', 'XmodForQuestionAnswering', 'XmodForSequenceClassification', 'XmodForTokenClassification', 'XmodModel', 'XmodPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
1
import argparse import os.path as osp import re import torch from safetensors.torch import load_file, save_file # =================# # UNet Conversion # # =================# _lowerCamelCase = [ # (stable-diffusion, HF Diffusers) ('time_embed.0.weight', 'time_embedding.linear_1.weight'), ('time_embed.0.bias', 'time_embedding.linear_1.bias'), ('time_embed.2.weight', 'time_embedding.linear_2.weight'), ('time_embed.2.bias', 'time_embedding.linear_2.bias'), ('input_blocks.0.0.weight', 'conv_in.weight'), ('input_blocks.0.0.bias', 'conv_in.bias'), ('out.0.weight', 'conv_norm_out.weight'), ('out.0.bias', 'conv_norm_out.bias'), ('out.2.weight', 'conv_out.weight'), ('out.2.bias', 'conv_out.bias'), ] _lowerCamelCase = [ # (stable-diffusion, HF Diffusers) ('in_layers.0', 'norm1'), ('in_layers.2', 'conv1'), ('out_layers.0', 'norm2'), ('out_layers.3', 'conv2'), ('emb_layers.1', 'time_emb_proj'), ('skip_connection', 'conv_shortcut'), ] _lowerCamelCase = [] # hardcoded number of downblocks and resnets/attentions... # would need smarter logic for other networks. for i in range(4): # loop over downblocks/upblocks for j in range(2): # loop over resnets/attentions for downblocks _lowerCamelCase = F'''down_blocks.{i}.resnets.{j}.''' _lowerCamelCase = F'''input_blocks.{3*i + j + 1}.0.''' unet_conversion_map_layer.append((sd_down_res_prefix, hf_down_res_prefix)) if i < 3: # no attention layers in down_blocks.3 _lowerCamelCase = F'''down_blocks.{i}.attentions.{j}.''' _lowerCamelCase = F'''input_blocks.{3*i + j + 1}.1.''' unet_conversion_map_layer.append((sd_down_atn_prefix, hf_down_atn_prefix)) for j in range(3): # loop over resnets/attentions for upblocks _lowerCamelCase = F'''up_blocks.{i}.resnets.{j}.''' _lowerCamelCase = F'''output_blocks.{3*i + j}.0.''' unet_conversion_map_layer.append((sd_up_res_prefix, hf_up_res_prefix)) if i > 0: # no attention layers in up_blocks.0 _lowerCamelCase = F'''up_blocks.{i}.attentions.{j}.''' _lowerCamelCase = F'''output_blocks.{3*i + j}.1.''' unet_conversion_map_layer.append((sd_up_atn_prefix, hf_up_atn_prefix)) if i < 3: # no downsample in down_blocks.3 _lowerCamelCase = F'''down_blocks.{i}.downsamplers.0.conv.''' _lowerCamelCase = F'''input_blocks.{3*(i+1)}.0.op.''' unet_conversion_map_layer.append((sd_downsample_prefix, hf_downsample_prefix)) # no upsample in up_blocks.3 _lowerCamelCase = F'''up_blocks.{i}.upsamplers.0.''' _lowerCamelCase = F'''output_blocks.{3*i + 2}.{1 if i == 0 else 2}.''' unet_conversion_map_layer.append((sd_upsample_prefix, hf_upsample_prefix)) _lowerCamelCase = 'mid_block.attentions.0.' _lowerCamelCase = 'middle_block.1.' unet_conversion_map_layer.append((sd_mid_atn_prefix, hf_mid_atn_prefix)) for j in range(2): _lowerCamelCase = F'''mid_block.resnets.{j}.''' _lowerCamelCase = F'''middle_block.{2*j}.''' unet_conversion_map_layer.append((sd_mid_res_prefix, hf_mid_res_prefix)) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): # buyer beware: this is a *brittle* function, # and correct output requires that all of these pieces interact in # the exact order in which I have arranged them. SCREAMING_SNAKE_CASE__ = {k: k for k in unet_state_dict.keys()} for sd_name, hf_name in unet_conversion_map: SCREAMING_SNAKE_CASE__ = sd_name for k, v in mapping.items(): if "resnets" in k: for sd_part, hf_part in unet_conversion_map_resnet: SCREAMING_SNAKE_CASE__ = v.replace(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = v for k, v in mapping.items(): for sd_part, hf_part in unet_conversion_map_layer: SCREAMING_SNAKE_CASE__ = v.replace(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = v SCREAMING_SNAKE_CASE__ = {v: unet_state_dict[k] for k, v in mapping.items()} return new_state_dict # ================# # VAE Conversion # # ================# _lowerCamelCase = [ # (stable-diffusion, HF Diffusers) ('nin_shortcut', 'conv_shortcut'), ('norm_out', 'conv_norm_out'), ('mid.attn_1.', 'mid_block.attentions.0.'), ] for i in range(4): # down_blocks have two resnets for j in range(2): _lowerCamelCase = F'''encoder.down_blocks.{i}.resnets.{j}.''' _lowerCamelCase = F'''encoder.down.{i}.block.{j}.''' vae_conversion_map.append((sd_down_prefix, hf_down_prefix)) if i < 3: _lowerCamelCase = F'''down_blocks.{i}.downsamplers.0.''' _lowerCamelCase = F'''down.{i}.downsample.''' vae_conversion_map.append((sd_downsample_prefix, hf_downsample_prefix)) _lowerCamelCase = F'''up_blocks.{i}.upsamplers.0.''' _lowerCamelCase = F'''up.{3-i}.upsample.''' vae_conversion_map.append((sd_upsample_prefix, hf_upsample_prefix)) # up_blocks have three resnets # also, up blocks in hf are numbered in reverse from sd for j in range(3): _lowerCamelCase = F'''decoder.up_blocks.{i}.resnets.{j}.''' _lowerCamelCase = F'''decoder.up.{3-i}.block.{j}.''' vae_conversion_map.append((sd_up_prefix, hf_up_prefix)) # this part accounts for mid blocks in both the encoder and the decoder for i in range(2): _lowerCamelCase = F'''mid_block.resnets.{i}.''' _lowerCamelCase = F'''mid.block_{i+1}.''' vae_conversion_map.append((sd_mid_res_prefix, hf_mid_res_prefix)) _lowerCamelCase = [ # (stable-diffusion, HF Diffusers) ('norm.', 'group_norm.'), ('q.', 'query.'), ('k.', 'key.'), ('v.', 'value.'), ('proj_out.', 'proj_attn.'), ] def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str ): # convert HF linear weights to SD conv2d weights return w.reshape(*w.shape , 1 , 1 ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[str, Any] ): SCREAMING_SNAKE_CASE__ = {k: k for k in vae_state_dict.keys()} for k, v in mapping.items(): for sd_part, hf_part in vae_conversion_map: SCREAMING_SNAKE_CASE__ = v.replace(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = v for k, v in mapping.items(): if "attentions" in k: for sd_part, hf_part in vae_conversion_map_attn: SCREAMING_SNAKE_CASE__ = v.replace(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = v SCREAMING_SNAKE_CASE__ = {v: vae_state_dict[k] for k, v in mapping.items()} SCREAMING_SNAKE_CASE__ = ["""q""", """k""", """v""", """proj_out"""] for k, v in new_state_dict.items(): for weight_name in weights_to_convert: if f'''mid.attn_1.{weight_name}.weight''' in k: print(f'''Reshaping {k} for SD format''' ) SCREAMING_SNAKE_CASE__ = reshape_weight_for_sd(UpperCamelCase__ ) return new_state_dict # =========================# # Text Encoder Conversion # # =========================# _lowerCamelCase = [ # (stable-diffusion, HF Diffusers) ('resblocks.', 'text_model.encoder.layers.'), ('ln_1', 'layer_norm1'), ('ln_2', 'layer_norm2'), ('.c_fc.', '.fc1.'), ('.c_proj.', '.fc2.'), ('.attn', '.self_attn'), ('ln_final.', 'transformer.text_model.final_layer_norm.'), ('token_embedding.weight', 'transformer.text_model.embeddings.token_embedding.weight'), ('positional_embedding', 'transformer.text_model.embeddings.position_embedding.weight'), ] _lowerCamelCase = {re.escape(x[1]): x[0] for x in textenc_conversion_lst} _lowerCamelCase = re.compile('|'.join(protected.keys())) # Ordering is from https://github.com/pytorch/pytorch/blob/master/test/cpp/api/modules.cpp _lowerCamelCase = {'q': 0, 'k': 1, 'v': 2} def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[str, Any] ): SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = {} for k, v in text_enc_dict.items(): if ( k.endswith(""".self_attn.q_proj.weight""" ) or k.endswith(""".self_attn.k_proj.weight""" ) or k.endswith(""".self_attn.v_proj.weight""" ) ): SCREAMING_SNAKE_CASE__ = k[: -len(""".q_proj.weight""" )] SCREAMING_SNAKE_CASE__ = k[-len("""q_proj.weight""" )] if k_pre not in capture_qkv_weight: SCREAMING_SNAKE_CASE__ = [None, None, None] SCREAMING_SNAKE_CASE__ = v continue if ( k.endswith(""".self_attn.q_proj.bias""" ) or k.endswith(""".self_attn.k_proj.bias""" ) or k.endswith(""".self_attn.v_proj.bias""" ) ): SCREAMING_SNAKE_CASE__ = k[: -len(""".q_proj.bias""" )] SCREAMING_SNAKE_CASE__ = k[-len("""q_proj.bias""" )] if k_pre not in capture_qkv_bias: SCREAMING_SNAKE_CASE__ = [None, None, None] SCREAMING_SNAKE_CASE__ = v continue SCREAMING_SNAKE_CASE__ = textenc_pattern.sub(lambda UpperCamelCase__ : protected[re.escape(m.group(0 ) )] , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = v for k_pre, tensors in capture_qkv_weight.items(): if None in tensors: raise Exception("""CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing""" ) SCREAMING_SNAKE_CASE__ = textenc_pattern.sub(lambda UpperCamelCase__ : protected[re.escape(m.group(0 ) )] , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = torch.cat(UpperCamelCase__ ) for k_pre, tensors in capture_qkv_bias.items(): if None in tensors: raise Exception("""CORRUPTED MODEL: one of the q-k-v values for the text encoder was missing""" ) SCREAMING_SNAKE_CASE__ = textenc_pattern.sub(lambda UpperCamelCase__ : protected[re.escape(m.group(0 ) )] , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = torch.cat(UpperCamelCase__ ) return new_state_dict def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[Any] ): return text_enc_dict if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument('--model_path', default=None, type=str, required=True, help='Path to the model to convert.') parser.add_argument('--checkpoint_path', default=None, type=str, required=True, help='Path to the output model.') parser.add_argument('--half', action='store_true', help='Save weights in half precision.') parser.add_argument( '--use_safetensors', action='store_true', help='Save weights use safetensors, default is ckpt.' ) _lowerCamelCase = parser.parse_args() assert args.model_path is not None, "Must provide a model path!" assert args.checkpoint_path is not None, "Must provide a checkpoint path!" # Path for safetensors _lowerCamelCase = osp.join(args.model_path, 'unet', 'diffusion_pytorch_model.safetensors') _lowerCamelCase = osp.join(args.model_path, 'vae', 'diffusion_pytorch_model.safetensors') _lowerCamelCase = osp.join(args.model_path, 'text_encoder', 'model.safetensors') # Load models from safetensors if it exists, if it doesn't pytorch if osp.exists(unet_path): _lowerCamelCase = load_file(unet_path, device='cpu') else: _lowerCamelCase = osp.join(args.model_path, 'unet', 'diffusion_pytorch_model.bin') _lowerCamelCase = torch.load(unet_path, map_location='cpu') if osp.exists(vae_path): _lowerCamelCase = load_file(vae_path, device='cpu') else: _lowerCamelCase = osp.join(args.model_path, 'vae', 'diffusion_pytorch_model.bin') _lowerCamelCase = torch.load(vae_path, map_location='cpu') if osp.exists(text_enc_path): _lowerCamelCase = load_file(text_enc_path, device='cpu') else: _lowerCamelCase = osp.join(args.model_path, 'text_encoder', 'pytorch_model.bin') _lowerCamelCase = torch.load(text_enc_path, map_location='cpu') # Convert the UNet model _lowerCamelCase = convert_unet_state_dict(unet_state_dict) _lowerCamelCase = {'model.diffusion_model.' + k: v for k, v in unet_state_dict.items()} # Convert the VAE model _lowerCamelCase = convert_vae_state_dict(vae_state_dict) _lowerCamelCase = {'first_stage_model.' + k: v for k, v in vae_state_dict.items()} # Easiest way to identify v2.0 model seems to be that the text encoder (OpenCLIP) is deeper _lowerCamelCase = 'text_model.encoder.layers.22.layer_norm2.bias' in text_enc_dict if is_vaa_model: # Need to add the tag 'transformer' in advance so we can knock it out from the final layer-norm _lowerCamelCase = {'transformer.' + k: v for k, v in text_enc_dict.items()} _lowerCamelCase = convert_text_enc_state_dict_vaa(text_enc_dict) _lowerCamelCase = {'cond_stage_model.model.' + k: v for k, v in text_enc_dict.items()} else: _lowerCamelCase = convert_text_enc_state_dict(text_enc_dict) _lowerCamelCase = {'cond_stage_model.transformer.' + k: v for k, v in text_enc_dict.items()} # Put together new checkpoint _lowerCamelCase = {**unet_state_dict, **vae_state_dict, **text_enc_dict} if args.half: _lowerCamelCase = {k: v.half() for k, v in state_dict.items()} if args.use_safetensors: save_file(state_dict, args.checkpoint_path) else: _lowerCamelCase = {'state_dict': state_dict} torch.save(state_dict, args.checkpoint_path)
59
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "timm_backbone" def __init__( self :Union[str, Any] , __A :str=None , __A :Union[str, Any]=3 , __A :str=True , __A :Any=True , __A :Optional[Any]=None , **__A :List[str] , ) -> Tuple: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = backbone SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = features_only SCREAMING_SNAKE_CASE__ = use_pretrained_backbone SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = out_indices if out_indices is not None else (-1,)
59
1
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str ): # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError("""The given input must be positive""" ) # get the generated string sequence SCREAMING_SNAKE_CASE__ = gray_code_sequence_string(SCREAMING_SNAKE_CASE_ ) # # convert them to integers for i in range(len(SCREAMING_SNAKE_CASE_ ) ): SCREAMING_SNAKE_CASE__ = int(sequence[i] , 2 ) return sequence def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[str, Any] ): # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] SCREAMING_SNAKE_CASE__ = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits SCREAMING_SNAKE_CASE__ = gray_code_sequence_string(bit_count - 1 ) SCREAMING_SNAKE_CASE__ = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): SCREAMING_SNAKE_CASE__ = """0""" + smaller_sequence[i] sequence.append(SCREAMING_SNAKE_CASE_ ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): SCREAMING_SNAKE_CASE__ = """1""" + smaller_sequence[i] sequence.append(SCREAMING_SNAKE_CASE_ ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
700
from math import pow, sqrt def SCREAMING_SNAKE_CASE__ ( *UpperCamelCase__: float ): SCREAMING_SNAKE_CASE__ = len(UpperCamelCase__ ) > 0 and all(value > 0.0 for value in values ) return result def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ ) else ValueError("""Input Error: Molar mass values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(effusion_rate * sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(effusion_rate / sqrt(molar_mass_a / molar_mass_a ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(molar_mass / pow(effusion_rate_a / effusion_rate_a , 2 ) , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: float , UpperCamelCase__: float , UpperCamelCase__: float ): return ( round(pow(effusion_rate_a / effusion_rate_a , 2 ) / molar_mass , 6 ) if validate(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else ValueError( """Input Error: Molar mass and effusion rate values must greater than 0.""" ) )
59
0
'''simple docstring''' import os from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen, xsplitext from ..table import array_cast from ..utils.py_utils import no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: from .features import FeatureType _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = False, False, False @dataclass class UpperCamelCase_ : lowerCamelCase_ = None lowerCamelCase_ = True lowerCamelCase_ = True lowerCamelCase_ = None # Automatically constructed lowerCamelCase_ = "dict" lowerCamelCase_ = pa.struct({"bytes": pa.binary(), "path": pa.string()} ) lowerCamelCase_ = field(default="Audio" , init=_snake_case , repr=_snake_case ) def __call__( self :List[str] ) -> List[Any]: """simple docstring""" return self.pa_type def _snake_case ( self :str , __A :Union[str, bytes, dict] ) -> dict: """simple docstring""" try: import soundfile as sf # soundfile is a dependency of librosa, needed to decode audio files. except ImportError as err: raise ImportError("""To support encoding audio data, please install 'soundfile'.""" ) from err if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): return {"bytes": None, "path": value} elif isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): return {"bytes": value, "path": None} elif "array" in value: # convert the audio array to wav bytes SCREAMING_SNAKE_CASE__ = BytesIO() sf.write(lowerCAmelCase__ , value["""array"""] , value["""sampling_rate"""] , format="""wav""" ) return {"bytes": buffer.getvalue(), "path": None} elif value.get("""path""" ) is not None and os.path.isfile(value["""path"""] ): # we set "bytes": None to not duplicate the data if they're already available locally if value["path"].endswith("""pcm""" ): # "PCM" only has raw audio bytes if value.get("""sampling_rate""" ) is None: # At least, If you want to convert "PCM-byte" to "WAV-byte", you have to know sampling rate raise KeyError("""To use PCM files, please specify a 'sampling_rate' in Audio object""" ) if value.get("""bytes""" ): # If we already had PCM-byte, we don`t have to make "read file, make bytes" (just use it!) SCREAMING_SNAKE_CASE__ = np.frombuffer(value["""bytes"""] , dtype=np.intaa ).astype(np.floataa ) / 3_2767 else: SCREAMING_SNAKE_CASE__ = np.memmap(value["""path"""] , dtype="""h""" , mode="""r""" ).astype(np.floataa ) / 3_2767 SCREAMING_SNAKE_CASE__ = BytesIO(bytes() ) sf.write(lowerCAmelCase__ , lowerCAmelCase__ , value["""sampling_rate"""] , format="""wav""" ) return {"bytes": buffer.getvalue(), "path": None} else: return {"bytes": None, "path": value.get("""path""" )} elif value.get("""bytes""" ) is not None or value.get("""path""" ) is not None: # store the audio bytes, and path is used to infer the audio format using the file extension return {"bytes": value.get("""bytes""" ), "path": value.get("""path""" )} else: raise ValueError( f'''An audio sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.''' ) def _snake_case ( self :Any , __A :dict , __A :Optional[Dict[str, Union[str, bool, None]]] = None ) -> dict: """simple docstring""" if not self.decode: raise RuntimeError("""Decoding is disabled for this feature. Please use Audio(decode=True) instead.""" ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = (value["""path"""], BytesIO(value["""bytes"""] )) if value["""bytes"""] is not None else (value["""path"""], None) if path is None and file is None: raise ValueError(f'''An audio sample should have one of \'path\' or \'bytes\' but both are None in {value}.''' ) try: import librosa import soundfile as sf except ImportError as err: raise ImportError("""To support decoding audio files, please install 'librosa' and 'soundfile'.""" ) from err SCREAMING_SNAKE_CASE__ = xsplitext(lowerCAmelCase__ )[1][1:].lower() if path is not None else None if not config.IS_OPUS_SUPPORTED and audio_format == "opus": raise RuntimeError( """Decoding 'opus' files requires system library 'libsndfile'>=1.0.31, """ """You can try to update `soundfile` python library: `pip install \"soundfile>=0.12.1\"`. """ ) elif not config.IS_MP3_SUPPORTED and audio_format == "mp3": raise RuntimeError( """Decoding 'mp3' files requires system library 'libsndfile'>=1.1.0, """ """You can try to update `soundfile` python library: `pip install \"soundfile>=0.12.1\"`. """ ) if file is None: SCREAMING_SNAKE_CASE__ = token_per_repo_id or {} SCREAMING_SNAKE_CASE__ = path.split("""::""" )[-1] try: SCREAMING_SNAKE_CASE__ = string_to_dict(lowerCAmelCase__ , config.HUB_DATASETS_URL )["""repo_id"""] SCREAMING_SNAKE_CASE__ = token_per_repo_id[repo_id] except (ValueError, KeyError): SCREAMING_SNAKE_CASE__ = None with xopen(lowerCAmelCase__ , """rb""" , use_auth_token=lowerCAmelCase__ ) as f: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = sf.read(lowerCAmelCase__ ) else: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = sf.read(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE__ = array.T if self.mono: SCREAMING_SNAKE_CASE__ = librosa.to_mono(lowerCAmelCase__ ) if self.sampling_rate and self.sampling_rate != sampling_rate: SCREAMING_SNAKE_CASE__ = librosa.resample(lowerCAmelCase__ , orig_sr=lowerCAmelCase__ , target_sr=self.sampling_rate ) SCREAMING_SNAKE_CASE__ = self.sampling_rate return {"path": path, "array": array, "sampling_rate": sampling_rate} def _snake_case ( self :Any ) -> Union["FeatureType", Dict[str, "FeatureType"]]: """simple docstring""" from .features import Value if self.decode: raise ValueError("""Cannot flatten a decoded Audio feature.""" ) return { "bytes": Value("""binary""" ), "path": Value("""string""" ), } def _snake_case ( self :Dict , __A :Union[pa.StringArray, pa.StructArray] ) -> pa.StructArray: """simple docstring""" if pa.types.is_string(storage.type ): SCREAMING_SNAKE_CASE__ = pa.array([None] * len(lowerCAmelCase__ ) , type=pa.binary() ) SCREAMING_SNAKE_CASE__ = pa.StructArray.from_arrays([bytes_array, storage] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): SCREAMING_SNAKE_CASE__ = pa.array([None] * len(lowerCAmelCase__ ) , type=pa.string() ) SCREAMING_SNAKE_CASE__ = pa.StructArray.from_arrays([storage, path_array] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ) and storage.type.get_all_field_indices("""array""" ): SCREAMING_SNAKE_CASE__ = pa.array([Audio().encode_example(lowerCAmelCase__ ) if x is not None else None for x in storage.to_pylist()] ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("""bytes""" ) >= 0: SCREAMING_SNAKE_CASE__ = storage.field("""bytes""" ) else: SCREAMING_SNAKE_CASE__ = pa.array([None] * len(lowerCAmelCase__ ) , type=pa.binary() ) if storage.type.get_field_index("""path""" ) >= 0: SCREAMING_SNAKE_CASE__ = storage.field("""path""" ) else: SCREAMING_SNAKE_CASE__ = pa.array([None] * len(lowerCAmelCase__ ) , type=pa.string() ) SCREAMING_SNAKE_CASE__ = pa.StructArray.from_arrays([bytes_array, path_array] , ["""bytes""", """path"""] , mask=storage.is_null() ) return array_cast(lowerCAmelCase__ , self.pa_type ) def _snake_case ( self :Dict , __A :pa.StructArray ) -> pa.StructArray: """simple docstring""" @no_op_if_value_is_null def path_to_bytes(__A :List[Any] ): with xopen(lowerCAmelCase__ , """rb""" ) as f: SCREAMING_SNAKE_CASE__ = f.read() return bytes_ SCREAMING_SNAKE_CASE__ = pa.array( [ (path_to_bytes(x["""path"""] ) if x["""bytes"""] is None else x["""bytes"""]) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) SCREAMING_SNAKE_CASE__ = pa.array( [os.path.basename(lowerCAmelCase__ ) if path is not None else None for path in storage.field("""path""" ).to_pylist()] , type=pa.string() , ) SCREAMING_SNAKE_CASE__ = pa.StructArray.from_arrays([bytes_array, path_array] , ["""bytes""", """path"""] , mask=bytes_array.is_null() ) return array_cast(lowerCAmelCase__ , self.pa_type )
701
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'shi-labs/nat-mini-in1k-224': 'https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json', # See all Nat models at https://huggingface.co/models?filter=nat } class UpperCamelCase_ ( UpperCamelCase__ , UpperCamelCase__ ): lowerCamelCase_ = "nat" lowerCamelCase_ = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self :List[Any] , __A :Optional[Any]=4 , __A :Any=3 , __A :Optional[int]=64 , __A :Optional[int]=[3, 4, 6, 5] , __A :Union[str, Any]=[2, 4, 8, 16] , __A :Optional[Any]=7 , __A :Optional[Any]=3.0 , __A :List[Any]=True , __A :int=0.0 , __A :Dict=0.0 , __A :Optional[Any]=0.1 , __A :str="gelu" , __A :Optional[Any]=0.0_2 , __A :Optional[int]=1E-5 , __A :Optional[int]=0.0 , __A :Optional[Any]=None , __A :Union[str, Any]=None , **__A :Union[str, Any] , ) -> Optional[int]: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = patch_size SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = embed_dim SCREAMING_SNAKE_CASE__ = depths SCREAMING_SNAKE_CASE__ = len(__A ) SCREAMING_SNAKE_CASE__ = num_heads SCREAMING_SNAKE_CASE__ = kernel_size SCREAMING_SNAKE_CASE__ = mlp_ratio SCREAMING_SNAKE_CASE__ = qkv_bias SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = drop_path_rate SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model SCREAMING_SNAKE_CASE__ = int(embed_dim * 2 ** (len(__A ) - 1) ) SCREAMING_SNAKE_CASE__ = layer_scale_init_value SCREAMING_SNAKE_CASE__ = ["""stem"""] + [f'''stage{idx}''' for idx in range(1 , len(__A ) + 1 )] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = get_aligned_output_features_output_indices( out_features=__A , out_indices=__A , stage_names=self.stage_names )
59
0
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=snake_case__ ) class UpperCamelCase_ ( snake_case__ ): lowerCamelCase_ = field(default="automatic-speech-recognition" , metadata={"include_in_asdict_even_if_is_default": True} ) lowerCamelCase_ = Features({"audio": Audio()} ) lowerCamelCase_ = Features({"transcription": Value("string" )} ) lowerCamelCase_ = "audio" lowerCamelCase_ = "transcription" def _snake_case ( self :List[Any] , __A :Any ) -> int: """simple docstring""" if self.audio_column not in features: raise ValueError(f'''Column {self.audio_column} is not present in features.''' ) if not isinstance(features[self.audio_column] , UpperCAmelCase_ ): raise ValueError(f'''Column {self.audio_column} is not an Audio type.''' ) SCREAMING_SNAKE_CASE__ = copy.deepcopy(self ) SCREAMING_SNAKE_CASE__ = self.input_schema.copy() SCREAMING_SNAKE_CASE__ = features[self.audio_column] SCREAMING_SNAKE_CASE__ = input_schema return task_template @property def _snake_case ( self :List[str] ) -> Dict[str, str]: """simple docstring""" return {self.audio_column: "audio", self.transcription_column: "transcription"}
702
from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any ): return {key.lstrip("""-""" ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = ArgumentParser( """HuggingFace Datasets CLI tool""" , usage="""datasets-cli <command> [<args>]""" , allow_abbrev=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = parser.add_subparsers(help="""datasets-cli command helpers""" ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(UpperCamelCase__ ) EnvironmentCommand.register_subcommand(UpperCamelCase__ ) TestCommand.register_subcommand(UpperCamelCase__ ) RunBeamCommand.register_subcommand(UpperCamelCase__ ) DummyDataCommand.register_subcommand(UpperCamelCase__ ) # Parse args SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = parser.parse_known_args() if not hasattr(UpperCamelCase__ , """func""" ): parser.print_help() exit(1 ) SCREAMING_SNAKE_CASE__ = parse_unknown_args(UpperCamelCase__ ) # Run SCREAMING_SNAKE_CASE__ = args.func(UpperCamelCase__ , **UpperCamelCase__ ) service.run() if __name__ == "__main__": main()
59
0
from typing import List from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { "snap-research/efficientformer-l1-300": ( "https://huggingface.co/snap-research/efficientformer-l1-300/resolve/main/config.json" ), } class UpperCamelCase_ ( _SCREAMING_SNAKE_CASE ): lowerCamelCase_ = "efficientformer" def __init__( self :Dict , __A :Optional[int] = [3, 2, 6, 4] , __A :Optional[Any] = [48, 96, 224, 448] , __A :str = [True, True, True, True] , __A :Any = 448 , __A :Union[str, Any] = 32 , __A :Union[str, Any] = 4 , __A :int = 7 , __A :Dict = 5 , __A :Any = 8 , __A :List[Any] = 4 , __A :Optional[Any] = 0.0 , __A :Union[str, Any] = 16 , __A :str = 3 , __A :Union[str, Any] = 3 , __A :List[str] = 3 , __A :str = 2 , __A :Dict = 1 , __A :Any = 0.0 , __A :List[Any] = 1 , __A :Any = True , __A :Optional[Any] = True , __A :List[str] = 1E-5 , __A :Union[str, Any] = "gelu" , __A :List[Any] = 0.0_2 , __A :int = 1E-12 , __A :Any = 224 , __A :Optional[Any] = 1E-05 , **__A :Optional[int] , ) -> None: """simple docstring""" super().__init__(**A_ ) SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = hidden_sizes SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = patch_size SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = depths SCREAMING_SNAKE_CASE__ = mlp_expansion_ratio SCREAMING_SNAKE_CASE__ = downsamples SCREAMING_SNAKE_CASE__ = dim SCREAMING_SNAKE_CASE__ = key_dim SCREAMING_SNAKE_CASE__ = attention_ratio SCREAMING_SNAKE_CASE__ = resolution SCREAMING_SNAKE_CASE__ = pool_size SCREAMING_SNAKE_CASE__ = downsample_patch_size SCREAMING_SNAKE_CASE__ = downsample_stride SCREAMING_SNAKE_CASE__ = downsample_pad SCREAMING_SNAKE_CASE__ = drop_path_rate SCREAMING_SNAKE_CASE__ = num_metaad_blocks SCREAMING_SNAKE_CASE__ = distillation SCREAMING_SNAKE_CASE__ = use_layer_scale SCREAMING_SNAKE_CASE__ = layer_scale_init_value SCREAMING_SNAKE_CASE__ = image_size SCREAMING_SNAKE_CASE__ = batch_norm_eps
703
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase__ ): def __init__( self :List[Any] , *__A :Tuple , **__A :Dict ) -> None: """simple docstring""" warnings.warn( """The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use LayoutLMv2ImageProcessor instead.""" , __A , ) super().__init__(*__A , **__A )
59
0
from datetime import datetime as dt import os from github import Github _lowerCamelCase = [ '''good first issue''', '''good second issue''', '''good difficult issue''', '''feature request''', '''new model''', '''wip''', ] def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = Github(os.environ["""GITHUB_TOKEN"""] ) SCREAMING_SNAKE_CASE__ = g.get_repo("""huggingface/transformers""" ) SCREAMING_SNAKE_CASE__ = repo.get_issues(state="""open""" ) for issue in open_issues: SCREAMING_SNAKE_CASE__ = sorted([comment for comment in issue.get_comments()] , key=lambda UpperCamelCase__ : i.created_at , reverse=_lowercase ) SCREAMING_SNAKE_CASE__ = comments[0] if len(_lowercase ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state="""closed""" ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( """This issue has been automatically marked as stale because it has not had """ """recent activity. If you think this still needs to be addressed """ """please comment on this thread.\n\nPlease note that issues that do not follow the """ """[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) """ """are likely to be ignored.""" ) if __name__ == "__main__": main()
704
import unittest from datasets import load_dataset from transformers.pipelines import pipeline from transformers.testing_utils import is_pipeline_test, nested_simplify, require_torch, slow @is_pipeline_test @require_torch class UpperCamelCase_ ( unittest.TestCase ): @require_torch def _snake_case ( self :Dict ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = pipeline( task="""zero-shot-audio-classification""" , model="""hf-internal-testing/tiny-clap-htsat-unfused""" ) SCREAMING_SNAKE_CASE__ = load_dataset("""ashraq/esc50""" ) SCREAMING_SNAKE_CASE__ = dataset["""train"""]["""audio"""][-1]["""array"""] SCREAMING_SNAKE_CASE__ = audio_classifier(__A , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__A ) , [{"""score""": 0.5_0_1, """label""": """Sound of a dog"""}, {"""score""": 0.4_9_9, """label""": """Sound of vaccum cleaner"""}] , ) @unittest.skip("""No models are available in TF""" ) def _snake_case ( self :Dict ) -> List[str]: """simple docstring""" pass @slow @require_torch def _snake_case ( self :Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = pipeline( task="""zero-shot-audio-classification""" , model="""laion/clap-htsat-unfused""" , ) # This is an audio of a dog SCREAMING_SNAKE_CASE__ = load_dataset("""ashraq/esc50""" ) SCREAMING_SNAKE_CASE__ = dataset["""train"""]["""audio"""][-1]["""array"""] SCREAMING_SNAKE_CASE__ = audio_classifier(__A , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__A ) , [ {"""score""": 0.9_9_9, """label""": """Sound of a dog"""}, {"""score""": 0.0_0_1, """label""": """Sound of vaccum cleaner"""}, ] , ) SCREAMING_SNAKE_CASE__ = audio_classifier([audio] * 5 , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] ) self.assertEqual( nested_simplify(__A ) , [ [ {"""score""": 0.9_9_9, """label""": """Sound of a dog"""}, {"""score""": 0.0_0_1, """label""": """Sound of vaccum cleaner"""}, ], ] * 5 , ) SCREAMING_SNAKE_CASE__ = audio_classifier( [audio] * 5 , candidate_labels=["""Sound of a dog""", """Sound of vaccum cleaner"""] , batch_size=5 ) self.assertEqual( nested_simplify(__A ) , [ [ {"""score""": 0.9_9_9, """label""": """Sound of a dog"""}, {"""score""": 0.0_0_1, """label""": """Sound of vaccum cleaner"""}, ], ] * 5 , ) @unittest.skip("""No models are available in TF""" ) def _snake_case ( self :str ) -> Optional[int]: """simple docstring""" pass
59
0
from functools import reduce _lowerCamelCase = ( "73167176531330624919225119674426574742355349194934" "96983520312774506326239578318016984801869478851843" "85861560789112949495459501737958331952853208805511" "12540698747158523863050715693290963295227443043557" "66896648950445244523161731856403098711121722383113" "62229893423380308135336276614282806444486645238749" "30358907296290491560440772390713810515859307960866" "70172427121883998797908792274921901699720888093776" "65727333001053367881220235421809751254540594752243" "52584907711670556013604839586446706324415722155397" "53697817977846174064955149290862569321978468622482" "83972241375657056057490261407972968652414535100474" "82166370484403199890008895243450658541227588666881" "16427171479924442928230863465674813919123162824586" "17866458359124566529476545682848912883142607690042" "24219022671055626321111109370544217506941658960408" "07198403850962455444362981230987879927244284909188" "84580156166097919133875499200524063689912560717606" "05886116467109405077541002256983155200055935729725" "71636269561882670428252483600823257530420752963450" ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str = N ): return max( # mypy cannot properly interpret reduce int(reduce(lambda UpperCamelCase__ , UpperCamelCase__ : str(int(_A ) * int(_A ) ) , n[i : i + 13] ) ) for i in range(len(_A ) - 12 ) ) if __name__ == "__main__": print(F'''{solution() = }''')
705
import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 _lowerCamelCase = data_utils.TransfoXLTokenizer _lowerCamelCase = data_utils.TransfoXLCorpus _lowerCamelCase = data_utils _lowerCamelCase = data_utils def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Dict , UpperCamelCase__: Any , UpperCamelCase__: Any , UpperCamelCase__: Tuple ): if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(UpperCamelCase__ , """rb""" ) as fp: SCREAMING_SNAKE_CASE__ = pickle.load(UpperCamelCase__ , encoding="""latin1""" ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) SCREAMING_SNAKE_CASE__ = pytorch_dump_folder_path + """/""" + VOCAB_FILES_NAMES["""pretrained_vocab_file"""] print(f'''Save vocabulary to {pytorch_vocab_dump_path}''' ) SCREAMING_SNAKE_CASE__ = corpus.vocab.__dict__ torch.save(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = corpus.__dict__ corpus_dict_no_vocab.pop("""vocab""" , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = pytorch_dump_folder_path + """/""" + CORPUS_NAME print(f'''Save dataset to {pytorch_dataset_dump_path}''' ) torch.save(UpperCamelCase__ , UpperCamelCase__ ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model SCREAMING_SNAKE_CASE__ = os.path.abspath(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = os.path.abspath(UpperCamelCase__ ) print(f'''Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.''' ) # Initialise PyTorch model if transfo_xl_config_file == "": SCREAMING_SNAKE_CASE__ = TransfoXLConfig() else: SCREAMING_SNAKE_CASE__ = TransfoXLConfig.from_json_file(UpperCamelCase__ ) print(f'''Building PyTorch model from configuration: {config}''' ) SCREAMING_SNAKE_CASE__ = TransfoXLLMHeadModel(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = load_tf_weights_in_transfo_xl(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # Save pytorch-model SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) print(f'''Save PyTorch model to {os.path.abspath(UpperCamelCase__ )}''' ) torch.save(model.state_dict() , UpperCamelCase__ ) print(f'''Save configuration file to {os.path.abspath(UpperCamelCase__ )}''' ) with open(UpperCamelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--tf_checkpoint_path', default='', type=str, help='An optional path to a TensorFlow checkpoint path to be converted.', ) parser.add_argument( '--transfo_xl_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained BERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--transfo_xl_dataset_file', default='', type=str, help='An optional dataset file to be converted in a vocabulary.', ) _lowerCamelCase = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
59
0
_lowerCamelCase = {'a': ['c', 'b'], 'b': ['d', 'e'], 'c': [], 'd': [], 'e': []} _lowerCamelCase = ['a', 'b', 'c', 'd', 'e'] def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Dict ): SCREAMING_SNAKE_CASE__ = start # add current to visited visited.append(lowerCamelCase_ ) SCREAMING_SNAKE_CASE__ = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: SCREAMING_SNAKE_CASE__ = topological_sort(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # if all neighbors visited add current to sort sort.append(lowerCamelCase_ ) # if all vertices haven't been visited select a new one to visit if len(lowerCamelCase_ ) != len(lowerCamelCase_ ): for vertice in vertices: if vertice not in visited: SCREAMING_SNAKE_CASE__ = topological_sort(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # return sort return sort if __name__ == "__main__": _lowerCamelCase = topological_sort('a', [], []) print(sort)
706
import argparse import tensorflow as tf import torch from transformers import BertConfig, BertForMaskedLM from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertPooler, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging logging.set_verbosity_info() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str , UpperCamelCase__: str , UpperCamelCase__: str ): def get_masked_lm_array(UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = f'''masked_lm/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) def get_encoder_array(UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = f'''encoder/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) def get_encoder_layer_array(UpperCamelCase__: int , UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = f'''encoder/_transformer_layers/{layer_index}/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) def get_encoder_attention_layer_array(UpperCamelCase__: int , UpperCamelCase__: str , UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = f'''encoder/_transformer_layers/{layer_index}/_attention_layer/{name}/.ATTRIBUTES/VARIABLE_VALUE''' SCREAMING_SNAKE_CASE__ = tf.train.load_variable(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = array.reshape(UpperCamelCase__ ) if "kernel" in name: SCREAMING_SNAKE_CASE__ = array.transpose() return torch.from_numpy(UpperCamelCase__ ) print(f'''Loading model based on config from {config_path}...''' ) SCREAMING_SNAKE_CASE__ = BertConfig.from_json_file(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = BertForMaskedLM(UpperCamelCase__ ) # Layers for layer_index in range(0 , config.num_hidden_layers ): SCREAMING_SNAKE_CASE__ = model.bert.encoder.layer[layer_index] # Self-attention SCREAMING_SNAKE_CASE__ = layer.attention.self SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_query_dense/kernel""" , self_attn.query.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_query_dense/bias""" , self_attn.query.bias.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_key_dense/kernel""" , self_attn.key.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_key_dense/bias""" , self_attn.key.bias.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_value_dense/kernel""" , self_attn.value.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_value_dense/bias""" , self_attn.value.bias.data.shape ) # Self-attention Output SCREAMING_SNAKE_CASE__ = layer.attention.output SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_output_dense/kernel""" , self_output.dense.weight.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_attention_layer_array( UpperCamelCase__ , """_output_dense/bias""" , self_output.dense.bias.data.shape ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_attention_layer_norm/gamma""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_attention_layer_norm/beta""" ) # Intermediate SCREAMING_SNAKE_CASE__ = layer.intermediate SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_intermediate_dense/kernel""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_intermediate_dense/bias""" ) # Output SCREAMING_SNAKE_CASE__ = layer.output SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_dense/kernel""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_dense/bias""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_layer_norm/gamma""" ) SCREAMING_SNAKE_CASE__ = get_encoder_layer_array(UpperCamelCase__ , """_output_layer_norm/beta""" ) # Embeddings SCREAMING_SNAKE_CASE__ = get_encoder_array("""_position_embedding_layer/embeddings""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_type_embedding_layer/embeddings""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_embedding_norm_layer/gamma""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_embedding_norm_layer/beta""" ) # LM Head SCREAMING_SNAKE_CASE__ = model.cls.predictions.transform SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""dense/kernel""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""dense/bias""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""layer_norm/gamma""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""layer_norm/beta""" ) SCREAMING_SNAKE_CASE__ = get_masked_lm_array("""embedding_table""" ) # Pooling SCREAMING_SNAKE_CASE__ = BertPooler(config=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_pooler_layer/kernel""" ) SCREAMING_SNAKE_CASE__ = get_encoder_array("""_pooler_layer/bias""" ) # Export final model model.save_pretrained(UpperCamelCase__ ) # Integration test - should load without any errors ;) SCREAMING_SNAKE_CASE__ = BertForMaskedLM.from_pretrained(UpperCamelCase__ ) print(new_model.eval() ) print("""Model conversion was done sucessfully!""" ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '--tf_checkpoint_path', type=str, required=True, help='Path to the TensorFlow Token Dropping checkpoint path.' ) parser.add_argument( '--bert_config_file', type=str, required=True, help='The config json file corresponding to the BERT model. This specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', type=str, required=True, help='Path to the output PyTorch model.', ) _lowerCamelCase = parser.parse_args() convert_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
59
0
import re def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str ): if len(re.findall("""[ATCG]""" , UpperCamelCase__ ) ) != len(UpperCamelCase__ ): raise ValueError("""Invalid Strand""" ) return dna.translate(dna.maketrans("""ATCG""" , """TAGC""" ) ) if __name__ == "__main__": import doctest doctest.testmod()
707
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _lowerCamelCase = '\\n Text data.\n Second line of data.' _lowerCamelCase = 'file' @pytest.fixture(scope="""session""" ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / (FILE_PATH + """.zstd""") SCREAMING_SNAKE_CASE__ = bytes(UpperCamelCase__ , """utf-8""" ) with zstd.open(UpperCamelCase__ , """wb""" ) as f: f.write(UpperCamelCase__ ) return path @pytest.fixture def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): with open(os.path.join(tmpfs.local_root_dir , UpperCamelCase__ ) , """w""" ) as f: f.write(UpperCamelCase__ ) return FILE_PATH @pytest.mark.parametrize("""compression_format""" , ["""gzip""", """xz""", """zstd"""] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int , UpperCamelCase__: Dict , UpperCamelCase__: int , UpperCamelCase__: str , UpperCamelCase__: Optional[int] , UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = {"""gzip""": gz_file, """xz""": xz_file, """zstd""": zstd_path} SCREAMING_SNAKE_CASE__ = input_paths[compression_format] SCREAMING_SNAKE_CASE__ = tmp_path / """cache""" SCREAMING_SNAKE_CASE__ = DownloadConfig(cache_dir=UpperCamelCase__ , extract_compressed_file=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = cached_path(UpperCamelCase__ , download_config=UpperCamelCase__ ) with open(UpperCamelCase__ ) as f: SCREAMING_SNAKE_CASE__ = f.read() with open(UpperCamelCase__ ) as f: SCREAMING_SNAKE_CASE__ = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("""default_extracted""" , [True, False] ) @pytest.mark.parametrize("""default_cache_dir""" , [True, False] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Tuple , UpperCamelCase__: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Any , UpperCamelCase__: Union[str, Any] ): SCREAMING_SNAKE_CASE__ = """custom_cache""" SCREAMING_SNAKE_CASE__ = """custom_extracted_dir""" SCREAMING_SNAKE_CASE__ = tmp_path / """custom_extracted_path""" if default_extracted: SCREAMING_SNAKE_CASE__ = ("""downloads""" if default_cache_dir else custom_cache_dir, """extracted""") else: monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_DIR""" , UpperCamelCase__ ) monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_PATH""" , str(UpperCamelCase__ ) ) SCREAMING_SNAKE_CASE__ = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) SCREAMING_SNAKE_CASE__ = xz_file SCREAMING_SNAKE_CASE__ = ( DownloadConfig(extract_compressed_file=UpperCamelCase__ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=UpperCamelCase__ ) ) SCREAMING_SNAKE_CASE__ = cached_path(UpperCamelCase__ , download_config=UpperCamelCase__ ) assert Path(UpperCamelCase__ ).parent.parts[-2:] == expected def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[int] ): # absolute path SCREAMING_SNAKE_CASE__ = str(Path(UpperCamelCase__ ).resolve() ) assert cached_path(UpperCamelCase__ ) == text_file # relative path SCREAMING_SNAKE_CASE__ = str(Path(UpperCamelCase__ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(UpperCamelCase__ ) == text_file def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): # absolute path SCREAMING_SNAKE_CASE__ = str(tmp_path.resolve() / """__missing_file__.txt""" ) with pytest.raises(UpperCamelCase__ ): cached_path(UpperCamelCase__ ) # relative path SCREAMING_SNAKE_CASE__ = """./__missing_file__.txt""" with pytest.raises(UpperCamelCase__ ): cached_path(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): SCREAMING_SNAKE_CASE__ = get_from_cache(f'''tmp://{tmpfs_file}''' ) with open(UpperCamelCase__ ) as f: SCREAMING_SNAKE_CASE__ = f.read() assert output_file_content == FILE_CONTENT @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( ): with pytest.raises(UpperCamelCase__ ): cached_path("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(UpperCamelCase__ ): http_get("""https://huggingface.co""" , temp_file=UpperCamelCase__ ) with pytest.raises(UpperCamelCase__ ): http_head("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[Any] ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(UpperCamelCase__ ): ftp_get("""ftp://huggingface.co""" , temp_file=UpperCamelCase__ ) with pytest.raises(UpperCamelCase__ ): ftp_head("""ftp://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(UpperCamelCase__ ): fsspec_get("""s3://huggingface.co""" , temp_file=UpperCamelCase__ ) with pytest.raises(UpperCamelCase__ ): fsspec_head("""s3://huggingface.co""" )
59
0
'''simple docstring''' import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt'} # See all LED models at https://huggingface.co/models?filter=LED _lowerCamelCase = { 'vocab_file': { 'allenai/led-base-16384': 'https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json', }, 'merges_file': { 'allenai/led-base-16384': 'https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt', }, 'tokenizer_file': { 'allenai/led-base-16384': 'https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json', }, } _lowerCamelCase = { 'allenai/led-base-16384': 16384, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = ( list(range(ord("""!""" ) , ord("""~""" ) + 1 ) ) + list(range(ord("""¡""" ) , ord("""¬""" ) + 1 ) ) + list(range(ord("""®""" ) , ord("""ÿ""" ) + 1 ) ) ) SCREAMING_SNAKE_CASE__ = bs[:] SCREAMING_SNAKE_CASE__ = 0 for b in range(2**8 ): if b not in bs: bs.append(__A ) cs.append(2**8 + n ) n += 1 SCREAMING_SNAKE_CASE__ = [chr(__A ) for n in cs] return dict(zip(__A , __A ) ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = set() SCREAMING_SNAKE_CASE__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) SCREAMING_SNAKE_CASE__ = char return pairs class UpperCamelCase_ ( __A ): lowerCamelCase_ = VOCAB_FILES_NAMES lowerCamelCase_ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase_ = ["input_ids", "attention_mask"] def __init__( self :Tuple , __A :int , __A :str , __A :str="replace" , __A :int="<s>" , __A :Optional[int]="</s>" , __A :Optional[int]="</s>" , __A :List[Any]="<s>" , __A :str="<unk>" , __A :Dict="<pad>" , __A :Union[str, Any]="<mask>" , __A :str=False , **__A :int , ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = AddedToken(__A , lstrip=__A , rstrip=__A ) if isinstance(__A , __A ) else bos_token SCREAMING_SNAKE_CASE__ = AddedToken(__A , lstrip=__A , rstrip=__A ) if isinstance(__A , __A ) else eos_token SCREAMING_SNAKE_CASE__ = AddedToken(__A , lstrip=__A , rstrip=__A ) if isinstance(__A , __A ) else sep_token SCREAMING_SNAKE_CASE__ = AddedToken(__A , lstrip=__A , rstrip=__A ) if isinstance(__A , __A ) else cls_token SCREAMING_SNAKE_CASE__ = AddedToken(__A , lstrip=__A , rstrip=__A ) if isinstance(__A , __A ) else unk_token SCREAMING_SNAKE_CASE__ = AddedToken(__A , lstrip=__A , rstrip=__A ) if isinstance(__A , __A ) else pad_token # Mask token behave like a normal word, i.e. include the space before it SCREAMING_SNAKE_CASE__ = AddedToken(__A , lstrip=__A , rstrip=__A ) if isinstance(__A , __A ) else mask_token super().__init__( errors=__A , bos_token=__A , eos_token=__A , unk_token=__A , sep_token=__A , cls_token=__A , pad_token=__A , mask_token=__A , add_prefix_space=__A , **__A , ) with open(__A , encoding="""utf-8""" ) as vocab_handle: SCREAMING_SNAKE_CASE__ = json.load(__A ) SCREAMING_SNAKE_CASE__ = {v: k for k, v in self.encoder.items()} SCREAMING_SNAKE_CASE__ = errors # how to handle errors in decoding SCREAMING_SNAKE_CASE__ = bytes_to_unicode() SCREAMING_SNAKE_CASE__ = {v: k for k, v in self.byte_encoder.items()} with open(__A , encoding="""utf-8""" ) as merges_handle: SCREAMING_SNAKE_CASE__ = merges_handle.read().split("""\n""" )[1:-1] SCREAMING_SNAKE_CASE__ = [tuple(merge.split() ) for merge in bpe_merges] SCREAMING_SNAKE_CASE__ = dict(zip(__A , range(len(__A ) ) ) ) SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions SCREAMING_SNAKE_CASE__ = re.compile(r"""\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+""" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def _snake_case ( self :Union[str, Any] ) -> Optional[int]: """simple docstring""" return len(self.encoder ) def _snake_case ( self :Union[str, Any] ) -> Union[str, Any]: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def _snake_case ( self :int , __A :str ) -> Optional[Any]: """simple docstring""" if token in self.cache: return self.cache[token] SCREAMING_SNAKE_CASE__ = tuple(__A ) SCREAMING_SNAKE_CASE__ = get_pairs(__A ) if not pairs: return token while True: SCREAMING_SNAKE_CASE__ = min(__A , key=lambda __A : self.bpe_ranks.get(__A , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break SCREAMING_SNAKE_CASE__ = bigram SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = 0 while i < len(__A ): try: SCREAMING_SNAKE_CASE__ = word.index(__A , __A ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) SCREAMING_SNAKE_CASE__ = j if word[i] == first and i < len(__A ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 SCREAMING_SNAKE_CASE__ = tuple(__A ) SCREAMING_SNAKE_CASE__ = new_word if len(__A ) == 1: break else: SCREAMING_SNAKE_CASE__ = get_pairs(__A ) SCREAMING_SNAKE_CASE__ = ''' '''.join(__A ) SCREAMING_SNAKE_CASE__ = word return word def _snake_case ( self :int , __A :Optional[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = [] for token in re.findall(self.pat , __A ): SCREAMING_SNAKE_CASE__ = ''''''.join( self.byte_encoder[b] for b in token.encode("""utf-8""" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(__A ).split(""" """ ) ) return bpe_tokens def _snake_case ( self :Union[str, Any] , __A :Union[str, Any] ) -> List[str]: """simple docstring""" return self.encoder.get(__A , self.encoder.get(self.unk_token ) ) def _snake_case ( self :Dict , __A :Optional[int] ) -> Any: """simple docstring""" return self.decoder.get(__A ) def _snake_case ( self :List[str] , __A :List[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ''''''.join(__A ) SCREAMING_SNAKE_CASE__ = bytearray([self.byte_decoder[c] for c in text] ).decode("""utf-8""" , errors=self.errors ) return text def _snake_case ( self :Any , __A :str , __A :Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__A ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return SCREAMING_SNAKE_CASE__ = os.path.join( __A , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join( __A , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(__A , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=__A , ensure_ascii=__A ) + """\n""" ) SCREAMING_SNAKE_CASE__ = 0 with open(__A , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda __A : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' """ Please check that the tokenizer is not corrupted!""" ) SCREAMING_SNAKE_CASE__ = token_index writer.write(""" """.join(__A ) + """\n""" ) index += 1 return vocab_file, merge_file def _snake_case ( self :str , __A :List[int] , __A :Optional[List[int]] = None ) -> List[int]: """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] SCREAMING_SNAKE_CASE__ = [self.cls_token_id] SCREAMING_SNAKE_CASE__ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _snake_case ( self :Optional[int] , __A :List[int] , __A :Optional[List[int]] = None , __A :bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__A , token_ids_a=__A , already_has_special_tokens=__A ) if token_ids_a is None: return [1] + ([0] * len(__A )) + [1] return [1] + ([0] * len(__A )) + [1, 1] + ([0] * len(__A )) + [1] def _snake_case ( self :Any , __A :List[int] , __A :Optional[List[int]] = None ) -> List[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = [self.sep_token_id] SCREAMING_SNAKE_CASE__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _snake_case ( self :Optional[int] , __A :Optional[Any] , __A :Optional[int]=False , **__A :Optional[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = kwargs.pop("""add_prefix_space""" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(__A ) > 0 and not text[0].isspace()): SCREAMING_SNAKE_CASE__ = ''' ''' + text return (text, kwargs) def _snake_case ( self :Any , __A :Union[Dict[str, EncodedInput], BatchEncoding] , __A :Optional[int] = None , __A :PaddingStrategy = PaddingStrategy.DO_NOT_PAD , __A :Optional[int] = None , __A :Optional[bool] = None , ) -> dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = super()._pad( encoded_inputs=__A , max_length=__A , padding_strategy=__A , pad_to_multiple_of=__A , return_attention_mask=__A , ) # Load from model defaults if return_attention_mask is None: SCREAMING_SNAKE_CASE__ = '''attention_mask''' in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: SCREAMING_SNAKE_CASE__ = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. SCREAMING_SNAKE_CASE__ = len(encoded_inputs["""global_attention_mask"""] ) != len(__A ) if needs_to_be_padded: SCREAMING_SNAKE_CASE__ = len(__A ) - len(encoded_inputs["""global_attention_mask"""] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` SCREAMING_SNAKE_CASE__ = ( encoded_inputs['''global_attention_mask'''] + [-1] * difference ) elif self.padding_side == "left": SCREAMING_SNAKE_CASE__ = [-1] * difference + encoded_inputs[ '''global_attention_mask''' ] else: raise ValueError("""Invalid padding strategy:""" + str(self.padding_side ) ) return encoded_inputs
708
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _lowerCamelCase = logging.getLogger(__name__) def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser( description="""Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset.""" ) parser.add_argument( """--dataset_name""" , type=UpperCamelCase__ , default="""wikitext""" , help="""Name of the training. Explore datasets at: hf.co/datasets.""" , ) parser.add_argument( """--dataset_config""" , type=UpperCamelCase__ , default="""wikitext-103-raw-v1""" , help="""Configuration name of the dataset.""" ) parser.add_argument( """--tokenizer_name_or_path""" , type=UpperCamelCase__ , default="""sayakpaul/unigram-tokenizer-wikitext""" , help="""Tokenizer identifier. Can be a local filepath or a Hub identifier.""" , ) parser.add_argument( """--shard_size""" , type=UpperCamelCase__ , default=1_000 , help="""Number of entries to go in a single shard.""" , ) parser.add_argument("""--split""" , type=UpperCamelCase__ , default="""train""" , choices=["""train""", """test""", """validation"""] ) parser.add_argument( """--limit""" , default=UpperCamelCase__ , type=UpperCamelCase__ , help="""Limit the number of shards (used for debugging).""" , ) parser.add_argument( """--max_length""" , type=UpperCamelCase__ , default=512 , help="""Maximum sequence length. For training on TPUs, it helps to have a maximum""" """ sequence length that is a multiple of 8.""" , ) parser.add_argument( """--output_dir""" , default="""tf-tpu""" , type=UpperCamelCase__ , help="""Output directory where the TFRecord shards will be saved. If the""" """ path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord""" """ shards will be directly saved to a Google Cloud Storage bucket.""" , ) SCREAMING_SNAKE_CASE__ = parser.parse_args() return args def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[Any] ): def fn(UpperCamelCase__: Any ): return tokenizer(examples["""text"""] ) return fn def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = [] for i in range(len(tokenized_data["""input_ids"""] ) ): SCREAMING_SNAKE_CASE__ = { """input_ids""": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["""input_ids"""][i] ) ), """attention_mask""": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["""attention_mask"""][i] ) ), } SCREAMING_SNAKE_CASE__ = tf.train.Features(feature=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = tf.train.Example(features=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = example.SerializeToString() records.append(UpperCamelCase__ ) return records def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): SCREAMING_SNAKE_CASE__ = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: SCREAMING_SNAKE_CASE__ = min(len(UpperCamelCase__ ) , args.limit ) SCREAMING_SNAKE_CASE__ = dataset.select(range(UpperCamelCase__ ) ) print(f'''Limiting the dataset to {args.limit} entries.''' ) SCREAMING_SNAKE_CASE__ = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) SCREAMING_SNAKE_CASE__ = os.path.join(args.output_dir , args.split ) if not os.path.exists(UpperCamelCase__ ): os.makedirs(UpperCamelCase__ ) else: SCREAMING_SNAKE_CASE__ = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. SCREAMING_SNAKE_CASE__ = tokenize_function(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = dataset.map(UpperCamelCase__ , batched=UpperCamelCase__ , num_proc=4 , remove_columns=["""text"""] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(UpperCamelCase__: int ): # Concatenate all texts. SCREAMING_SNAKE_CASE__ = {k: sum(examples[k] , [] ) for k in examples.keys()} SCREAMING_SNAKE_CASE__ = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 SCREAMING_SNAKE_CASE__ = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. SCREAMING_SNAKE_CASE__ = { k: [t[i : i + args.max_length] for i in range(0 , UpperCamelCase__ , args.max_length )] for k, t in concatenated_examples.items() } return result SCREAMING_SNAKE_CASE__ = dataset_tokenized.map(UpperCamelCase__ , batched=UpperCamelCase__ , batch_size=1_000 , num_proc=4 ) SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 0 for shard in range(0 , len(UpperCamelCase__ ) , args.shard_size ): SCREAMING_SNAKE_CASE__ = grouped_dataset[shard : shard + args.shard_size] SCREAMING_SNAKE_CASE__ = len(dataset_snapshot["""input_ids"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(UpperCamelCase__ , f'''dataset-{shard_count}-{records_containing}.tfrecord''' ) SCREAMING_SNAKE_CASE__ = get_serialized_examples(UpperCamelCase__ ) with tf.io.TFRecordWriter(UpperCamelCase__ ) as out_file: for i in range(len(UpperCamelCase__ ) ): SCREAMING_SNAKE_CASE__ = serialized_examples[i] out_file.write(UpperCamelCase__ ) print("""Wrote file {} containing {} records""".format(UpperCamelCase__ , UpperCamelCase__ ) ) shard_count += 1 total_records += records_containing with open(f'''split-{args.split}-records-count.txt''' , """w""" ) as f: print(f'''Total {args.split} records: {total_records}''' , file=UpperCamelCase__ ) if __name__ == "__main__": _lowerCamelCase = parse_args() main(args)
59
0
import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_MAPPING, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerModel, ) from transformers.models.segformer.modeling_segformer import SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import SegformerImageProcessor class UpperCamelCase_ ( UpperCamelCase__ ): def _snake_case ( self :Union[str, Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_lowerCAmelCase , """hidden_sizes""" ) ) self.parent.assertTrue(hasattr(_lowerCAmelCase , """num_attention_heads""" ) ) self.parent.assertTrue(hasattr(_lowerCAmelCase , """num_encoder_blocks""" ) ) class UpperCamelCase_ : def __init__( self :int , __A :int , __A :Optional[int]=13 , __A :Union[str, Any]=64 , __A :Tuple=3 , __A :int=4 , __A :List[Any]=[2, 2, 2, 2] , __A :List[str]=[8, 4, 2, 1] , __A :str=[16, 32, 64, 128] , __A :Optional[int]=[1, 4, 8, 16] , __A :Optional[Any]=[1, 2, 4, 8] , __A :List[Any]=True , __A :Tuple=True , __A :Optional[int]="gelu" , __A :Tuple=0.1 , __A :int=0.1 , __A :Any=0.0_2 , __A :Optional[int]=3 , __A :Dict=None , ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = parent SCREAMING_SNAKE_CASE__ = batch_size SCREAMING_SNAKE_CASE__ = image_size SCREAMING_SNAKE_CASE__ = num_channels SCREAMING_SNAKE_CASE__ = num_encoder_blocks SCREAMING_SNAKE_CASE__ = sr_ratios SCREAMING_SNAKE_CASE__ = depths SCREAMING_SNAKE_CASE__ = hidden_sizes SCREAMING_SNAKE_CASE__ = downsampling_rates SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = is_training SCREAMING_SNAKE_CASE__ = use_labels SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = num_labels SCREAMING_SNAKE_CASE__ = scope def _snake_case ( self :Any ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE__ = None if self.use_labels: SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) SCREAMING_SNAKE_CASE__ = self.get_config() return config, pixel_values, labels def _snake_case ( self :List[str] ) -> int: """simple docstring""" return SegformerConfig( image_size=self.image_size , num_channels=self.num_channels , num_encoder_blocks=self.num_encoder_blocks , depths=self.depths , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def _snake_case ( self :Tuple , __A :Optional[int] , __A :Any , __A :str ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = SegformerModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE__ = model(_lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = SCREAMING_SNAKE_CASE__ = self.image_size // (self.downsampling_rates[-1] * 2) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], expected_height, expected_width) ) def _snake_case ( self :List[str] , __A :Dict , __A :Union[str, Any] , __A :Any ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.num_labels SCREAMING_SNAKE_CASE__ = SegformerForSemanticSegmentation(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE__ = model(_lowerCAmelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) SCREAMING_SNAKE_CASE__ = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size // 4, self.image_size // 4) ) self.parent.assertGreater(result.loss , 0.0 ) def _snake_case ( self :Union[str, Any] , __A :Optional[Any] , __A :List[Any] , __A :Any ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = SegformerForSemanticSegmentation(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE__ = torch.randint(0 , 1 , (self.batch_size, self.image_size, self.image_size) ).to(_lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertGreater(result.loss , 0.0 ) def _snake_case ( self :Tuple ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = config_and_inputs SCREAMING_SNAKE_CASE__ = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class UpperCamelCase_ ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = ( ( SegformerModel, SegformerForSemanticSegmentation, SegformerForImageClassification, ) if is_torch_available() else () ) lowerCamelCase_ = ( { "feature-extraction": SegformerModel, "image-classification": SegformerForImageClassification, "image-segmentation": SegformerForSemanticSegmentation, } if is_torch_available() else {} ) lowerCamelCase_ = True lowerCamelCase_ = False lowerCamelCase_ = False lowerCamelCase_ = False def _snake_case ( self :Dict ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = SegformerModelTester(self ) SCREAMING_SNAKE_CASE__ = SegformerConfigTester(self , config_class=_lowerCAmelCase ) def _snake_case ( self :List[str] ) -> str: """simple docstring""" self.config_tester.run_common_tests() def _snake_case ( self :int ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def _snake_case ( self :int ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_binary_image_segmentation(*_lowerCAmelCase ) def _snake_case ( self :List[str] ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_segmentation(*_lowerCAmelCase ) @unittest.skip("""SegFormer does not use inputs_embeds""" ) def _snake_case ( self :Optional[Any] ) -> Any: """simple docstring""" pass @unittest.skip("""SegFormer does not have get_input_embeddings method and get_output_embeddings methods""" ) def _snake_case ( self :Optional[Any] ) -> int: """simple docstring""" pass def _snake_case ( self :List[str] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ = model_class(_lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE__ = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE__ = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def _snake_case ( self :List[Any] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE__ = True for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) SCREAMING_SNAKE_CASE__ = outputs.attentions SCREAMING_SNAKE_CASE__ = sum(self.model_tester.depths ) self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) # check that output_attentions also work using config del inputs_dict["output_attentions"] SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) SCREAMING_SNAKE_CASE__ = outputs.attentions self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) # verify the first attentions (first block, first layer) SCREAMING_SNAKE_CASE__ = (self.model_tester.image_size // 4) ** 2 SCREAMING_SNAKE_CASE__ = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) # verify the last attentions (last block, last layer) SCREAMING_SNAKE_CASE__ = (self.model_tester.image_size // 32) ** 2 SCREAMING_SNAKE_CASE__ = (self.model_tester.image_size // (32 * self.model_tester.sr_ratios[-1])) ** 2 self.assertListEqual( list(attentions[-1].shape[-3:] ) , [self.model_tester.num_attention_heads[-1], expected_seq_len, expected_reduced_seq_len] , ) SCREAMING_SNAKE_CASE__ = len(_lowerCAmelCase ) # Check attention is always last and order is fine SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) self.assertEqual(out_len + 1 , len(_lowerCAmelCase ) ) SCREAMING_SNAKE_CASE__ = outputs.attentions self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) # verify the first attentions (first block, first layer) SCREAMING_SNAKE_CASE__ = (self.model_tester.image_size // 4) ** 2 SCREAMING_SNAKE_CASE__ = (self.model_tester.image_size // (4 * self.model_tester.sr_ratios[0])) ** 2 self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads[0], expected_seq_len, expected_reduced_seq_len] , ) def _snake_case ( self :int ) -> Any: """simple docstring""" def check_hidden_states_output(__A :int , __A :Optional[int] , __A :List[str] ): SCREAMING_SNAKE_CASE__ = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) SCREAMING_SNAKE_CASE__ = outputs.hidden_states SCREAMING_SNAKE_CASE__ = self.model_tester.num_encoder_blocks self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.hidden_sizes[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] SCREAMING_SNAKE_CASE__ = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def _snake_case ( self :Any ) -> Dict: """simple docstring""" if not self.model_tester.is_training: return SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE__ = True for model_class in self.all_model_classes: if model_class in get_values(_lowerCAmelCase ): continue SCREAMING_SNAKE_CASE__ = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.train() SCREAMING_SNAKE_CASE__ = self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = model(**_lowerCAmelCase ).loss loss.backward() @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def _snake_case ( self :Any ) -> Union[str, Any]: """simple docstring""" pass @slow def _snake_case ( self :int ) -> int: """simple docstring""" for model_name in SEGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE__ = SegformerModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch class UpperCamelCase_ ( unittest.TestCase ): @slow def _snake_case ( self :Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_lowerCAmelCase , align=_lowerCAmelCase , do_random_crop=_lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = SegformerForSemanticSegmentation.from_pretrained("""nvidia/segformer-b0-finetuned-ade-512-512""" ).to( _lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = prepare_img() SCREAMING_SNAKE_CASE__ = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ) SCREAMING_SNAKE_CASE__ = encoded_inputs.pixel_values.to(_lowerCAmelCase ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(_lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = torch.tensor( [ [[-4.6_3_1_0, -5.5_2_3_2, -6.2_3_5_6], [-5.1_9_2_1, -6.1_4_4_4, -6.5_9_9_6], [-5.4_4_2_4, -6.2_7_9_0, -6.7_5_7_4]], [[-1_2.1_3_9_1, -1_3.3_1_2_2, -1_3.9_5_5_4], [-1_2.8_7_3_2, -1_3.9_3_5_2, -1_4.3_5_6_3], [-1_2.9_4_3_8, -1_3.8_2_2_6, -1_4.2_5_1_3]], [[-1_2.5_1_3_4, -1_3.4_6_8_6, -1_4.4_9_1_5], [-1_2.8_6_6_9, -1_4.4_3_4_3, -1_4.7_7_5_8], [-1_3.2_5_2_3, -1_4.5_8_1_9, -1_5.0_6_9_4]], ] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , _lowerCAmelCase , atol=1E-4 ) ) @slow def _snake_case ( self :Dict ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_lowerCAmelCase , align=_lowerCAmelCase , do_random_crop=_lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = SegformerForSemanticSegmentation.from_pretrained( """nvidia/segformer-b1-finetuned-cityscapes-1024-1024""" ).to(_lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = prepare_img() SCREAMING_SNAKE_CASE__ = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ) SCREAMING_SNAKE_CASE__ = encoded_inputs.pixel_values.to(_lowerCAmelCase ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(_lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = torch.Size((1, model.config.num_labels, 128, 128) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = torch.tensor( [ [[-1_3.5_7_4_8, -1_3.9_1_1_1, -1_2.6_5_0_0], [-1_4.3_5_0_0, -1_5.3_6_8_3, -1_4.2_3_2_8], [-1_4.7_5_3_2, -1_6.0_4_2_4, -1_5.6_0_8_7]], [[-1_7.1_6_5_1, -1_5.8_7_2_5, -1_2.9_6_5_3], [-1_7.2_5_8_0, -1_7.3_7_1_8, -1_4.8_2_2_3], [-1_6.6_0_5_8, -1_6.8_7_8_3, -1_6.7_4_5_2]], [[-3.6_4_5_6, -3.0_2_0_9, -1.4_2_0_3], [-3.0_7_9_7, -3.1_9_5_9, -2.0_0_0_0], [-1.8_7_5_7, -1.9_2_1_7, -1.6_9_9_7]], ] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3, :3] , _lowerCAmelCase , atol=1E-1 ) ) @slow def _snake_case ( self :List[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_lowerCAmelCase , align=_lowerCAmelCase , do_random_crop=_lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = SegformerForSemanticSegmentation.from_pretrained("""nvidia/segformer-b0-finetuned-ade-512-512""" ).to( _lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = prepare_img() SCREAMING_SNAKE_CASE__ = image_processor(images=_lowerCAmelCase , return_tensors="""pt""" ) SCREAMING_SNAKE_CASE__ = encoded_inputs.pixel_values.to(_lowerCAmelCase ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(_lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = outputs.logits.detach().cpu() SCREAMING_SNAKE_CASE__ = image_processor.post_process_semantic_segmentation(outputs=_lowerCAmelCase , target_sizes=[(500, 300)] ) SCREAMING_SNAKE_CASE__ = torch.Size((500, 300) ) self.assertEqual(segmentation[0].shape , _lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = image_processor.post_process_semantic_segmentation(outputs=_lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = torch.Size((128, 128) ) self.assertEqual(segmentation[0].shape , _lowerCAmelCase )
709
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] ): SCREAMING_SNAKE_CASE__ = [] for data in source_data: for i, el in enumerate(UpperCamelCase__ ): if len(UpperCamelCase__ ) < i + 1: data_lists.append([] ) data_lists[i].append(float(UpperCamelCase__ ) ) return data_lists def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] , UpperCamelCase__: list[int] ): SCREAMING_SNAKE_CASE__ = [] for dlist, weight in zip(UpperCamelCase__ , UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = min(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = max(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = [] # for weight 0 score is 1 - actual score if weight == 0: for item in dlist: try: score.append(1 - ((item - mind) / (maxd - mind)) ) except ZeroDivisionError: score.append(1 ) elif weight == 1: for item in dlist: try: score.append((item - mind) / (maxd - mind) ) except ZeroDivisionError: score.append(0 ) # weight not 0 or 1 else: SCREAMING_SNAKE_CASE__ = f'''Invalid weight of {weight:f} provided''' raise ValueError(UpperCamelCase__ ) score_lists.append(UpperCamelCase__ ) return score_lists def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] ): SCREAMING_SNAKE_CASE__ = [0 for i in range(len(score_lists[0] ) )] for slist in score_lists: for j, ele in enumerate(UpperCamelCase__ ): SCREAMING_SNAKE_CASE__ = final_scores[j] + ele return final_scores def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: list[list[float]] , UpperCamelCase__: list[int] ): SCREAMING_SNAKE_CASE__ = get_data(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = calculate_each_score(UpperCamelCase__ , UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = generate_final_scores(UpperCamelCase__ ) # append scores to source data for i, ele in enumerate(UpperCamelCase__ ): source_data[i].append(UpperCamelCase__ ) return source_data
59
0
from typing import List from .keymap import KEYMAP, get_character def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[str, Any] ): def decorator(UpperCamelCase__: List[Any] ): SCREAMING_SNAKE_CASE__ = getattr(lowerCamelCase__ , """handle_key""" , [] ) handle += [key] setattr(lowerCamelCase__ , """handle_key""" , lowerCamelCase__ ) return func return decorator def SCREAMING_SNAKE_CASE__ ( *UpperCamelCase__: Any ): def decorator(UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = getattr(lowerCamelCase__ , """handle_key""" , [] ) handle += keys setattr(lowerCamelCase__ , """handle_key""" , lowerCamelCase__ ) return func return decorator class UpperCamelCase_ ( UpperCamelCase__ ): def __new__( cls :Any , __A :List[str] , __A :int , __A :Union[str, Any] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = super().__new__(cls , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) if not hasattr(UpperCamelCase_ , """key_handler""" ): setattr(UpperCamelCase_ , """key_handler""" , {} ) setattr(UpperCamelCase_ , """handle_input""" , KeyHandler.handle_input ) for value in attrs.values(): SCREAMING_SNAKE_CASE__ = getattr(UpperCamelCase_ , """handle_key""" , [] ) for key in handled_keys: SCREAMING_SNAKE_CASE__ = value return new_cls @staticmethod def _snake_case ( cls :Tuple ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = get_character() if char != KEYMAP["undefined"]: SCREAMING_SNAKE_CASE__ = ord(UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = cls.key_handler.get(UpperCamelCase_ ) if handler: SCREAMING_SNAKE_CASE__ = char return handler(cls ) else: return None def SCREAMING_SNAKE_CASE__ ( cls: List[Any] ): return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy() )
710
import warnings from functools import wraps from typing import Callable def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Callable ): @wraps(UpperCamelCase__ ) def _inner_fn(*UpperCamelCase__: Dict , **UpperCamelCase__: Any ): warnings.warn( (f'''\'{fn.__name__}\' is experimental and might be subject to breaking changes in the future.''') , UpperCamelCase__ , ) return fn(*UpperCamelCase__ , **UpperCamelCase__ ) return _inner_fn
59
0
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "AutoTokenizer" lowerCamelCase_ = ["tokenizer"] lowerCamelCase_ = { "semantic_prompt": 1, "coarse_prompt": 2, "fine_prompt": 2, } def __init__( self :str , __A :int , __A :int=None ) -> Dict: """simple docstring""" super().__init__(__A ) SCREAMING_SNAKE_CASE__ = speaker_embeddings @classmethod def _snake_case ( cls :List[Any] , __A :List[Any] , __A :Optional[Any]="speaker_embeddings_path.json" , **__A :Union[str, Any] ) -> List[Any]: """simple docstring""" if speaker_embeddings_dict_path is not None: SCREAMING_SNAKE_CASE__ = get_file_from_repo( __A , __A , subfolder=kwargs.pop("""subfolder""" , __A ) , cache_dir=kwargs.pop("""cache_dir""" , __A ) , force_download=kwargs.pop("""force_download""" , __A ) , proxies=kwargs.pop("""proxies""" , __A ) , resume_download=kwargs.pop("""resume_download""" , __A ) , local_files_only=kwargs.pop("""local_files_only""" , __A ) , use_auth_token=kwargs.pop("""use_auth_token""" , __A ) , revision=kwargs.pop("""revision""" , __A ) , ) if speaker_embeddings_path is None: logger.warning( f'''`{os.path.join(__A , __A )}` does not exists\n , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json\n dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.''' ) SCREAMING_SNAKE_CASE__ = None else: with open(__A ) as speaker_embeddings_json: SCREAMING_SNAKE_CASE__ = json.load(__A ) else: SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = AutoTokenizer.from_pretrained(__A , **__A ) return cls(tokenizer=__A , speaker_embeddings=__A ) def _snake_case ( self :Optional[Any] , __A :List[str] , __A :Optional[Any]="speaker_embeddings_path.json" , __A :int="speaker_embeddings" , __A :bool = False , **__A :Tuple , ) -> Dict: """simple docstring""" if self.speaker_embeddings is not None: os.makedirs(os.path.join(__A , __A , """v2""" ) , exist_ok=__A ) SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": SCREAMING_SNAKE_CASE__ = self._load_voice_preset(__A ) SCREAMING_SNAKE_CASE__ = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict["""repo_or_path"""] , __A , f'''{prompt_key}_{key}''' ) , voice_preset[key] , allow_pickle=__A , ) SCREAMING_SNAKE_CASE__ = os.path.join(__A , f'''{prompt_key}_{key}.npy''' ) SCREAMING_SNAKE_CASE__ = tmp_dict with open(os.path.join(__A , __A ) , """w""" ) as fp: json.dump(__A , __A ) super().save_pretrained(__A , __A , **__A ) def _snake_case ( self :Any , __A :str = None , **__A :Optional[int] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.speaker_embeddings[voice_preset] SCREAMING_SNAKE_CASE__ = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( f'''Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].''' ) SCREAMING_SNAKE_CASE__ = get_file_from_repo( self.speaker_embeddings.get("""repo_or_path""" , """/""" ) , voice_preset_paths[key] , subfolder=kwargs.pop("""subfolder""" , __A ) , cache_dir=kwargs.pop("""cache_dir""" , __A ) , force_download=kwargs.pop("""force_download""" , __A ) , proxies=kwargs.pop("""proxies""" , __A ) , resume_download=kwargs.pop("""resume_download""" , __A ) , local_files_only=kwargs.pop("""local_files_only""" , __A ) , use_auth_token=kwargs.pop("""use_auth_token""" , __A ) , revision=kwargs.pop("""revision""" , __A ) , ) if path is None: raise ValueError( f'''`{os.path.join(self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] )}` does not exists\n , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset}\n embeddings.''' ) SCREAMING_SNAKE_CASE__ = np.load(__A ) return voice_preset_dict def _snake_case ( self :Any , __A :Optional[dict] = None ) -> List[str]: """simple docstring""" for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(f'''Voice preset unrecognized, missing {key} as a key.''' ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(f'''{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.''' ) def __call__( self :Tuple , __A :List[Any]=None , __A :Optional[int]=None , __A :str="pt" , __A :Tuple=256 , __A :Dict=False , __A :List[Any]=True , __A :Dict=False , **__A :Tuple , ) -> Optional[Any]: """simple docstring""" if voice_preset is not None and not isinstance(__A , __A ): if ( isinstance(__A , __A ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): SCREAMING_SNAKE_CASE__ = self._load_voice_preset(__A ) else: if isinstance(__A , __A ) and not voice_preset.endswith(""".npz""" ): SCREAMING_SNAKE_CASE__ = voice_preset + ".npz" SCREAMING_SNAKE_CASE__ = np.load(__A ) if voice_preset is not None: self._validate_voice_preset_dict(__A , **__A ) SCREAMING_SNAKE_CASE__ = BatchFeature(data=__A , tensor_type=__A ) SCREAMING_SNAKE_CASE__ = self.tokenizer( __A , return_tensors=__A , padding="""max_length""" , max_length=__A , return_attention_mask=__A , return_token_type_ids=__A , add_special_tokens=__A , **__A , ) if voice_preset is not None: SCREAMING_SNAKE_CASE__ = voice_preset return encoded_text
711
import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = RoCBertTokenizer lowerCamelCase_ = None lowerCamelCase_ = False lowerCamelCase_ = True lowerCamelCase_ = filter_non_english def _snake_case ( self :List[Any] ) -> List[Any]: """simple docstring""" super().setUp() SCREAMING_SNAKE_CASE__ = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """你""", """好""", """是""", """谁""", """a""", """b""", """c""", """d"""] SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = {} for i, value in enumerate(__A ): SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""word_shape_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""word_pronunciation_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) with open(self.word_shape_file , """w""" , encoding="""utf-8""" ) as word_shape_writer: json.dump(__A , __A , ensure_ascii=__A ) with open(self.word_pronunciation_file , """w""" , encoding="""utf-8""" ) as word_pronunciation_writer: json.dump(__A , __A , ensure_ascii=__A ) def _snake_case ( self :List[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) SCREAMING_SNAKE_CASE__ = tokenizer.tokenize("""你好[SEP]你是谁""" ) self.assertListEqual(__A , ["""你""", """好""", """[SEP]""", """你""", """是""", """谁"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_shape_ids(__A ) , [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_pronunciation_ids(__A ) , [5, 6, 2, 5, 7, 8] ) def _snake_case ( self :List[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer() self.assertListEqual(tokenizer.tokenize("""ah\u535A\u63A8zz""" ) , ["""ah""", """\u535A""", """\u63A8""", """zz"""] ) def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hällo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""h\u00E9llo"""] ) def _snake_case ( self :Any ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :List[str] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ) , ["""hello"""] ) def _snake_case ( self :Union[str, Any] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :int ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HäLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :Union[str, Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , strip_accents=__A ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ) , ["""HaLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def _snake_case ( self :List[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = RoCBertBasicTokenizer(do_lower_case=__A , never_split=["""[UNK]"""] ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? [UNK]""" ) , ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?""", """[UNK]"""] ) def _snake_case ( self :Any ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing"""] SCREAMING_SNAKE_CASE__ = {} for i, token in enumerate(__A ): SCREAMING_SNAKE_CASE__ = i SCREAMING_SNAKE_CASE__ = RoCBertWordpieceTokenizer(vocab=__A , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""unwanted running""" ) , ["""un""", """##want""", """##ed""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.tokenize("""unwantedX running""" ) , ["""[UNK]""", """runn""", """##ing"""] ) def _snake_case ( self :Any ) -> str: """simple docstring""" self.assertTrue(_is_whitespace(""" """ ) ) self.assertTrue(_is_whitespace("""\t""" ) ) self.assertTrue(_is_whitespace("""\r""" ) ) self.assertTrue(_is_whitespace("""\n""" ) ) self.assertTrue(_is_whitespace("""\u00A0""" ) ) self.assertFalse(_is_whitespace("""A""" ) ) self.assertFalse(_is_whitespace("""-""" ) ) def _snake_case ( self :int ) -> str: """simple docstring""" self.assertTrue(_is_control("""\u0005""" ) ) self.assertFalse(_is_control("""A""" ) ) self.assertFalse(_is_control(""" """ ) ) self.assertFalse(_is_control("""\t""" ) ) self.assertFalse(_is_control("""\r""" ) ) def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" self.assertTrue(_is_punctuation("""-""" ) ) self.assertTrue(_is_punctuation("""$""" ) ) self.assertTrue(_is_punctuation("""`""" ) ) self.assertTrue(_is_punctuation(""".""" ) ) self.assertFalse(_is_punctuation("""A""" ) ) self.assertFalse(_is_punctuation(""" """ ) ) def _snake_case ( self :str ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(__A ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) if self.test_rust_tokenizer: SCREAMING_SNAKE_CASE__ = self.get_rust_tokenizer() self.assertListEqual( [rust_tokenizer.tokenize(__A ) for t in ["""Test""", """\xad""", """test"""]] , [["""[UNK]"""], [], ["""[UNK]"""]] ) def _snake_case ( self :int ) -> Any: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = f'''A, naïve {tokenizer_r.mask_token} AllenNLP sentence.''' SCREAMING_SNAKE_CASE__ = tokenizer_r.encode_plus( __A , return_attention_mask=__A , return_token_type_ids=__A , return_offsets_mapping=__A , add_special_tokens=__A , ) SCREAMING_SNAKE_CASE__ = tokenizer_r.do_lower_case if hasattr(__A , """do_lower_case""" ) else False SCREAMING_SNAKE_CASE__ = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """A"""), ((1, 2), ""","""), ((3, 5), """na"""), ((5, 6), """##ï"""), ((6, 8), """##ve"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """Allen"""), ((21, 23), """##NL"""), ((23, 24), """##P"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """a"""), ((1, 2), ""","""), ((3, 8), """naive"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """allen"""), ((21, 23), """##nl"""), ((23, 24), """##p"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["""input_ids"""] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["""offset_mapping"""] ) def _snake_case ( self :Optional[int] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""的""", """人""", """有"""] SCREAMING_SNAKE_CASE__ = """""".join(__A ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.convert_ids_to_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.convert_ids_to_tokens(__A ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(__A , __A ) self.assertListEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained(__A , **__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.encode(__A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer_r.convert_ids_to_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer_p.convert_ids_to_tokens(__A ) # it is expected that only the first Chinese character is not preceded by "##". SCREAMING_SNAKE_CASE__ = [ f'''##{token}''' if idx != 0 else token for idx, token in enumerate(__A ) ] self.assertListEqual(__A , __A ) self.assertListEqual(__A , __A ) @slow def _snake_case ( self :Union[str, Any] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file ) SCREAMING_SNAKE_CASE__ = tokenizer.encode("""你好""" , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.encode("""你是谁""" , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.build_inputs_with_special_tokens(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.build_inputs_with_special_tokens(__A , __A ) assert encoded_sentence == [1] + text + [2] assert encoded_pair == [1] + text + [2] + text_a + [2] def _snake_case ( self :List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_tokenizers(do_lower_case=__A ) for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE__ = """你好,你是谁""" SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_shape_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_pronunciation_ids(__A ) SCREAMING_SNAKE_CASE__ = tokenizer.prepare_for_model( __A , __A , __A , add_special_tokens=__A ) SCREAMING_SNAKE_CASE__ = tokenizer.encode_plus(__A , add_special_tokens=__A ) self.assertEqual(__A , __A )
59
0
import os # noqa: this is just for tests import os as renamed_os # noqa: this is just for tests from os import path # noqa: this is just for tests from os import path as renamed_path # noqa: this is just for tests from os.path import join # noqa: this is just for tests from os.path import join as renamed_join # noqa: this is just for tests _lowerCamelCase = open # noqa: we just need to have a builtin inside this module to test it properly
712
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[str, Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Optional[Any] , UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = original_name.split(""".""" )[0] SCREAMING_SNAKE_CASE__ = key.split(""".""" ) SCREAMING_SNAKE_CASE__ = int(key_list[key_list.index(UpperCamelCase__ ) - 2] ) SCREAMING_SNAKE_CASE__ = int(key_list[key_list.index(UpperCamelCase__ ) - 1] ) SCREAMING_SNAKE_CASE__ = orig_block_num - offset SCREAMING_SNAKE_CASE__ = key.replace(f'''{orig_block_num}.{layer_num}.{original_name}''' , f'''block.{new_block_num}.{layer_num}.{new_name}''' ) return key def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = OrderedDict() SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 0, 0 for key, value in state_dict.items(): if key.startswith("""network""" ): SCREAMING_SNAKE_CASE__ = key.replace("""network""" , """poolformer.encoder""" ) if "proj" in key: # Works for the first embedding as well as the internal embedding layers if key.endswith("""bias""" ) and "patch_embed" not in key: patch_emb_offset += 1 SCREAMING_SNAKE_CASE__ = key[: key.find("""proj""" )] SCREAMING_SNAKE_CASE__ = key.replace(UpperCamelCase__ , f'''patch_embeddings.{total_embed_found}.''' ) SCREAMING_SNAKE_CASE__ = key.replace("""proj""" , """projection""" ) if key.endswith("""bias""" ): total_embed_found += 1 if "patch_embeddings" in key: SCREAMING_SNAKE_CASE__ = """poolformer.encoder.""" + key if "mlp.fc1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """mlp.fc1""" , """output.conv1""" ) if "mlp.fc2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """mlp.fc2""" , """output.conv2""" ) if "norm1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """norm1""" , """before_norm""" ) if "norm2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """norm2""" , """after_norm""" ) if "layer_scale_1" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """layer_scale_1""" , """layer_scale_1""" ) if "layer_scale_2" in key: SCREAMING_SNAKE_CASE__ = replace_key_with_offset(UpperCamelCase__ , UpperCamelCase__ , """layer_scale_2""" , """layer_scale_2""" ) if "head" in key: SCREAMING_SNAKE_CASE__ = key.replace("""head""" , """classifier""" ) SCREAMING_SNAKE_CASE__ = value return new_state_dict def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = """http://images.cocodataset.org/val2017/000000039769.jpg""" SCREAMING_SNAKE_CASE__ = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return image @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] , UpperCamelCase__: Optional[int] , UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = PoolFormerConfig() # set attributes based on model_name SCREAMING_SNAKE_CASE__ = """huggingface/label-files""" SCREAMING_SNAKE_CASE__ = model_name[-3:] SCREAMING_SNAKE_CASE__ = 1_000 SCREAMING_SNAKE_CASE__ = """imagenet-1k-id2label.json""" SCREAMING_SNAKE_CASE__ = (1, 1_000) # set config attributes SCREAMING_SNAKE_CASE__ = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) SCREAMING_SNAKE_CASE__ = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE__ = idalabel SCREAMING_SNAKE_CASE__ = {v: k for k, v in idalabel.items()} if size == "s12": SCREAMING_SNAKE_CASE__ = [2, 2, 6, 2] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "s24": SCREAMING_SNAKE_CASE__ = [4, 4, 12, 4] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "s36": SCREAMING_SNAKE_CASE__ = [6, 6, 18, 6] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9 elif size == "m36": SCREAMING_SNAKE_CASE__ = [6, 6, 18, 6] SCREAMING_SNAKE_CASE__ = [96, 192, 384, 768] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9_5 elif size == "m48": SCREAMING_SNAKE_CASE__ = [8, 8, 24, 8] SCREAMING_SNAKE_CASE__ = [96, 192, 384, 768] SCREAMING_SNAKE_CASE__ = 4.0 SCREAMING_SNAKE_CASE__ = 1e-6 SCREAMING_SNAKE_CASE__ = 0.9_5 else: raise ValueError(f'''Size {size} not supported''' ) # load image processor SCREAMING_SNAKE_CASE__ = PoolFormerImageProcessor(crop_pct=UpperCamelCase__ ) # Prepare image SCREAMING_SNAKE_CASE__ = prepare_img() SCREAMING_SNAKE_CASE__ = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ).pixel_values logger.info(f'''Converting model {model_name}...''' ) # load original state dict SCREAMING_SNAKE_CASE__ = torch.load(UpperCamelCase__ , map_location=torch.device("""cpu""" ) ) # rename keys SCREAMING_SNAKE_CASE__ = rename_keys(UpperCamelCase__ ) # create HuggingFace model and load state dict SCREAMING_SNAKE_CASE__ = PoolFormerForImageClassification(UpperCamelCase__ ) model.load_state_dict(UpperCamelCase__ ) model.eval() # Define image processor SCREAMING_SNAKE_CASE__ = PoolFormerImageProcessor(crop_pct=UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = image_processor(images=prepare_img() , return_tensors="""pt""" ).pixel_values # forward pass SCREAMING_SNAKE_CASE__ = model(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = outputs.logits # define expected logit slices for different models if size == "s12": SCREAMING_SNAKE_CASE__ = torch.tensor([-0.3_0_4_5, -0.6_7_5_8, -0.4_8_6_9] ) elif size == "s24": SCREAMING_SNAKE_CASE__ = torch.tensor([0.4_4_0_2, -0.1_3_7_4, -0.8_0_4_5] ) elif size == "s36": SCREAMING_SNAKE_CASE__ = torch.tensor([-0.6_0_8_0, -0.5_1_3_3, -0.5_8_9_8] ) elif size == "m36": SCREAMING_SNAKE_CASE__ = torch.tensor([0.3_9_5_2, 0.2_2_6_3, -1.2_6_6_8] ) elif size == "m48": SCREAMING_SNAKE_CASE__ = torch.tensor([0.1_1_6_7, -0.0_6_5_6, -0.3_4_2_3] ) else: raise ValueError(f'''Size {size} not supported''' ) # verify logits assert logits.shape == expected_shape assert torch.allclose(logits[0, :3] , UpperCamelCase__ , atol=1e-2 ) # finally, save model and image processor logger.info(f'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) model.save_pretrained(UpperCamelCase__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() parser.add_argument( '--model_name', default='poolformer_s12', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) _lowerCamelCase = parser.parse_args() convert_poolformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
59
0
'''simple docstring''' from collections.abc import Sequence from queue import Queue class UpperCamelCase_ : def __init__( self :str , __A :Tuple , __A :Optional[int] , __A :List[str] , __A :Optional[int]=None , __A :List[str]=None ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = start SCREAMING_SNAKE_CASE__ = end SCREAMING_SNAKE_CASE__ = val SCREAMING_SNAKE_CASE__ = (start + end) // 2 SCREAMING_SNAKE_CASE__ = left SCREAMING_SNAKE_CASE__ = right def __repr__( self :List[str] ) -> Dict: """simple docstring""" return f'''SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})''' class UpperCamelCase_ : def __init__( self :int , __A :Sequence , __A :int ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = collection SCREAMING_SNAKE_CASE__ = function if self.collection: SCREAMING_SNAKE_CASE__ = self._build_tree(0 , len(__A ) - 1 ) def _snake_case ( self :Optional[Any] , __A :Tuple , __A :int ) -> Any: """simple docstring""" self._update_tree(self.root , __A , __A ) def _snake_case ( self :Dict , __A :List[Any] , __A :List[Any] ) -> Union[str, Any]: """simple docstring""" return self._query_range(self.root , __A , __A ) def _snake_case ( self :str , __A :Dict , __A :Optional[Any] ) -> Tuple: """simple docstring""" if start == end: return SegmentTreeNode(__A , __A , self.collection[start] ) SCREAMING_SNAKE_CASE__ = (start + end) // 2 SCREAMING_SNAKE_CASE__ = self._build_tree(__A , __A ) SCREAMING_SNAKE_CASE__ = self._build_tree(mid + 1 , __A ) return SegmentTreeNode(__A , __A , self.fn(left.val , right.val ) , __A , __A ) def _snake_case ( self :Dict , __A :Optional[Any] , __A :str , __A :str ) -> Dict: """simple docstring""" if node.start == i and node.end == i: SCREAMING_SNAKE_CASE__ = val return if i <= node.mid: self._update_tree(node.left , __A , __A ) else: self._update_tree(node.right , __A , __A ) SCREAMING_SNAKE_CASE__ = self.fn(node.left.val , node.right.val ) def _snake_case ( self :int , __A :List[Any] , __A :List[str] , __A :str ) -> Optional[int]: """simple docstring""" if node.start == i and node.end == j: return node.val if i <= node.mid: if j <= node.mid: # range in left child tree return self._query_range(node.left , __A , __A ) else: # range in left child tree and right child tree return self.fn( self._query_range(node.left , __A , node.mid ) , self._query_range(node.right , node.mid + 1 , __A ) , ) else: # range in right child tree return self._query_range(node.right , __A , __A ) def _snake_case ( self :Optional[int] ) -> Optional[int]: """simple docstring""" if self.root is not None: SCREAMING_SNAKE_CASE__ = Queue() queue.put(self.root ) while not queue.empty(): SCREAMING_SNAKE_CASE__ = queue.get() yield node if node.left is not None: queue.put(node.left ) if node.right is not None: queue.put(node.right ) if __name__ == "__main__": import operator for fn in [operator.add, max, min]: print('*' * 50) _lowerCamelCase = SegmentTree([2, 1, 5, 3, 4], fn) for node in arr.traverse(): print(node) print() arr.update(1, 5) for node in arr.traverse(): print(node) print() print(arr.query_range(3, 4)) # 7 print(arr.query_range(2, 2)) # 5 print(arr.query_range(1, 3)) # 13 print()
713
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'google/pix2struct-textcaps-base': ( 'https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json' ), } class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct_text_model" lowerCamelCase_ = ["past_key_values"] lowerCamelCase_ = { "hidden_size": "hidden_size", "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self :Union[str, Any] , __A :Any=5_0244 , __A :Optional[Any]=768 , __A :Tuple=64 , __A :List[str]=2048 , __A :int=12 , __A :str=12 , __A :Any=32 , __A :Tuple=128 , __A :int=0.1 , __A :str=1E-6 , __A :Optional[Any]=1.0 , __A :Union[str, Any]="gelu_new" , __A :Any=0 , __A :List[str]=False , __A :Optional[Any]=0 , __A :int=1 , __A :Optional[int]=False , __A :Optional[Any]=True , **__A :List[Any] , ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = d_kv SCREAMING_SNAKE_CASE__ = d_ff SCREAMING_SNAKE_CASE__ = num_layers SCREAMING_SNAKE_CASE__ = num_heads SCREAMING_SNAKE_CASE__ = relative_attention_num_buckets SCREAMING_SNAKE_CASE__ = relative_attention_max_distance SCREAMING_SNAKE_CASE__ = dropout_rate SCREAMING_SNAKE_CASE__ = layer_norm_epsilon SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = use_cache SCREAMING_SNAKE_CASE__ = eos_token_id SCREAMING_SNAKE_CASE__ = decoder_start_token_id # for backwards compatibility SCREAMING_SNAKE_CASE__ = dense_act_fn super().__init__( pad_token_id=__A , eos_token_id=__A , decoder_start_token_id=__A , tie_word_embeddings=__A , is_decoder=__A , **__A , ) @classmethod def _snake_case ( cls :Optional[int] , __A :Union[str, os.PathLike] , **__A :Optional[int] ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = cls.get_config_dict(__A , **__A ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get("""model_type""" ) == "pix2struct": SCREAMING_SNAKE_CASE__ = config_dict["""text_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct_vision_model" def __init__( self :Optional[int] , __A :int=768 , __A :Optional[Any]=768 , __A :Union[str, Any]=2048 , __A :int=64 , __A :Union[str, Any]=12 , __A :str=12 , __A :Any="gelu_new" , __A :List[Any]=1E-6 , __A :Dict=0.0 , __A :int=0.0 , __A :int=1E-10 , __A :Dict=1.0 , __A :int=4096 , __A :int=32 , __A :int=128 , **__A :Tuple , ) -> str: """simple docstring""" super().__init__(**__A ) SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = patch_embed_hidden_size SCREAMING_SNAKE_CASE__ = d_ff SCREAMING_SNAKE_CASE__ = dropout_rate SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = attention_dropout SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = dense_act_fn SCREAMING_SNAKE_CASE__ = seq_len SCREAMING_SNAKE_CASE__ = relative_attention_num_buckets SCREAMING_SNAKE_CASE__ = relative_attention_max_distance SCREAMING_SNAKE_CASE__ = d_kv @classmethod def _snake_case ( cls :str , __A :Union[str, os.PathLike] , **__A :str ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__A ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = cls.get_config_dict(__A , **__A ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get("""model_type""" ) == "pix2struct": SCREAMING_SNAKE_CASE__ = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__A , **__A ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "pix2struct" lowerCamelCase_ = True def __init__( self :str , __A :Optional[Any]=None , __A :List[str]=None , __A :Optional[Any]=1.0 , __A :Optional[Any]=0.0_2 , __A :Any=False , __A :Tuple=False , __A :Any=True , **__A :Dict , ) -> Union[str, Any]: """simple docstring""" super().__init__(tie_word_embeddings=__A , is_encoder_decoder=__A , **__A ) if text_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""text_config is None. Initializing the Pix2StructTextConfig with default values.""" ) if vision_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""vision_config is None. Initializing the Pix2StructVisionConfig with default values.""" ) SCREAMING_SNAKE_CASE__ = PixaStructTextConfig(**__A ) SCREAMING_SNAKE_CASE__ = PixaStructVisionConfig(**__A ) SCREAMING_SNAKE_CASE__ = self.text_config.decoder_start_token_id SCREAMING_SNAKE_CASE__ = self.text_config.pad_token_id SCREAMING_SNAKE_CASE__ = self.text_config.eos_token_id SCREAMING_SNAKE_CASE__ = initializer_factor SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = self.initializer_range SCREAMING_SNAKE_CASE__ = self.initializer_range SCREAMING_SNAKE_CASE__ = is_vqa @classmethod def _snake_case ( cls :Union[str, Any] , __A :PixaStructTextConfig , __A :PixaStructVisionConfig , **__A :Optional[int] ) -> Optional[Any]: """simple docstring""" return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__A ) def _snake_case ( self :str ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = copy.deepcopy(self.__dict__ ) SCREAMING_SNAKE_CASE__ = self.text_config.to_dict() SCREAMING_SNAKE_CASE__ = self.vision_config.to_dict() SCREAMING_SNAKE_CASE__ = self.__class__.model_type return output
59
0
def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Tuple , UpperCamelCase__: str ): SCREAMING_SNAKE_CASE__ = len(lowerCAmelCase__ ) print("""The following activities are selected:""" ) # The first activity is always selected SCREAMING_SNAKE_CASE__ = 0 print(lowerCAmelCase__ , end=""",""" ) # Consider rest of the activities for j in range(lowerCAmelCase__ ): # If this activity has start time greater than # or equal to the finish time of previously # selected activity, then select it if start[j] >= finish[i]: print(lowerCAmelCase__ , end=""",""" ) SCREAMING_SNAKE_CASE__ = j if __name__ == "__main__": import doctest doctest.testmod() _lowerCamelCase = [1, 3, 0, 5, 8, 5] _lowerCamelCase = [2, 4, 6, 7, 9, 9] print_max_activities(start, finish)
714
import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = inspect.getfile(accelerate.test_utils ) SCREAMING_SNAKE_CASE__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_script.py"""] ) SCREAMING_SNAKE_CASE__ = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_distributed_data_loop.py"""] ) SCREAMING_SNAKE_CASE__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_ops.py"""] ) @require_multi_gpu def _snake_case ( self :Optional[Any] ) -> Tuple: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Tuple ) -> Optional[Any]: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices.''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.operation_file_path] print(f'''Command: {cmd}''' ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Optional[Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__A , env=os.environ.copy() ) @require_multi_gpu def _snake_case ( self :Optional[int] ) -> str: """simple docstring""" print(f'''Found {torch.cuda.device_count()} devices, using 2 devices only''' ) SCREAMING_SNAKE_CASE__ = ["""torchrun""", f'''--nproc_per_node={torch.cuda.device_count()}''', self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices="""0,1""" ): execute_subprocess_async(__A , env=os.environ.copy() ) if __name__ == "__main__": _lowerCamelCase = Accelerator() _lowerCamelCase = (accelerator.state.process_index + 2, 10) _lowerCamelCase = torch.randint(0, 10, shape).to(accelerator.device) _lowerCamelCase = '' _lowerCamelCase = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." _lowerCamelCase = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += F"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." _lowerCamelCase = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
59
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCamelCase = { "configuration_whisper": ["WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP", "WhisperConfig", "WhisperOnnxConfig"], "feature_extraction_whisper": ["WhisperFeatureExtractor"], "processing_whisper": ["WhisperProcessor"], "tokenization_whisper": ["WhisperTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ["WhisperTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ "WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "WhisperForConditionalGeneration", "WhisperModel", "WhisperPreTrainedModel", "WhisperForAudioClassification", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ "TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFWhisperForConditionalGeneration", "TFWhisperModel", "TFWhisperPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ "FlaxWhisperForConditionalGeneration", "FlaxWhisperModel", "FlaxWhisperPreTrainedModel", "FlaxWhisperForAudioClassification", ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
715
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _lowerCamelCase = { 'configuration_layoutlmv3': [ 'LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv3Config', 'LayoutLMv3OnnxConfig', ], 'processing_layoutlmv3': ['LayoutLMv3Processor'], 'tokenization_layoutlmv3': ['LayoutLMv3Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['LayoutLMv3TokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv3ForQuestionAnswering', 'LayoutLMv3ForSequenceClassification', 'LayoutLMv3ForTokenClassification', 'LayoutLMv3Model', 'LayoutLMv3PreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLayoutLMv3ForQuestionAnswering', 'TFLayoutLMv3ForSequenceClassification', 'TFLayoutLMv3ForTokenClassification', 'TFLayoutLMv3Model', 'TFLayoutLMv3PreTrainedModel', ] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = ['LayoutLMv3FeatureExtractor'] _lowerCamelCase = ['LayoutLMv3ImageProcessor'] if TYPE_CHECKING: from .configuration_layoutlmva import ( LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig, LayoutLMvaOnnxConfig, ) from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_layoutlmva import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, TFLayoutLMvaPreTrainedModel, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor from .image_processing_layoutlmva import LayoutLMvaImageProcessor else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
0
import flax.linen as nn import jax import jax.numpy as jnp class UpperCamelCase_ ( nn.Module ): lowerCamelCase_ = 42 lowerCamelCase_ = jnp.floataa def _snake_case ( self :Tuple ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self :Union[str, Any] , __A :List[Any] ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ = hidden_states.shape SCREAMING_SNAKE_CASE__ = jax.image.resize( _A , shape=(batch, height * 2, width * 2, channels) , method="""nearest""" , ) SCREAMING_SNAKE_CASE__ = self.conv(_A ) return hidden_states class UpperCamelCase_ ( nn.Module ): lowerCamelCase_ = 42 lowerCamelCase_ = jnp.floataa def _snake_case ( self :Tuple ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self :Optional[Any] , __A :Any ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.conv(_A ) return hidden_states class UpperCamelCase_ ( nn.Module ): lowerCamelCase_ = 42 lowerCamelCase_ = None lowerCamelCase_ = 0.0 lowerCamelCase_ = None lowerCamelCase_ = jnp.floataa def _snake_case ( self :List[str] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.in_channels if self.out_channels is None else self.out_channels SCREAMING_SNAKE_CASE__ = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) SCREAMING_SNAKE_CASE__ = nn.Conv( _A , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) SCREAMING_SNAKE_CASE__ = nn.Dense(_A , dtype=self.dtype ) SCREAMING_SNAKE_CASE__ = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) SCREAMING_SNAKE_CASE__ = nn.Dropout(self.dropout_prob ) SCREAMING_SNAKE_CASE__ = nn.Conv( _A , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) SCREAMING_SNAKE_CASE__ = self.in_channels != out_channels if self.use_nin_shortcut is None else self.use_nin_shortcut SCREAMING_SNAKE_CASE__ = None if use_nin_shortcut: SCREAMING_SNAKE_CASE__ = nn.Conv( _A , kernel_size=(1, 1) , strides=(1, 1) , padding="""VALID""" , dtype=self.dtype , ) def __call__( self :Optional[int] , __A :Dict , __A :List[str] , __A :int=True ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = hidden_states SCREAMING_SNAKE_CASE__ = self.norma(_A ) SCREAMING_SNAKE_CASE__ = nn.swish(_A ) SCREAMING_SNAKE_CASE__ = self.conva(_A ) SCREAMING_SNAKE_CASE__ = self.time_emb_proj(nn.swish(_A ) ) SCREAMING_SNAKE_CASE__ = jnp.expand_dims(jnp.expand_dims(_A , 1 ) , 1 ) SCREAMING_SNAKE_CASE__ = hidden_states + temb SCREAMING_SNAKE_CASE__ = self.norma(_A ) SCREAMING_SNAKE_CASE__ = nn.swish(_A ) SCREAMING_SNAKE_CASE__ = self.dropout(_A , _A ) SCREAMING_SNAKE_CASE__ = self.conva(_A ) if self.conv_shortcut is not None: SCREAMING_SNAKE_CASE__ = self.conv_shortcut(_A ) return hidden_states + residual
716
import inspect import unittest class UpperCamelCase_ ( unittest.TestCase ): def _snake_case ( self :str ) -> Union[str, Any]: """simple docstring""" try: import diffusers # noqa: F401 except ImportError: assert False def _snake_case ( self :Any ) -> Any: """simple docstring""" import diffusers from diffusers.dependency_versions_table import deps SCREAMING_SNAKE_CASE__ = inspect.getmembers(__A , inspect.isclass ) for cls_name, cls_module in all_classes: if "dummy_" in cls_module.__module__: for backend in cls_module._backends: if backend == "k_diffusion": SCREAMING_SNAKE_CASE__ = """k-diffusion""" elif backend == "invisible_watermark": SCREAMING_SNAKE_CASE__ = """invisible-watermark""" assert backend in deps, f'''{backend} is not in the deps table!'''
59
0
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'salesforce/blip2-opt-2.7b': 'https://huggingface.co/salesforce/blip2-opt-2.7b/resolve/main/config.json', } class UpperCamelCase_ ( _A ): lowerCamelCase_ = "blip_2_vision_model" def __init__( self :Tuple , __A :Any=1408 , __A :Union[str, Any]=6144 , __A :Any=39 , __A :Union[str, Any]=16 , __A :Any=224 , __A :Tuple=14 , __A :Dict="gelu" , __A :str=0.0_0_0_0_1 , __A :Dict=0.0 , __A :Any=1E-10 , __A :List[Any]=True , **__A :Union[str, Any] , ) -> List[Any]: """simple docstring""" super().__init__(**__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = patch_size SCREAMING_SNAKE_CASE__ = image_size SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = attention_dropout SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = qkv_bias @classmethod def _snake_case ( cls :str , __A :Union[str, os.PathLike] , **__A :Optional[Any] ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = cls.get_config_dict(__lowerCamelCase , **__lowerCamelCase ) # get the vision config dict if we are loading from Blip2Config if config_dict.get("""model_type""" ) == "blip-2": SCREAMING_SNAKE_CASE__ = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__lowerCamelCase , **__lowerCamelCase ) class UpperCamelCase_ ( _A ): lowerCamelCase_ = "blip_2_qformer" def __init__( self :Tuple , __A :Union[str, Any]=3_0522 , __A :Union[str, Any]=768 , __A :Tuple=12 , __A :Dict=12 , __A :Optional[int]=3072 , __A :int="gelu" , __A :Any=0.1 , __A :int=0.1 , __A :str=512 , __A :Any=0.0_2 , __A :Any=1E-12 , __A :Optional[Any]=0 , __A :List[Any]="absolute" , __A :Any=2 , __A :str=1408 , **__A :int , ) -> Optional[Any]: """simple docstring""" super().__init__(pad_token_id=__lowerCamelCase , **__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = position_embedding_type SCREAMING_SNAKE_CASE__ = cross_attention_frequency SCREAMING_SNAKE_CASE__ = encoder_hidden_size @classmethod def _snake_case ( cls :Union[str, Any] , __A :Union[str, os.PathLike] , **__A :Optional[int] ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = cls.get_config_dict(__lowerCamelCase , **__lowerCamelCase ) # get the qformer config dict if we are loading from Blip2Config if config_dict.get("""model_type""" ) == "blip-2": SCREAMING_SNAKE_CASE__ = config_dict["qformer_config"] if "model_type" in config_dict and hasattr(cls , """model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(__lowerCamelCase , **__lowerCamelCase ) class UpperCamelCase_ ( _A ): lowerCamelCase_ = "blip-2" lowerCamelCase_ = True def __init__( self :Union[str, Any] , __A :Any=None , __A :Optional[int]=None , __A :Union[str, Any]=None , __A :Tuple=32 , **__A :Any ) -> Optional[int]: """simple docstring""" super().__init__(**__lowerCamelCase ) if vision_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""vision_config is None. initializing the Blip2VisionConfig with default values.""" ) if qformer_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""qformer_config is None. Initializing the Blip2QFormerConfig with default values.""" ) if text_config is None: SCREAMING_SNAKE_CASE__ = {} logger.info("""text_config is None. Initializing the text config with default values (`OPTConfig`).""" ) SCREAMING_SNAKE_CASE__ = BlipaVisionConfig(**__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = BlipaQFormerConfig(**__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = text_config["model_type"] if "model_type" in text_config else "opt" SCREAMING_SNAKE_CASE__ = CONFIG_MAPPING[text_model_type](**__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = self.text_config.tie_word_embeddings SCREAMING_SNAKE_CASE__ = self.text_config.is_encoder_decoder SCREAMING_SNAKE_CASE__ = num_query_tokens SCREAMING_SNAKE_CASE__ = self.vision_config.hidden_size SCREAMING_SNAKE_CASE__ = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES SCREAMING_SNAKE_CASE__ = 1.0 SCREAMING_SNAKE_CASE__ = 0.0_2 @classmethod def _snake_case ( cls :Optional[int] , __A :BlipaVisionConfig , __A :BlipaQFormerConfig , __A :PretrainedConfig , **__A :Tuple , ) -> Tuple: """simple docstring""" return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **__lowerCamelCase , ) def _snake_case ( self :Tuple ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = copy.deepcopy(self.__dict__ ) SCREAMING_SNAKE_CASE__ = self.vision_config.to_dict() SCREAMING_SNAKE_CASE__ = self.qformer_config.to_dict() SCREAMING_SNAKE_CASE__ = self.text_config.to_dict() SCREAMING_SNAKE_CASE__ = self.__class__.model_type return output
717
import warnings from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch from ...models import UNetaDModel from ...schedulers import RePaintScheduler from ...utils import PIL_INTERPOLATION, logging, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput _lowerCamelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[List, PIL.Image.Image, torch.Tensor] ): warnings.warn( """The preprocess method is deprecated and will be removed in a future version. Please""" """ use VaeImageProcessor.preprocess instead""" , UpperCamelCase__ , ) if isinstance(UpperCamelCase__ , torch.Tensor ): return image elif isinstance(UpperCamelCase__ , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ = [image] if isinstance(image[0] , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = image[0].size SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = (x - x % 8 for x in (w, h)) # resize to integer multiple of 8 SCREAMING_SNAKE_CASE__ = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION["""lanczos"""] ) )[None, :] for i in image] SCREAMING_SNAKE_CASE__ = np.concatenate(UpperCamelCase__ , axis=0 ) SCREAMING_SNAKE_CASE__ = np.array(UpperCamelCase__ ).astype(np.floataa ) / 2_5_5.0 SCREAMING_SNAKE_CASE__ = image.transpose(0 , 3 , 1 , 2 ) SCREAMING_SNAKE_CASE__ = 2.0 * image - 1.0 SCREAMING_SNAKE_CASE__ = torch.from_numpy(UpperCamelCase__ ) elif isinstance(image[0] , torch.Tensor ): SCREAMING_SNAKE_CASE__ = torch.cat(UpperCamelCase__ , dim=0 ) return image def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Union[List, PIL.Image.Image, torch.Tensor] ): if isinstance(UpperCamelCase__ , torch.Tensor ): return mask elif isinstance(UpperCamelCase__ , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ = [mask] if isinstance(mask[0] , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = mask[0].size SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 SCREAMING_SNAKE_CASE__ = [np.array(m.convert("""L""" ).resize((w, h) , resample=PIL_INTERPOLATION["""nearest"""] ) )[None, :] for m in mask] SCREAMING_SNAKE_CASE__ = np.concatenate(UpperCamelCase__ , axis=0 ) SCREAMING_SNAKE_CASE__ = mask.astype(np.floataa ) / 2_5_5.0 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = torch.from_numpy(UpperCamelCase__ ) elif isinstance(mask[0] , torch.Tensor ): SCREAMING_SNAKE_CASE__ = torch.cat(UpperCamelCase__ , dim=0 ) return mask class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = 42 lowerCamelCase_ = 42 def __init__( self :Any , __A :List[Any] , __A :Optional[Any] ) -> Union[str, Any]: """simple docstring""" super().__init__() self.register_modules(unet=__A , scheduler=__A ) @torch.no_grad() def __call__( self :str , __A :Union[torch.Tensor, PIL.Image.Image] , __A :Union[torch.Tensor, PIL.Image.Image] , __A :int = 250 , __A :float = 0.0 , __A :int = 10 , __A :int = 10 , __A :Optional[Union[torch.Generator, List[torch.Generator]]] = None , __A :Optional[str] = "pil" , __A :bool = True , ) -> Union[ImagePipelineOutput, Tuple]: """simple docstring""" SCREAMING_SNAKE_CASE__ = image SCREAMING_SNAKE_CASE__ = _preprocess_image(__A ) SCREAMING_SNAKE_CASE__ = original_image.to(device=self.device , dtype=self.unet.dtype ) SCREAMING_SNAKE_CASE__ = _preprocess_mask(__A ) SCREAMING_SNAKE_CASE__ = mask_image.to(device=self.device , dtype=self.unet.dtype ) SCREAMING_SNAKE_CASE__ = original_image.shape[0] # sample gaussian noise to begin the loop if isinstance(__A , __A ) and len(__A ) != batch_size: raise ValueError( f'''You have passed a list of generators of length {len(__A )}, but requested an effective batch''' f''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) SCREAMING_SNAKE_CASE__ = original_image.shape SCREAMING_SNAKE_CASE__ = randn_tensor(__A , generator=__A , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(__A , __A , __A , self.device ) SCREAMING_SNAKE_CASE__ = eta SCREAMING_SNAKE_CASE__ = self.scheduler.timesteps[0] + 1 SCREAMING_SNAKE_CASE__ = generator[0] if isinstance(__A , __A ) else generator for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): if t < t_last: # predict the noise residual SCREAMING_SNAKE_CASE__ = self.unet(__A , __A ).sample # compute previous image: x_t -> x_t-1 SCREAMING_SNAKE_CASE__ = self.scheduler.step(__A , __A , __A , __A , __A , __A ).prev_sample else: # compute the reverse: x_t-1 -> x_t SCREAMING_SNAKE_CASE__ = self.scheduler.undo_step(__A , __A , __A ) SCREAMING_SNAKE_CASE__ = t SCREAMING_SNAKE_CASE__ = (image / 2 + 0.5).clamp(0 , 1 ) SCREAMING_SNAKE_CASE__ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": SCREAMING_SNAKE_CASE__ = self.numpy_to_pil(__A ) if not return_dict: return (image,) return ImagePipelineOutput(images=__A )
59
0
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int , UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): SCREAMING_SNAKE_CASE__ = n - k # Calculate C(n,k) for i in range(a_ ): result *= n - i result //= i + 1 return result def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): return binomial_coefficient(2 * node_count , a_ ) // (node_count + 1) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): if n < 0: raise ValueError("""factorial() not defined for negative values""" ) SCREAMING_SNAKE_CASE__ = 1 for i in range(1 , n + 1 ): result *= i return result def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: int ): return catalan_number(a_ ) * factorial(a_ ) if __name__ == "__main__": _lowerCamelCase = int(input('Enter the number of nodes: ').strip() or 0) if node_count <= 0: raise ValueError('We need some nodes to work with.') print( F'''Given {node_count} nodes, there are {binary_tree_count(node_count)} ''' F'''binary trees and {catalan_number(node_count)} binary search trees.''' )
718
import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ , unittest.TestCase ): lowerCamelCase_ = OpenAIGPTTokenizer lowerCamelCase_ = OpenAIGPTTokenizerFast lowerCamelCase_ = True lowerCamelCase_ = False def _snake_case ( self :Optional[Any] ) -> Dict: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt SCREAMING_SNAKE_CASE__ = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """w</w>""", """r</w>""", """t</w>""", """lo""", """low""", """er</w>""", """low</w>""", """lowest</w>""", """newer</w>""", """wider</w>""", """<unk>""", ] SCREAMING_SNAKE_CASE__ = dict(zip(__A , range(len(__A ) ) ) ) SCREAMING_SNAKE_CASE__ = ["""#version: 0.2""", """l o""", """lo w""", """e r</w>""", """"""] SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" ) as fp: fp.write(json.dumps(__A ) ) with open(self.merges_file , """w""" ) as fp: fp.write("""\n""".join(__A ) ) def _snake_case ( self :Union[str, Any] , __A :str ) -> List[Any]: """simple docstring""" return "lower newer", "lower newer" def _snake_case ( self :Optional[Any] ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ = OpenAIGPTTokenizer(self.vocab_file , self.merges_file ) SCREAMING_SNAKE_CASE__ = """lower""" SCREAMING_SNAKE_CASE__ = ["""low""", """er</w>"""] SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(__A ) self.assertListEqual(__A , __A ) SCREAMING_SNAKE_CASE__ = tokens + ["""<unk>"""] SCREAMING_SNAKE_CASE__ = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , __A ) def _snake_case ( self :Optional[Any] , __A :Optional[Any]=15 ) -> Any: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): SCREAMING_SNAKE_CASE__ = self.rust_tokenizer_class.from_pretrained(__A , **__A ) # Simple input SCREAMING_SNAKE_CASE__ = """This is a simple input""" SCREAMING_SNAKE_CASE__ = ["""This is a simple input 1""", """This is a simple input 2"""] SCREAMING_SNAKE_CASE__ = ("""This is a simple input""", """This is a pair""") SCREAMING_SNAKE_CASE__ = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests self.assertRaises(__A , tokenizer_r.encode , __A , max_length=__A , padding="""max_length""" ) # Simple input self.assertRaises(__A , tokenizer_r.encode_plus , __A , max_length=__A , padding="""max_length""" ) # Simple input self.assertRaises( __A , tokenizer_r.batch_encode_plus , __A , max_length=__A , padding="""max_length""" , ) # Pair input self.assertRaises(__A , tokenizer_r.encode , __A , max_length=__A , padding="""max_length""" ) # Pair input self.assertRaises(__A , tokenizer_r.encode_plus , __A , max_length=__A , padding="""max_length""" ) # Pair input self.assertRaises( __A , tokenizer_r.batch_encode_plus , __A , max_length=__A , padding="""max_length""" , ) def _snake_case ( self :Dict ) -> List[Any]: """simple docstring""" pass @require_ftfy @require_spacy @require_tokenizers class UpperCamelCase_ ( UpperCamelCase__ ): pass
59
0
import importlib.util import os import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import ( is_accelerate_available, is_flax_available, is_safetensors_available, is_tf_available, is_torch_available, ) from . import BaseTransformersCLICommand def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Tuple ): return EnvironmentCommand() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[Any] ): return EnvironmentCommand(args.accelerate_config_file ) class UpperCamelCase_ ( _UpperCAmelCase ): @staticmethod def _snake_case ( __A :ArgumentParser ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = parser.add_parser("""env""" ) download_parser.set_defaults(func=__UpperCamelCase ) download_parser.add_argument( """--accelerate-config_file""" , default=__UpperCamelCase , help="""The accelerate config file to use for the default values in the launching script.""" , ) download_parser.set_defaults(func=__UpperCamelCase ) def __init__( self :Any , __A :List[Any] , *__A :Any ) -> None: """simple docstring""" SCREAMING_SNAKE_CASE__ = accelerate_config_file def _snake_case ( self :Union[str, Any] ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = """not installed""" if is_safetensors_available(): import safetensors SCREAMING_SNAKE_CASE__ = safetensors.__version__ elif importlib.util.find_spec("""safetensors""" ) is not None: import safetensors SCREAMING_SNAKE_CASE__ = f'''{safetensors.__version__} but is ignored because of PyTorch version too old.''' SCREAMING_SNAKE_CASE__ = """not installed""" SCREAMING_SNAKE_CASE__ = SCREAMING_SNAKE_CASE__ = """not found""" if is_accelerate_available(): import accelerate from accelerate.commands.config import default_config_file, load_config_from_file SCREAMING_SNAKE_CASE__ = accelerate.__version__ # Get the default from the config file. if self._accelerate_config_file is not None or os.path.isfile(__UpperCamelCase ): SCREAMING_SNAKE_CASE__ = load_config_from_file(self._accelerate_config_file ).to_dict() SCREAMING_SNAKE_CASE__ = ( """\n""".join([f'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else f'''\t{accelerate_config}''' ) SCREAMING_SNAKE_CASE__ = """not installed""" SCREAMING_SNAKE_CASE__ = """NA""" if is_torch_available(): import torch SCREAMING_SNAKE_CASE__ = torch.__version__ SCREAMING_SNAKE_CASE__ = torch.cuda.is_available() SCREAMING_SNAKE_CASE__ = """not installed""" SCREAMING_SNAKE_CASE__ = """NA""" if is_tf_available(): import tensorflow as tf SCREAMING_SNAKE_CASE__ = tf.__version__ try: # deprecated in v2.1 SCREAMING_SNAKE_CASE__ = tf.test.is_gpu_available() except AttributeError: # returns list of devices, convert to bool SCREAMING_SNAKE_CASE__ = bool(tf.config.list_physical_devices("""GPU""" ) ) SCREAMING_SNAKE_CASE__ = """not installed""" SCREAMING_SNAKE_CASE__ = """not installed""" SCREAMING_SNAKE_CASE__ = """not installed""" SCREAMING_SNAKE_CASE__ = """NA""" if is_flax_available(): import flax import jax import jaxlib SCREAMING_SNAKE_CASE__ = flax.__version__ SCREAMING_SNAKE_CASE__ = jax.__version__ SCREAMING_SNAKE_CASE__ = jaxlib.__version__ SCREAMING_SNAKE_CASE__ = jax.lib.xla_bridge.get_backend().platform SCREAMING_SNAKE_CASE__ = { """`transformers` version""": version, """Platform""": platform.platform(), """Python version""": platform.python_version(), """Huggingface_hub version""": huggingface_hub.__version__, """Safetensors version""": f'''{safetensors_version}''', """Accelerate version""": f'''{accelerate_version}''', """Accelerate config""": f'''{accelerate_config_str}''', """PyTorch version (GPU?)""": f'''{pt_version} ({pt_cuda_available})''', """Tensorflow version (GPU?)""": f'''{tf_version} ({tf_cuda_available})''', """Flax version (CPU?/GPU?/TPU?)""": f'''{flax_version} ({jax_backend})''', """Jax version""": f'''{jax_version}''', """JaxLib version""": f'''{jaxlib_version}''', """Using GPU in script?""": """<fill in>""", """Using distributed or parallel set-up in script?""": """<fill in>""", } print("""\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n""" ) print(self.format_dict(__UpperCamelCase ) ) return info @staticmethod def _snake_case ( __A :Union[str, Any] ) -> Dict: """simple docstring""" return "\n".join([f'''- {prop}: {val}''' for prop, val in d.items()] ) + "\n"
719
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=UpperCamelCase__ ) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = field(default="image-classification" , metadata={"include_in_asdict_even_if_is_default": True} ) lowerCamelCase_ = Features({"image": Image()} ) lowerCamelCase_ = Features({"labels": ClassLabel} ) lowerCamelCase_ = "image" lowerCamelCase_ = "labels" def _snake_case ( self :List[str] , __A :Tuple ) -> Tuple: """simple docstring""" if self.label_column not in features: raise ValueError(f'''Column {self.label_column} is not present in features.''' ) if not isinstance(features[self.label_column] , __A ): raise ValueError(f'''Column {self.label_column} is not a ClassLabel.''' ) SCREAMING_SNAKE_CASE__ = copy.deepcopy(self ) SCREAMING_SNAKE_CASE__ = self.label_schema.copy() SCREAMING_SNAKE_CASE__ = features[self.label_column] SCREAMING_SNAKE_CASE__ = label_schema return task_template @property def _snake_case ( self :Dict ) -> Dict[str, str]: """simple docstring""" return { self.image_column: "image", self.label_column: "labels", }
59
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = torch.device('cpu') def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = """http://images.cocodataset.org/val2017/000000039769.jpg""" SCREAMING_SNAKE_CASE__ = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return im def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Dict ): if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.1_7_0_3e0_0, 2.1_1_0_7e0_0, -2.0_8_1_1e0_0, 8.8_6_8_5e-0_1, 2.4_3_6_0e-0_1] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.9_6_3_6e-0_1, 2.3_4_7_8e-0_1, -1.6_9_6_3e0_0, -1.7_3_8_1e0_0, -8.6_3_3_7e-0_1] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.2_7_6_8e-0_1, -4.7_4_2_9e-0_1, -1.0_8_9_7e0_0, -1.0_2_4_8e0_0, 3.5_5_2_3e-0_2] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.5_3_3_0e-0_1, 2.4_2_1_1e-0_1, -6.0_1_8_5e-0_1, -8.2_7_8_9e-0_1, -6.0_4_4_6e-0_2] ) def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Any , UpperCamelCase__: Union[str, Any] , UpperCamelCase__: Any ): SCREAMING_SNAKE_CASE__ = dct.pop(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = val def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: List[str] ): SCREAMING_SNAKE_CASE__ = [] for k in state_dict.keys(): SCREAMING_SNAKE_CASE__ = k if ".pwconv" in k: SCREAMING_SNAKE_CASE__ = k_new.replace(""".pwconv""" , """.point_wise_conv""" ) if ".dwconv" in k: SCREAMING_SNAKE_CASE__ = k_new.replace(""".dwconv""" , """.depth_wise_conv""" ) if ".Proj." in k: SCREAMING_SNAKE_CASE__ = k_new.replace(""".Proj.""" , """.proj.""" ) if "patch_embed" in k_new: SCREAMING_SNAKE_CASE__ = k_new.replace("""patch_embed""" , """swiftformer.patch_embed.patch_embedding""" ) if "network" in k_new: SCREAMING_SNAKE_CASE__ = k_new.split(""".""" ) if ls[2].isdigit(): SCREAMING_SNAKE_CASE__ = """swiftformer.encoder.network.""" + ls[1] + """.blocks.""" + ls[2] + """.""" + """.""".join(ls[3:] ) else: SCREAMING_SNAKE_CASE__ = k_new.replace("""network""" , """swiftformer.encoder.network""" ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Tuple , UpperCamelCase__: Any , UpperCamelCase__: int ): SCREAMING_SNAKE_CASE__ = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size SCREAMING_SNAKE_CASE__ = 1_000 SCREAMING_SNAKE_CASE__ = """huggingface/label-files""" SCREAMING_SNAKE_CASE__ = """imagenet-1k-id2label.json""" SCREAMING_SNAKE_CASE__ = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) SCREAMING_SNAKE_CASE__ = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE__ = idalabel SCREAMING_SNAKE_CASE__ = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": SCREAMING_SNAKE_CASE__ = [3, 3, 6, 4] SCREAMING_SNAKE_CASE__ = [48, 56, 112, 220] elif swiftformer_name == "swiftformer_s": SCREAMING_SNAKE_CASE__ = [3, 3, 9, 6] SCREAMING_SNAKE_CASE__ = [48, 64, 168, 224] elif swiftformer_name == "swiftformer_l1": SCREAMING_SNAKE_CASE__ = [4, 3, 10, 5] SCREAMING_SNAKE_CASE__ = [48, 96, 192, 384] elif swiftformer_name == "swiftformer_l3": SCREAMING_SNAKE_CASE__ = [4, 4, 12, 6] SCREAMING_SNAKE_CASE__ = [64, 128, 320, 512] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith("""https""" ): SCREAMING_SNAKE_CASE__ = torch.hub.load_state_dict_from_url(UpperCamelCase__ , map_location="""cpu""" , check_hash=UpperCamelCase__ ) else: SCREAMING_SNAKE_CASE__ = torch.load(UpperCamelCase__ , map_location="""cpu""" ) SCREAMING_SNAKE_CASE__ = checkpoint SCREAMING_SNAKE_CASE__ = create_rename_keys(UpperCamelCase__ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # load HuggingFace model SCREAMING_SNAKE_CASE__ = SwiftFormerForImageClassification(UpperCamelCase__ ).eval() hf_model.load_state_dict(UpperCamelCase__ ) # prepare test inputs SCREAMING_SNAKE_CASE__ = prepare_img() SCREAMING_SNAKE_CASE__ = ViTImageProcessor.from_pretrained("""preprocessor_config""" ) SCREAMING_SNAKE_CASE__ = processor(images=UpperCamelCase__ , return_tensors="""pt""" ) # compare outputs from both models SCREAMING_SNAKE_CASE__ = get_expected_output(UpperCamelCase__ ) SCREAMING_SNAKE_CASE__ = hf_model(inputs["""pixel_values"""] ).logits assert hf_logits.shape == torch.Size([1, 1_000] ) assert torch.allclose(hf_logits[0, 0:5] , UpperCamelCase__ , atol=1e-3 ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) print(f'''Saving model {swiftformer_name} to {pytorch_dump_folder_path}''' ) hf_model.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--swiftformer_name', default='swiftformer_xs', choices=['swiftformer_xs', 'swiftformer_s', 'swiftformer_l1', 'swiftformer_l3'], type=str, help='Name of the SwiftFormer model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='./converted_outputs/', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--original_ckpt', default=None, type=str, help='Path to the original model checkpoint.') _lowerCamelCase = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
720
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCamelCase = { 'configuration_xmod': [ 'XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XmodConfig', 'XmodOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase = [ 'XMOD_PRETRAINED_MODEL_ARCHIVE_LIST', 'XmodForCausalLM', 'XmodForMaskedLM', 'XmodForMultipleChoice', 'XmodForQuestionAnswering', 'XmodForSequenceClassification', 'XmodForTokenClassification', 'XmodModel', 'XmodPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys _lowerCamelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
59
0