code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py _SCREAMING_SNAKE_CASE : List[str] = '''src/diffusers''' _SCREAMING_SNAKE_CASE : Optional[Any] = '''.''' # This is to make sure the diffusers module imported is the one in the repo. _SCREAMING_SNAKE_CASE : Dict = importlib.util.spec_from_file_location( '''diffusers''', os.path.join(DIFFUSERS_PATH, '''__init__.py'''), submodule_search_locations=[DIFFUSERS_PATH], ) _SCREAMING_SNAKE_CASE : Optional[int] = spec.loader.load_module() def lowerCamelCase__ ( _lowerCamelCase : Dict , _lowerCamelCase : Tuple ) -> Union[str, Any]: return line.startswith(__UpperCAmelCase ) or len(__UpperCAmelCase ) <= 1 or re.search(r'^\s*\)(\s*->.*:|:)\s*$' , __UpperCAmelCase ) is not None def lowerCamelCase__ ( _lowerCamelCase : str ) -> str: lowerCamelCase_ = object_name.split('.' ) lowerCamelCase_ = 0 # First let's find the module where our object lives. lowerCamelCase_ = parts[i] while i < len(__UpperCAmelCase ) and not os.path.isfile(os.path.join(__UpperCAmelCase , F'''{module}.py''' ) ): i += 1 if i < len(__UpperCAmelCase ): lowerCamelCase_ = os.path.join(__UpperCAmelCase , parts[i] ) if i >= len(__UpperCAmelCase ): raise ValueError(F'''`object_name` should begin with the name of a module of diffusers but got {object_name}.''' ) with open(os.path.join(__UpperCAmelCase , F'''{module}.py''' ) , 'r' , encoding='utf-8' , newline='\n' ) as f: lowerCamelCase_ = f.readlines() # Now let's find the class / func in the code! lowerCamelCase_ = '''''' lowerCamelCase_ = 0 for name in parts[i + 1 :]: while ( line_index < len(__UpperCAmelCase ) and re.search(rF'''^{indent}(class|def)\s+{name}(\(|\:)''' , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(__UpperCAmelCase ): raise ValueError(F''' {object_name} does not match any function or class in {module}.''' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). lowerCamelCase_ = line_index while line_index < len(__UpperCAmelCase ) and _should_continue(lines[line_index] , __UpperCAmelCase ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 lowerCamelCase_ = lines[start_index:line_index] return "".join(__UpperCAmelCase ) _SCREAMING_SNAKE_CASE : Any = re.compile(R'''^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)''') _SCREAMING_SNAKE_CASE : Union[str, Any] = re.compile(R'''^\s*(\S+)->(\S+)(\s+.*|$)''') _SCREAMING_SNAKE_CASE : str = re.compile(R'''<FILL\s+[^>]*>''') def lowerCamelCase__ ( _lowerCamelCase : List[str] ) -> int: lowerCamelCase_ = code.split('\n' ) lowerCamelCase_ = 0 while idx < len(__UpperCAmelCase ) and len(lines[idx] ) == 0: idx += 1 if idx < len(__UpperCAmelCase ): return re.search(r'^(\s*)\S' , lines[idx] ).groups()[0] return "" def lowerCamelCase__ ( _lowerCamelCase : int ) -> Optional[Any]: lowerCamelCase_ = len(get_indent(__UpperCAmelCase ) ) > 0 if has_indent: lowerCamelCase_ = F'''class Bla:\n{code}''' lowerCamelCase_ = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=__UpperCAmelCase ) lowerCamelCase_ = black.format_str(__UpperCAmelCase , mode=__UpperCAmelCase ) lowerCamelCase_ = style_docstrings_in_code(__UpperCAmelCase ) return result[len('class Bla:\n' ) :] if has_indent else result def lowerCamelCase__ ( _lowerCamelCase : Dict , _lowerCamelCase : Dict=False ) -> List[Any]: with open(__UpperCAmelCase , 'r' , encoding='utf-8' , newline='\n' ) as f: lowerCamelCase_ = f.readlines() lowerCamelCase_ = [] lowerCamelCase_ = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(__UpperCAmelCase ): lowerCamelCase_ = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. lowerCamelCase_ = search.groups() lowerCamelCase_ = find_code_in_diffusers(__UpperCAmelCase ) lowerCamelCase_ = get_indent(__UpperCAmelCase ) lowerCamelCase_ = line_index + 1 if indent == theoretical_indent else line_index + 2 lowerCamelCase_ = theoretical_indent lowerCamelCase_ = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. lowerCamelCase_ = True while line_index < len(__UpperCAmelCase ) and should_continue: line_index += 1 if line_index >= len(__UpperCAmelCase ): break lowerCamelCase_ = lines[line_index] lowerCamelCase_ = _should_continue(__UpperCAmelCase , __UpperCAmelCase ) and re.search(F'''^{indent}# End copy''' , __UpperCAmelCase ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 lowerCamelCase_ = lines[start_index:line_index] lowerCamelCase_ = ''''''.join(__UpperCAmelCase ) # Remove any nested `Copied from` comments to avoid circular copies lowerCamelCase_ = [line for line in theoretical_code.split('\n' ) if _re_copy_warning.search(__UpperCAmelCase ) is None] lowerCamelCase_ = '''\n'''.join(__UpperCAmelCase ) # Before comparing, use the `replace_pattern` on the original code. if len(__UpperCAmelCase ) > 0: lowerCamelCase_ = replace_pattern.replace('with' , '' ).split(',' ) lowerCamelCase_ = [_re_replace_pattern.search(__UpperCAmelCase ) for p in patterns] for pattern in patterns: if pattern is None: continue lowerCamelCase_ = pattern.groups() lowerCamelCase_ = re.sub(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) if option.strip() == "all-casing": lowerCamelCase_ = re.sub(obja.lower() , obja.lower() , __UpperCAmelCase ) lowerCamelCase_ = re.sub(obja.upper() , obja.upper() , __UpperCAmelCase ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line lowerCamelCase_ = blackify(lines[start_index - 1] + theoretical_code ) lowerCamelCase_ = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: lowerCamelCase_ = lines[:start_index] + [theoretical_code] + lines[line_index:] lowerCamelCase_ = start_index + 1 if overwrite and len(__UpperCAmelCase ) > 0: # Warn the user a file has been modified. print(F'''Detected changes, rewriting {filename}.''' ) with open(__UpperCAmelCase , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(__UpperCAmelCase ) return diffs def lowerCamelCase__ ( _lowerCamelCase : Union[str, Any] = False ) -> Tuple: lowerCamelCase_ = glob.glob(os.path.join(__UpperCAmelCase , '**/*.py' ) , recursive=__UpperCAmelCase ) lowerCamelCase_ = [] for filename in all_files: lowerCamelCase_ = is_copy_consistent(__UpperCAmelCase , __UpperCAmelCase ) diffs += [F'''- {filename}: copy does not match {d[0]} at line {d[1]}''' for d in new_diffs] if not overwrite and len(__UpperCAmelCase ) > 0: lowerCamelCase_ = '''\n'''.join(__UpperCAmelCase ) raise Exception( 'Found the following copy inconsistencies:\n' + diff + '\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.' ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE : Optional[int] = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') _SCREAMING_SNAKE_CASE : Any = parser.parse_args() check_copies(args.fix_and_overwrite)
183
"""simple docstring""" def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase ) -> List[str]: lowercase__: Optional[Any] = 0 lowercase__: List[Any] = len(__UpperCAmelCase ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None lowercase__: Tuple = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(__UpperCAmelCase ): return None lowercase__: Optional[int] = sorted_collection[point] if current_item == item: return point else: if point < left: lowercase__: List[Any] = left lowercase__: int = point elif point > right: lowercase__: Dict = right lowercase__: List[str] = point else: if item < current_item: lowercase__: int = point - 1 else: lowercase__: int = point + 1 return None def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[str]: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None lowercase__: Optional[int] = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(__UpperCAmelCase ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) elif point > right: return interpolation_search_by_recursion(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , point - 1 ) else: return interpolation_search_by_recursion( __UpperCAmelCase , __UpperCAmelCase , point + 1 , __UpperCAmelCase ) def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ) -> List[Any]: if collection != sorted(__UpperCAmelCase ): raise ValueError('''Collection must be ascending sorted''' ) return True if __name__ == "__main__": import sys __A = 0 if debug == 1: __A = [1_0, 3_0, 4_0, 4_5, 5_0, 6_6, 7_7, 9_3] try: __assert_sorted(collection) except ValueError: sys.exit("Sequence must be ascending sorted to apply interpolation search") __A = 6_7 __A = interpolation_search(collection, target) if result is not None: print(f'''{target} found at positions: {result}''') else: print("Not found")
177
0
'''simple docstring''' def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' return price * (1 + tax_rate) if __name__ == "__main__": print(F"{price_plus_tax(1_0_0, 0.25) = }") print(F"{price_plus_tax(1_2_5.5_0, 0.05) = }")
367
'''simple docstring''' from maths.is_square_free import is_square_free from maths.prime_factors import prime_factors def a ( lowerCamelCase__ ): '''simple docstring''' A_ : List[Any] = prime_factors(lowerCamelCase__ ) if is_square_free(lowerCamelCase__ ): return -1 if len(lowerCamelCase__ ) % 2 else 1 return 0 if __name__ == "__main__": import doctest doctest.testmod()
135
0
"""simple docstring""" import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _a : List[str] = '▁' _a : str = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class __A ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): _UpperCamelCase : Tuple = BigBirdTokenizer _UpperCamelCase : Tuple = BigBirdTokenizerFast _UpperCamelCase : Optional[int] = True _UpperCamelCase : List[str] = True def __A ( self ): super().setUp() _lowerCAmelCase : Dict = self.tokenizer_class(lowerCAmelCase__ , keep_accents=lowerCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def __A ( self ): _lowerCAmelCase : Union[str, Any] = """<s>""" _lowerCAmelCase : List[str] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCAmelCase__ ) , lowerCAmelCase__ ) def __A ( self ): _lowerCAmelCase : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<unk>""" ) self.assertEqual(vocab_keys[1] , """<s>""" ) self.assertEqual(vocab_keys[-1] , """[MASK]""" ) self.assertEqual(len(lowerCAmelCase__ ) , 1004 ) def __A ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1000 ) def __A ( self ): if not self.test_rust_tokenizer: return _lowerCAmelCase : Optional[Any] = self.get_tokenizer() _lowerCAmelCase : Optional[Any] = self.get_rust_tokenizer() _lowerCAmelCase : Union[str, Any] = """I was born in 92000, and this is falsé.""" _lowerCAmelCase : List[Any] = tokenizer.tokenize(lowerCAmelCase__ ) _lowerCAmelCase : int = rust_tokenizer.tokenize(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) _lowerCAmelCase : Optional[Any] = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) _lowerCAmelCase : Union[str, Any] = rust_tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) _lowerCAmelCase : Dict = self.get_rust_tokenizer() _lowerCAmelCase : str = tokenizer.encode(lowerCAmelCase__ ) _lowerCAmelCase : str = rust_tokenizer.encode(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) def __A ( self ): _lowerCAmelCase : Optional[int] = BigBirdTokenizer(lowerCAmelCase__ , keep_accents=lowerCAmelCase__ ) _lowerCAmelCase : int = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(lowerCAmelCase__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [285, 46, 10, 170, 382] , ) _lowerCAmelCase : Any = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( lowerCAmelCase__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) _lowerCAmelCase : Tuple = tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) self.assertListEqual( lowerCAmelCase__ , [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) _lowerCAmelCase : List[str] = tokenizer.convert_ids_to_tokens(lowerCAmelCase__ ) self.assertListEqual( lowerCAmelCase__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) @cached_property def __A ( self ): return BigBirdTokenizer.from_pretrained("""google/bigbird-roberta-base""" ) @slow def __A ( self ): _lowerCAmelCase : Any = """Hello World!""" _lowerCAmelCase : List[Any] = [65, 18536, 2260, 101, 66] self.assertListEqual(lowerCAmelCase__ , self.big_tokenizer.encode(lowerCAmelCase__ ) ) @slow def __A ( self ): _lowerCAmelCase : str = ( """This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will""" """ add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth""" ) # fmt: off _lowerCAmelCase : Optional[int] = [65, 871, 419, 358, 946, 991, 2521, 452, 358, 1357, 387, 7751, 3536, 112, 985, 456, 126, 865, 938, 5400, 5734, 458, 1368, 467, 786, 2462, 5246, 1159, 633, 865, 4519, 457, 582, 852, 2557, 427, 916, 508, 405, 34324, 497, 391, 408, 11342, 1244, 385, 100, 938, 985, 456, 574, 362, 12597, 3200, 3129, 1172, 66] # noqa: E231 # fmt: on self.assertListEqual(lowerCAmelCase__ , self.big_tokenizer.encode(lowerCAmelCase__ ) ) @require_torch @slow def __A ( self ): import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence _lowerCAmelCase : Any = list(self.big_tokenizer.get_vocab().keys() )[:10] _lowerCAmelCase : Tuple = """ """.join(lowerCAmelCase__ ) _lowerCAmelCase : int = self.big_tokenizer.encode_plus(lowerCAmelCase__ , return_tensors="""pt""" , return_token_type_ids=lowerCAmelCase__ ) _lowerCAmelCase : Tuple = self.big_tokenizer.batch_encode_plus( [sequence + """ """ + sequence] , return_tensors="""pt""" , return_token_type_ids=lowerCAmelCase__ ) _lowerCAmelCase : Optional[int] = BigBirdConfig(attention_type="""original_full""" ) _lowerCAmelCase : Dict = BigBirdModel(lowerCAmelCase__ ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**lowerCAmelCase__ ) model(**lowerCAmelCase__ ) @slow def __A ( self ): _lowerCAmelCase : int = BigBirdTokenizer.from_pretrained("""google/bigbird-roberta-base""" ) _lowerCAmelCase : Any = tokenizer.decode(tokenizer("""Paris is the [MASK].""" ).input_ids ) self.assertTrue(decoded_text == """[CLS] Paris is the[MASK].[SEP]""" ) @slow def __A ( self ): _lowerCAmelCase : Any = {"""input_ids""": [[65, 39286, 458, 36335, 2001, 456, 13073, 13266, 455, 113, 7746, 1741, 11157, 391, 13073, 13266, 455, 113, 3967, 35412, 113, 4936, 109, 3870, 2377, 113, 30084, 45720, 458, 134, 17496, 112, 503, 11672, 113, 118, 112, 5665, 13347, 38687, 112, 1496, 31389, 112, 3268, 47264, 134, 962, 112, 16377, 8035, 23130, 430, 12169, 15518, 28592, 458, 146, 41697, 109, 391, 12169, 15518, 16689, 458, 146, 41358, 109, 452, 726, 4034, 111, 763, 35412, 5082, 388, 1903, 111, 9051, 391, 2870, 48918, 1900, 1123, 550, 998, 112, 9586, 15985, 455, 391, 410, 22955, 37636, 114, 66], [65, 448, 17496, 419, 3663, 385, 763, 113, 27533, 2870, 3283, 13043, 1639, 24713, 523, 656, 24013, 18550, 2521, 517, 27014, 21244, 420, 1212, 1465, 391, 927, 4833, 388, 578, 11786, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 484, 2169, 7687, 21932, 18146, 726, 363, 17032, 3391, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCAmelCase__ , model_name="""google/bigbird-roberta-base""" , revision="""215c99f1600e06f83acce68422f2035b2b5c3510""" , )
44
'''simple docstring''' import itertools import string from collections.abc import Generator, Iterable def a__ ( lowercase : Iterable[str], lowercase : int ) -> Generator[tuple[str, ...], None, None]: """simple docstring""" _UpperCamelCase = iter(lowercase ) while True: _UpperCamelCase = tuple(itertools.islice(lowercase, lowercase ) ) if not chunk: return yield chunk def a__ ( lowercase : str ) -> str: """simple docstring""" _UpperCamelCase = ''''''.join([c.upper() for c in dirty if c in string.ascii_letters] ) _UpperCamelCase = '''''' if len(lowercase ) < 2: return dirty for i in range(len(lowercase ) - 1 ): clean += dirty[i] if dirty[i] == dirty[i + 1]: clean += "X" clean += dirty[-1] if len(lowercase ) & 1: clean += "X" return clean def a__ ( lowercase : str ) -> list[str]: """simple docstring""" _UpperCamelCase = '''ABCDEFGHIKLMNOPQRSTUVWXYZ''' # we're using a list instead of a '2d' array because it makes the math # for setting up the table and doing the actual encoding/decoding simpler _UpperCamelCase = [] # copy key chars into the table if they are in `alphabet` ignoring duplicates for char in key.upper(): if char not in table and char in alphabet: table.append(lowercase ) # fill the rest of the table in with the remaining alphabet chars for char in alphabet: if char not in table: table.append(lowercase ) return table def a__ ( lowercase : str, lowercase : str ) -> str: """simple docstring""" _UpperCamelCase = generate_table(lowercase ) _UpperCamelCase = prepare_input(lowercase ) _UpperCamelCase = '''''' # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(lowercase, 2 ): _UpperCamelCase , _UpperCamelCase = divmod(table.index(lowercase ), 5 ) _UpperCamelCase , _UpperCamelCase = divmod(table.index(lowercase ), 5 ) if rowa == rowa: ciphertext += table[rowa * 5 + (cola + 1) % 5] ciphertext += table[rowa * 5 + (cola + 1) % 5] elif cola == cola: ciphertext += table[((rowa + 1) % 5) * 5 + cola] ciphertext += table[((rowa + 1) % 5) * 5 + cola] else: # rectangle ciphertext += table[rowa * 5 + cola] ciphertext += table[rowa * 5 + cola] return ciphertext def a__ ( lowercase : str, lowercase : str ) -> str: """simple docstring""" _UpperCamelCase = generate_table(lowercase ) _UpperCamelCase = '''''' # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(lowercase, 2 ): _UpperCamelCase , _UpperCamelCase = divmod(table.index(lowercase ), 5 ) _UpperCamelCase , _UpperCamelCase = divmod(table.index(lowercase ), 5 ) if rowa == rowa: plaintext += table[rowa * 5 + (cola - 1) % 5] plaintext += table[rowa * 5 + (cola - 1) % 5] elif cola == cola: plaintext += table[((rowa - 1) % 5) * 5 + cola] plaintext += table[((rowa - 1) % 5) * 5 + cola] else: # rectangle plaintext += table[rowa * 5 + cola] plaintext += table[rowa * 5 + cola] return plaintext
324
0
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import GLPNImageProcessor class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def __init__( self : Dict , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : int=7 , SCREAMING_SNAKE_CASE_ : Any=3 , SCREAMING_SNAKE_CASE_ : List[str]=18 , SCREAMING_SNAKE_CASE_ : Tuple=30 , SCREAMING_SNAKE_CASE_ : List[Any]=4_00 , SCREAMING_SNAKE_CASE_ : List[Any]=True , SCREAMING_SNAKE_CASE_ : int=32 , SCREAMING_SNAKE_CASE_ : Any=True , ) -> List[Any]: '''simple docstring''' A: str = parent A: Any = batch_size A: List[str] = num_channels A: Union[str, Any] = image_size A: Optional[int] = min_resolution A: int = max_resolution A: Dict = do_resize A: List[Any] = size_divisor A: int = do_rescale def _snake_case ( self : Dict ) -> str: '''simple docstring''' return { "do_resize": self.do_resize, "size_divisor": self.size_divisor, "do_rescale": self.do_rescale, } @require_torch @require_vision class lowerCAmelCase_ ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' UpperCamelCase_ : int = GLPNImageProcessor if is_vision_available() else None def _snake_case ( self : int ) -> Optional[Any]: '''simple docstring''' A: str = GLPNImageProcessingTester(self ) @property def _snake_case ( self : Any ) -> List[str]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def _snake_case ( self : Optional[Any] ) -> Dict: '''simple docstring''' A: Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''do_resize''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''size_divisor''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''resample''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''do_rescale''' ) ) def _snake_case ( self : Any ) -> List[Any]: '''simple docstring''' pass def _snake_case ( self : Union[str, Any] ) -> int: '''simple docstring''' A: List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A: Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE_ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE_ , Image.Image ) # Test not batched input (GLPNImageProcessor doesn't support batching) A: int = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def _snake_case ( self : List[Any] ) -> Optional[int]: '''simple docstring''' A: Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A: List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE_ , numpify=SCREAMING_SNAKE_CASE_ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE_ , np.ndarray ) # Test not batched input (GLPNImageProcessor doesn't support batching) A: Tuple = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def _snake_case ( self : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' A: Any = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A: Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE_ , torchify=SCREAMING_SNAKE_CASE_ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE_ , torch.Tensor ) # Test not batched input (GLPNImageProcessor doesn't support batching) A: Dict = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 )
334
'''simple docstring''' import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() UpperCamelCase = logging.get_logger('''transformers.models.encodec''') UpperCamelCase = { '''quantizer.vq.layers.*._codebook.inited''': '''quantizer.layers.*.codebook.inited''', '''quantizer.vq.layers.*._codebook.cluster_size''': '''quantizer.layers.*.codebook.cluster_size''', '''quantizer.vq.layers.*._codebook.embed''': '''quantizer.layers.*.codebook.embed''', '''quantizer.vq.layers.*._codebook.embed_avg''': '''quantizer.layers.*.codebook.embed_avg''', } UpperCamelCase = { '''encoder.model.0.conv.conv''': '''encoder.layers.0.conv''', '''encoder.model.1.block.1.conv.conv''': '''encoder.layers.1.block.1.conv''', '''encoder.model.1.block.3.conv.conv''': '''encoder.layers.1.block.3.conv''', '''encoder.model.1.shortcut.conv.conv''': '''encoder.layers.1.shortcut.conv''', '''encoder.model.3.conv.conv''': '''encoder.layers.3.conv''', '''encoder.model.4.block.1.conv.conv''': '''encoder.layers.4.block.1.conv''', '''encoder.model.4.block.3.conv.conv''': '''encoder.layers.4.block.3.conv''', '''encoder.model.4.shortcut.conv.conv''': '''encoder.layers.4.shortcut.conv''', '''encoder.model.6.conv.conv''': '''encoder.layers.6.conv''', '''encoder.model.7.block.1.conv.conv''': '''encoder.layers.7.block.1.conv''', '''encoder.model.7.block.3.conv.conv''': '''encoder.layers.7.block.3.conv''', '''encoder.model.7.shortcut.conv.conv''': '''encoder.layers.7.shortcut.conv''', '''encoder.model.9.conv.conv''': '''encoder.layers.9.conv''', '''encoder.model.10.block.1.conv.conv''': '''encoder.layers.10.block.1.conv''', '''encoder.model.10.block.3.conv.conv''': '''encoder.layers.10.block.3.conv''', '''encoder.model.10.shortcut.conv.conv''': '''encoder.layers.10.shortcut.conv''', '''encoder.model.12.conv.conv''': '''encoder.layers.12.conv''', '''encoder.model.13.lstm''': '''encoder.layers.13.lstm''', '''encoder.model.15.conv.conv''': '''encoder.layers.15.conv''', } UpperCamelCase = { '''encoder.model.0.conv.norm''': '''encoder.layers.0.norm''', '''encoder.model.1.block.1.conv.norm''': '''encoder.layers.1.block.1.norm''', '''encoder.model.1.block.3.conv.norm''': '''encoder.layers.1.block.3.norm''', '''encoder.model.1.shortcut.conv.norm''': '''encoder.layers.1.shortcut.norm''', '''encoder.model.3.conv.norm''': '''encoder.layers.3.norm''', '''encoder.model.4.block.1.conv.norm''': '''encoder.layers.4.block.1.norm''', '''encoder.model.4.block.3.conv.norm''': '''encoder.layers.4.block.3.norm''', '''encoder.model.4.shortcut.conv.norm''': '''encoder.layers.4.shortcut.norm''', '''encoder.model.6.conv.norm''': '''encoder.layers.6.norm''', '''encoder.model.7.block.1.conv.norm''': '''encoder.layers.7.block.1.norm''', '''encoder.model.7.block.3.conv.norm''': '''encoder.layers.7.block.3.norm''', '''encoder.model.7.shortcut.conv.norm''': '''encoder.layers.7.shortcut.norm''', '''encoder.model.9.conv.norm''': '''encoder.layers.9.norm''', '''encoder.model.10.block.1.conv.norm''': '''encoder.layers.10.block.1.norm''', '''encoder.model.10.block.3.conv.norm''': '''encoder.layers.10.block.3.norm''', '''encoder.model.10.shortcut.conv.norm''': '''encoder.layers.10.shortcut.norm''', '''encoder.model.12.conv.norm''': '''encoder.layers.12.norm''', '''encoder.model.15.conv.norm''': '''encoder.layers.15.norm''', } UpperCamelCase = { '''decoder.model.0.conv.conv''': '''decoder.layers.0.conv''', '''decoder.model.1.lstm''': '''decoder.layers.1.lstm''', '''decoder.model.3.convtr.convtr''': '''decoder.layers.3.conv''', '''decoder.model.4.block.1.conv.conv''': '''decoder.layers.4.block.1.conv''', '''decoder.model.4.block.3.conv.conv''': '''decoder.layers.4.block.3.conv''', '''decoder.model.4.shortcut.conv.conv''': '''decoder.layers.4.shortcut.conv''', '''decoder.model.6.convtr.convtr''': '''decoder.layers.6.conv''', '''decoder.model.7.block.1.conv.conv''': '''decoder.layers.7.block.1.conv''', '''decoder.model.7.block.3.conv.conv''': '''decoder.layers.7.block.3.conv''', '''decoder.model.7.shortcut.conv.conv''': '''decoder.layers.7.shortcut.conv''', '''decoder.model.9.convtr.convtr''': '''decoder.layers.9.conv''', '''decoder.model.10.block.1.conv.conv''': '''decoder.layers.10.block.1.conv''', '''decoder.model.10.block.3.conv.conv''': '''decoder.layers.10.block.3.conv''', '''decoder.model.10.shortcut.conv.conv''': '''decoder.layers.10.shortcut.conv''', '''decoder.model.12.convtr.convtr''': '''decoder.layers.12.conv''', '''decoder.model.13.block.1.conv.conv''': '''decoder.layers.13.block.1.conv''', '''decoder.model.13.block.3.conv.conv''': '''decoder.layers.13.block.3.conv''', '''decoder.model.13.shortcut.conv.conv''': '''decoder.layers.13.shortcut.conv''', '''decoder.model.15.conv.conv''': '''decoder.layers.15.conv''', } UpperCamelCase = { '''decoder.model.0.conv.norm''': '''decoder.layers.0.norm''', '''decoder.model.3.convtr.norm''': '''decoder.layers.3.norm''', '''decoder.model.4.block.1.conv.norm''': '''decoder.layers.4.block.1.norm''', '''decoder.model.4.block.3.conv.norm''': '''decoder.layers.4.block.3.norm''', '''decoder.model.4.shortcut.conv.norm''': '''decoder.layers.4.shortcut.norm''', '''decoder.model.6.convtr.norm''': '''decoder.layers.6.norm''', '''decoder.model.7.block.1.conv.norm''': '''decoder.layers.7.block.1.norm''', '''decoder.model.7.block.3.conv.norm''': '''decoder.layers.7.block.3.norm''', '''decoder.model.7.shortcut.conv.norm''': '''decoder.layers.7.shortcut.norm''', '''decoder.model.9.convtr.norm''': '''decoder.layers.9.norm''', '''decoder.model.10.block.1.conv.norm''': '''decoder.layers.10.block.1.norm''', '''decoder.model.10.block.3.conv.norm''': '''decoder.layers.10.block.3.norm''', '''decoder.model.10.shortcut.conv.norm''': '''decoder.layers.10.shortcut.norm''', '''decoder.model.12.convtr.norm''': '''decoder.layers.12.norm''', '''decoder.model.13.block.1.conv.norm''': '''decoder.layers.13.block.1.norm''', '''decoder.model.13.block.3.conv.norm''': '''decoder.layers.13.block.3.norm''', '''decoder.model.13.shortcut.conv.norm''': '''decoder.layers.13.shortcut.norm''', '''decoder.model.15.conv.norm''': '''decoder.layers.15.norm''', } UpperCamelCase = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } UpperCamelCase = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } UpperCamelCase = [] UpperCamelCase = [] def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> Dict: for attribute in key.split('''.''' ): A: Union[str, Any] = getattr(__lowercase , __lowercase ) if weight_type is not None: A: Tuple = getattr(__lowercase , __lowercase ).shape else: A: str = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": A: Dict = value elif weight_type == "weight_g": A: Tuple = value elif weight_type == "weight_v": A: Any = value elif weight_type == "bias": A: str = value elif weight_type == "running_mean": A: List[Any] = value elif weight_type == "running_var": A: Dict = value elif weight_type == "num_batches_tracked": A: List[str] = value elif weight_type == "weight_ih_l0": A: Dict = value elif weight_type == "weight_hh_l0": A: Optional[int] = value elif weight_type == "bias_ih_l0": A: List[Any] = value elif weight_type == "bias_hh_l0": A: str = value elif weight_type == "weight_ih_l1": A: Optional[int] = value elif weight_type == "weight_hh_l1": A: int = value elif weight_type == "bias_ih_l1": A: Optional[Any] = value elif weight_type == "bias_hh_l1": A: str = value else: A: Optional[int] = value logger.info(F"""{key + ('.' + weight_type if weight_type is not None else '')} was initialized from {full_name}.""" ) def SCREAMING_SNAKE_CASE( __lowercase , __lowercase ) -> Optional[Any]: for key in ignore_keys: if key.endswith('''.*''' ): if name.startswith(key[:-1] ): return True elif ".*." in key: A , A: Any = key.split('''.*.''' ) if prefix in name and suffix in name: return True elif key in name: return True return False def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase ) -> Tuple: A: Any = [] if model_name == "encodec_24khz" or "encodec_32khz": A: List[str] = MAPPING_24K elif model_name == "encodec_48khz": A: List[Any] = MAPPING_48K else: raise ValueError(F"""Unsupported model: {model_name}""" ) for name, value in orig_dict.items(): if should_ignore(__lowercase , __lowercase ): logger.info(F"""{name} was ignored""" ) continue A: Optional[int] = False for key, mapped_key in MAPPING.items(): if "*" in key: A , A: Optional[int] = key.split('''.*.''' ) if prefix in name and suffix in name: A: str = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith('''embed''' ) and name.endswith('''embed_avg''' ): continue A: Optional[Any] = True if "*" in mapped_key: A: Any = name.split(__lowercase )[0].split('''.''' )[-2] A: Tuple = mapped_key.replace('''*''' , __lowercase ) if "weight_g" in name: A: str = '''weight_g''' elif "weight_v" in name: A: List[Any] = '''weight_v''' elif "weight_ih_l0" in name: A: Dict = '''weight_ih_l0''' elif "weight_hh_l0" in name: A: int = '''weight_hh_l0''' elif "bias_ih_l0" in name: A: Union[str, Any] = '''bias_ih_l0''' elif "bias_hh_l0" in name: A: Tuple = '''bias_hh_l0''' elif "weight_ih_l1" in name: A: int = '''weight_ih_l1''' elif "weight_hh_l1" in name: A: Optional[Any] = '''weight_hh_l1''' elif "bias_ih_l1" in name: A: Dict = '''bias_ih_l1''' elif "bias_hh_l1" in name: A: str = '''bias_hh_l1''' elif "bias" in name: A: Union[str, Any] = '''bias''' elif "weight" in name: A: Dict = '''weight''' elif "running_mean" in name: A: Tuple = '''running_mean''' elif "running_var" in name: A: Any = '''running_var''' elif "num_batches_tracked" in name: A: str = '''num_batches_tracked''' else: A: Tuple = None set_recursively(__lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) continue if not is_used: unused_weights.append(__lowercase ) logger.warning(F"""Unused weights: {unused_weights}""" ) @torch.no_grad() def SCREAMING_SNAKE_CASE( __lowercase , __lowercase , __lowercase , __lowercase=None , __lowercase=None , ) -> Dict: if config_path is not None: A: Tuple = EncodecConfig.from_pretrained(__lowercase ) else: A: Union[str, Any] = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": A: Union[str, Any] = [8, 5, 4, 4] A: Dict = [2.2] A: List[Any] = 6_4 A: Optional[Any] = 3_2_0_0_0 A: List[Any] = 2_0_4_8 A: Optional[Any] = False A: int = False A: Union[str, Any] = False elif model_name == "encodec_48khz": A: Optional[int] = [8, 5, 4, 2] A: List[Any] = [3.0, 6.0, 1_2.0, 2_4.0] A: List[Any] = 4_8_0_0_0 A: int = 2 A: List[Any] = False A: Any = '''time_group_norm''' A: Optional[Any] = True A: Any = 1.0 A: Any = 0.0_1 else: raise ValueError(F"""Unknown model name: {model_name}""" ) A: str = EncodecModel(__lowercase ) A: Optional[Any] = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(__lowercase ) A: Union[str, Any] = torch.load(__lowercase ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights A: Optional[int] = original_checkpoint['''best_state'''] recursively_load_weights(__lowercase , __lowercase , __lowercase ) model.save_pretrained(__lowercase ) if repo_id: print('''Pushing to the hub...''' ) feature_extractor.push_to_hub(__lowercase ) model.push_to_hub(__lowercase ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() parser.add_argument( '''--model''', default='''encodec_24khz''', type=str, help='''The model to convert. Should be one of \'encodec_24khz\', \'encodec_32khz\', \'encodec_48khz\'.''', ) parser.add_argument('''--checkpoint_path''', required=True, default=None, type=str, help='''Path to original checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, default=None, type=str, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) UpperCamelCase = parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
334
1
"""simple docstring""" import argparse import collections import json from pathlib import Path import requests import torch import yaml from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTImageProcessor, MobileViTVaConfig, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, ) from transformers.utils import logging logging.set_verbosity_info() __UpperCamelCase = logging.get_logger(__name__) def lowercase (SCREAMING_SNAKE_CASE_ : Dict ) -> Optional[Any]: print('Loading config file...' ) def flatten_yaml_as_dict(SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Tuple="" , SCREAMING_SNAKE_CASE_ : str="." ): SCREAMING_SNAKE_CASE = [] for k, v in d.items(): SCREAMING_SNAKE_CASE = parent_key + sep + k if parent_key else k if isinstance(SCREAMING_SNAKE_CASE_ , collections.abc.MutableMapping ): items.extend(flatten_yaml_as_dict(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , sep=SCREAMING_SNAKE_CASE_ ).items() ) else: items.append((new_key, v) ) return dict(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = argparse.Namespace() with open(SCREAMING_SNAKE_CASE_ , 'r' ) as yaml_file: try: SCREAMING_SNAKE_CASE = yaml.load(SCREAMING_SNAKE_CASE_ , Loader=yaml.FullLoader ) SCREAMING_SNAKE_CASE = flatten_yaml_as_dict(SCREAMING_SNAKE_CASE_ ) for k, v in flat_cfg.items(): setattr(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) except yaml.YAMLError as exc: logger.error('Error while loading config file: {}. Error message: {}'.format(SCREAMING_SNAKE_CASE_ , str(SCREAMING_SNAKE_CASE_ ) ) ) return config def lowercase (SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Any ) -> List[Any]: SCREAMING_SNAKE_CASE = MobileViTVaConfig() SCREAMING_SNAKE_CASE = False # dataset if task_name.startswith('imagenet1k_' ): SCREAMING_SNAKE_CASE = 10_00 if int(task_name.strip().split('_' )[-1] ) == 3_84: SCREAMING_SNAKE_CASE = 3_84 else: SCREAMING_SNAKE_CASE = 2_56 SCREAMING_SNAKE_CASE = 'imagenet-1k-id2label.json' elif task_name.startswith('imagenet21k_to_1k_' ): SCREAMING_SNAKE_CASE = 2_10_00 if int(task_name.strip().split('_' )[-1] ) == 3_84: SCREAMING_SNAKE_CASE = 3_84 else: SCREAMING_SNAKE_CASE = 2_56 SCREAMING_SNAKE_CASE = 'imagenet-22k-id2label.json' elif task_name.startswith('ade20k_' ): SCREAMING_SNAKE_CASE = 1_51 SCREAMING_SNAKE_CASE = 5_12 SCREAMING_SNAKE_CASE = 'ade20k-id2label.json' SCREAMING_SNAKE_CASE = True elif task_name.startswith('voc_' ): SCREAMING_SNAKE_CASE = 21 SCREAMING_SNAKE_CASE = 5_12 SCREAMING_SNAKE_CASE = 'pascal-voc-id2label.json' SCREAMING_SNAKE_CASE = True # orig_config SCREAMING_SNAKE_CASE = load_orig_config_file(SCREAMING_SNAKE_CASE_ ) assert getattr(SCREAMING_SNAKE_CASE_ , 'model.classification.name' , -1 ) == "mobilevit_v2", "Invalid model" SCREAMING_SNAKE_CASE = getattr(SCREAMING_SNAKE_CASE_ , 'model.classification.mitv2.width_multiplier' , 1.0 ) assert ( getattr(SCREAMING_SNAKE_CASE_ , 'model.classification.mitv2.attn_norm_layer' , -1 ) == "layer_norm_2d" ), "Norm layers other than layer_norm_2d is not supported" SCREAMING_SNAKE_CASE = getattr(SCREAMING_SNAKE_CASE_ , 'model.classification.activation.name' , 'swish' ) # config.image_size == getattr(orig_config, 'sampler.bs.crop_size_width', 256) if is_segmentation_model: SCREAMING_SNAKE_CASE = getattr(SCREAMING_SNAKE_CASE_ , 'model.segmentation.output_stride' , 16 ) if "_deeplabv3" in task_name: SCREAMING_SNAKE_CASE = getattr(SCREAMING_SNAKE_CASE_ , 'model.segmentation.deeplabv3.aspp_rates' , [12, 24, 36] ) SCREAMING_SNAKE_CASE = getattr(SCREAMING_SNAKE_CASE_ , 'model.segmentation.deeplabv3.aspp_out_channels' , 5_12 ) SCREAMING_SNAKE_CASE = getattr(SCREAMING_SNAKE_CASE_ , 'model.segmentation.deeplabv3.aspp_dropout' , 0.1 ) # id2label SCREAMING_SNAKE_CASE = 'huggingface/label-files' SCREAMING_SNAKE_CASE = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , repo_type='dataset' ) , 'r' ) ) SCREAMING_SNAKE_CASE = {int(SCREAMING_SNAKE_CASE_ ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE = idalabel SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} return config def lowercase (SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : str ) -> List[str]: SCREAMING_SNAKE_CASE = dct.pop(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = val def lowercase (SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Dict=False ) -> List[Any]: if base_model: SCREAMING_SNAKE_CASE = '' else: SCREAMING_SNAKE_CASE = 'mobilevitv2.' SCREAMING_SNAKE_CASE = [] for k in state_dict.keys(): if k[:8] == "encoder.": SCREAMING_SNAKE_CASE = k[8:] else: SCREAMING_SNAKE_CASE = k if ".block." in k: SCREAMING_SNAKE_CASE = k_new.replace('.block.' , '.' ) if ".conv." in k: SCREAMING_SNAKE_CASE = k_new.replace('.conv.' , '.convolution.' ) if ".norm." in k: SCREAMING_SNAKE_CASE = k_new.replace('.norm.' , '.normalization.' ) if "conv_1." in k: SCREAMING_SNAKE_CASE = k_new.replace('conv_1.' , F'{model_prefix}conv_stem.' ) for i in [1, 2]: if F'layer_{i}.' in k: SCREAMING_SNAKE_CASE = k_new.replace(F'layer_{i}.' , F'{model_prefix}encoder.layer.{i-1}.layer.' ) if ".exp_1x1." in k: SCREAMING_SNAKE_CASE = k_new.replace('.exp_1x1.' , '.expand_1x1.' ) if ".red_1x1." in k: SCREAMING_SNAKE_CASE = k_new.replace('.red_1x1.' , '.reduce_1x1.' ) for i in [3, 4, 5]: if F'layer_{i}.0.' in k: SCREAMING_SNAKE_CASE = k_new.replace(F'layer_{i}.0.' , F'{model_prefix}encoder.layer.{i-1}.downsampling_layer.' ) if F'layer_{i}.1.local_rep.0.' in k: SCREAMING_SNAKE_CASE = k_new.replace(F'layer_{i}.1.local_rep.0.' , F'{model_prefix}encoder.layer.{i-1}.conv_kxk.' ) if F'layer_{i}.1.local_rep.1.' in k: SCREAMING_SNAKE_CASE = k_new.replace(F'layer_{i}.1.local_rep.1.' , F'{model_prefix}encoder.layer.{i-1}.conv_1x1.' ) for i in [3, 4, 5]: if i == 3: SCREAMING_SNAKE_CASE = [0, 1] elif i == 4: SCREAMING_SNAKE_CASE = [0, 1, 2, 3] elif i == 5: SCREAMING_SNAKE_CASE = [0, 1, 2] for j in j_in: if F'layer_{i}.1.global_rep.{j}.' in k: SCREAMING_SNAKE_CASE = k_new.replace( F'layer_{i}.1.global_rep.{j}.' , F'{model_prefix}encoder.layer.{i-1}.transformer.layer.{j}.' ) if F'layer_{i}.1.global_rep.{j+1}.' in k: SCREAMING_SNAKE_CASE = k_new.replace( F'layer_{i}.1.global_rep.{j+1}.' , F'{model_prefix}encoder.layer.{i-1}.layernorm.' ) if F'layer_{i}.1.conv_proj.' in k: SCREAMING_SNAKE_CASE = k_new.replace(F'layer_{i}.1.conv_proj.' , F'{model_prefix}encoder.layer.{i-1}.conv_projection.' ) if "pre_norm_attn.0." in k: SCREAMING_SNAKE_CASE = k_new.replace('pre_norm_attn.0.' , 'layernorm_before.' ) if "pre_norm_attn.1." in k: SCREAMING_SNAKE_CASE = k_new.replace('pre_norm_attn.1.' , 'attention.' ) if "pre_norm_ffn.0." in k: SCREAMING_SNAKE_CASE = k_new.replace('pre_norm_ffn.0.' , 'layernorm_after.' ) if "pre_norm_ffn.1." in k: SCREAMING_SNAKE_CASE = k_new.replace('pre_norm_ffn.1.' , 'ffn.conv1.' ) if "pre_norm_ffn.3." in k: SCREAMING_SNAKE_CASE = k_new.replace('pre_norm_ffn.3.' , 'ffn.conv2.' ) if "classifier.1." in k: SCREAMING_SNAKE_CASE = k_new.replace('classifier.1.' , 'classifier.' ) if "seg_head." in k: SCREAMING_SNAKE_CASE = k_new.replace('seg_head.' , 'segmentation_head.' ) if ".aspp_layer." in k: SCREAMING_SNAKE_CASE = k_new.replace('.aspp_layer.' , '.' ) if ".aspp_pool." in k: SCREAMING_SNAKE_CASE = k_new.replace('.aspp_pool.' , '.' ) rename_keys.append((k, k_new) ) return rename_keys def lowercase (SCREAMING_SNAKE_CASE_ : List[str] ) -> Tuple: SCREAMING_SNAKE_CASE = [] for k in state_dict.keys(): if k.startswith('seg_head.aux_head.' ): keys_to_ignore.append(SCREAMING_SNAKE_CASE_ ) for k in keys_to_ignore: state_dict.pop(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def lowercase () -> Any: SCREAMING_SNAKE_CASE = 'http://images.cocodataset.org/val2017/000000039769.jpg' # url = "https://cdn.britannica.com/86/141086-050-9D7C75EE/Gulfstream-G450-business-jet-passengers.jpg" SCREAMING_SNAKE_CASE = Image.open(requests.get(SCREAMING_SNAKE_CASE_ , stream=SCREAMING_SNAKE_CASE_ ).raw ) return im @torch.no_grad() def lowercase (SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : str ) -> List[Any]: SCREAMING_SNAKE_CASE = get_mobilevitva_config(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # load original state_dict SCREAMING_SNAKE_CASE = torch.load(SCREAMING_SNAKE_CASE_ , map_location='cpu' ) # load huggingface model if task_name.startswith('ade20k_' ) or task_name.startswith('voc_' ): SCREAMING_SNAKE_CASE = MobileViTVaForSemanticSegmentation(SCREAMING_SNAKE_CASE_ ).eval() SCREAMING_SNAKE_CASE = False else: SCREAMING_SNAKE_CASE = MobileViTVaForImageClassification(SCREAMING_SNAKE_CASE_ ).eval() SCREAMING_SNAKE_CASE = False # remove and rename some keys of load the original model SCREAMING_SNAKE_CASE = checkpoint remove_unused_keys(SCREAMING_SNAKE_CASE_ ) SCREAMING_SNAKE_CASE = create_rename_keys(SCREAMING_SNAKE_CASE_ , base_model=SCREAMING_SNAKE_CASE_ ) for rename_key_src, rename_key_dest in rename_keys: rename_key(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # load modified state_dict model.load_state_dict(SCREAMING_SNAKE_CASE_ ) # Check outputs on an image, prepared by MobileViTImageProcessor SCREAMING_SNAKE_CASE = MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 32 ) SCREAMING_SNAKE_CASE = image_processor(images=prepare_img() , return_tensors='pt' ) SCREAMING_SNAKE_CASE = model(**SCREAMING_SNAKE_CASE_ ) # verify classification model if task_name.startswith('imagenet' ): SCREAMING_SNAKE_CASE = outputs.logits SCREAMING_SNAKE_CASE = logits.argmax(-1 ).item() print('Predicted class:' , model.config.idalabel[predicted_class_idx] ) if task_name.startswith('imagenet1k_256' ) and config.width_multiplier == 1.0: # expected_logits for base variant SCREAMING_SNAKE_CASE = torch.tensor([-1.6336E00, -7.3204E-02, -5.1883E-01] ) assert torch.allclose(logits[0, :3] , SCREAMING_SNAKE_CASE_ , atol=1E-4 ) Path(SCREAMING_SNAKE_CASE_ ).mkdir(exist_ok=SCREAMING_SNAKE_CASE_ ) print(F'Saving model {task_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) print(F'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": __UpperCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--task''', default='''imagenet1k_256''', type=str, help=( '''Name of the task for which the MobileViTV2 model you\'d like to convert is trained on . ''' ''' Classification (ImageNet-1k) - MobileViTV2 (256x256) : imagenet1k_256 - MobileViTV2 (Trained on 256x256 and Finetuned on 384x384) : imagenet1k_384 - MobileViTV2 (Trained on ImageNet-21k and Finetuned on ImageNet-1k 256x256) : imagenet21k_to_1k_256 - MobileViTV2 (Trained on ImageNet-21k, Finetuned on ImageNet-1k 256x256, and Finetuned on ImageNet-1k 384x384) : imagenet21k_to_1k_384 Segmentation - ADE20K Dataset : ade20k_deeplabv3 - Pascal VOC 2012 Dataset: voc_deeplabv3 ''' ), choices=[ '''imagenet1k_256''', '''imagenet1k_384''', '''imagenet21k_to_1k_256''', '''imagenet21k_to_1k_384''', '''ade20k_deeplabv3''', '''voc_deeplabv3''', ], ) parser.add_argument( '''--orig_checkpoint_path''', required=True, type=str, help='''Path to the original state dict (.pt file).''' ) parser.add_argument('''--orig_config_path''', required=True, type=str, help='''Path to the original config file.''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, type=str, help='''Path to the output PyTorch model directory.''' ) __UpperCamelCase = parser.parse_args() convert_mobilevitva_checkpoint( args.task, args.orig_checkpoint_path, args.orig_config_path, args.pytorch_dump_folder_path )
113
"""simple docstring""" def lowercase (SCREAMING_SNAKE_CASE_ : int = 10_00 ) -> int: SCREAMING_SNAKE_CASE = 2**power SCREAMING_SNAKE_CASE = 0 while n: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = r + n % 10, n // 10 return r if __name__ == "__main__": print(solution(int(str(input()).strip())))
113
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A__ : List[str] = { 'configuration_clipseg': [ 'CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP', 'CLIPSegConfig', 'CLIPSegTextConfig', 'CLIPSegVisionConfig', ], 'processing_clipseg': ['CLIPSegProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : int = [ 'CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST', 'CLIPSegModel', 'CLIPSegPreTrainedModel', 'CLIPSegTextModel', 'CLIPSegVisionModel', 'CLIPSegForImageSegmentation', ] if TYPE_CHECKING: from .configuration_clipseg import ( CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPSegConfig, CLIPSegTextConfig, CLIPSegVisionConfig, ) from .processing_clipseg import CLIPSegProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clipseg import ( CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPSegForImageSegmentation, CLIPSegModel, CLIPSegPreTrainedModel, CLIPSegTextModel, CLIPSegVisionModel, ) else: import sys A__ : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
358
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class lowercase__ ( unittest.TestCase ): def UpperCAmelCase__ ( self : Any ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def UpperCAmelCase__ ( self : Dict ): lowerCamelCase_ : Optional[Any] =1 lowerCamelCase_ : Union[str, Any] =3 lowerCamelCase_ : Dict =(32, 32) lowerCamelCase_ : List[Any] =floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(snake_case__ ) return image @property def UpperCAmelCase__ ( self : Union[str, Any] ): torch.manual_seed(0 ) lowerCamelCase_ : Dict =UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) return model @property def UpperCAmelCase__ ( self : List[Any] ): torch.manual_seed(0 ) lowerCamelCase_ : Union[str, Any] =AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) return model @property def UpperCAmelCase__ ( self : Optional[Any] ): torch.manual_seed(0 ) lowerCamelCase_ : Any =RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5006 , ) return RobertaSeriesModelWithTransformation(snake_case__ ) @property def UpperCAmelCase__ ( self : int ): def extract(*snake_case__ : Dict , **snake_case__ : int ): class lowercase__ : def __init__( self : Optional[Any] ): lowerCamelCase_ : Union[str, Any] =torch.ones([0] ) def UpperCAmelCase__ ( self : List[Any] , snake_case__ : Any ): self.pixel_values.to(snake_case__ ) return self return Out() return extract def UpperCAmelCase__ ( self : Optional[Any] ): lowerCamelCase_ : Dict ="cpu" # ensure determinism for the device-dependent torch.Generator lowerCamelCase_ : Dict =self.dummy_cond_unet lowerCamelCase_ : List[str] =PNDMScheduler(skip_prk_steps=snake_case__ ) lowerCamelCase_ : List[Any] =self.dummy_vae lowerCamelCase_ : Any =self.dummy_text_encoder lowerCamelCase_ : List[Any] =XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) lowerCamelCase_ : List[str] =77 lowerCamelCase_ : Optional[int] =self.dummy_image.to(snake_case__ ) lowerCamelCase_ : Any =init_image / 2 + 0.5 # make sure here that pndm scheduler skips prk lowerCamelCase_ : List[Any] =AltDiffusionImgaImgPipeline( unet=snake_case__ , scheduler=snake_case__ , vae=snake_case__ , text_encoder=snake_case__ , tokenizer=snake_case__ , safety_checker=snake_case__ , feature_extractor=self.dummy_extractor , ) lowerCamelCase_ : int =VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=snake_case__ ) lowerCamelCase_ : Optional[int] =alt_pipe.to(snake_case__ ) alt_pipe.set_progress_bar_config(disable=snake_case__ ) lowerCamelCase_ : Any ="A painting of a squirrel eating a burger" lowerCamelCase_ : Union[str, Any] =torch.Generator(device=snake_case__ ).manual_seed(0 ) lowerCamelCase_ : str =alt_pipe( [prompt] , generator=snake_case__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=snake_case__ , ) lowerCamelCase_ : List[Any] =output.images lowerCamelCase_ : str =torch.Generator(device=snake_case__ ).manual_seed(0 ) lowerCamelCase_ : Optional[Any] =alt_pipe( [prompt] , generator=snake_case__ , guidance_scale=6.0 , num_inference_steps=2 , output_type="np" , image=snake_case__ , return_dict=snake_case__ , )[0] lowerCamelCase_ : Tuple =image[0, -3:, -3:, -1] lowerCamelCase_ : Union[str, Any] =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCamelCase_ : List[str] =np.array([0.4_427, 0.3_731, 0.4_249, 0.4_941, 0.4_546, 0.4_148, 0.4_193, 0.4_666, 0.4_499] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 5E-3 @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def UpperCAmelCase__ ( self : Tuple ): lowerCamelCase_ : Dict =self.dummy_cond_unet lowerCamelCase_ : Any =PNDMScheduler(skip_prk_steps=snake_case__ ) lowerCamelCase_ : Tuple =self.dummy_vae lowerCamelCase_ : Union[str, Any] =self.dummy_text_encoder lowerCamelCase_ : Optional[Any] =XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) lowerCamelCase_ : Tuple =77 lowerCamelCase_ : str =self.dummy_image.to(snake_case__ ) # put models in fp16 lowerCamelCase_ : Optional[Any] =unet.half() lowerCamelCase_ : Dict =vae.half() lowerCamelCase_ : str =bert.half() # make sure here that pndm scheduler skips prk lowerCamelCase_ : Optional[int] =AltDiffusionImgaImgPipeline( unet=snake_case__ , scheduler=snake_case__ , vae=snake_case__ , text_encoder=snake_case__ , tokenizer=snake_case__ , safety_checker=snake_case__ , feature_extractor=self.dummy_extractor , ) lowerCamelCase_ : List[Any] =VaeImageProcessor(vae_scale_factor=alt_pipe.vae_scale_factor , do_normalize=snake_case__ ) lowerCamelCase_ : int =alt_pipe.to(snake_case__ ) alt_pipe.set_progress_bar_config(disable=snake_case__ ) lowerCamelCase_ : Tuple ="A painting of a squirrel eating a burger" lowerCamelCase_ : Tuple =torch.manual_seed(0 ) lowerCamelCase_ : List[Any] =alt_pipe( [prompt] , generator=snake_case__ , num_inference_steps=2 , output_type="np" , image=snake_case__ , ).images assert image.shape == (1, 32, 32, 3) @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def UpperCAmelCase__ ( self : Union[str, Any] ): lowerCamelCase_ : int =load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) # resize to resolution that is divisible by 8 but not 16 or 32 lowerCamelCase_ : Any =init_image.resize((760, 504) ) lowerCamelCase_ : List[str] ="BAAI/AltDiffusion" lowerCamelCase_ : List[str] =AltDiffusionImgaImgPipeline.from_pretrained( snake_case__ , safety_checker=snake_case__ , ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) pipe.enable_attention_slicing() lowerCamelCase_ : Optional[int] ="A fantasy landscape, trending on artstation" lowerCamelCase_ : List[Any] =torch.manual_seed(0 ) lowerCamelCase_ : Optional[int] =pipe( prompt=snake_case__ , image=snake_case__ , strength=0.75 , guidance_scale=7.5 , generator=snake_case__ , output_type="np" , ) lowerCamelCase_ : Optional[int] =output.images[0] lowerCamelCase_ : Tuple =image[255:258, 383:386, -1] assert image.shape == (504, 760, 3) lowerCamelCase_ : Optional[int] =np.array([0.9_358, 0.9_397, 0.9_599, 0.9_901, 1.0_000, 1.0_000, 0.9_882, 1.0_000, 1.0_000] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class lowercase__ ( unittest.TestCase ): def UpperCAmelCase__ ( self : str ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self : List[Any] ): lowerCamelCase_ : str =load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) lowerCamelCase_ : Any =init_image.resize((768, 512) ) lowerCamelCase_ : Optional[int] =load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/img2img/fantasy_landscape_alt.npy" ) lowerCamelCase_ : Dict ="BAAI/AltDiffusion" lowerCamelCase_ : List[str] =AltDiffusionImgaImgPipeline.from_pretrained( snake_case__ , safety_checker=snake_case__ , ) pipe.to(snake_case__ ) pipe.set_progress_bar_config(disable=snake_case__ ) pipe.enable_attention_slicing() lowerCamelCase_ : Optional[Any] ="A fantasy landscape, trending on artstation" lowerCamelCase_ : Dict =torch.manual_seed(0 ) lowerCamelCase_ : int =pipe( prompt=snake_case__ , image=snake_case__ , strength=0.75 , guidance_scale=7.5 , generator=snake_case__ , output_type="np" , ) lowerCamelCase_ : int =output.images[0] assert image.shape == (512, 768, 3) # img2img is flaky across GPUs even in fp32, so using MAE here assert np.abs(expected_image - image ).max() < 1E-2
209
0
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class _a ( __a ): __a : Optional[int] = ["""image_processor""", """tokenizer"""] __a : str = """CLIPImageProcessor""" __a : Optional[Any] = ("""CLIPTokenizer""", """CLIPTokenizerFast""") def __init__( self : Optional[int] , lowercase : int=None , lowercase : Tuple=None , **lowercase : List[str] ): '''simple docstring''' UpperCAmelCase = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , lowercase , ) UpperCAmelCase = kwargs.pop('''feature_extractor''' ) UpperCAmelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(lowercase , lowercase ) def __call__( self : int , lowercase : List[str]=None , lowercase : Dict=None , lowercase : Tuple=None , **lowercase : int ): '''simple docstring''' if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: UpperCAmelCase = self.tokenizer(lowercase , return_tensors=lowercase , **lowercase ) if images is not None: UpperCAmelCase = self.image_processor(lowercase , return_tensors=lowercase , **lowercase ) if text is not None and images is not None: UpperCAmelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**lowercase ) , tensor_type=lowercase ) def A ( self : int , *lowercase : Any , **lowercase : Union[str, Any] ): '''simple docstring''' return self.tokenizer.batch_decode(*lowercase , **lowercase ) def A ( self : Optional[Any] , *lowercase : Any , **lowercase : Union[str, Any] ): '''simple docstring''' return self.tokenizer.decode(*lowercase , **lowercase ) @property def A ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase = self.tokenizer.model_input_names UpperCAmelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def A ( self : List[Any] ): '''simple docstring''' warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , lowercase , ) return self.image_processor_class @property def A ( self : Any ): '''simple docstring''' warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , lowercase , ) return self.image_processor
34
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable lowerCAmelCase = {'''configuration_gpt_neox''': ['''GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTNeoXConfig''']} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase = ['''GPTNeoXTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase = [ '''GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTNeoXForCausalLM''', '''GPTNeoXForQuestionAnswering''', '''GPTNeoXForSequenceClassification''', '''GPTNeoXForTokenClassification''', '''GPTNeoXLayer''', '''GPTNeoXModel''', '''GPTNeoXPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_neox import GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_neox_fast import GPTNeoXTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox import ( GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXLayer, GPTNeoXModel, GPTNeoXPreTrainedModel, ) else: import sys lowerCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
295
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig class snake_case ( __lowerCamelCase): __UpperCamelCase = """bert-generation""" def __init__( self : Dict , a__ : str=5_03_58 , a__ : int=10_24 , a__ : List[str]=24 , a__ : Optional[Any]=16 , a__ : List[str]=40_96 , a__ : Union[str, Any]="gelu" , a__ : int=0.1 , a__ : Optional[Any]=0.1 , a__ : Any=5_12 , a__ : Any=0.0_2 , a__ : int=1E-1_2 , a__ : List[Any]=0 , a__ : List[str]=2 , a__ : Any=1 , a__ : Tuple="absolute" , a__ : List[Any]=True , **a__ : int , ) -> List[Any]: '''simple docstring''' super().__init__(pad_token_id=__lowercase , bos_token_id=__lowercase , eos_token_id=__lowercase , **__lowercase ) _A = vocab_size _A = hidden_size _A = num_hidden_layers _A = num_attention_heads _A = hidden_act _A = intermediate_size _A = hidden_dropout_prob _A = attention_probs_dropout_prob _A = max_position_embeddings _A = initializer_range _A = layer_norm_eps _A = position_embedding_type _A = use_cache
350
"""simple docstring""" import logging import os import sys import warnings from dataclasses import dataclass, field from random import randint from typing import Optional import datasets import evaluate import numpy as np from datasets import DatasetDict, load_dataset import transformers from transformers import ( AutoConfig, AutoFeatureExtractor, AutoModelForAudioClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version a_ = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("4.31.0") require_version("datasets>=1.14.0", "To fix: pip install -r examples/pytorch/audio-classification/requirements.txt") def a__ ( __lowercase , __lowercase , __lowercase = 1_6000 ) -> List[str]: _A = int(round(sample_rate * max_length ) ) if len(__lowercase ) <= sample_length: return wav _A = randint(0 , len(__lowercase ) - sample_length - 1 ) return wav[random_offset : random_offset + sample_length] @dataclass class snake_case : __UpperCamelCase = field(default=_UpperCamelCase , metadata={'help': 'Name of a dataset from the datasets package'}) __UpperCamelCase = field( default=_UpperCamelCase , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'}) __UpperCamelCase = field( default=_UpperCamelCase , metadata={'help': 'A file containing the training audio paths and labels.'}) __UpperCamelCase = field( default=_UpperCamelCase , metadata={'help': 'A file containing the validation audio paths and labels.'}) __UpperCamelCase = field( default='train' , metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' } , ) __UpperCamelCase = field( default='validation' , metadata={ 'help': ( 'The name of the training data set split to use (via the datasets library). Defaults to \'validation\'' ) } , ) __UpperCamelCase = field( default='audio' , metadata={'help': 'The name of the dataset column containing the audio data. Defaults to \'audio\''} , ) __UpperCamelCase = field( default='label' , metadata={'help': 'The name of the dataset column containing the labels. Defaults to \'label\''}) __UpperCamelCase = field( default=_UpperCamelCase , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) __UpperCamelCase = field( default=_UpperCamelCase , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of evaluation examples to this ' 'value if set.' ) } , ) __UpperCamelCase = field( default=20 , metadata={'help': 'Audio clips will be randomly cut to this length during training if the value is set.'} , ) @dataclass class snake_case : __UpperCamelCase = field( default='facebook/wav2vec2-base' , metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} , ) __UpperCamelCase = field( default=_UpperCamelCase , metadata={'help': 'Pretrained config name or path if not the same as model_name'}) __UpperCamelCase = field( default=_UpperCamelCase , metadata={'help': 'Where do you want to store the pretrained models downloaded from the Hub'}) __UpperCamelCase = field( default='main' , metadata={'help': 'The specific model version to use (can be a branch name, tag name or commit id).'} , ) __UpperCamelCase = field( default=_UpperCamelCase , metadata={'help': 'Name or path of preprocessor config.'}) __UpperCamelCase = field( default=_UpperCamelCase , metadata={'help': 'Whether to freeze the feature encoder layers of the model.'}) __UpperCamelCase = field( default=_UpperCamelCase , metadata={'help': 'Whether to generate an attention mask in the feature extractor.'}) __UpperCamelCase = field( default=_UpperCamelCase , metadata={ 'help': ( 'Will use the token generated when running `huggingface-cli login` (necessary to use this script ' 'with private models).' ) } , ) __UpperCamelCase = field( default=_UpperCamelCase , metadata={'help': 'Whether to freeze the feature extractor layers of the model.'}) __UpperCamelCase = field( default=_UpperCamelCase , metadata={'help': 'Will enable to load a pretrained model whose head dimensions are different.'} , ) def a_ ( self : List[str] ) -> List[str]: '''simple docstring''' if not self.freeze_feature_extractor and self.freeze_feature_encoder: warnings.warn( "The argument `--freeze_feature_extractor` is deprecated and " "will be removed in a future version. Use `--freeze_feature_encoder`" "instead. Setting `freeze_feature_encoder==True`." , a__ , ) if self.freeze_feature_extractor and not self.freeze_feature_encoder: raise ValueError( "The argument `--freeze_feature_extractor` is deprecated and " "should not be used in combination with `--freeze_feature_encoder`." "Only make use of `--freeze_feature_encoder`." ) def a__ ( ) -> Tuple: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _A = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _A , _A , _A = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _A , _A , _A = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_audio_classification" , __lowercase , __lowercase ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _A = training_args.get_process_log_level() logger.setLevel(__lowercase ) transformers.utils.logging.set_verbosity(__lowercase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu} """ + f"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(f"""Training/evaluation parameters {training_args}""" ) # Set seed before initializing model. set_seed(training_args.seed ) # Detecting last checkpoint. _A = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _A = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. """ "Use --overwrite_output_dir to train from scratch." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Initialize our dataset and prepare it for the audio classification task. _A = DatasetDict() _A = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.train_split_name , use_auth_token=True if model_args.use_auth_token else None , ) _A = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.eval_split_name , use_auth_token=True if model_args.use_auth_token else None , ) if data_args.audio_column_name not in raw_datasets["train"].column_names: raise ValueError( f"""--audio_column_name {data_args.audio_column_name} not found in dataset '{data_args.dataset_name}'. """ "Make sure to set `--audio_column_name` to the correct audio column - one of " f"""{', '.join(raw_datasets['train'].column_names )}.""" ) if data_args.label_column_name not in raw_datasets["train"].column_names: raise ValueError( f"""--label_column_name {data_args.label_column_name} not found in dataset '{data_args.dataset_name}'. """ "Make sure to set `--label_column_name` to the correct text column - one of " f"""{', '.join(raw_datasets['train'].column_names )}.""" ) # Setting `return_attention_mask=True` is the way to get a correctly masked mean-pooling over # transformer outputs in the classifier, but it doesn't always lead to better accuracy _A = AutoFeatureExtractor.from_pretrained( model_args.feature_extractor_name or model_args.model_name_or_path , return_attention_mask=model_args.attention_mask , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # `datasets` takes care of automatically loading and resampling the audio, # so we just need to set the correct target sampling rate. _A = raw_datasets.cast_column( data_args.audio_column_name , datasets.features.Audio(sampling_rate=feature_extractor.sampling_rate ) ) _A = feature_extractor.model_input_names[0] def train_transforms(__lowercase ): _A = [] for audio in batch[data_args.audio_column_name]: _A = random_subsample( audio["array"] , max_length=data_args.max_length_seconds , sample_rate=feature_extractor.sampling_rate ) subsampled_wavs.append(__lowercase ) _A = feature_extractor(__lowercase , sampling_rate=feature_extractor.sampling_rate ) _A = {model_input_name: inputs.get(__lowercase )} _A = list(batch[data_args.label_column_name] ) return output_batch def val_transforms(__lowercase ): _A = [audio["array"] for audio in batch[data_args.audio_column_name]] _A = feature_extractor(__lowercase , sampling_rate=feature_extractor.sampling_rate ) _A = {model_input_name: inputs.get(__lowercase )} _A = list(batch[data_args.label_column_name] ) return output_batch # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. _A = raw_datasets["train"].features[data_args.label_column_name].names _A , _A = {}, {} for i, label in enumerate(__lowercase ): _A = str(__lowercase ) _A = label # Load the accuracy metric from the datasets package _A = evaluate.load("accuracy" ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with # `predictions` and `label_ids` fields) and has to return a dictionary string to float. def compute_metrics(__lowercase ): _A = np.argmax(eval_pred.predictions , axis=1 ) return metric.compute(predictions=__lowercase , references=eval_pred.label_ids ) _A = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(__lowercase ) , labelaid=__lowercase , idalabel=__lowercase , finetuning_task="audio-classification" , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _A = AutoModelForAudioClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=__lowercase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) # freeze the convolutional waveform encoder if model_args.freeze_feature_encoder: model.freeze_feature_encoder() if training_args.do_train: if data_args.max_train_samples is not None: _A = ( raw_datasets["train"].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms raw_datasets["train"].set_transform(__lowercase , output_all_columns=__lowercase ) if training_args.do_eval: if data_args.max_eval_samples is not None: _A = ( raw_datasets["eval"].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms raw_datasets["eval"].set_transform(__lowercase , output_all_columns=__lowercase ) # Initialize our trainer _A = Trainer( model=__lowercase , args=__lowercase , train_dataset=raw_datasets["train"] if training_args.do_train else None , eval_dataset=raw_datasets["eval"] if training_args.do_eval else None , compute_metrics=__lowercase , tokenizer=__lowercase , ) # Training if training_args.do_train: _A = None if training_args.resume_from_checkpoint is not None: _A = training_args.resume_from_checkpoint elif last_checkpoint is not None: _A = last_checkpoint _A = trainer.train(resume_from_checkpoint=__lowercase ) trainer.save_model() trainer.log_metrics("train" , train_result.metrics ) trainer.save_metrics("train" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _A = trainer.evaluate() trainer.log_metrics("eval" , __lowercase ) trainer.save_metrics("eval" , __lowercase ) # Write model card and (optionally) push to hub _A = { "finetuned_from": model_args.model_name_or_path, "tasks": "audio-classification", "dataset": data_args.dataset_name, "tags": ["audio-classification"], } if training_args.push_to_hub: trainer.push_to_hub(**__lowercase ) else: trainer.create_model_card(**__lowercase ) if __name__ == "__main__": main()
163
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _A = { """configuration_bloom""": ["""BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BloomConfig""", """BloomOnnxConfig"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = ["""BloomTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ """BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST""", """BloomForCausalLM""", """BloomModel""", """BloomPreTrainedModel""", """BloomForSequenceClassification""", """BloomForTokenClassification""", """BloomForQuestionAnswering""", ] if TYPE_CHECKING: from .configuration_bloom import BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP, BloomConfig, BloomOnnxConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bloom_fast import BloomTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bloom import ( BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST, BloomForCausalLM, BloomForQuestionAnswering, BloomForSequenceClassification, BloomForTokenClassification, BloomModel, BloomPreTrainedModel, ) else: import sys _A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
242
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''facebook/data2vec-vision-base-ft''': ( '''https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json''' ), } class SCREAMING_SNAKE_CASE__ ( lowercase ): """simple docstring""" a : List[Any] ="data2vec-vision" def __init__( self , snake_case__=768 , snake_case__=12 , snake_case__=12 , snake_case__=3_072 , snake_case__="gelu" , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.02 , snake_case__=1e-12 , snake_case__=224 , snake_case__=16 , snake_case__=3 , snake_case__=False , snake_case__=False , snake_case__=False , snake_case__=False , snake_case__=0.1 , snake_case__=0.1 , snake_case__=True , snake_case__=[3, 5, 7, 11] , snake_case__=[1, 2, 3, 6] , snake_case__=True , snake_case__=0.4 , snake_case__=256 , snake_case__=1 , snake_case__=False , snake_case__=255 , **snake_case__ , ): """simple docstring""" super().__init__(**snake_case__ ) lowerCAmelCase : Tuple = hidden_size lowerCAmelCase : List[Any] = num_hidden_layers lowerCAmelCase : Tuple = num_attention_heads lowerCAmelCase : Optional[int] = intermediate_size lowerCAmelCase : Optional[int] = hidden_act lowerCAmelCase : Dict = hidden_dropout_prob lowerCAmelCase : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase : int = initializer_range lowerCAmelCase : Dict = layer_norm_eps lowerCAmelCase : Optional[int] = image_size lowerCAmelCase : Optional[Any] = patch_size lowerCAmelCase : Optional[Any] = num_channels lowerCAmelCase : Union[str, Any] = use_mask_token lowerCAmelCase : str = use_absolute_position_embeddings lowerCAmelCase : Any = use_relative_position_bias lowerCAmelCase : List[str] = use_shared_relative_position_bias lowerCAmelCase : str = layer_scale_init_value lowerCAmelCase : Union[str, Any] = drop_path_rate lowerCAmelCase : Any = use_mean_pooling # decode head attributes (semantic segmentation) lowerCAmelCase : Optional[int] = out_indices lowerCAmelCase : Union[str, Any] = pool_scales # auxiliary head attributes (semantic segmentation) lowerCAmelCase : str = use_auxiliary_head lowerCAmelCase : int = auxiliary_loss_weight lowerCAmelCase : Tuple = auxiliary_channels lowerCAmelCase : List[str] = auxiliary_num_convs lowerCAmelCase : Tuple = auxiliary_concat_input lowerCAmelCase : List[str] = semantic_loss_ignore_index class SCREAMING_SNAKE_CASE__ ( lowercase ): """simple docstring""" a : Union[str, Any] =version.parse("1.11" ) @property def lowercase__ ( self ): """simple docstring""" return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def lowercase__ ( self ): """simple docstring""" return 1e-4
108
0
class UpperCAmelCase : def __init__(self : int , snake_case__ : int ) -> str: '''simple docstring''' snake_case : Union[str, Any] = n snake_case : Any = [None] * self.n snake_case : List[Any] = 0 # index of the first element snake_case : Tuple = 0 snake_case : Dict = 0 def __len__(self : Dict ) -> int: '''simple docstring''' return self.size def _SCREAMING_SNAKE_CASE (self : List[str] ) -> bool: '''simple docstring''' return self.size == 0 def _SCREAMING_SNAKE_CASE (self : Union[str, Any] ) -> str: '''simple docstring''' return False if self.is_empty() else self.array[self.front] def _SCREAMING_SNAKE_CASE (self : Optional[int] , snake_case__ : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' if self.size >= self.n: raise Exception("QUEUE IS FULL" ) snake_case : Union[str, Any] = data snake_case : Union[str, Any] = (self.rear + 1) % self.n self.size += 1 return self def _SCREAMING_SNAKE_CASE (self : Any ) -> List[str]: '''simple docstring''' if self.size == 0: raise Exception("UNDERFLOW" ) snake_case : int = self.array[self.front] snake_case : Optional[Any] = None snake_case : Any = (self.front + 1) % self.n self.size -= 1 return temp
10
def UpperCamelCase ( __lowerCamelCase : str ): snake_case : Union[str, Any] = 0 # if input_string is "aba" than new_input_string become "a|b|a" snake_case : Tuple = "" snake_case : Optional[int] = "" # append each character + "|" in new_string for range(0, length-1) for i in input_string[: len(__lowerCamelCase ) - 1]: new_input_string += i + "|" # append last character new_input_string += input_string[-1] # we will store the starting and ending of previous furthest ending palindromic # substring snake_case , snake_case : Tuple = 0, 0 # length[i] shows the length of palindromic substring with center i snake_case : Any = [1 for i in range(len(__lowerCamelCase ) )] # for each character in new_string find corresponding palindromic string snake_case : int = 0 for j in range(len(__lowerCamelCase ) ): snake_case : Optional[Any] = 1 if j > r else min(length[l + r - j] // 2 , r - j + 1 ) while ( j - k >= 0 and j + k < len(__lowerCamelCase ) and new_input_string[k + j] == new_input_string[j - k] ): k += 1 snake_case : str = 2 * k - 1 # does this string is ending after the previously explored end (that is r) ? # if yes the update the new r to the last index of this if j + k - 1 > r: snake_case : List[str] = j - k + 1 # noqa: E741 snake_case : Dict = j + k - 1 # update max_length and start position if max_length < length[j]: snake_case : Optional[Any] = length[j] snake_case : int = j # create that string snake_case : Any = new_input_string[start - max_length // 2 : start + max_length // 2 + 1] for i in s: if i != "|": output_string += i return output_string if __name__ == "__main__": import doctest doctest.testmod()
10
1
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('>=', '4.25.0')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, ) else: from .modeling_text_unet import UNetFlatConditionModel from .pipeline_versatile_diffusion import VersatileDiffusionPipeline from .pipeline_versatile_diffusion_dual_guided import VersatileDiffusionDualGuidedPipeline from .pipeline_versatile_diffusion_image_variation import VersatileDiffusionImageVariationPipeline from .pipeline_versatile_diffusion_text_to_image import VersatileDiffusionTextToImagePipeline
129
import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class lowerCamelCase__ ( lowerCamelCase__): '''simple docstring''' snake_case_ ="""Speech2TextFeatureExtractor""" snake_case_ ="""Speech2TextTokenizer""" def __init__(self ,__lowerCamelCase ,__lowerCamelCase ) -> str: """simple docstring""" super().__init__(__lowerCamelCase ,__lowerCamelCase ) lowerCAmelCase__ : int = self.feature_extractor lowerCAmelCase__ : List[str] = False def __call__(self ,*__lowerCamelCase ,**__lowerCamelCase ) -> Dict: """simple docstring""" if self._in_target_context_manager: return self.current_processor(*__lowerCamelCase ,**__lowerCamelCase ) if "raw_speech" in kwargs: warnings.warn('''Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.''' ) lowerCAmelCase__ : Optional[Any] = kwargs.pop('''raw_speech''' ) else: lowerCAmelCase__ : str = kwargs.pop('''audio''' ,__lowerCamelCase ) lowerCAmelCase__ : List[str] = kwargs.pop('''sampling_rate''' ,__lowerCamelCase ) lowerCAmelCase__ : List[str] = kwargs.pop('''text''' ,__lowerCamelCase ) if len(__lowerCamelCase ) > 0: lowerCAmelCase__ : Union[str, Any] = args[0] lowerCAmelCase__ : str = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if audio is not None: lowerCAmelCase__ : str = self.feature_extractor(__lowerCamelCase ,*__lowerCamelCase ,sampling_rate=__lowerCamelCase ,**__lowerCamelCase ) if text is not None: lowerCAmelCase__ : Any = self.tokenizer(__lowerCamelCase ,**__lowerCamelCase ) if text is None: return inputs elif audio is None: return encodings else: lowerCAmelCase__ : str = encodings['''input_ids'''] return inputs def lowerCAmelCase__ (self ,*__lowerCamelCase ,**__lowerCamelCase ) -> List[str]: """simple docstring""" return self.tokenizer.batch_decode(*__lowerCamelCase ,**__lowerCamelCase ) def lowerCAmelCase__ (self ,*__lowerCamelCase ,**__lowerCamelCase ) -> Tuple: """simple docstring""" return self.tokenizer.decode(*__lowerCamelCase ,**__lowerCamelCase ) @contextmanager def lowerCAmelCase__ (self ) -> Any: """simple docstring""" warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your audio inputs, or in a separate call.''' ) lowerCAmelCase__ : int = True lowerCAmelCase__ : Union[str, Any] = self.tokenizer yield lowerCAmelCase__ : List[str] = self.feature_extractor lowerCAmelCase__ : Any = False
129
1
import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class UpperCAmelCase : '''simple docstring''' def __init__( self : int , lowerCAmelCase_ : int , lowerCAmelCase_ : Dict=1_3 , lowerCAmelCase_ : Dict=7 , lowerCAmelCase_ : str=6 , lowerCAmelCase_ : Optional[Any]=1_7 , lowerCAmelCase_ : Union[str, Any]=2_3 , lowerCAmelCase_ : List[Any]=1_1 , lowerCAmelCase_ : List[str]=True , ): """simple docstring""" _A: Any = parent _A: str = batch_size _A: Optional[Any] = seq_length _A: List[str] = act_dim _A: Union[str, Any] = state_dim _A: str = hidden_size _A: List[str] = max_length _A: Tuple = is_training def __magic_name__ ( self : int ): """simple docstring""" _A: str = floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) _A: Any = floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) _A: List[Any] = floats_tensor((self.batch_size, self.seq_length, 1) ) _A: Optional[Any] = floats_tensor((self.batch_size, self.seq_length, 1) ) _A: Dict = ids_tensor((self.batch_size, self.seq_length) , vocab_size=1_0_0_0 ) _A: str = random_attention_mask((self.batch_size, self.seq_length) ) _A: Optional[Any] = self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def __magic_name__ ( self : List[Any] ): """simple docstring""" return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def __magic_name__ ( self : Tuple , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : Dict , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Optional[int] , ): """simple docstring""" _A: str = DecisionTransformerModel(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() _A: Union[str, Any] = model(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def __magic_name__ ( self : List[Any] ): """simple docstring""" _A: Tuple = self.prepare_config_and_inputs() ( ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ): str = config_and_inputs _A: Optional[int] = { '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): '''simple docstring''' __UpperCamelCase : Dict = (DecisionTransformerModel,) if is_torch_available() else () __UpperCamelCase : Dict = () __UpperCamelCase : str = {'''feature-extraction''': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids __UpperCamelCase : Any = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features __UpperCamelCase : Dict = False __UpperCamelCase : Any = False __UpperCamelCase : Tuple = False __UpperCamelCase : Optional[Any] = False __UpperCamelCase : List[Any] = False __UpperCamelCase : List[str] = False __UpperCamelCase : List[str] = False __UpperCamelCase : str = False __UpperCamelCase : str = False def __magic_name__ ( self : str ): """simple docstring""" _A: str = DecisionTransformerModelTester(self ) _A: List[str] = ConfigTester(self , config_class=lowerCAmelCase_ , hidden_size=3_7 ) def __magic_name__ ( self : Optional[Any] ): """simple docstring""" self.config_tester.run_common_tests() def __magic_name__ ( self : Dict ): """simple docstring""" _A: Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase_ ) @slow def __magic_name__ ( self : List[Any] ): """simple docstring""" for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _A: Dict = DecisionTransformerModel.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) def __magic_name__ ( self : int ): """simple docstring""" _A , _A: Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A: Any = model_class(lowerCAmelCase_ ) _A: List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _A: Union[str, Any] = [*signature.parameters.keys()] _A: Tuple = [ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(lowerCAmelCase_ )] , lowerCAmelCase_ ) @require_torch class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def __magic_name__ ( self : int ): """simple docstring""" _A: Any = 2 # number of steps of autoregressive prediction we will perform _A: Tuple = 1_0 # defined by the RL environment, may be normalized _A: List[str] = DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) _A: Dict = model.to(lowerCAmelCase_ ) _A: int = model.config torch.manual_seed(0 ) _A: int = torch.randn(1 , 1 , config.state_dim ).to(device=lowerCAmelCase_ , dtype=torch.floataa ) # env.reset() _A: Optional[Any] = torch.tensor( [[0.242793, -0.28693074, 0.8742613], [0.67815274, -0.08101085, -0.12952147]] , device=lowerCAmelCase_ ) _A: Dict = torch.tensor(lowerCAmelCase_ , device=lowerCAmelCase_ , dtype=torch.floataa ).reshape(1 , 1 , 1 ) _A: Optional[Any] = state _A: Union[str, Any] = torch.zeros(1 , 0 , config.act_dim , device=lowerCAmelCase_ , dtype=torch.floataa ) _A: Any = torch.zeros(1 , 0 , device=lowerCAmelCase_ , dtype=torch.floataa ) _A: Tuple = torch.tensor(0 , device=lowerCAmelCase_ , dtype=torch.long ).reshape(1 , 1 ) for step in range(lowerCAmelCase_ ): _A: str = torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=lowerCAmelCase_ )] , dim=1 ) _A: List[str] = torch.cat([rewards, torch.zeros(1 , 1 , device=lowerCAmelCase_ )] , dim=1 ) _A: int = torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): _A , _A , _A: str = model( states=lowerCAmelCase_ , actions=lowerCAmelCase_ , rewards=lowerCAmelCase_ , returns_to_go=lowerCAmelCase_ , timesteps=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , return_dict=lowerCAmelCase_ , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1e-4 ) ) _A , _A , _A , _A: Optional[Any] = ( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=lowerCAmelCase_ , dtype=torch.floataa ), 1.0, False, {}, ) _A: Any = action_pred[0, -1] _A: List[str] = torch.cat([states, state] , dim=1 ) _A: Dict = returns_to_go[0, -1] - reward _A: int = torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) _A: List[str] = torch.cat( [timesteps, torch.ones((1, 1) , device=lowerCAmelCase_ , dtype=torch.long ) * (step + 1)] , dim=1 )
301
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase__ : Tuple = logging.get_logger(__name__) UpperCAmelCase__ : Union[str, Any] = { 'google/mobilenet_v1_1.0_224': 'https://huggingface.co/google/mobilenet_v1_1.0_224/resolve/main/config.json', 'google/mobilenet_v1_0.75_192': 'https://huggingface.co/google/mobilenet_v1_0.75_192/resolve/main/config.json', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 } class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' __UpperCamelCase : Optional[int] = '''mobilenet_v1''' def __init__( self : Optional[int] , lowerCAmelCase_ : Any=3 , lowerCAmelCase_ : str=2_2_4 , lowerCAmelCase_ : List[str]=1.0 , lowerCAmelCase_ : Any=8 , lowerCAmelCase_ : Tuple="relu6" , lowerCAmelCase_ : List[Any]=True , lowerCAmelCase_ : Optional[int]=0.999 , lowerCAmelCase_ : List[str]=0.02 , lowerCAmelCase_ : List[Any]=0.001 , **lowerCAmelCase_ : Optional[Any] , ): """simple docstring""" super().__init__(**lowerCAmelCase_ ) if depth_multiplier <= 0: raise ValueError('''depth_multiplier must be greater than zero.''' ) _A: Any = num_channels _A: Optional[int] = image_size _A: Optional[Any] = depth_multiplier _A: Tuple = min_depth _A: Any = hidden_act _A: Dict = tf_padding _A: List[Any] = classifier_dropout_prob _A: Tuple = initializer_range _A: Tuple = layer_norm_eps class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' __UpperCamelCase : Dict = version.parse('''1.11''' ) @property def __magic_name__ ( self : Union[str, Any] ): """simple docstring""" return OrderedDict([('''pixel_values''', {0: '''batch'''})] ) @property def __magic_name__ ( self : Optional[Any] ): """simple docstring""" if self.task == "image-classification": return OrderedDict([('''logits''', {0: '''batch'''})] ) else: return OrderedDict([('''last_hidden_state''', {0: '''batch'''}), ('''pooler_output''', {0: '''batch'''})] ) @property def __magic_name__ ( self : Dict ): """simple docstring""" return 1e-4
301
1
import requests def snake_case( __magic_name__ , __magic_name__ ) -> None: '''simple docstring''' lowercase : str = {'''Content-Type''': '''application/json'''} lowercase : Union[str, Any] = requests.post(__magic_name__ , json={'''text''': message_body} , headers=__magic_name__ ) if response.status_code != 2_00: lowercase : int = ( '''Request to slack returned an error ''' F"""{response.status_code}, the response is:\n{response.text}""" ) raise ValueError(__magic_name__ ) if __name__ == "__main__": # Set the slack url to the one provided by Slack when you create the webhook at # https://my.slack.com/services/new/incoming-webhook/ send_slack_message('<YOUR MESSAGE BODY>', '<SLACK CHANNEL URL>')
308
import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = {'vocab_file': 'vocab.txt'} lowerCAmelCase_ = { 'vocab_file': { 'openbmb/cpm-ant-10b': 'https://huggingface.co/openbmb/cpm-ant-10b/blob/main/vocab.txt', }, } lowerCAmelCase_ = { 'openbmb/cpm-ant-10b': 10_24, } def snake_case( __magic_name__ ) -> int: '''simple docstring''' lowercase : Optional[int] = collections.OrderedDict() with open(__magic_name__ , '''r''' , encoding='''utf-8''' ) as reader: lowercase : str = reader.readlines() for index, token in enumerate(__magic_name__ ): lowercase : Union[str, Any] = token.rstrip('''\n''' ) lowercase : List[Any] = index return vocab class _A ( _lowerCamelCase ): def __init__( self : List[str] , _A : Any , _A : List[str]="<unk>" , _A : Union[str, Any]=200 ) -> List[Any]: """simple docstring""" lowercase : Optional[int] = vocab lowercase : List[str] = unk_token lowercase : Any = max_input_chars_per_word def __a ( self : List[str] , _A : Tuple ) -> str: """simple docstring""" lowercase : Dict = list(_A ) if len(_A ) > self.max_input_chars_per_word: return [self.unk_token] lowercase : int = 0 lowercase : Dict = [] while start < len(_A ): lowercase : Optional[Any] = len(_A ) lowercase : List[str] = None while start < end: lowercase : List[Any] = ''''''.join(chars[start:end] ) if substr in self.vocab: lowercase : Union[str, Any] = substr break end -= 1 if cur_substr is None: sub_tokens.append(self.unk_token ) start += 1 else: sub_tokens.append(_A ) lowercase : Dict = end return sub_tokens class _A ( _lowerCamelCase ): _UpperCamelCase : List[str] = VOCAB_FILES_NAMES _UpperCamelCase : Optional[int] = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCamelCase : List[Any] = ['''input_ids''', '''attention_mask'''] _UpperCamelCase : int = False def __init__( self : List[str] , _A : int , _A : Optional[Any]="<d>" , _A : Any="</d>" , _A : Optional[Any]="<s>" , _A : Any="</s>" , _A : Any="<pad>" , _A : List[Any]="<unk>" , _A : Optional[Any]="</n>" , _A : List[str]="</_>" , _A : Optional[Any]="left" , **_A : str , ) -> Tuple: """simple docstring""" requires_backends(self , ['''jieba'''] ) super().__init__( bod_token=_A , eod_token=_A , bos_token=_A , eos_token=_A , pad_token=_A , unk_token=_A , line_token=_A , space_token=_A , padding_side=_A , **_A , ) lowercase : str = bod_token lowercase : str = eod_token lowercase : Any = load_vocab(_A ) lowercase : List[Any] = self.encoder[space_token] lowercase : Tuple = self.encoder[line_token] del self.encoder[space_token] del self.encoder[line_token] lowercase : Any = collections.OrderedDict(sorted(self.encoder.items() , key=lambda _A : x[1] ) ) lowercase : int = {v: k for k, v in self.encoder.items()} lowercase : Optional[Any] = WordpieceTokenizer(vocab=self.encoder , unk_token=self.unk_token ) @property def __a ( self : Dict ) -> Optional[int]: """simple docstring""" return self.encoder[self.bod_token] @property def __a ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" return self.encoder[self.eod_token] @property def __a ( self : List[str] ) -> List[str]: """simple docstring""" return self.encoder["\n"] @property def __a ( self : List[Any] ) -> int: """simple docstring""" return len(self.encoder ) def __a ( self : Union[str, Any] ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def __a ( self : str , _A : List[str] ) -> Tuple: """simple docstring""" lowercase : int = [] for x in jieba.cut(_A , cut_all=_A ): output_tokens.extend(self.wordpiece_tokenizer.tokenize(_A ) ) return output_tokens def __a ( self : List[Any] , _A : Tuple , **_A : Optional[int] ) -> Any: """simple docstring""" lowercase : List[str] = [i for i in token_ids if i >= 0] lowercase : Any = [ x for x in token_ids if x != self.pad_token_id and x != self.eos_token_id and x != self.bos_token_id ] return super()._decode(_A , **_A ) def __a ( self : List[Any] , _A : int ) -> Optional[Any]: """simple docstring""" return token in self.encoder def __a ( self : Dict , _A : List[str] ) -> str: """simple docstring""" return "".join(_A ) def __a ( self : List[str] , _A : List[str] ) -> Any: """simple docstring""" return self.encoder.get(_A , self.encoder.get(self.unk_token ) ) def __a ( self : Tuple , _A : Union[str, Any] ) -> Tuple: """simple docstring""" return self.decoder.get(_A , self.unk_token ) def __a ( self : List[Any] , _A : str , _A : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if os.path.isdir(_A ): lowercase : str = os.path.join( _A , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) else: lowercase : Optional[int] = (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory lowercase : Any = 0 if " " in self.encoder: lowercase : List[Any] = self.encoder[''' '''] del self.encoder[" "] if "\n" in self.encoder: lowercase : Dict = self.encoder['''\n'''] del self.encoder["\n"] lowercase : Union[str, Any] = collections.OrderedDict(sorted(self.encoder.items() , key=lambda _A : x[1] ) ) with open(_A , '''w''' , encoding='''utf-8''' ) as writer: for token, token_index in self.encoder.items(): if index != token_index: logger.warning( f"""Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.""" ''' Please check that the vocabulary is not corrupted!''' ) lowercase : Any = token_index writer.write(token + '''\n''' ) index += 1 return (vocab_file,) def __a ( self : str , _A : List[int] , _A : List[int] = None ) -> List[int]: """simple docstring""" if token_ids_a is None: return [self.bos_token_id] + token_ids_a return [self.bos_token_id] + token_ids_a + [self.bos_token_id] + token_ids_a def __a ( self : int , _A : List[int] , _A : Optional[List[int]] = None , _A : bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_A , token_ids_a=_A , already_has_special_tokens=_A ) if token_ids_a is not None: return [1] + ([0] * len(_A )) + [1] + ([0] * len(_A )) return [1] + ([0] * len(_A ))
308
1
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import StableDiffusionKDiffusionPipeline from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() @slow @require_torch_gpu class lowercase ( unittest.TestCase ): """simple docstring""" def _snake_case ( self ) -> Union[str, Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _snake_case ( self ) -> List[str]: _UpperCAmelCase : Union[str, Any] = StableDiffusionKDiffusionPipeline.from_pretrained("""CompVis/stable-diffusion-v1-4""" ) _UpperCAmelCase : Optional[int] = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) sd_pipe.set_scheduler("""sample_euler""" ) _UpperCAmelCase : Union[str, Any] = """A painting of a squirrel eating a burger""" _UpperCAmelCase : Tuple = torch.manual_seed(0 ) _UpperCAmelCase : Optional[Any] = sd_pipe([prompt] ,generator=a_ ,guidance_scale=9.0 ,num_inference_steps=20 ,output_type="""np""" ) _UpperCAmelCase : int = output.images _UpperCAmelCase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCAmelCase : int = np.array([0.0447, 0.0492, 0.0468, 0.0408, 0.0383, 0.0408, 0.0354, 0.0380, 0.0339] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _snake_case ( self ) -> Optional[int]: _UpperCAmelCase : str = StableDiffusionKDiffusionPipeline.from_pretrained("""stabilityai/stable-diffusion-2-1-base""" ) _UpperCAmelCase : Tuple = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) sd_pipe.set_scheduler("""sample_euler""" ) _UpperCAmelCase : Optional[Any] = """A painting of a squirrel eating a burger""" _UpperCAmelCase : Any = torch.manual_seed(0 ) _UpperCAmelCase : List[str] = sd_pipe([prompt] ,generator=a_ ,guidance_scale=9.0 ,num_inference_steps=20 ,output_type="""np""" ) _UpperCAmelCase : Optional[Any] = output.images _UpperCAmelCase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCAmelCase : Optional[int] = np.array([0.1237, 0.1320, 0.1438, 0.1359, 0.1390, 0.1132, 0.1277, 0.1175, 0.1112] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-1 def _snake_case ( self ) -> Optional[int]: _UpperCAmelCase : Optional[Any] = StableDiffusionKDiffusionPipeline.from_pretrained("""stabilityai/stable-diffusion-2-1-base""" ) _UpperCAmelCase : List[Any] = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) sd_pipe.set_scheduler("""sample_dpmpp_2m""" ) _UpperCAmelCase : int = """A painting of a squirrel eating a burger""" _UpperCAmelCase : Union[str, Any] = torch.manual_seed(0 ) _UpperCAmelCase : Any = sd_pipe( [prompt] ,generator=a_ ,guidance_scale=7.5 ,num_inference_steps=15 ,output_type="""np""" ,use_karras_sigmas=a_ ,) _UpperCAmelCase : int = output.images _UpperCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) _UpperCAmelCase : Dict = np.array( [0.1138_1689, 0.1211_2921, 0.138_9457, 0.1254_9606, 0.124_4964, 0.1083_1517, 0.1156_2866, 0.1086_7816, 0.1049_9048] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
360
'''simple docstring''' from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class lowercase ( _lowerCamelCase ): """simple docstring""" def __init__( self ,a_ ,a_ = None ,a_ = None ,a_ = True ,a_ = None ,a_ = False ,a_ = None ,a_ = True ,a_ = "arrow" ,**a_ ,) -> str: super().__init__( split=a_ ,features=a_ ,cache_dir=a_ ,keep_in_memory=a_ ,streaming=a_ ,**a_ ,) _UpperCAmelCase : Any = load_from_cache_file _UpperCAmelCase : Optional[int] = file_format _UpperCAmelCase : int = Spark( df=a_ ,features=a_ ,cache_dir=a_ ,working_dir=a_ ,**a_ ,) def _snake_case ( self ) -> int: if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) _UpperCAmelCase : str = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=a_ ,file_format=self._file_format ,) return self.builder.as_dataset(split=self.split )
349
0
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union _a : List[str]= TypeVar("T") _a : int= Union[List[T], Tuple[T, ...]] _a : Union[str, Any]= Union[T, List[T], Dict[str, T]] _a : int= Union[str, bytes, os.PathLike]
172
"""simple docstring""" import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline _a : int= datasets.utils.logging.get_logger(__name__) @dataclass class UpperCamelCase ( datasets.BuilderConfig ): UpperCAmelCase : Optional[datasets.Features] = None UpperCAmelCase : str = "utf-8" UpperCAmelCase : Optional[str] = None UpperCAmelCase : Optional[str] = None UpperCAmelCase : bool = True # deprecated UpperCAmelCase : Optional[int] = None # deprecated UpperCAmelCase : int = 10 << 20 # 10MB UpperCAmelCase : Optional[bool] = None class UpperCamelCase ( datasets.ArrowBasedBuilder ): UpperCAmelCase : int = JsonConfig def _lowercase (self : int) -> List[str]: if self.config.block_size is not None: logger.warning('The JSON loader parameter `block_size` is deprecated. Please use `chunksize` instead') __snake_case : Any = self.config.block_size if self.config.use_threads is not True: logger.warning( 'The JSON loader parameter `use_threads` is deprecated and doesn\'t have any effect anymore.') if self.config.newlines_in_values is not None: raise ValueError('The JSON loader parameter `newlines_in_values` is no longer supported') return datasets.DatasetInfo(features=self.config.features) def _lowercase (self : Dict , _A : Any) -> Optional[Any]: if not self.config.data_files: raise ValueError(f"At least one data file must be specified, but got data_files={self.config.data_files}") __snake_case : Dict = dl_manager.download_and_extract(self.config.data_files) if isinstance(_A , (str, list, tuple)): __snake_case : str = data_files if isinstance(_A , _A): __snake_case : int = [files] __snake_case : Tuple = [dl_manager.iter_files(_A) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'files': files})] __snake_case : str = [] for split_name, files in data_files.items(): if isinstance(_A , _A): __snake_case : Optional[int] = [files] __snake_case : int = [dl_manager.iter_files(_A) for file in files] splits.append(datasets.SplitGenerator(name=_A , gen_kwargs={'files': files})) return splits def _lowercase (self : Optional[Any] , _A : pa.Table) -> pa.Table: if self.config.features is not None: # adding missing columns for column_name in set(self.config.features) - set(pa_table.column_names): __snake_case : List[Any] = self.config.features.arrow_schema.field(_A).type __snake_case : Any = pa_table.append_column(_A , pa.array([None] * len(_A) , type=_A)) # more expensive cast to support nested structures with keys in a different order # allows str <-> int/float or str to Audio for example __snake_case : List[str] = table_cast(_A , self.config.features.arrow_schema) return pa_table def _lowercase (self : Dict , _A : Any) -> Union[str, Any]: for file_idx, file in enumerate(itertools.chain.from_iterable(_A)): # If the file is one json object and if we need to look at the list of items in one specific field if self.config.field is not None: with open(_A , encoding=self.config.encoding , errors=self.config.encoding_errors) as f: __snake_case : Tuple = json.load(_A) # We keep only the field we are interested in __snake_case : List[str] = dataset[self.config.field] # We accept two format: a list of dicts or a dict of lists if isinstance(_A , (list, tuple)): __snake_case : Optional[int] = set().union(*[row.keys() for row in dataset]) __snake_case : List[str] = {col: [row.get(_A) for row in dataset] for col in keys} else: __snake_case : Optional[int] = dataset __snake_case : Tuple = pa.Table.from_pydict(_A) yield file_idx, self._cast_table(_A) # If the file has one json object per line else: with open(_A , 'rb') as f: __snake_case : int = 0 # Use block_size equal to the chunk size divided by 32 to leverage multithreading # Set a default minimum value of 16kB if the chunk size is really small __snake_case : Tuple = max(self.config.chunksize // 32 , 16 << 10) __snake_case : str = ( self.config.encoding_errors if self.config.encoding_errors is not None else 'strict' ) while True: __snake_case : Union[str, Any] = f.read(self.config.chunksize) if not batch: break # Finish current line try: batch += f.readline() except (AttributeError, io.UnsupportedOperation): batch += readline(_A) # PyArrow only accepts utf-8 encoded bytes if self.config.encoding != "utf-8": __snake_case : int = batch.decode(self.config.encoding , errors=_A).encode('utf-8') try: while True: try: __snake_case : Tuple = paj.read_json( io.BytesIO(_A) , read_options=paj.ReadOptions(block_size=_A)) break except (pa.ArrowInvalid, pa.ArrowNotImplementedError) as e: if ( isinstance(_A , pa.ArrowInvalid) and "straddling" not in str(_A) or block_size > len(_A) ): raise else: # Increase the block size in case it was too small. # The block size will be reset for the next file. logger.debug( f"Batch of {len(_A)} bytes couldn't be parsed with block_size={block_size}. Retrying with block_size={block_size * 2}.") block_size *= 2 except pa.ArrowInvalid as e: try: with open( _A , encoding=self.config.encoding , errors=self.config.encoding_errors) as f: __snake_case : List[Any] = json.load(_A) except json.JSONDecodeError: logger.error(f"Failed to read file '{file}' with error {type(_A)}: {e}") raise e # If possible, parse the file as a list of json objects and exit the loop if isinstance(_A , _A): # list is the only sequence type supported in JSON try: __snake_case : List[str] = set().union(*[row.keys() for row in dataset]) __snake_case : List[str] = {col: [row.get(_A) for row in dataset] for col in keys} __snake_case : List[str] = pa.Table.from_pydict(_A) except (pa.ArrowInvalid, AttributeError) as e: logger.error(f"Failed to read file '{file}' with error {type(_A)}: {e}") raise ValueError(f"Not able to read records in the JSON file at {file}.") from None yield file_idx, self._cast_table(_A) break else: logger.error(f"Failed to read file '{file}' with error {type(_A)}: {e}") raise ValueError( f"Not able to read records in the JSON file at {file}. " f"You should probably indicate the field of the JSON file containing your records. " f"This JSON file contain the following fields: {str(list(dataset.keys()))}. " f"Select the correct one and provide it as `field='XXX'` to the dataset loading method. ") from None # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(_A) batch_idx += 1
172
1
"""simple docstring""" import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets a :Union[str, Any] = "\\n@inproceedings{pillutla-etal:mauve:neurips2021,\n title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers},\n author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid},\n booktitle = {NeurIPS},\n year = {2021}\n}\n\n" a :List[Any] = "\\nMAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure.\n\nMAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences.\n\nFor details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021).\n\nThis metrics is a wrapper around the official implementation of MAUVE:\nhttps://github.com/krishnap25/mauve\n" a :str = "\nCalculates MAUVE scores between two lists of generated text and reference text.\nArgs:\n predictions: list of generated text to score. Each predictions\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\nOptional Args:\n num_buckets: the size of the histogram to quantize P and Q. Options: 'auto' (default) or an integer\n pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1\n kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9\n kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5\n kmeans_max_iter: maximum number of k-means iterations. Default 500\n featurize_model_name: name of the model from which features are obtained. Default 'gpt2-large' Use one of ['gpt2', 'gpt2-medium', 'gpt2-large', 'gpt2-xl'].\n device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU\n max_text_length: maximum number of tokens to consider. Default 1024\n divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25\n mauve_scaling_factor: \"c\" from the paper. Default 5.\n verbose: If True (default), print running time updates\n seed: random seed to initialize k-means cluster assignments.\nReturns:\n mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer,\n frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer,\n divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve,\n p_hist: a discrete distribution, which is a quantized version of the text distribution p_text,\n q_hist: same as above, but with q_text.\nExamples:\n\n >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest\n >>> import datasets\n >>> mauve = datasets.load_metric('mauve')\n >>> predictions = [\"hello there\", \"general kenobi\"]\n >>> references = [\"hello there\", \"general kenobi\"]\n >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP\n >>> print(out.mauve) # doctest: +SKIP\n 1.0\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class __a (datasets.Metric): '''simple docstring''' def _a ( self ) -> List[Any]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="""https://github.com/krishnap25/mauve""" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/krishnap25/mauve"""] , reference_urls=[ """https://arxiv.org/abs/2102.01454""", """https://github.com/krishnap25/mauve""", ] , ) def _a ( self , _a , _a , _a=None , _a=None , _a=None , _a=None , _a="auto" , _a=-1 , _a=0.9 , _a=5 , _a=500 , _a="gpt2-large" , _a=-1 , _a=1_024 , _a=25 , _a=5 , _a=True , _a=25 , ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = compute_mauve( p_text=_a , q_text=_a , p_features=_a , q_features=_a , p_tokens=_a , q_tokens=_a , num_buckets=_a , pca_max_data=_a , kmeans_explained_var=_a , kmeans_num_redo=_a , kmeans_max_iter=_a , featurize_model_name=_a , device_id=_a , max_text_length=_a , divergence_curve_discretization_size=_a , mauve_scaling_factor=_a , verbose=_a , seed=_a , ) return out
56
"""simple docstring""" from math import loga def _lowercase ( __lowerCAmelCase ) -> int: if a < 0: raise ValueError("""Input value must be a positive integer""" ) elif isinstance(__lowerCAmelCase , __lowerCAmelCase ): raise TypeError("""Input value must be a 'int' type""" ) return 0 if (a == 0) else int(loga(a & -a ) ) if __name__ == "__main__": import doctest doctest.testmod()
56
1
"""simple docstring""" import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class _UpperCAmelCase ( __snake_case, __snake_case, __snake_case ): '''simple docstring''' @register_to_config def __init__(self , a_ , a_ , a_ , a_ , a_ , a_ , a_ , a_ , a_ , a_ = False , ): '''simple docstring''' super().__init__() __snake_case : Optional[int] = nn.Embedding(a_ , a_ ) __snake_case : Union[str, Any] = nn.Embedding(a_ , a_ ) __snake_case : List[str] = False __snake_case : List[Any] = nn.Dropout(p=a_ ) __snake_case : Tuple = TaConfig( vocab_size=a_ , d_model=a_ , num_heads=a_ , d_kv=a_ , d_ff=a_ , dropout_rate=a_ , feed_forward_proj=a_ , is_decoder=a_ , is_encoder_decoder=a_ , ) __snake_case : Union[str, Any] = nn.ModuleList() for lyr_num in range(a_ ): __snake_case : Tuple = TaBlock(a_ ) self.encoders.append(a_ ) __snake_case : Optional[int] = TaLayerNorm(a_ ) __snake_case : List[str] = nn.Dropout(p=a_ ) def SCREAMING_SNAKE_CASE (self , a_ , a_ ): '''simple docstring''' __snake_case : Optional[Any] = self.token_embedder(a_ ) __snake_case : Any = encoder_input_tokens.shape[1] __snake_case : str = torch.arange(a_ , device=encoder_input_tokens.device ) x += self.position_encoding(a_ ) __snake_case : Any = self.dropout_pre(a_ ) # inverted the attention mask __snake_case : Dict = encoder_input_tokens.size() __snake_case : Optional[int] = self.get_extended_attention_mask(a_ , a_ ) for lyr in self.encoders: __snake_case : Any = lyr(a_ , a_ )[0] __snake_case : Dict = self.layer_norm(a_ ) return self.dropout_post(a_ ), encoder_inputs_mask
102
"""simple docstring""" import numpy as np from PIL import Image def lowercase_ ( _lowerCamelCase: np.ndarray , _lowerCamelCase: int , _lowerCamelCase: int ) -> np.ndarray: '''simple docstring''' __lowerCamelCase : Dict = np.array(_lowerCamelCase ) if arr.shape[0] != arr.shape[1]: raise ValueError("The input array is not a square matrix" ) __lowerCamelCase : Optional[int] = 0 __lowerCamelCase : Optional[int] = 0 __lowerCamelCase : List[Any] = 0 __lowerCamelCase : Union[str, Any] = 0 # compute the shape of the output matrix __lowerCamelCase : Any = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape maxpool_shape __lowerCamelCase : Dict = np.zeros((maxpool_shape, maxpool_shape) ) while i < arr.shape[0]: if i + size > arr.shape[0]: # if the end of the matrix is reached, break break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the maximum of the pooling matrix __lowerCamelCase : int = np.max(arr[i : i + size, j : j + size] ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __lowerCamelCase : Any = 0 __lowerCamelCase : Optional[int] = 0 return updated_arr def lowercase_ ( _lowerCamelCase: np.ndarray , _lowerCamelCase: int , _lowerCamelCase: int ) -> np.ndarray: '''simple docstring''' __lowerCamelCase : Union[str, Any] = np.array(_lowerCamelCase ) if arr.shape[0] != arr.shape[1]: raise ValueError("The input array is not a square matrix" ) __lowerCamelCase : List[Any] = 0 __lowerCamelCase : Tuple = 0 __lowerCamelCase : Optional[int] = 0 __lowerCamelCase : str = 0 # compute the shape of the output matrix __lowerCamelCase : List[Any] = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape avgpool_shape __lowerCamelCase : List[str] = np.zeros((avgpool_shape, avgpool_shape) ) while i < arr.shape[0]: # if the end of the matrix is reached, break if i + size > arr.shape[0]: break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the average of the pooling matrix __lowerCamelCase : Optional[Any] = int(np.average(arr[i : i + size, j : j + size] ) ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __lowerCamelCase : List[Any] = 0 __lowerCamelCase : Tuple = 0 return updated_arr # Main Function if __name__ == "__main__": from doctest import testmod testmod(name='''avgpooling''', verbose=True) # Loading the image __A = Image.open('''path_to_image''') # Converting the image to numpy array and maxpooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(maxpooling(np.array(image), size=3, stride=2)).show() # Converting the image to numpy array and averagepooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(avgpooling(np.array(image), size=3, stride=2)).show()
135
0
import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatchSampler, SkipDataLoader, skip_first_batches, ) class lowerCamelCase_ ( _A ): '''simple docstring''' def __init__( self : Optional[int] , __lowerCamelCase : int=0.01 , __lowerCamelCase : Tuple=10_00 ) -> Optional[int]: A : Union[str, Any] = p_stop A : Optional[Any] = max_length def __iter__( self : Union[str, Any] ) -> List[str]: A : int = 0 A : List[str] = False while not stop and count < self.max_length: yield count count += 1 A : List[Any] = random.random() < self.p_stop class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , __lowerCamelCase : List[str] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : Optional[Any]=True ) -> List[Any]: A : str = [ BatchSamplerShard(__lowerCamelCase , 2 , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) for i in range(2 ) ] A : str = [list(__lowerCamelCase ) for batch_sampler_shard in batch_sampler_shards] if not split_batches: self.assertListEqual([len(__lowerCamelCase ) for shard in batch_sampler_shards] , [len(__lowerCamelCase ) for e in expected] ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> Optional[Any]: # Check the shards when the dataset is a round multiple of total batch size. A : Dict = BatchSampler(range(24 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : Any = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) A : Tuple = BatchSampler(range(24 ) , batch_size=3 , drop_last=__lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. A : Tuple = BatchSampler(range(21 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : Optional[int] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [0, 1, 2]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) A : List[Any] = BatchSampler(range(21 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : Any = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. A : Optional[int] = BatchSampler(range(22 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : Union[str, Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 0, 1]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) A : Optional[int] = BatchSampler(range(22 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : int = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. A : List[Any] = BatchSampler(range(20 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : Dict = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 0]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [1, 2, 3]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) A : str = BatchSampler(range(20 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : Optional[Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) # Check the shards when the dataset is very small. A : Union[str, Any] = BatchSampler(range(2 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : str = [[[0, 1, 0]], [[1, 0, 1]]] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) A : Dict = BatchSampler(range(2 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : Optional[int] = [[], []] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> str: # Check the shards when the dataset is a round multiple of batch size. A : Tuple = BatchSampler(range(24 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : str = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) A : Tuple = BatchSampler(range(24 ) , batch_size=4 , drop_last=__lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size. A : Optional[Any] = BatchSampler(range(22 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : int = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [0, 1]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) A : Optional[Any] = BatchSampler(range(22 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : Any = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. A : Any = BatchSampler(range(21 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : str = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 0]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [1, 2]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) A : List[str] = BatchSampler(range(21 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : Optional[int] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) # Check the shards when the dataset is very small. A : List[str] = BatchSampler(range(2 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : List[Any] = [[[0, 1]], [[0, 1]]] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) A : Optional[int] = BatchSampler(range(2 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : List[Any] = [[], []] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self : Any ) -> Any: # Check the shards when the dataset is a round multiple of total batch size. A : Tuple = BatchSampler(range(24 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : Tuple = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) A : List[Any] = BatchSampler(range(24 ) , batch_size=3 , drop_last=__lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. A : List[str] = BatchSampler(range(21 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : Any = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) A : Union[str, Any] = BatchSampler(range(21 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : str = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. A : str = BatchSampler(range(22 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : Union[str, Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) A : Dict = BatchSampler(range(22 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : Tuple = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. A : Dict = BatchSampler(range(20 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : int = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) A : List[str] = BatchSampler(range(20 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : Dict = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) # Check the shards when the dataset is very small. A : Dict = BatchSampler(range(2 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : Optional[int] = [[[0, 1]], []] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) A : str = BatchSampler(range(2 ) , batch_size=3 , drop_last=__lowerCamelCase ) A : Optional[int] = [[], []] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , even_batches=__lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> Any: # Check the shards when the dataset is a round multiple of batch size. A : List[Any] = BatchSampler(range(24 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : int = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) A : List[Any] = BatchSampler(range(24 ) , batch_size=4 , drop_last=__lowerCamelCase ) # Expected shouldn't change self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size. A : List[str] = BatchSampler(range(22 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : List[Any] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) A : Any = BatchSampler(range(22 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : Optional[Any] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. A : Tuple = BatchSampler(range(21 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : int = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) A : Optional[int] = BatchSampler(range(21 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : List[Any] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) # Check the shards when the dataset is very small. A : Dict = BatchSampler(range(2 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : Tuple = [[[0, 1]], []] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) A : int = BatchSampler(range(2 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : int = [[], []] self.check_batch_sampler_shards(__lowerCamelCase , __lowerCamelCase , split_batches=__lowerCamelCase , even_batches=__lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ) -> Optional[int]: A : Optional[int] = [[0, 1, 2], [3, 4], [5, 6, 7, 8], [9, 10, 11], [12, 13]] A : Optional[Any] = [BatchSamplerShard(__lowerCamelCase , 2 , __lowerCamelCase , even_batches=__lowerCamelCase ) for i in range(2 )] self.assertEqual(len(batch_sampler_shards[0] ) , 3 ) self.assertEqual(len(batch_sampler_shards[1] ) , 2 ) self.assertListEqual(list(batch_sampler_shards[0] ) , [[0, 1, 2], [5, 6, 7, 8], [12, 13]] ) self.assertListEqual(list(batch_sampler_shards[1] ) , [[3, 4], [9, 10, 11]] ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , __lowerCamelCase : List[str] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : int , __lowerCamelCase : Dict=False , __lowerCamelCase : List[str]=2 , __lowerCamelCase : Union[str, Any]=False ) -> Optional[Any]: random.seed(__lowerCamelCase ) A : Tuple = list(__lowerCamelCase ) A : Any = [ IterableDatasetShard( __lowerCamelCase , batch_size=__lowerCamelCase , drop_last=__lowerCamelCase , num_processes=__lowerCamelCase , process_index=__lowerCamelCase , split_batches=__lowerCamelCase , ) for i in range(__lowerCamelCase ) ] A : Optional[Any] = [] for iterable_dataset_shard in iterable_dataset_shards: # Since our random iterable dataset will be... random... we need to use a seed to get reproducible results. random.seed(__lowerCamelCase ) iterable_dataset_lists.append(list(__lowerCamelCase ) ) A : Optional[int] = batch_size // num_processes if split_batches else batch_size # All iterable dataset shard should have the same length, a round multiple of shard_batch_size A : Optional[int] = iterable_dataset_lists[0] for l in iterable_dataset_lists[1:]: self.assertEqual(len(__lowerCamelCase ) , len(__lowerCamelCase ) ) self.assertTrue(len(__lowerCamelCase ) % shard_batch_size == 0 ) A : str = [] for idx in range(0 , len(__lowerCamelCase ) , __lowerCamelCase ): for l in iterable_dataset_lists: observed += l[idx : idx + shard_batch_size] if not drop_last: while len(__lowerCamelCase ) < len(__lowerCamelCase ): reference += reference self.assertListEqual(__lowerCamelCase , reference[: len(__lowerCamelCase )] ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> Dict: A : Optional[int] = 42 A : Union[str, Any] = RandomIterableDataset() self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) # Edge case with a very small dataset A : List[Any] = RandomIterableDataset(max_length=2 ) self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) self.check_iterable_dataset_shards(__lowerCamelCase , __lowerCamelCase , batch_size=4 , drop_last=__lowerCamelCase , split_batches=__lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self : str ) -> int: A : Tuple = BatchSampler(range(16 ) , batch_size=4 , drop_last=__lowerCamelCase ) A : List[str] = SkipBatchSampler(__lowerCamelCase , 2 ) self.assertListEqual(list(__lowerCamelCase ) , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> Tuple: A : Tuple = SkipDataLoader(list(range(16 ) ) , batch_size=4 , skip_batches=2 ) self.assertListEqual([t.tolist() for t in dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def SCREAMING_SNAKE_CASE__ ( self : Dict ) -> List[Any]: A : int = DataLoader(list(range(16 ) ) , batch_size=4 ) A : Dict = skip_first_batches(__lowerCamelCase , num_batches=2 ) self.assertListEqual([t.tolist() for t in new_dataloader] , [[8, 9, 10, 11], [12, 13, 14, 15]] ) def SCREAMING_SNAKE_CASE__ ( self : Any ) -> Union[str, Any]: A : List[str] = DataLoaderShard(list(range(16 ) ) , batch_size=4 ) for idx, _ in enumerate(__lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(__lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> Union[str, Any]: Accelerator() A : int = DataLoaderDispatcher(range(16 ) , batch_size=4 ) for idx, _ in enumerate(__lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(__lowerCamelCase ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 )
256
import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging __SCREAMING_SNAKE_CASE = logging.get_logger(__name__) def UpperCAmelCase ( _lowerCamelCase , _lowerCamelCase ): A : Any = set() A : int = [] def parse_line(_lowerCamelCase ): for line in fp: if isinstance(_lowerCamelCase , _lowerCamelCase ): A : Any = line.decode("UTF-8" ) if "warnings summary (final)" in line: continue # This means we are outside the body of a warning elif not line.startswith(" " ): # process a single warning and move it to `selected_warnings`. if len(_lowerCamelCase ) > 0: A : Union[str, Any] = "\n".join(_lowerCamelCase ) # Only keep the warnings specified in `targets` if any(f""": {x}: """ in warning for x in targets ): selected_warnings.add(_lowerCamelCase ) buffer.clear() continue else: A : Union[str, Any] = line.strip() buffer.append(_lowerCamelCase ) if from_gh: for filename in os.listdir(_lowerCamelCase ): A : Tuple = os.path.join(_lowerCamelCase , _lowerCamelCase ) if not os.path.isdir(_lowerCamelCase ): # read the file if filename != "warnings.txt": continue with open(_lowerCamelCase ) as fp: parse_line(_lowerCamelCase ) else: try: with zipfile.ZipFile(_lowerCamelCase ) as z: for filename in z.namelist(): if not os.path.isdir(_lowerCamelCase ): # read the file if filename != "warnings.txt": continue with z.open(_lowerCamelCase ) as fp: parse_line(_lowerCamelCase ) except Exception: logger.warning( f"""{artifact_path} is either an invalid zip file or something else wrong. This file is skipped.""" ) return selected_warnings def UpperCAmelCase ( _lowerCamelCase , _lowerCamelCase ): A : Tuple = set() A : Union[str, Any] = [os.path.join(_lowerCamelCase , _lowerCamelCase ) for p in os.listdir(_lowerCamelCase ) if (p.endswith(".zip" ) or from_gh)] for p in paths: selected_warnings.update(extract_warnings_from_single_artifact(_lowerCamelCase , _lowerCamelCase ) ) return selected_warnings if __name__ == "__main__": def UpperCAmelCase ( _lowerCamelCase ): return values.split("," ) __SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument("""--workflow_run_id""", type=str, required=True, help="""A GitHub Actions workflow run id.""") parser.add_argument( """--output_dir""", type=str, required=True, help="""Where to store the downloaded artifacts and other result files.""", ) parser.add_argument("""--token""", default=None, type=str, help="""A token that has actions:read permission.""") # optional parameters parser.add_argument( """--targets""", default="""DeprecationWarning,UserWarning,FutureWarning""", type=list_str, help="""Comma-separated list of target warning(s) which we want to extract.""", ) parser.add_argument( """--from_gh""", action="""store_true""", help="""If running from a GitHub action workflow and collecting warnings from its artifacts.""", ) __SCREAMING_SNAKE_CASE = parser.parse_args() __SCREAMING_SNAKE_CASE = args.from_gh if from_gh: # The artifacts have to be downloaded using `actions/download-artifact@v3` pass else: os.makedirs(args.output_dir, exist_ok=True) # get download links __SCREAMING_SNAKE_CASE = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, """artifacts.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) # download artifacts for idx, (name, url) in enumerate(artifacts.items()): print(name) print(url) print("""=""" * 80) download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) # extract warnings from artifacts __SCREAMING_SNAKE_CASE = extract_warnings(args.output_dir, args.targets) __SCREAMING_SNAKE_CASE = sorted(selected_warnings) with open(os.path.join(args.output_dir, """selected_warnings.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(selected_warnings, fp, ensure_ascii=False, indent=4)
256
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import GLPNImageProcessor class a_ ( unittest.TestCase ): """simple docstring""" def __init__( self : Any ,snake_case : Optional[int] ,snake_case : List[str]=7 ,snake_case : Any=3 ,snake_case : Tuple=18 ,snake_case : List[str]=30 ,snake_case : int=400 ,snake_case : Tuple=True ,snake_case : List[Any]=32 ,snake_case : Tuple=True ,): SCREAMING_SNAKE_CASE =parent SCREAMING_SNAKE_CASE =batch_size SCREAMING_SNAKE_CASE =num_channels SCREAMING_SNAKE_CASE =image_size SCREAMING_SNAKE_CASE =min_resolution SCREAMING_SNAKE_CASE =max_resolution SCREAMING_SNAKE_CASE =do_resize SCREAMING_SNAKE_CASE =size_divisor SCREAMING_SNAKE_CASE =do_rescale def _lowerCAmelCase ( self : Optional[int] ): return { "do_resize": self.do_resize, "size_divisor": self.size_divisor, "do_rescale": self.do_rescale, } @require_torch @require_vision class a_ ( lowerCamelCase_ , unittest.TestCase ): """simple docstring""" __UpperCAmelCase = GLPNImageProcessor if is_vision_available() else None def _lowerCAmelCase ( self : List[str] ): SCREAMING_SNAKE_CASE =GLPNImageProcessingTester(self ) @property def _lowerCAmelCase ( self : str ): return self.image_processor_tester.prepare_image_processor_dict() def _lowerCAmelCase ( self : int ): SCREAMING_SNAKE_CASE =self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case ,'do_resize' ) ) self.assertTrue(hasattr(snake_case ,'size_divisor' ) ) self.assertTrue(hasattr(snake_case ,'resample' ) ) self.assertTrue(hasattr(snake_case ,'do_rescale' ) ) def _lowerCAmelCase ( self : Dict ): pass def _lowerCAmelCase ( self : Optional[int] ): # Initialize image_processing SCREAMING_SNAKE_CASE =self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE =prepare_image_inputs(self.image_processor_tester ,equal_resolution=snake_case ) for image in image_inputs: self.assertIsInstance(snake_case ,Image.Image ) # Test not batched input (GLPNImageProcessor doesn't support batching) SCREAMING_SNAKE_CASE =image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def _lowerCAmelCase ( self : Union[str, Any] ): # Initialize image_processing SCREAMING_SNAKE_CASE =self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors SCREAMING_SNAKE_CASE =prepare_image_inputs(self.image_processor_tester ,equal_resolution=snake_case ,numpify=snake_case ) for image in image_inputs: self.assertIsInstance(snake_case ,np.ndarray ) # Test not batched input (GLPNImageProcessor doesn't support batching) SCREAMING_SNAKE_CASE =image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 ) def _lowerCAmelCase ( self : Union[str, Any] ): # Initialize image_processing SCREAMING_SNAKE_CASE =self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE =prepare_image_inputs(self.image_processor_tester ,equal_resolution=snake_case ,torchify=snake_case ) for image in image_inputs: self.assertIsInstance(snake_case ,torch.Tensor ) # Test not batched input (GLPNImageProcessor doesn't support batching) SCREAMING_SNAKE_CASE =image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0 ) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0 )
334
def snake_case__ ( lowerCAmelCase_ ): """simple docstring""" return " ".join( ''.join(word[::-1] ) if len(lowerCAmelCase_ ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words("Hey wollef sroirraw"))
334
1
"""simple docstring""" import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __A = logging.get_logger(__name__) __A = """▁""" __A = {"""vocab_file""": """vocab.txt""", """sentencepiece_model_ckpt""": """sentencepiece.bpe.model"""} __A = { """sentencepiece_model_file""": """sentencepiece.bpe.model""", """vocab_file""": """vocab.txt""", } __A = { """vocab_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", }, """sentencepiece_model_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", }, } __A = { """ernie-m-base""": 514, """ernie-m-large""": 514, } __A = { """ernie-m-base""": {"""do_lower_case""": False}, """ernie-m-large""": {"""do_lower_case""": False}, } class _lowerCAmelCase ( __lowercase ): """simple docstring""" __magic_name__ :List[str] = ["input_ids"] __magic_name__ :Tuple = VOCAB_FILES_NAMES __magic_name__ :Union[str, Any] = PRETRAINED_INIT_CONFIGURATION __magic_name__ :Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ :List[str] = PRETRAINED_VOCAB_FILES_MAP __magic_name__ :Dict = RESOURCE_FILES_NAMES def __init__( self , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=False , __UpperCAmelCase="utf8" , __UpperCAmelCase="[UNK]" , __UpperCAmelCase="[SEP]" , __UpperCAmelCase="[PAD]" , __UpperCAmelCase="[CLS]" , __UpperCAmelCase="[MASK]" , __UpperCAmelCase = None , **__UpperCAmelCase , ): '''simple docstring''' lowerCAmelCase__ :Tuple = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , vocab_file=UpperCAmelCase__ , encoding=UpperCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCAmelCase__ , ) lowerCAmelCase__ :str = do_lower_case lowerCAmelCase__ :List[str] = sentencepiece_model_ckpt lowerCAmelCase__ :Union[str, Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCAmelCase__ ) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: lowerCAmelCase__ :List[Any] = self.load_vocab(filepath=UpperCAmelCase__ ) else: lowerCAmelCase__ :int = {self.sp_model.id_to_piece(UpperCAmelCase__ ): id for id in range(self.sp_model.get_piece_size() )} lowerCAmelCase__ :Union[str, Any] = {v: k for k, v in self.vocab.items()} def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' if text is None: return None lowerCAmelCase__ :Any = self.tokenize(UpperCAmelCase__ ) lowerCAmelCase__ , lowerCAmelCase__ :int = '', [] for i, ch in enumerate(UpperCAmelCase__ ): if ch in self.SP_CHAR_MAPPING: lowerCAmelCase__ :str = self.SP_CHAR_MAPPING.get(UpperCAmelCase__ ) else: lowerCAmelCase__ :List[Any] = unicodedata.normalize('NFKC' , UpperCAmelCase__ ) if self.is_whitespace(UpperCAmelCase__ ): continue normalized_text += ch char_mapping.extend([i] * len(UpperCAmelCase__ ) ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ :Union[str, Any] = normalized_text, [], 0 if self.do_lower_case: lowerCAmelCase__ :Tuple = text.lower() for token in split_tokens: if token[:1] == "▁": lowerCAmelCase__ :Tuple = token[1:] lowerCAmelCase__ :int = text[offset:].index(UpperCAmelCase__ ) + offset lowerCAmelCase__ :Optional[Any] = start + len(UpperCAmelCase__ ) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1) ) lowerCAmelCase__ :Optional[int] = end return token_mapping @property def snake_case ( self ): '''simple docstring''' return len(self.vocab ) def snake_case ( self ): '''simple docstring''' return dict(self.vocab , **self.added_tokens_encoder ) def __getstate__( self ): '''simple docstring''' lowerCAmelCase__ :Dict = self.__dict__.copy() lowerCAmelCase__ :Optional[int] = None return state def __setstate__( self , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): lowerCAmelCase__ :Union[str, Any] = {} lowerCAmelCase__ :Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.sentencepiece_model_ckpt ) def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' return "".join((self.SP_CHAR_MAPPING.get(UpperCAmelCase__ , UpperCAmelCase__ ) for c in text) ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase=False , __UpperCAmelCase=6_4 , __UpperCAmelCase=0.1 ): '''simple docstring''' if self.sp_model_kwargs.get('enable_sampling' ) is True: lowerCAmelCase__ :int = True if self.sp_model_kwargs.get('alpha' ) is not None: lowerCAmelCase__ :Optional[Any] = self.sp_model_kwargs.get('alpha' ) if self.sp_model_kwargs.get('nbest_size' ) is not None: lowerCAmelCase__ :Tuple = self.sp_model_kwargs.get('nbest_size' ) if not enable_sampling: lowerCAmelCase__ :Any = self.sp_model.EncodeAsPieces(UpperCAmelCase__ ) else: lowerCAmelCase__ :Tuple = self.sp_model.SampleEncodeAsPieces(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) lowerCAmelCase__ :str = [] for pi, piece in enumerate(UpperCAmelCase__ ): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(UpperCAmelCase__ ) and pi != 0: new_pieces.append(UpperCAmelCase__ ) continue else: continue lowerCAmelCase__ :List[Any] = 0 for i, chunk in enumerate(UpperCAmelCase__ ): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(UpperCAmelCase__ ) or self.is_punct(UpperCAmelCase__ ): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) new_pieces.append(UpperCAmelCase__ ) lowerCAmelCase__ :Optional[Any] = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowerCAmelCase__ :Optional[int] = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) lowerCAmelCase__ :int = i if len(UpperCAmelCase__ ) > lst_i: new_pieces.append(piece[lst_i:] ) return new_pieces def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Optional[Any] = ''.join(UpperCAmelCase__ ).replace(UpperCAmelCase__ , ' ' ).strip() return out_string def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :List[Any] = self.convert_ids_to_tokens(UpperCAmelCase__ ) lowerCAmelCase__ :Optional[Any] = ''.join(UpperCAmelCase__ ).replace(UpperCAmelCase__ , ' ' ).strip() return out_string def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' return self.vocab.get(UpperCAmelCase__ , self.vocab.get(self.unk_token ) ) def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' return self.reverse_vocab.get(UpperCAmelCase__ , self.unk_token ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase=None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase__ :List[Any] = [self.cls_token_id] lowerCAmelCase__ :Optional[int] = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase=None ): '''simple docstring''' if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase=False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(UpperCAmelCase__ )) + [1, 1] + ([0] * len(UpperCAmelCase__ )) + [1] return [1] + ([0] * len(UpperCAmelCase__ )) + [1] def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' if token_ids_a is None: # [CLS] X [SEP] return (len(UpperCAmelCase__ ) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(UpperCAmelCase__ ) + 1) + [1] * (len(UpperCAmelCase__ ) + 3) def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' if "\u4e00" <= char <= "\u9fff": return True return False def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' if char in ",;:.?!~,;:。?!《》【】": return True return False def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(UpperCAmelCase__ ) == 1: lowerCAmelCase__ :List[Any] = unicodedata.category(UpperCAmelCase__ ) if cat == "Zs": return True return False def snake_case ( self , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :List[str] = {} with io.open(UpperCAmelCase__ , 'r' , encoding='utf-8' ) as f: for index, line in enumerate(UpperCAmelCase__ ): lowerCAmelCase__ :Any = line.rstrip('\n' ) lowerCAmelCase__ :Tuple = int(UpperCAmelCase__ ) return token_to_idx def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' lowerCAmelCase__ :Any = 0 if os.path.isdir(UpperCAmelCase__ ): lowerCAmelCase__ :Optional[Any] = os.path.join( UpperCAmelCase__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) else: lowerCAmelCase__ :str = (filename_prefix + '-' if filename_prefix else '') + save_directory with open(UpperCAmelCase__ , 'w' , encoding='utf-8' ) as writer: for token, token_index in sorted(self.vocab.items() , key=lambda __UpperCAmelCase : kv[1] ): if index != token_index: logger.warning( F"Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive." ' Please check that the vocabulary is not corrupted!' ) lowerCAmelCase__ :List[str] = token_index writer.write(token + '\n' ) index += 1 lowerCAmelCase__ :List[str] = os.path.join(UpperCAmelCase__ , 'sentencepiece.bpe.model' ) with open(UpperCAmelCase__ , 'wb' ) as fi: lowerCAmelCase__ :Dict = self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase__ ) return (vocab_file,)
361
"""simple docstring""" import json from typing import Iterator, List, Union from tokenizers import AddedToken, Regex, Tokenizer, decoders, normalizers, pre_tokenizers, trainers from tokenizers.implementations.base_tokenizer import BaseTokenizer from tokenizers.models import Unigram from tokenizers.processors import TemplateProcessing class _lowerCAmelCase ( a ): """simple docstring""" def __init__( self , __UpperCAmelCase = "▁" , __UpperCAmelCase = True , __UpperCAmelCase = "<unk>" , __UpperCAmelCase = "</s>" , __UpperCAmelCase = "<pad>" , ): '''simple docstring''' lowerCAmelCase__ :Tuple = { 'pad': {'id': 0, 'token': pad_token}, 'eos': {'id': 1, 'token': eos_token}, 'unk': {'id': 2, 'token': unk_token}, } lowerCAmelCase__ :Optional[int] = [None] * len(self.special_tokens ) for token_dict in self.special_tokens.values(): lowerCAmelCase__ :Any = token_dict['token'] lowerCAmelCase__ :int = Tokenizer(Unigram() ) lowerCAmelCase__ :Tuple = normalizers.Sequence( [ normalizers.Nmt(), normalizers.NFKC(), normalizers.Replace(Regex(' {2,}' ) , ' ' ), normalizers.Lowercase(), ] ) lowerCAmelCase__ :Any = pre_tokenizers.Sequence( [ pre_tokenizers.Metaspace(replacement=__UpperCAmelCase , add_prefix_space=__UpperCAmelCase ), pre_tokenizers.Digits(individual_digits=__UpperCAmelCase ), pre_tokenizers.Punctuation(), ] ) lowerCAmelCase__ :List[str] = decoders.Metaspace(replacement=__UpperCAmelCase , add_prefix_space=__UpperCAmelCase ) lowerCAmelCase__ :Tuple = TemplateProcessing( single=F"$A {self.special_tokens['eos']['token']}" , special_tokens=[(self.special_tokens['eos']['token'], self.special_tokens['eos']['id'])] , ) lowerCAmelCase__ :Optional[int] = { 'model': 'SentencePieceUnigram', 'replacement': replacement, 'add_prefix_space': add_prefix_space, } super().__init__(__UpperCAmelCase , __UpperCAmelCase ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase = 8_0_0_0 , __UpperCAmelCase = True , ): '''simple docstring''' lowerCAmelCase__ :int = trainers.UnigramTrainer( vocab_size=__UpperCAmelCase , special_tokens=self.special_tokens_list , show_progress=__UpperCAmelCase , ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ): lowerCAmelCase__ :int = [files] self._tokenizer.train(__UpperCAmelCase , trainer=__UpperCAmelCase ) self.add_unk_id() def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase = 8_0_0_0 , __UpperCAmelCase = True , ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = trainers.UnigramTrainer( vocab_size=__UpperCAmelCase , special_tokens=self.special_tokens_list , show_progress=__UpperCAmelCase , ) self._tokenizer.train_from_iterator(__UpperCAmelCase , trainer=__UpperCAmelCase ) self.add_unk_id() def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = json.loads(self._tokenizer.to_str() ) lowerCAmelCase__ :List[str] = self.special_tokens['unk']['id'] lowerCAmelCase__ :Union[str, Any] = Tokenizer.from_str(json.dumps(__UpperCAmelCase ) )
254
0
'''simple docstring''' from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef import datasets lowerCamelCase : List[Any] = "\\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n" lowerCamelCase : Optional[int] = "\\nGLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n" lowerCamelCase : Dict = "\nCompute GLUE evaluation metric associated to each GLUE dataset.\nArgs:\n predictions: list of predictions to score.\n Each translation should be tokenized into a list of tokens.\n references: list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\nReturns: depending on the GLUE subset, one or several of:\n \"accuracy\": Accuracy\n \"f1\": F1 score\n \"pearson\": Pearson Correlation\n \"spearmanr\": Spearman Correlation\n \"matthews_correlation\": Matthew Correlation\nExamples:\n\n >>> glue_metric = datasets.load_metric('glue', 'sst2') # 'sst2' or any of [\"mnli\", \"mnli_mismatched\", \"mnli_matched\", \"qnli\", \"rte\", \"wnli\", \"hans\"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'mrpc') # 'mrpc' or 'qqp'\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0, 'f1': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'stsb')\n >>> references = [0., 1., 2., 3., 4., 5.]\n >>> predictions = [0., 1., 2., 3., 4., 5.]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print({\"pearson\": round(results[\"pearson\"], 2), \"spearmanr\": round(results[\"spearmanr\"], 2)})\n {'pearson': 1.0, 'spearmanr': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'cola')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'matthews_correlation': 1.0}\n" def _lowerCAmelCase ( _UpperCamelCase : List[Any] , _UpperCamelCase : Optional[Any] ) -> Dict: """simple docstring""" return float((preds == labels).mean() ) def _lowerCAmelCase ( _UpperCamelCase : Tuple , _UpperCamelCase : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE =simple_accuracy(_UpperCamelCase , _UpperCamelCase ) _SCREAMING_SNAKE_CASE =float(fa_score(y_true=_UpperCamelCase , y_pred=_UpperCamelCase ) ) return { "accuracy": acc, "f1": fa, } def _lowerCAmelCase ( _UpperCamelCase : List[Any] , _UpperCamelCase : str ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE =float(pearsonr(_UpperCamelCase , _UpperCamelCase )[0] ) _SCREAMING_SNAKE_CASE =float(spearmanr(_UpperCamelCase , _UpperCamelCase )[0] ) return { "pearson": pearson_corr, "spearmanr": spearman_corr, } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): def A ( self : List[str] ) -> List[str]: '''simple docstring''' if self.config_name not in [ "sst2", "mnli", "mnli_mismatched", "mnli_matched", "cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans", ]: raise KeyError( 'You should supply a configuration name selected in ' '["sst2", "mnli", "mnli_mismatched", "mnli_matched", ' '"cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans"]' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('int64' if self.config_name != 'stsb' else 'float32' ), 'references': datasets.Value('int64' if self.config_name != 'stsb' else 'float32' ), } ) , codebase_urls=[] , reference_urls=[] , format='numpy' , ) def A ( self : int , _a : Any , _a : int ) -> Tuple: '''simple docstring''' if self.config_name == "cola": return {"matthews_correlation": matthews_corrcoef(_a , _a )} elif self.config_name == "stsb": return pearson_and_spearman(_a , _a ) elif self.config_name in ["mrpc", "qqp"]: return acc_and_fa(_a , _a ) elif self.config_name in ["sst2", "mnli", "mnli_mismatched", "mnli_matched", "qnli", "rte", "wnli", "hans"]: return {"accuracy": simple_accuracy(_a , _a )} else: raise KeyError( 'You should supply a configuration name selected in ' '["sst2", "mnli", "mnli_mismatched", "mnli_matched", ' '"cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans"]' )
47
import pytest from datasets.parallel import ParallelBackendConfig, parallel_backend from datasets.utils.py_utils import map_nested from .utils import require_dill_gt_0_3_2, require_joblibspark, require_not_windows def lowerCAmelCase__(__snake_case ) -> int: # picklable for multiprocessing '''simple docstring''' return i + 1 @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows def lowerCAmelCase__() -> Any: '''simple docstring''' with parallel_backend('''spark''' ): assert ParallelBackendConfig.backend_name == "spark" lowerCamelCase__ = [1, 2, 3] with pytest.raises(__snake_case ): with parallel_backend('''unsupported backend''' ): map_nested(__snake_case ,__snake_case ,num_proc=2 ) with pytest.raises(__snake_case ): with parallel_backend('''unsupported backend''' ): map_nested(__snake_case ,__snake_case ,num_proc=-1 ) @require_dill_gt_0_3_2 @require_joblibspark @require_not_windows @pytest.mark.parametrize('''num_proc''' ,[2, -1] ) def lowerCAmelCase__(__snake_case ) -> Tuple: '''simple docstring''' lowerCamelCase__ = [1, 2] lowerCamelCase__ = {'''a''': 1, '''b''': 2} lowerCamelCase__ = {'''a''': [1, 2], '''b''': [3, 4]} lowerCamelCase__ = {'''a''': {'''1''': 1}, '''b''': 2} lowerCamelCase__ = {'''a''': 1, '''b''': 2, '''c''': 3, '''d''': 4} lowerCamelCase__ = [2, 3] lowerCamelCase__ = {'''a''': 2, '''b''': 3} lowerCamelCase__ = {'''a''': [2, 3], '''b''': [4, 5]} lowerCamelCase__ = {'''a''': {'''1''': 2}, '''b''': 3} lowerCamelCase__ = {'''a''': 2, '''b''': 3, '''c''': 4, '''d''': 5} with parallel_backend('''spark''' ): assert map_nested(__snake_case ,__snake_case ,num_proc=__snake_case ) == expected_map_nested_sa assert map_nested(__snake_case ,__snake_case ,num_proc=__snake_case ) == expected_map_nested_sa assert map_nested(__snake_case ,__snake_case ,num_proc=__snake_case ) == expected_map_nested_sa assert map_nested(__snake_case ,__snake_case ,num_proc=__snake_case ) == expected_map_nested_sa assert map_nested(__snake_case ,__snake_case ,num_proc=__snake_case ) == expected_map_nested_sa
209
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging __A = logging.get_logger(__name__) __A = '''▁''' __A = {'''vocab_file''': '''sentencepiece.bpe.model''', '''monolingual_vocab_file''': '''dict.txt'''} __A = { '''vocab_file''': { '''vinai/bartpho-syllable''': '''https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model''', }, '''monolingual_vocab_file''': { '''vinai/bartpho-syllable''': '''https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt''', }, } __A = {'''vinai/bartpho-syllable''': 1024} class lowercase ( snake_case__): """simple docstring""" a__ : List[str] = VOCAB_FILES_NAMES a__ : Tuple = PRETRAINED_VOCAB_FILES_MAP a__ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ : int = ["input_ids", "attention_mask"] def __init__( self : Optional[int] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Dict="<s>" , __UpperCAmelCase : List[str]="</s>" , __UpperCAmelCase : str="</s>" , __UpperCAmelCase : List[Any]="<s>" , __UpperCAmelCase : Union[str, Any]="<unk>" , __UpperCAmelCase : List[str]="<pad>" , __UpperCAmelCase : Any="<mask>" , __UpperCAmelCase : Optional[Dict[str, Any]] = None , **__UpperCAmelCase : str , ) -> None: # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase_= AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token UpperCAmelCase_= {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) UpperCAmelCase_= vocab_file UpperCAmelCase_= monolingual_vocab_file UpperCAmelCase_= spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__UpperCAmelCase ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility UpperCAmelCase_= {} UpperCAmelCase_= 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(__UpperCAmelCase ) not in self.fairseq_tokens_to_ids: UpperCAmelCase_= cnt cnt += 1 with open(__UpperCAmelCase , """r""" , encoding="""utf-8""" ) as f: for line in f.readlines(): UpperCAmelCase_= line.strip().split()[0] UpperCAmelCase_= len(self.fairseq_tokens_to_ids ) if str(__UpperCAmelCase ) not in self.fairseq_tokens_to_ids: UpperCAmelCase_= len(self.fairseq_tokens_to_ids ) UpperCAmelCase_= {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Union[str, Any] ) -> Optional[Any]: UpperCAmelCase_= self.__dict__.copy() UpperCAmelCase_= None UpperCAmelCase_= self.sp_model.serialized_model_proto() return state def __setstate__( self : Optional[int] , __UpperCAmelCase : Optional[Any] ) -> Optional[int]: UpperCAmelCase_= d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): UpperCAmelCase_= {} UpperCAmelCase_= spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def _SCREAMING_SNAKE_CASE ( self : Dict , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase_= [self.cls_token_id] UpperCAmelCase_= [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _SCREAMING_SNAKE_CASE ( self : List[str] , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None , __UpperCAmelCase : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(__UpperCAmelCase )) + [1] return [1] + ([0] * len(__UpperCAmelCase )) + [1, 1] + ([0] * len(__UpperCAmelCase )) + [1] def _SCREAMING_SNAKE_CASE ( self : Tuple , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: UpperCAmelCase_= [self.sep_token_id] UpperCAmelCase_= [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> str: return len(self.fairseq_ids_to_tokens ) def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Dict: UpperCAmelCase_= {self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _SCREAMING_SNAKE_CASE ( self : Any , __UpperCAmelCase : str ) -> List[str]: return self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase ) def _SCREAMING_SNAKE_CASE ( self : List[Any] , __UpperCAmelCase : str ) -> List[str]: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , __UpperCAmelCase : Any ) -> Tuple: return self.fairseq_ids_to_tokens[index] def _SCREAMING_SNAKE_CASE ( self : List[str] , __UpperCAmelCase : int ) -> List[str]: UpperCAmelCase_= """""".join(__UpperCAmelCase ).replace(__UpperCAmelCase , """ """ ).strip() return out_string def _SCREAMING_SNAKE_CASE ( self : Optional[int] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase_= os.path.join( __UpperCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCAmelCase_= os.path.join( __UpperCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""monolingual_vocab_file"""] , ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase , """wb""" ) as fi: UpperCAmelCase_= self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( __UpperCAmelCase ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(__UpperCAmelCase , """w""" , encoding="""utf-8""" ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(F"""{str(__UpperCAmelCase )} \n""" ) return out_vocab_file, out_monolingual_vocab_file
277
from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class lowercase ( snake_case__): """simple docstring""" a__ : str = ["vqvae"] def __init__( self : List[Any] , __UpperCAmelCase : AutoencoderKL , __UpperCAmelCase : UNetaDConditionModel , __UpperCAmelCase : Mel , __UpperCAmelCase : Union[DDIMScheduler, DDPMScheduler] , ) -> str: super().__init__() self.register_modules(unet=__UpperCAmelCase , scheduler=__UpperCAmelCase , mel=__UpperCAmelCase , vqvae=__UpperCAmelCase ) def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> int: return 50 if isinstance(self.scheduler , __UpperCAmelCase ) else 1_000 @torch.no_grad() def __call__( self : List[Any] , __UpperCAmelCase : int = 1 , __UpperCAmelCase : str = None , __UpperCAmelCase : np.ndarray = None , __UpperCAmelCase : int = 0 , __UpperCAmelCase : int = 0 , __UpperCAmelCase : int = None , __UpperCAmelCase : torch.Generator = None , __UpperCAmelCase : float = 0 , __UpperCAmelCase : float = 0 , __UpperCAmelCase : torch.Generator = None , __UpperCAmelCase : float = 0 , __UpperCAmelCase : torch.Tensor = None , __UpperCAmelCase : torch.Tensor = None , __UpperCAmelCase : Union[str, Any]=True , ) -> Union[ Union[AudioPipelineOutput, ImagePipelineOutput], Tuple[List[Image.Image], Tuple[int, List[np.ndarray]]], ]: UpperCAmelCase_= steps or self.get_default_steps() self.scheduler.set_timesteps(__UpperCAmelCase ) UpperCAmelCase_= step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: UpperCAmelCase_= (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: UpperCAmelCase_= randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=__UpperCAmelCase , device=self.device , ) UpperCAmelCase_= noise UpperCAmelCase_= None if audio_file is not None or raw_audio is not None: self.mel.load_audio(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase_= self.mel.audio_slice_to_image(__UpperCAmelCase ) UpperCAmelCase_= np.frombuffer(input_image.tobytes() , dtype="""uint8""" ).reshape( (input_image.height, input_image.width) ) UpperCAmelCase_= (input_image / 255) * 2 - 1 UpperCAmelCase_= torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: UpperCAmelCase_= self.vqvae.encode(torch.unsqueeze(__UpperCAmelCase , 0 ) ).latent_dist.sample( generator=__UpperCAmelCase )[0] UpperCAmelCase_= self.vqvae.config.scaling_factor * input_images if start_step > 0: UpperCAmelCase_= self.scheduler.add_noise(__UpperCAmelCase , __UpperCAmelCase , self.scheduler.timesteps[start_step - 1] ) UpperCAmelCase_= ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) UpperCAmelCase_= int(mask_start_secs * pixels_per_second ) UpperCAmelCase_= int(mask_end_secs * pixels_per_second ) UpperCAmelCase_= self.scheduler.add_noise(__UpperCAmelCase , __UpperCAmelCase , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , __UpperCAmelCase ): UpperCAmelCase_= self.unet(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase )["""sample"""] else: UpperCAmelCase_= self.unet(__UpperCAmelCase , __UpperCAmelCase )["""sample"""] if isinstance(self.scheduler , __UpperCAmelCase ): UpperCAmelCase_= self.scheduler.step( model_output=__UpperCAmelCase , timestep=__UpperCAmelCase , sample=__UpperCAmelCase , eta=__UpperCAmelCase , generator=__UpperCAmelCase , )["""prev_sample"""] else: UpperCAmelCase_= self.scheduler.step( model_output=__UpperCAmelCase , timestep=__UpperCAmelCase , sample=__UpperCAmelCase , generator=__UpperCAmelCase , )["""prev_sample"""] if mask is not None: if mask_start > 0: UpperCAmelCase_= mask[:, step, :, :mask_start] if mask_end > 0: UpperCAmelCase_= mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance UpperCAmelCase_= 1 / self.vqvae.config.scaling_factor * images UpperCAmelCase_= self.vqvae.decode(__UpperCAmelCase )["""sample"""] UpperCAmelCase_= (images / 2 + 0.5).clamp(0 , 1 ) UpperCAmelCase_= images.cpu().permute(0 , 2 , 3 , 1 ).numpy() UpperCAmelCase_= (images * 255).round().astype("""uint8""" ) UpperCAmelCase_= list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(__UpperCAmelCase , mode="""RGB""" ).convert("""L""" ) for _ in images) ) UpperCAmelCase_= [self.mel.image_to_audio(__UpperCAmelCase ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(__UpperCAmelCase )[:, np.newaxis, :] ) , **ImagePipelineOutput(__UpperCAmelCase ) ) @torch.no_grad() def _SCREAMING_SNAKE_CASE ( self : List[str] , __UpperCAmelCase : List[Image.Image] , __UpperCAmelCase : int = 50 ) -> np.ndarray: assert isinstance(self.scheduler , __UpperCAmelCase ) self.scheduler.set_timesteps(__UpperCAmelCase ) UpperCAmelCase_= np.array( [np.frombuffer(image.tobytes() , dtype="""uint8""" ).reshape((1, image.height, image.width) ) for image in images] ) UpperCAmelCase_= (sample / 255) * 2 - 1 UpperCAmelCase_= torch.Tensor(__UpperCAmelCase ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): UpperCAmelCase_= t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps UpperCAmelCase_= self.scheduler.alphas_cumprod[t] UpperCAmelCase_= ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) UpperCAmelCase_= 1 - alpha_prod_t UpperCAmelCase_= self.unet(__UpperCAmelCase , __UpperCAmelCase )["""sample"""] UpperCAmelCase_= (1 - alpha_prod_t_prev) ** 0.5 * model_output UpperCAmelCase_= (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) UpperCAmelCase_= sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def _SCREAMING_SNAKE_CASE ( __UpperCAmelCase : torch.Tensor , __UpperCAmelCase : torch.Tensor , __UpperCAmelCase : float ) -> torch.Tensor: UpperCAmelCase_= acos(torch.dot(torch.flatten(__UpperCAmelCase ) , torch.flatten(__UpperCAmelCase ) ) / torch.norm(__UpperCAmelCase ) / torch.norm(__UpperCAmelCase ) ) return sin((1 - alpha) * theta ) * xa / sin(__UpperCAmelCase ) + sin(alpha * theta ) * xa / sin(__UpperCAmelCase )
277
1
'''simple docstring''' import datasets from .evaluate import evaluate _A : Tuple ='''\ @inproceedings{Rajpurkar2016SQuAD10, title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text}, author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang}, booktitle={EMNLP}, year={2016} } ''' _A : int =''' This metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD). Stanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by crowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span, from the corresponding reading passage, or the question might be unanswerable. ''' _A : Dict =''' Computes SQuAD scores (F1 and EM). Args: predictions: List of question-answers dictionaries with the following key-values: - \'id\': id of the question-answer pair as given in the references (see below) - \'prediction_text\': the text of the answer references: List of question-answers dictionaries with the following key-values: - \'id\': id of the question-answer pair (see above), - \'answers\': a Dict in the SQuAD dataset format { \'text\': list of possible texts for the answer, as a list of strings \'answer_start\': list of start positions for the answer, as a list of ints } Note that answer_start values are not taken into account to compute the metric. Returns: \'exact_match\': Exact match (the normalized answer exactly match the gold answer) \'f1\': The F-score of predicted tokens versus the gold answer Examples: >>> predictions = [{\'prediction_text\': \'1976\', \'id\': \'56e10a3be3433e1400422b22\'}] >>> references = [{\'answers\': {\'answer_start\': [97], \'text\': [\'1976\']}, \'id\': \'56e10a3be3433e1400422b22\'}] >>> squad_metric = datasets.load_metric("squad") >>> results = squad_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 100.0, \'f1\': 100.0} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowercase ( datasets.Metric ): def lowerCamelCase_ ( self: List[str] ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": {"""id""": datasets.Value("""string""" ), """prediction_text""": datasets.Value("""string""" )}, """references""": { """id""": datasets.Value("""string""" ), """answers""": datasets.features.Sequence( { """text""": datasets.Value("""string""" ), """answer_start""": datasets.Value("""int32""" ), } ), }, } ) , codebase_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , reference_urls=["""https://rajpurkar.github.io/SQuAD-explorer/"""] , ) def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase__: List[Any] , UpperCamelCase__: Tuple ): lowerCamelCase__ : List[str] = {prediction["""id"""]: prediction["""prediction_text"""] for prediction in predictions} lowerCamelCase__ : Tuple = [ { """paragraphs""": [ { """qas""": [ { """answers""": [{"""text""": answer_text} for answer_text in ref["""answers"""]["""text"""]], """id""": ref["""id"""], } for ref in references ] } ] } ] lowerCamelCase__ : Any = evaluate(dataset=UpperCamelCase__ , predictions=UpperCamelCase__ ) return score
41
'''simple docstring''' from __future__ import annotations def _UpperCamelCase ( UpperCamelCase__ ): return [ord(UpperCamelCase__ ) - 9_6 for elem in plain] def _UpperCamelCase ( UpperCamelCase__ ): return "".join(chr(elem + 9_6 ) for elem in encoded ) def _UpperCamelCase ( ): UpperCAmelCase__ : int = encode(input("""-> """ ).strip().lower() ) print("""Encoded: """ , UpperCamelCase__ ) print("""Decoded:""" , decode(UpperCamelCase__ ) ) if __name__ == "__main__": main()
163
0
"""simple docstring""" import unittest from transformers import DebertaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, ) from transformers.models.deberta.modeling_deberta import DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST class lowercase__ ( snake_case__ ): def __init__( self : Any , snake_case__ : List[str] , snake_case__ : int=13 , snake_case__ : Dict=7 , snake_case__ : Dict=True , snake_case__ : Optional[int]=True , snake_case__ : Any=True , snake_case__ : List[Any]=True , snake_case__ : Tuple=99 , snake_case__ : Optional[int]=32 , snake_case__ : str=5 , snake_case__ : int=4 , snake_case__ : Dict=37 , snake_case__ : Union[str, Any]="gelu" , snake_case__ : Any=0.1 , snake_case__ : str=0.1 , snake_case__ : List[str]=512 , snake_case__ : int=16 , snake_case__ : Dict=2 , snake_case__ : Tuple=0.02 , snake_case__ : Tuple=False , snake_case__ : List[str]=True , snake_case__ : Tuple="None" , snake_case__ : List[Any]=3 , snake_case__ : Tuple=4 , snake_case__ : List[Any]=None , ): lowerCamelCase_ : List[Any] =parent lowerCamelCase_ : int =batch_size lowerCamelCase_ : List[Any] =seq_length lowerCamelCase_ : int =is_training lowerCamelCase_ : List[Any] =use_input_mask lowerCamelCase_ : str =use_token_type_ids lowerCamelCase_ : str =use_labels lowerCamelCase_ : List[str] =vocab_size lowerCamelCase_ : Union[str, Any] =hidden_size lowerCamelCase_ : List[str] =num_hidden_layers lowerCamelCase_ : Optional[Any] =num_attention_heads lowerCamelCase_ : Union[str, Any] =intermediate_size lowerCamelCase_ : Dict =hidden_act lowerCamelCase_ : List[str] =hidden_dropout_prob lowerCamelCase_ : Dict =attention_probs_dropout_prob lowerCamelCase_ : List[Any] =max_position_embeddings lowerCamelCase_ : int =type_vocab_size lowerCamelCase_ : Tuple =type_sequence_label_size lowerCamelCase_ : Tuple =initializer_range lowerCamelCase_ : Tuple =num_labels lowerCamelCase_ : Optional[Any] =num_choices lowerCamelCase_ : Dict =relative_attention lowerCamelCase_ : Optional[int] =position_biased_input lowerCamelCase_ : Dict =pos_att_type lowerCamelCase_ : Any =scope def UpperCAmelCase__ ( self : Tuple ): lowerCamelCase_ : Any =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase_ : List[str] =None if self.use_input_mask: lowerCamelCase_ : str =ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) lowerCamelCase_ : Optional[int] =None if self.use_token_type_ids: lowerCamelCase_ : Union[str, Any] =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCamelCase_ : Optional[Any] =None lowerCamelCase_ : int =None lowerCamelCase_ : Union[str, Any] =None if self.use_labels: lowerCamelCase_ : Union[str, Any] =ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase_ : Tuple =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCamelCase_ : str =ids_tensor([self.batch_size] , self.num_choices ) lowerCamelCase_ : List[str] =self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase__ ( self : int ): return DebertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def UpperCAmelCase__ ( self : Optional[int] ): lowerCamelCase_ : List[Any] =self.get_config() lowerCamelCase_ : Optional[Any] =300 return config def UpperCAmelCase__ ( self : Optional[int] , snake_case__ : List[Any] ): self.parent.assertListEqual(list(result.loss.size() ) , [] ) def UpperCAmelCase__ ( self : List[str] , snake_case__ : List[Any] , snake_case__ : str , snake_case__ : Optional[int] , snake_case__ : List[str] , snake_case__ : Optional[int] , snake_case__ : Union[str, Any] , snake_case__ : int ): lowerCamelCase_ : List[str] =DebertaModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() lowerCamelCase_ : Dict =model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ )[0] lowerCamelCase_ : Any =model(snake_case__ , token_type_ids=snake_case__ )[0] lowerCamelCase_ : Tuple =model(snake_case__ )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def UpperCAmelCase__ ( self : int , snake_case__ : Any , snake_case__ : int , snake_case__ : List[str] , snake_case__ : List[Any] , snake_case__ : List[str] , snake_case__ : int , snake_case__ : int ): lowerCamelCase_ : Optional[int] =DebertaForMaskedLM(config=snake_case__ ) model.to(snake_case__ ) model.eval() lowerCamelCase_ : Union[str, Any] =model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase__ ( self : str , snake_case__ : int , snake_case__ : Optional[int] , snake_case__ : Union[str, Any] , snake_case__ : Optional[Any] , snake_case__ : Optional[int] , snake_case__ : Union[str, Any] , snake_case__ : Optional[Any] ): lowerCamelCase_ : Optional[int] =self.num_labels lowerCamelCase_ : List[str] =DebertaForSequenceClassification(snake_case__ ) model.to(snake_case__ ) model.eval() lowerCamelCase_ : List[str] =model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(snake_case__ ) def UpperCAmelCase__ ( self : Optional[Any] , snake_case__ : int , snake_case__ : str , snake_case__ : Union[str, Any] , snake_case__ : List[Any] , snake_case__ : Union[str, Any] , snake_case__ : str , snake_case__ : str ): lowerCamelCase_ : Tuple =self.num_labels lowerCamelCase_ : Dict =DebertaForTokenClassification(config=snake_case__ ) model.to(snake_case__ ) model.eval() lowerCamelCase_ : int =model(snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase__ ( self : Union[str, Any] , snake_case__ : List[Any] , snake_case__ : Tuple , snake_case__ : Optional[int] , snake_case__ : Optional[Any] , snake_case__ : Union[str, Any] , snake_case__ : List[Any] , snake_case__ : List[Any] ): lowerCamelCase_ : List[Any] =DebertaForQuestionAnswering(config=snake_case__ ) model.to(snake_case__ ) model.eval() lowerCamelCase_ : Optional[Any] =model( snake_case__ , attention_mask=snake_case__ , token_type_ids=snake_case__ , start_positions=snake_case__ , end_positions=snake_case__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCAmelCase__ ( self : Any ): lowerCamelCase_ : List[Any] =self.prepare_config_and_inputs() ( lowerCamelCase_ ) : Dict =config_and_inputs lowerCamelCase_ : Tuple ={"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class lowercase__ ( snake_case__, snake_case__, unittest.TestCase ): _UpperCAmelCase :Dict = ( ( DebertaModel, DebertaForMaskedLM, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaForQuestionAnswering, ) if is_torch_available() else () ) _UpperCAmelCase :Optional[int] = ( { "feature-extraction": DebertaModel, "fill-mask": DebertaForMaskedLM, "question-answering": DebertaForQuestionAnswering, "text-classification": DebertaForSequenceClassification, "token-classification": DebertaForTokenClassification, "zero-shot": DebertaForSequenceClassification, } if is_torch_available() else {} ) _UpperCAmelCase :Any = True _UpperCAmelCase :Any = False _UpperCAmelCase :Tuple = False _UpperCAmelCase :Any = False _UpperCAmelCase :Optional[Any] = False def UpperCAmelCase__ ( self : Any ): lowerCamelCase_ : str =DebertaModelTester(self ) lowerCamelCase_ : int =ConfigTester(self , config_class=snake_case__ , hidden_size=37 ) def UpperCAmelCase__ ( self : List[str] ): self.config_tester.run_common_tests() def UpperCAmelCase__ ( self : int ): lowerCamelCase_ : int =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*snake_case__ ) def UpperCAmelCase__ ( self : Optional[int] ): lowerCamelCase_ : Optional[int] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*snake_case__ ) def UpperCAmelCase__ ( self : str ): lowerCamelCase_ : str =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*snake_case__ ) def UpperCAmelCase__ ( self : Any ): lowerCamelCase_ : int =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*snake_case__ ) def UpperCAmelCase__ ( self : Optional[int] ): lowerCamelCase_ : List[str] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*snake_case__ ) @slow def UpperCAmelCase__ ( self : List[Any] ): for model_name in DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase_ : Optional[int] =DebertaModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) @require_torch @require_sentencepiece @require_tokenizers class lowercase__ ( unittest.TestCase ): @unittest.skip(reason="Model not available yet" ) def UpperCAmelCase__ ( self : Dict ): pass @slow def UpperCAmelCase__ ( self : Union[str, Any] ): lowerCamelCase_ : Any =DebertaModel.from_pretrained("microsoft/deberta-base" ) lowerCamelCase_ : Dict =torch.tensor([[0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2]] ) lowerCamelCase_ : Dict =torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): lowerCamelCase_ : Dict =model(snake_case__ , attention_mask=snake_case__ )[0] # compare the actual values for a slice. lowerCamelCase_ : str =torch.tensor( [[[-0.5_986, -0.8_055, -0.8_462], [1.4_484, -0.9_348, -0.8_059], [0.3_123, 0.0_032, -1.4_131]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , snake_case__ , atol=1E-4 ) , F"""{output[:, 1:4, 1:4]}""" )
366
"""simple docstring""" from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer A__ : Dict = logging.get_logger(__name__) A__ : Dict = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } A__ : List[Any] = { 'vocab_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json' }, 'merges_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt' }, 'tokenizer_config_file': { 'facebook/blenderbot_small-90M': ( 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json' ) }, } A__ : Optional[int] = { 'facebook/blenderbot_small-90M': 512, } class lowercase__ ( snake_case__ ): _UpperCAmelCase :Optional[int] = VOCAB_FILES_NAMES _UpperCAmelCase :Tuple = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase :Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase :Tuple = BlenderbotSmallTokenizer def __init__( self : Tuple , snake_case__ : Optional[Any]=None , snake_case__ : str=None , snake_case__ : Any="<|endoftext|>" , snake_case__ : Tuple="<|endoftext|>" , snake_case__ : Tuple="<|endoftext|>" , snake_case__ : str=False , snake_case__ : int=True , **snake_case__ : Tuple , ): super().__init__( ByteLevelBPETokenizer( vocab=snake_case__ , merges=snake_case__ , add_prefix_space=snake_case__ , trim_offsets=snake_case__ , ) , bos_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , **snake_case__ , ) lowerCamelCase_ : Optional[int] =add_prefix_space def UpperCAmelCase__ ( self : Tuple , snake_case__ : Optional[Any] , snake_case__ : List[str]=None ): lowerCamelCase_ : Optional[Any] =[self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCAmelCase__ ( self : Tuple , snake_case__ : List[int] , snake_case__ : Optional[List[int]] = None ): lowerCamelCase_ : int =[self.sep_token_id] lowerCamelCase_ : List[Any] =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
209
0
class _SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__(self : Optional[Any] , UpperCAmelCase_ : int) ->Optional[int]: '''simple docstring''' lowerCamelCase__: Any =n lowerCamelCase__: Tuple =[None] * self.n lowerCamelCase__: str =0 # index of the first element lowerCamelCase__: Tuple =0 lowerCamelCase__: Optional[Any] =0 def __len__(self : str) ->int: '''simple docstring''' return self.size def SCREAMING_SNAKE_CASE_ (self : int) ->bool: '''simple docstring''' return self.size == 0 def SCREAMING_SNAKE_CASE_ (self : List[str]) ->str: '''simple docstring''' return False if self.is_empty() else self.array[self.front] def SCREAMING_SNAKE_CASE_ (self : int , UpperCAmelCase_ : Optional[int]) ->str: '''simple docstring''' if self.size >= self.n: raise Exception("QUEUE IS FULL") lowerCamelCase__: List[Any] =data lowerCamelCase__: Dict =(self.rear + 1) % self.n self.size += 1 return self def SCREAMING_SNAKE_CASE_ (self : Tuple) ->Tuple: '''simple docstring''' if self.size == 0: raise Exception("UNDERFLOW") lowerCamelCase__: Optional[Any] =self.array[self.front] lowerCamelCase__: Optional[int] =None lowerCamelCase__: Dict =(self.front + 1) % self.n self.size -= 1 return temp
10
import logging from transformers.configuration_utils import PretrainedConfig __A = logging.getLogger(__name__) class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowercase_ = "masked_bert" def __init__(self : Dict , UpperCAmelCase_ : Any=30_522 , UpperCAmelCase_ : List[Any]=768 , UpperCAmelCase_ : Optional[Any]=12 , UpperCAmelCase_ : str=12 , UpperCAmelCase_ : Tuple=3_072 , UpperCAmelCase_ : str="gelu" , UpperCAmelCase_ : Any=0.1 , UpperCAmelCase_ : Tuple=0.1 , UpperCAmelCase_ : Optional[Any]=512 , UpperCAmelCase_ : Union[str, Any]=2 , UpperCAmelCase_ : str=0.02 , UpperCAmelCase_ : str=1E-1_2 , UpperCAmelCase_ : Union[str, Any]=0 , UpperCAmelCase_ : str="topK" , UpperCAmelCase_ : List[str]="constant" , UpperCAmelCase_ : str=0.0 , **UpperCAmelCase_ : int , ) ->List[Any]: '''simple docstring''' super().__init__(pad_token_id=UpperCAmelCase_ , **UpperCAmelCase_) lowerCamelCase__: Optional[int] =vocab_size lowerCamelCase__: Dict =hidden_size lowerCamelCase__: Optional[int] =num_hidden_layers lowerCamelCase__: Any =num_attention_heads lowerCamelCase__: List[Any] =hidden_act lowerCamelCase__: str =intermediate_size lowerCamelCase__: Dict =hidden_dropout_prob lowerCamelCase__: str =attention_probs_dropout_prob lowerCamelCase__: int =max_position_embeddings lowerCamelCase__: Tuple =type_vocab_size lowerCamelCase__: str =initializer_range lowerCamelCase__: List[Any] =layer_norm_eps lowerCamelCase__: str =pruning_method lowerCamelCase__: Union[str, Any] =mask_init lowerCamelCase__: Optional[Any] =mask_scale
10
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case__ = { 'configuration_time_series_transformer': [ 'TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TimeSeriesTransformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ = [ 'TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TimeSeriesTransformerForPrediction', 'TimeSeriesTransformerModel', 'TimeSeriesTransformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimeSeriesTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimeSeriesTransformerForPrediction, TimeSeriesTransformerModel, TimeSeriesTransformerPreTrainedModel, ) else: import sys snake_case__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
360
'''simple docstring''' import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoImageProcessor, ViTImageProcessor from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_image_processing import CustomImageProcessor # noqa E402 snake_case__ = get_tests_dir("""fixtures""") class UpperCamelCase_ (unittest.TestCase ): """simple docstring""" def _a ( self : List[str] ): """simple docstring""" A_ : List[Any] = mock.Mock() A_ : List[str] = 500 A_ : Tuple = {} A_ : int = HTTPError A_ : Optional[Any] = {} # Download this model to make sure it's in the cache. A_ : Tuple = ViTImageProcessor.from_pretrained('''hf-internal-testing/tiny-random-vit''' ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch('''requests.Session.request''' , return_value=_lowerCamelCase ) as mock_head: A_ : List[Any] = ViTImageProcessor.from_pretrained('''hf-internal-testing/tiny-random-vit''' ) # This check we did call the fake head request mock_head.assert_called() def _a ( self : Tuple ): """simple docstring""" A_ : Tuple = ViTImageProcessor.from_pretrained( '''https://huggingface.co/hf-internal-testing/tiny-random-vit/resolve/main/preprocessor_config.json''' ) def _a ( self : Dict ): """simple docstring""" with self.assertRaises(_lowerCamelCase ): # config is in subfolder, the following should not work without specifying the subfolder A_ : Any = AutoImageProcessor.from_pretrained('''hf-internal-testing/stable-diffusion-all-variants''' ) A_ : Tuple = AutoImageProcessor.from_pretrained( '''hf-internal-testing/stable-diffusion-all-variants''' , subfolder='''feature_extractor''' ) self.assertIsNotNone(_lowerCamelCase ) @is_staging_test class UpperCamelCase_ (unittest.TestCase ): """simple docstring""" @classmethod def _a ( cls : Tuple ): """simple docstring""" A_ : int = TOKEN HfFolder.save_token(_lowerCamelCase ) @classmethod def _a ( cls : str ): """simple docstring""" try: delete_repo(token=cls._token , repo_id='''test-image-processor''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-image-processor-org''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''test-dynamic-image-processor''' ) except HTTPError: pass def _a ( self : List[Any] ): """simple docstring""" A_ : Dict = ViTImageProcessor.from_pretrained(_lowerCamelCase ) image_processor.push_to_hub('''test-image-processor''' , use_auth_token=self._token ) A_ : Optional[int] = ViTImageProcessor.from_pretrained(f'{USER}/test-image-processor' ) for k, v in image_processor.__dict__.items(): self.assertEqual(_lowerCamelCase , getattr(_lowerCamelCase , _lowerCamelCase ) ) # Reset repo delete_repo(token=self._token , repo_id='''test-image-processor''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained( _lowerCamelCase , repo_id='''test-image-processor''' , push_to_hub=_lowerCamelCase , use_auth_token=self._token ) A_ : List[Any] = ViTImageProcessor.from_pretrained(f'{USER}/test-image-processor' ) for k, v in image_processor.__dict__.items(): self.assertEqual(_lowerCamelCase , getattr(_lowerCamelCase , _lowerCamelCase ) ) def _a ( self : Optional[Any] ): """simple docstring""" A_ : int = ViTImageProcessor.from_pretrained(_lowerCamelCase ) image_processor.push_to_hub('''valid_org/test-image-processor''' , use_auth_token=self._token ) A_ : List[str] = ViTImageProcessor.from_pretrained('''valid_org/test-image-processor''' ) for k, v in image_processor.__dict__.items(): self.assertEqual(_lowerCamelCase , getattr(_lowerCamelCase , _lowerCamelCase ) ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-image-processor''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained( _lowerCamelCase , repo_id='''valid_org/test-image-processor-org''' , push_to_hub=_lowerCamelCase , use_auth_token=self._token ) A_ : Any = ViTImageProcessor.from_pretrained('''valid_org/test-image-processor-org''' ) for k, v in image_processor.__dict__.items(): self.assertEqual(_lowerCamelCase , getattr(_lowerCamelCase , _lowerCamelCase ) ) def _a ( self : Optional[Any] ): """simple docstring""" CustomImageProcessor.register_for_auto_class() A_ : Any = CustomImageProcessor.from_pretrained(_lowerCamelCase ) image_processor.push_to_hub('''test-dynamic-image-processor''' , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual( image_processor.auto_map , {'''AutoImageProcessor''': '''custom_image_processing.CustomImageProcessor'''} , ) A_ : str = AutoImageProcessor.from_pretrained( f'{USER}/test-dynamic-image-processor' , trust_remote_code=_lowerCamelCase ) # Can't make an isinstance check because the new_image_processor is from the CustomImageProcessor class of a dynamic module self.assertEqual(new_image_processor.__class__.__name__ , '''CustomImageProcessor''' )
4
0
"""simple docstring""" from collections import namedtuple import requests from lxml import html # type: ignore SCREAMING_SNAKE_CASE_ = namedtuple('''covid_data''', '''cases deaths recovered''') def lowercase (_lowerCAmelCase = "https://www.worldometers.info/coronavirus/" ): __lowerCAmelCase = """//div[@class = \"maincounter-number\"]/span/text()""" return covid_data(*html.fromstring(requests.get(_lowerCAmelCase ).content ).xpath(_lowerCAmelCase ) ) SCREAMING_SNAKE_CASE_ = '''Total COVID-19 cases in the world: {} Total deaths due to COVID-19 in the world: {} Total COVID-19 patients recovered in the world: {}''' print(fmt.format(*covid_stats()))
301
"""simple docstring""" def lowercase (_lowerCAmelCase = 100_0000 ): __lowerCAmelCase = 1 __lowerCAmelCase = 1 __lowerCAmelCase = {1: 1} for inputa in range(2 , _lowerCAmelCase ): __lowerCAmelCase = 0 __lowerCAmelCase = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: __lowerCAmelCase = (3 * number) + 1 counter += 1 if inputa not in counters: __lowerCAmelCase = counter if counter > pre_counter: __lowerCAmelCase = inputa __lowerCAmelCase = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
301
1
from math import pi, sqrt, tan def __lowercase ( __lowerCAmelCase : float ): if side_length < 0: raise ValueError('surface_area_cube() only accepts non-negative values' ) return 6 * side_length**2 def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float , __lowerCAmelCase : float ): if length < 0 or breadth < 0 or height < 0: raise ValueError('surface_area_cuboid() only accepts non-negative values' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def __lowercase ( __lowerCAmelCase : float ): if radius < 0: raise ValueError('surface_area_sphere() only accepts non-negative values' ) return 4 * pi * radius**2 def __lowercase ( __lowerCAmelCase : float ): if radius < 0: raise ValueError('surface_area_hemisphere() only accepts non-negative values' ) return 3 * pi * radius**2 def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float ): if radius < 0 or height < 0: raise ValueError('surface_area_cone() only accepts non-negative values' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float , __lowerCAmelCase : float ): if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( 'surface_area_conical_frustum() only accepts non-negative values' ) a__ = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float ): if radius < 0 or height < 0: raise ValueError('surface_area_cylinder() only accepts non-negative values' ) return 2 * pi * radius * (height + radius) def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float ): if torus_radius < 0 or tube_radius < 0: raise ValueError('surface_area_torus() only accepts non-negative values' ) if torus_radius < tube_radius: raise ValueError( 'surface_area_torus() does not support spindle or self intersecting tori' ) return 4 * pow(__lowerCAmelCase , 2 ) * torus_radius * tube_radius def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float ): if length < 0 or width < 0: raise ValueError('area_rectangle() only accepts non-negative values' ) return length * width def __lowercase ( __lowerCAmelCase : float ): if side_length < 0: raise ValueError('area_square() only accepts non-negative values' ) return side_length**2 def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float ): if base < 0 or height < 0: raise ValueError('area_triangle() only accepts non-negative values' ) return (base * height) / 2 def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float , __lowerCAmelCase : float ): if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('area_triangle_three_sides() only accepts non-negative values' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('Given three sides do not form a triangle' ) a__ = (sidea + sidea + sidea) / 2 a__ = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float ): if base < 0 or height < 0: raise ValueError('area_parallelogram() only accepts non-negative values' ) return base * height def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float , __lowerCAmelCase : float ): if basea < 0 or basea < 0 or height < 0: raise ValueError('area_trapezium() only accepts non-negative values' ) return 1 / 2 * (basea + basea) * height def __lowercase ( __lowerCAmelCase : float ): if radius < 0: raise ValueError('area_circle() only accepts non-negative values' ) return pi * radius**2 def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float ): if radius_x < 0 or radius_y < 0: raise ValueError('area_ellipse() only accepts non-negative values' ) return pi * radius_x * radius_y def __lowercase ( __lowerCAmelCase : float , __lowerCAmelCase : float ): if diagonal_a < 0 or diagonal_a < 0: raise ValueError('area_rhombus() only accepts non-negative values' ) return 1 / 2 * diagonal_a * diagonal_a def __lowercase ( __lowerCAmelCase : int , __lowerCAmelCase : float ): if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or sides < 3: raise ValueError( 'area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides' ) elif length < 0: raise ValueError( 'area_reg_polygon() only accepts non-negative values as \ length of a side' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('''[DEMO] Areas of various geometric shapes: \n''') print(f"""Rectangle: {area_rectangle(10, 20) = }""") print(f"""Square: {area_square(10) = }""") print(f"""Triangle: {area_triangle(10, 10) = }""") print(f"""Triangle: {area_triangle_three_sides(5, 12, 13) = }""") print(f"""Parallelogram: {area_parallelogram(10, 20) = }""") print(f"""Rhombus: {area_rhombus(10, 20) = }""") print(f"""Trapezium: {area_trapezium(10, 20, 30) = }""") print(f"""Circle: {area_circle(20) = }""") print(f"""Ellipse: {area_ellipse(10, 20) = }""") print('''\nSurface Areas of various geometric shapes: \n''') print(f"""Cube: {surface_area_cube(20) = }""") print(f"""Cuboid: {surface_area_cuboid(10, 20, 30) = }""") print(f"""Sphere: {surface_area_sphere(20) = }""") print(f"""Hemisphere: {surface_area_hemisphere(20) = }""") print(f"""Cone: {surface_area_cone(10, 20) = }""") print(f"""Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }""") print(f"""Cylinder: {surface_area_cylinder(10, 20) = }""") print(f"""Torus: {surface_area_torus(20, 10) = }""") print(f"""Equilateral Triangle: {area_reg_polygon(3, 10) = }""") print(f"""Square: {area_reg_polygon(4, 10) = }""") print(f"""Reqular Pentagon: {area_reg_polygon(5, 10) = }""")
109
from collections import defaultdict from math import ceil, sqrt def __lowercase ( __lowerCAmelCase : int = 1_0_0_0_0_0_0 , __lowerCAmelCase : int = 1_0 ): a__ = defaultdict(__lowerCAmelCase ) for outer_width in range(3 , (t_limit // 4) + 2 ): if outer_width * outer_width > t_limit: a__ = max( ceil(sqrt(outer_width * outer_width - t_limit ) ) , 1 ) else: a__ = 1 hole_width_lower_bound += (outer_width - hole_width_lower_bound) % 2 for hole_width in range(__lowerCAmelCase , outer_width - 1 , 2 ): count[outer_width * outer_width - hole_width * hole_width] += 1 return sum(1 for n in count.values() if 1 <= n <= 1_0 ) if __name__ == "__main__": print(f"""{solution() = }""")
109
1
from jiwer import compute_measures import datasets __A : str = '\\n@inproceedings{inproceedings,\n author = {Morris, Andrew and Maier, Viktoria and Green, Phil},\n year = {2004},\n month = {01},\n pages = {},\n title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.}\n}\n' __A : int = '\\nWord error rate (WER) is a common metric of the performance of an automatic speech recognition system.\n\nThe general difficulty of measuring performance lies in the fact that the recognized word sequence can have a different length from the reference word sequence (supposedly the correct one). The WER is derived from the Levenshtein distance, working at the word level instead of the phoneme level. The WER is a valuable tool for comparing different systems as well as for evaluating improvements within one system. This kind of measurement, however, provides no details on the nature of translation errors and further work is therefore required to identify the main source(s) of error and to focus any research effort.\n\nThis problem is solved by first aligning the recognized word sequence with the reference (spoken) word sequence using dynamic string alignment. Examination of this issue is seen through a theory called the power law that states the correlation between perplexity and word error rate.\n\nWord error rate can then be computed as:\n\nWER = (S + D + I) / N = (S + D + I) / (S + D + C)\n\nwhere\n\nS is the number of substitutions,\nD is the number of deletions,\nI is the number of insertions,\nC is the number of correct words,\nN is the number of words in the reference (N=S+D+C).\n\nThis value indicates the average number of errors per reference word. The lower the value, the better the\nperformance of the ASR system with a WER of 0 being a perfect score.\n' __A : Union[str, Any] = '\nCompute WER score of transcribed segments against references.\n\nArgs:\n references: List of references for each speech input.\n predictions: List of transcriptions to score.\n concatenate_texts (bool, default=False): Whether to concatenate all input texts or compute WER iteratively.\n\nReturns:\n (float): the word error rate\n\nExamples:\n\n >>> predictions = ["this is the prediction", "there is an other sample"]\n >>> references = ["this is the reference", "there is another one"]\n >>> wer = datasets.load_metric("wer")\n >>> wer_score = wer.compute(predictions=predictions, references=references)\n >>> print(wer_score)\n 0.5\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): def _lowercase ( self ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/jitsi/jiwer/'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/Word_error_rate''', ] , ) def _lowercase ( self , _A=None , _A=None , _A=False ): '''simple docstring''' if concatenate_texts: return compute_measures(_A , _A )["wer"] else: UpperCAmelCase = 0 UpperCAmelCase = 0 for prediction, reference in zip(_A , _A ): UpperCAmelCase = compute_measures(_A , _A ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
273
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) a__ : str = { 'configuration_trocr': ['TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TrOCRConfig'], 'processing_trocr': ['TrOCRProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : int = [ 'TROCR_PRETRAINED_MODEL_ARCHIVE_LIST', 'TrOCRForCausalLM', 'TrOCRPreTrainedModel', ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys a__ : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
349
0
'''simple docstring''' import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class _lowerCAmelCase ( __lowercase , unittest.TestCase ): """simple docstring""" snake_case_ = VideoToVideoSDPipeline snake_case_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({"video"} ) - {"image", "width", "height"} snake_case_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"video"} ) - {"image"} snake_case_ = PipelineTesterMixin.required_optional_params - {"latents"} snake_case_ = False # No `output_type`. snake_case_ = frozenset( [ "num_inference_steps", "generator", "latents", "return_dict", "callback", "callback_steps", ] ) def lowerCAmelCase ( self : Any )-> Optional[int]: torch.manual_seed(0 ) snake_case = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) snake_case = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=snake_case_ , set_alpha_to_one=snake_case_ , ) torch.manual_seed(0 ) snake_case = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) snake_case = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="""gelu""" , projection_dim=5_12 , ) snake_case = CLIPTextModel(snake_case_ ) snake_case = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) snake_case = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def lowerCAmelCase ( self : Optional[Any] , __snake_case : List[str] , __snake_case : Union[str, Any]=0 )-> Optional[int]: # 3 frames snake_case = floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) if str(snake_case_ ).startswith("""mps""" ): snake_case = torch.manual_seed(snake_case_ ) else: snake_case = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) snake_case = { '''prompt''': '''A painting of a squirrel eating a burger''', '''video''': video, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def lowerCAmelCase ( self : int )-> List[str]: snake_case = '''cpu''' # ensure determinism for the device-dependent torch.Generator snake_case = self.get_dummy_components() snake_case = VideoToVideoSDPipeline(**snake_case_ ) snake_case = sd_pipe.to(snake_case_ ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) snake_case = self.get_dummy_inputs(snake_case_ ) snake_case = '''np''' snake_case = sd_pipe(**snake_case_ ).frames snake_case = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) snake_case = np.array([1_06, 1_17, 1_13, 1_74, 1_37, 1_12, 1_48, 1_51, 1_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def lowerCAmelCase ( self : Tuple )-> Tuple: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=snake_case_ , expected_max_diff=5e-3 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def lowerCAmelCase ( self : Optional[int] )-> Optional[Any]: pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def lowerCAmelCase ( self : int )-> str: pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def lowerCAmelCase ( self : List[str] )-> Dict: pass def lowerCAmelCase ( self : Dict )-> List[Any]: return super().test_progress_bar() @slow @skip_mps class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def lowerCAmelCase ( self : List[Any] )-> Dict: snake_case = VideoToVideoSDPipeline.from_pretrained("""cerspense/zeroscope_v2_XL""" , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames snake_case = torch.Generator(device="""cpu""" ).manual_seed(0 ) snake_case = torch.randn((1, 10, 3, 10_24, 5_76) , generator=snake_case_ ) snake_case = video.to("""cuda""" ) snake_case = '''Spiderman is surfing''' snake_case = pipe(snake_case_ , video=snake_case_ , generator=snake_case_ , num_inference_steps=3 , output_type="""pt""" ).frames snake_case = np.array([-1.0_45_89_84, -1.1_27_92_97, -0.9_66_30_86, -0.91_50_39_06, -0.75_09_76_56] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1e-2
351
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : Any , __snake_case : Optional[Any] , __snake_case : List[Any]=7 , __snake_case : Optional[Any]=3 , __snake_case : str=18 , __snake_case : Union[str, Any]=30 , __snake_case : Union[str, Any]=4_00 , __snake_case : Optional[int]=True , __snake_case : Any=None , __snake_case : List[str]=True , )-> Optional[Any]: snake_case = size if size is not None else {"""height""": 18, """width""": 18} snake_case = parent snake_case = batch_size snake_case = num_channels snake_case = image_size snake_case = min_resolution snake_case = max_resolution snake_case = do_resize snake_case = size snake_case = apply_ocr def lowerCAmelCase ( self : List[Any] )-> List[str]: return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class _lowerCAmelCase ( A__ , unittest.TestCase ): """simple docstring""" snake_case_ = LayoutLMvaImageProcessor if is_pytesseract_available() else None def lowerCAmelCase ( self : int )-> Tuple: snake_case = LayoutLMvaImageProcessingTester(self ) @property def lowerCAmelCase ( self : Tuple )-> Tuple: return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase ( self : Union[str, Any] )-> Any: snake_case = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__snake_case , """do_resize""" ) ) self.assertTrue(hasattr(__snake_case , """size""" ) ) self.assertTrue(hasattr(__snake_case , """apply_ocr""" ) ) def lowerCAmelCase ( self : List[str] )-> List[Any]: snake_case = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 18} ) snake_case = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) def lowerCAmelCase ( self : Dict )-> Union[str, Any]: pass def lowerCAmelCase ( self : Tuple )-> Dict: # Initialize image_processing snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=__snake_case ) for image in image_inputs: self.assertIsInstance(__snake_case , Image.Image ) # Test not batched input snake_case = image_processing(image_inputs[0] , return_tensors="""pt""" ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) self.assertIsInstance(encoding.words , __snake_case ) self.assertIsInstance(encoding.boxes , __snake_case ) # Test batched snake_case = image_processing(__snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) def lowerCAmelCase ( self : int )-> str: # Initialize image_processing snake_case = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=__snake_case , numpify=__snake_case ) for image in image_inputs: self.assertIsInstance(__snake_case , np.ndarray ) # Test not batched input snake_case = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched snake_case = image_processing(__snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) def lowerCAmelCase ( self : List[Any] )-> Optional[Any]: # Initialize image_processing snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=__snake_case , torchify=__snake_case ) for image in image_inputs: self.assertIsInstance(__snake_case , torch.Tensor ) # Test not batched input snake_case = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched snake_case = image_processing(__snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) def lowerCAmelCase ( self : int )-> List[Any]: # with apply_OCR = True snake_case = LayoutLMvaImageProcessor() from datasets import load_dataset snake_case = load_dataset("""hf-internal-testing/fixtures_docvqa""" , split="""test""" ) snake_case = Image.open(ds[0]["""file"""] ).convert("""RGB""" ) snake_case = image_processing(__snake_case , return_tensors="""pt""" ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 2_24, 2_24) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 snake_case = [["""11:14""", """to""", """11:39""", """a.m""", """11:39""", """to""", """11:44""", """a.m.""", """11:44""", """a.m.""", """to""", """12:25""", """p.m.""", """12:25""", """to""", """12:58""", """p.m.""", """12:58""", """to""", """4:00""", """p.m.""", """2:00""", """to""", """5:00""", """p.m.""", """Coffee""", """Break""", """Coffee""", """will""", """be""", """served""", """for""", """men""", """and""", """women""", """in""", """the""", """lobby""", """adjacent""", """to""", """exhibit""", """area.""", """Please""", """move""", """into""", """exhibit""", """area.""", """(Exhibits""", """Open)""", """TRRF""", """GENERAL""", """SESSION""", """(PART""", """|)""", """Presiding:""", """Lee""", """A.""", """Waller""", """TRRF""", """Vice""", """President""", """“Introductory""", """Remarks”""", """Lee""", """A.""", """Waller,""", """TRRF""", """Vice""", """Presi-""", """dent""", """Individual""", """Interviews""", """with""", """TRRF""", """Public""", """Board""", """Members""", """and""", """Sci-""", """entific""", """Advisory""", """Council""", """Mem-""", """bers""", """Conducted""", """by""", """TRRF""", """Treasurer""", """Philip""", """G.""", """Kuehn""", """to""", """get""", """answers""", """which""", """the""", """public""", """refrigerated""", """warehousing""", """industry""", """is""", """looking""", """for.""", """Plus""", """questions""", """from""", """the""", """floor.""", """Dr.""", """Emil""", """M.""", """Mrak,""", """University""", """of""", """Cal-""", """ifornia,""", """Chairman,""", """TRRF""", """Board;""", """Sam""", """R.""", """Cecil,""", """University""", """of""", """Georgia""", """College""", """of""", """Agriculture;""", """Dr.""", """Stanley""", """Charm,""", """Tufts""", """University""", """School""", """of""", """Medicine;""", """Dr.""", """Robert""", """H.""", """Cotton,""", """ITT""", """Continental""", """Baking""", """Company;""", """Dr.""", """Owen""", """Fennema,""", """University""", """of""", """Wis-""", """consin;""", """Dr.""", """Robert""", """E.""", """Hardenburg,""", """USDA.""", """Questions""", """and""", """Answers""", """Exhibits""", """Open""", """Capt.""", """Jack""", """Stoney""", """Room""", """TRRF""", """Scientific""", """Advisory""", """Council""", """Meeting""", """Ballroom""", """Foyer"""]] # noqa: E231 snake_case = [[[1_41, 57, 2_14, 69], [2_28, 58, 2_52, 69], [1_41, 75, 2_16, 88], [2_30, 79, 2_80, 88], [1_42, 2_60, 2_18, 2_73], [2_30, 2_61, 2_55, 2_73], [1_43, 2_79, 2_18, 2_90], [2_31, 2_82, 2_90, 2_91], [1_43, 3_42, 2_18, 3_54], [2_31, 3_45, 2_89, 3_55], [2_02, 3_62, 2_27, 3_73], [1_43, 3_79, 2_20, 3_92], [2_31, 3_82, 2_91, 3_94], [1_44, 7_14, 2_20, 7_26], [2_31, 7_15, 2_56, 7_26], [1_44, 7_32, 2_20, 7_45], [2_32, 7_36, 2_91, 7_47], [1_44, 7_69, 2_18, 7_82], [2_31, 7_70, 2_56, 7_82], [1_41, 7_88, 2_02, 8_01], [2_15, 7_91, 2_74, 8_04], [1_43, 8_26, 2_04, 8_38], [2_15, 8_26, 2_40, 8_38], [1_42, 8_44, 2_02, 8_57], [2_15, 8_47, 2_74, 8_59], [3_34, 57, 4_27, 69], [4_40, 57, 5_22, 69], [3_69, 75, 4_61, 88], [4_69, 75, 5_16, 88], [5_28, 76, 5_62, 88], [5_70, 76, 6_67, 88], [6_75, 75, 7_11, 87], [7_21, 79, 7_78, 88], [7_89, 75, 8_40, 88], [3_69, 97, 4_70, 1_07], [4_84, 94, 5_07, 1_06], [5_18, 94, 5_62, 1_07], [5_76, 94, 6_55, 1_10], [6_68, 94, 7_92, 1_09], [8_04, 95, 8_29, 1_07], [3_69, 1_13, 4_65, 1_25], [4_77, 1_16, 5_47, 1_25], [5_62, 1_13, 6_58, 1_25], [6_71, 1_16, 7_48, 1_25], [7_61, 1_13, 8_11, 1_25], [3_69, 1_31, 4_65, 1_43], [4_77, 1_33, 5_48, 1_43], [5_63, 1_30, 6_98, 1_45], [7_10, 1_30, 8_02, 1_46], [3_36, 1_71, 4_12, 1_83], [4_23, 1_71, 5_72, 1_83], [5_82, 1_70, 7_16, 1_84], [7_28, 1_71, 8_17, 1_87], [8_29, 1_71, 8_44, 1_86], [3_38, 1_97, 4_82, 2_12], [5_07, 1_96, 5_57, 2_09], [5_69, 1_96, 5_95, 2_08], [6_10, 1_96, 7_02, 2_09], [5_05, 2_14, 5_83, 2_26], [5_95, 2_14, 6_56, 2_27], [6_70, 2_15, 8_07, 2_27], [3_35, 2_59, 5_43, 2_74], [5_56, 2_59, 7_08, 2_72], [3_72, 2_79, 4_22, 2_91], [4_35, 2_79, 4_60, 2_91], [4_74, 2_79, 5_74, 2_92], [5_87, 2_78, 6_64, 2_91], [6_76, 2_78, 7_38, 2_91], [7_51, 2_79, 8_34, 2_91], [3_72, 2_98, 4_34, 3_10], [3_35, 3_41, 4_83, 3_54], [4_97, 3_41, 6_55, 3_54], [6_67, 3_41, 7_28, 3_54], [7_40, 3_41, 8_25, 3_54], [3_35, 3_60, 4_30, 3_72], [4_42, 3_60, 5_34, 3_72], [5_45, 3_59, 6_87, 3_72], [6_97, 3_60, 7_54, 3_72], [7_65, 3_60, 8_23, 3_73], [3_34, 3_78, 4_28, 3_91], [4_40, 3_78, 5_77, 3_94], [5_90, 3_78, 7_05, 3_91], [7_20, 3_78, 8_01, 3_91], [3_34, 3_97, 4_00, 4_09], [3_70, 4_16, 5_29, 4_29], [5_44, 4_16, 5_76, 4_32], [5_87, 4_16, 6_65, 4_28], [6_77, 4_16, 8_14, 4_29], [3_72, 4_35, 4_52, 4_50], [4_65, 4_34, 4_95, 4_47], [5_11, 4_34, 6_00, 4_47], [6_11, 4_36, 6_37, 4_47], [6_49, 4_36, 6_94, 4_51], [7_05, 4_38, 8_24, 4_47], [3_69, 4_53, 4_52, 4_66], [4_64, 4_54, 5_09, 4_66], [5_22, 4_53, 6_11, 4_69], [6_25, 4_53, 7_92, 4_69], [3_70, 4_72, 5_56, 4_88], [5_70, 4_72, 6_84, 4_87], [6_97, 4_72, 7_18, 4_85], [7_32, 4_72, 8_35, 4_88], [3_69, 4_90, 4_11, 5_03], [4_25, 4_90, 4_84, 5_03], [4_96, 4_90, 6_35, 5_06], [6_45, 4_90, 7_07, 5_03], [7_18, 4_91, 7_61, 5_03], [7_71, 4_90, 8_40, 5_03], [3_36, 5_10, 3_74, 5_21], [3_88, 5_10, 4_47, 5_22], [4_60, 5_10, 4_89, 5_21], [5_03, 5_10, 5_80, 5_22], [5_92, 5_09, 7_36, 5_25], [7_45, 5_09, 7_70, 5_22], [7_81, 5_09, 8_40, 5_22], [3_38, 5_28, 4_34, 5_41], [4_48, 5_28, 5_96, 5_41], [6_09, 5_27, 6_87, 5_40], [7_00, 5_28, 7_92, 5_41], [3_36, 5_46, 3_97, 5_59], [4_07, 5_46, 4_31, 5_59], [4_43, 5_46, 5_25, 5_60], [5_37, 5_46, 6_80, 5_62], [6_88, 5_46, 7_14, 5_59], [7_22, 5_46, 8_37, 5_62], [3_36, 5_65, 4_49, 5_81], [4_61, 5_65, 4_85, 5_77], [4_97, 5_65, 6_65, 5_81], [6_81, 5_65, 7_18, 5_77], [7_32, 5_65, 8_37, 5_80], [3_37, 5_84, 4_38, 5_97], [4_52, 5_83, 5_21, 5_96], [5_35, 5_84, 6_77, 5_99], [6_90, 5_83, 7_87, 5_96], [8_01, 5_83, 8_25, 5_96], [3_38, 6_02, 4_78, 6_15], [4_92, 6_02, 5_30, 6_14], [5_43, 6_02, 6_38, 6_15], [6_50, 6_02, 6_76, 6_14], [6_88, 6_02, 7_88, 6_15], [8_02, 6_02, 8_43, 6_14], [3_37, 6_21, 5_02, 6_33], [5_16, 6_21, 6_15, 6_37], [6_29, 6_21, 7_74, 6_36], [7_89, 6_21, 8_27, 6_33], [3_37, 6_39, 4_18, 6_52], [4_32, 6_40, 5_71, 6_53], [5_87, 6_39, 7_31, 6_55], [7_43, 6_39, 7_69, 6_52], [7_80, 6_39, 8_41, 6_52], [3_38, 6_58, 4_40, 6_73], [4_55, 6_58, 4_91, 6_70], [5_08, 6_58, 6_02, 6_71], [6_16, 6_58, 6_38, 6_70], [6_54, 6_58, 8_35, 6_74], [3_37, 6_77, 4_29, 6_89], [3_37, 7_14, 4_82, 7_26], [4_95, 7_14, 5_48, 7_26], [5_61, 7_14, 6_83, 7_26], [3_38, 7_70, 4_61, 7_82], [4_74, 7_69, 5_54, 7_85], [4_89, 7_88, 5_62, 8_03], [5_76, 7_88, 6_43, 8_01], [6_56, 7_87, 7_51, 8_04], [7_64, 7_88, 8_44, 8_01], [3_34, 8_25, 4_21, 8_38], [4_30, 8_24, 5_74, 8_38], [5_84, 8_24, 7_23, 8_41], [3_35, 8_44, 4_50, 8_57], [4_64, 8_43, 5_83, 8_60], [6_28, 8_62, 7_55, 8_75], [7_69, 8_61, 8_48, 8_78]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , __snake_case ) self.assertListEqual(encoding.boxes , __snake_case ) # with apply_OCR = False snake_case = LayoutLMvaImageProcessor(apply_ocr=__snake_case ) snake_case = image_processing(__snake_case , return_tensors="""pt""" ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 2_24, 2_24) )
3
0
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from typing import Optional import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor from torchvision.transforms.functional import InterpolationMode import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, ViTImageProcessor, ViTMAEConfig, ViTMAEForPreTraining, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version a : List[Any] = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') require_version('datasets>=1.8.0', 'To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt') @dataclass class a : snake_case_ = field( default="cifar10" , metadata={"help": "Name of a dataset from the datasets package"} ) snake_case_ = field( default=_lowerCamelCase , metadata={"help": "The configuration name of the dataset to use (via the datasets library)."} ) snake_case_ = field( default=_lowerCamelCase , metadata={"help": "The column name of the images in the files."} ) snake_case_ = field(default=_lowerCamelCase , metadata={"help": "A folder containing the training data."} ) snake_case_ = field(default=_lowerCamelCase , metadata={"help": "A folder containing the validation data."} ) snake_case_ = field( default=0.15 , metadata={"help": "Percent to split off of train for validation."} ) snake_case_ = field( default=_lowerCamelCase , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) snake_case_ = field( default=_lowerCamelCase , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) def A_ ( self : Any ): snake_case_ = {} if self.train_dir is not None: snake_case_ = self.train_dir if self.validation_dir is not None: snake_case_ = self.validation_dir snake_case_ = data_files if data_files else None @dataclass class a : snake_case_ = field( default=_lowerCamelCase , metadata={ "help": ( "The model checkpoint for weights initialization.Don't set if you want to train a model from scratch." ) } , ) snake_case_ = field( default=_lowerCamelCase , metadata={"help": "Pretrained config name or path if not the same as model_name_or_path"} ) snake_case_ = field( default=_lowerCamelCase , metadata={ "help": ( "Override some existing default config settings when a model is trained from scratch. Example: " "n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index" ) } , ) snake_case_ = field( default=_lowerCamelCase , metadata={"help": "Where do you want to store the pretrained models downloaded from s3"} ) snake_case_ = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) snake_case_ = field(default=_lowerCamelCase , metadata={"help": "Name or path of preprocessor config."} ) snake_case_ = field( default=_lowerCamelCase , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) snake_case_ = field( default=0.75 , metadata={"help": "The ratio of the number of masked tokens in the input sequence."} ) snake_case_ = field( default=_lowerCamelCase , metadata={"help": "Whether or not to train with normalized pixel values as target."} ) @dataclass class a ( _lowerCamelCase ): snake_case_ = field( default=1e-3 , metadata={"help": "Base learning rate: absolute_lr = base_lr * total_batch_size / 256."} ) def __magic_name__ ( __UpperCAmelCase ) -> int: '''simple docstring''' snake_case_ = torch.stack([example['''pixel_values'''] for example in examples] ) return {"pixel_values": pixel_values} def __magic_name__ ( ) -> Optional[Any]: '''simple docstring''' snake_case_ = HfArgumentParser((ModelArguments, DataTrainingArguments, CustomTrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. snake_case_ ,snake_case_ ,snake_case_ = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: snake_case_ ,snake_case_ ,snake_case_ = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_mae''', __UpperCAmelCase, __UpperCAmelCase ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', handlers=[logging.StreamHandler(sys.stdout )], ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() snake_case_ = training_args.get_process_log_level() logger.setLevel(__UpperCAmelCase ) transformers.utils.logging.set_verbosity(__UpperCAmelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F"Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}" + F"distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}" ) logger.info(F"Training/evaluation parameters {training_args}" ) # Detecting last checkpoint. snake_case_ = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: snake_case_ = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"Output directory ({training_args.output_dir}) already exists and is not empty. " '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F"Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change " '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Initialize our dataset. snake_case_ = load_dataset( data_args.dataset_name, data_args.dataset_config_name, data_files=data_args.data_files, cache_dir=model_args.cache_dir, use_auth_token=True if model_args.use_auth_token else None, ) # If we don't have a validation split, split off a percentage of train as validation. snake_case_ = None if '''validation''' in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split, __UpperCAmelCase ) and data_args.train_val_split > 0.0: snake_case_ = ds['''train'''].train_test_split(data_args.train_val_split ) snake_case_ = split['''train'''] snake_case_ = split['''test'''] # Load pretrained model and image processor # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case_ = { '''cache_dir''': model_args.cache_dir, '''revision''': model_args.model_revision, '''use_auth_token''': True if model_args.use_auth_token else None, } if model_args.config_name: snake_case_ = ViTMAEConfig.from_pretrained(model_args.config_name, **__UpperCAmelCase ) elif model_args.model_name_or_path: snake_case_ = ViTMAEConfig.from_pretrained(model_args.model_name_or_path, **__UpperCAmelCase ) else: snake_case_ = ViTMAEConfig() logger.warning('''You are instantiating a new config instance from scratch.''' ) if model_args.config_overrides is not None: logger.info(F"Overriding config: {model_args.config_overrides}" ) config.update_from_string(model_args.config_overrides ) logger.info(F"New config: {config}" ) # adapt config config.update( { '''mask_ratio''': model_args.mask_ratio, '''norm_pix_loss''': model_args.norm_pix_loss, } ) # create image processor if model_args.image_processor_name: snake_case_ = ViTImageProcessor.from_pretrained(model_args.image_processor_name, **__UpperCAmelCase ) elif model_args.model_name_or_path: snake_case_ = ViTImageProcessor.from_pretrained(model_args.model_name_or_path, **__UpperCAmelCase ) else: snake_case_ = ViTImageProcessor() # create model if model_args.model_name_or_path: snake_case_ = ViTMAEForPreTraining.from_pretrained( model_args.model_name_or_path, from_tf=bool('''.ckpt''' in model_args.model_name_or_path ), config=__UpperCAmelCase, cache_dir=model_args.cache_dir, revision=model_args.model_revision, use_auth_token=True if model_args.use_auth_token else None, ) else: logger.info('''Training new model from scratch''' ) snake_case_ = ViTMAEForPreTraining(__UpperCAmelCase ) if training_args.do_train: snake_case_ = ds['''train'''].column_names else: snake_case_ = ds['''validation'''].column_names if data_args.image_column_name is not None: snake_case_ = data_args.image_column_name elif "image" in column_names: snake_case_ = '''image''' elif "img" in column_names: snake_case_ = '''img''' else: snake_case_ = column_names[0] # transformations as done in original MAE paper # source: https://github.com/facebookresearch/mae/blob/main/main_pretrain.py if "shortest_edge" in image_processor.size: snake_case_ = image_processor.size['''shortest_edge'''] else: snake_case_ = (image_processor.size['''height'''], image_processor.size['''width''']) snake_case_ = Compose( [ Lambda(lambda __UpperCAmelCase : img.convert('''RGB''' ) if img.mode != "RGB" else img ), RandomResizedCrop(__UpperCAmelCase, scale=(0.2, 1.0), interpolation=InterpolationMode.BICUBIC ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean, std=image_processor.image_std ), ] ) def preprocess_images(__UpperCAmelCase ): snake_case_ = [transforms(__UpperCAmelCase ) for image in examples[image_column_name]] return examples if training_args.do_train: if "train" not in ds: raise ValueError('''--do_train requires a train dataset''' ) if data_args.max_train_samples is not None: snake_case_ = ds['''train'''].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(__UpperCAmelCase ) if training_args.do_eval: if "validation" not in ds: raise ValueError('''--do_eval requires a validation dataset''' ) if data_args.max_eval_samples is not None: snake_case_ = ( ds['''validation'''].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(__UpperCAmelCase ) # Compute absolute learning rate snake_case_ = ( training_args.train_batch_size * training_args.gradient_accumulation_steps * training_args.world_size ) if training_args.base_learning_rate is not None: snake_case_ = training_args.base_learning_rate * total_train_batch_size / 256 # Initialize our trainer snake_case_ = Trainer( model=__UpperCAmelCase, args=__UpperCAmelCase, train_dataset=ds['''train'''] if training_args.do_train else None, eval_dataset=ds['''validation'''] if training_args.do_eval else None, tokenizer=__UpperCAmelCase, data_collator=__UpperCAmelCase, ) # Training if training_args.do_train: snake_case_ = None if training_args.resume_from_checkpoint is not None: snake_case_ = training_args.resume_from_checkpoint elif last_checkpoint is not None: snake_case_ = last_checkpoint snake_case_ = trainer.train(resume_from_checkpoint=__UpperCAmelCase ) trainer.save_model() trainer.log_metrics('''train''', train_result.metrics ) trainer.save_metrics('''train''', train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: snake_case_ = trainer.evaluate() trainer.log_metrics('''eval''', __UpperCAmelCase ) trainer.save_metrics('''eval''', __UpperCAmelCase ) # Write model card and (optionally) push to hub snake_case_ = { '''tasks''': '''masked-auto-encoding''', '''dataset''': data_args.dataset_name, '''tags''': ['''masked-auto-encoding'''], } if training_args.push_to_hub: trainer.push_to_hub(**__UpperCAmelCase ) else: trainer.create_model_card(**__UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' main() if __name__ == "__main__": main()
56
'''simple docstring''' from __future__ import annotations import math def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ) -> int: '''simple docstring''' if depth < 0: raise ValueError('''Depth cannot be less than 0''' ) if len(__UpperCAmelCase ) == 0: raise ValueError('''Scores cannot be empty''' ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1, node_index * 2, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ), minimax(depth + 1, node_index * 2 + 1, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ), ) return min( minimax(depth + 1, node_index * 2, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ), minimax(depth + 1, node_index * 2 + 1, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ), ) def __magic_name__ ( ) -> None: '''simple docstring''' snake_case_ = [90, 23, 6, 33, 21, 65, 123, 3_4423] snake_case_ = math.log(len(__UpperCAmelCase ), 2 ) print('''Optimal value : ''', end='''''' ) print(minimax(0, 0, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
56
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __A = logging.get_logger(__name__) __A = { "facebook/vit-mae-base": "https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json", # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" _UpperCAmelCase :Optional[Any] = "vit_mae" def __init__( self , _UpperCAmelCase=768 , _UpperCAmelCase=12 , _UpperCAmelCase=12 , _UpperCAmelCase=3072 , _UpperCAmelCase="gelu" , _UpperCAmelCase=0.0 , _UpperCAmelCase=0.0 , _UpperCAmelCase=0.02 , _UpperCAmelCase=1e-1_2 , _UpperCAmelCase=224 , _UpperCAmelCase=16 , _UpperCAmelCase=3 , _UpperCAmelCase=True , _UpperCAmelCase=16 , _UpperCAmelCase=512 , _UpperCAmelCase=8 , _UpperCAmelCase=2048 , _UpperCAmelCase=0.75 , _UpperCAmelCase=False , **_UpperCAmelCase , ): super().__init__(**_UpperCAmelCase ) lowercase__: str = hidden_size lowercase__: int = num_hidden_layers lowercase__: int = num_attention_heads lowercase__: List[str] = intermediate_size lowercase__: str = hidden_act lowercase__: Optional[Any] = hidden_dropout_prob lowercase__: Optional[int] = attention_probs_dropout_prob lowercase__: Tuple = initializer_range lowercase__: Tuple = layer_norm_eps lowercase__: int = image_size lowercase__: Optional[Any] = patch_size lowercase__: Dict = num_channels lowercase__: Tuple = qkv_bias lowercase__: List[str] = decoder_num_attention_heads lowercase__: List[Any] = decoder_hidden_size lowercase__: Dict = decoder_num_hidden_layers lowercase__: Dict = decoder_intermediate_size lowercase__: Optional[Any] = mask_ratio lowercase__: Optional[int] = norm_pix_loss
2
"""simple docstring""" import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger __A = "<<<<<<< This should probably be modified because it mentions: " __A = "=======\n>>>>>>>\n" __A = [ "TextEncoderConfig", "ByteTextEncoder", "SubwordTextEncoder", "encoder_config", "maybe_build_from_corpus", "manual_dir", ] __A = [ # (pattern, replacement) # Order is important here for some replacements (R"tfds\.core", R"datasets"), (R"tf\.io\.gfile\.GFile", R"open"), (R"tf\.([\w\d]+)", R"datasets.Value('\1')"), (R"tfds\.features\.Text\(\)", R"datasets.Value('string')"), (R"tfds\.features\.Text\(", R"datasets.Value('string'),"), (R"features\s*=\s*tfds.features.FeaturesDict\(", R"features=datasets.Features("), (R"tfds\.features\.FeaturesDict\(", R"dict("), (R"The TensorFlow Datasets Authors", R"The TensorFlow Datasets Authors and the HuggingFace Datasets Authors"), (R"tfds\.", R"datasets."), (R"dl_manager\.manual_dir", R"self.config.data_dir"), (R"self\.builder_config", R"self.config"), ] def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ) -> Tuple: return ConvertCommand(args.tfds_path , args.datasets_directory ) class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" @staticmethod def _snake_case ( _UpperCAmelCase ): lowercase__: int = parser.add_parser( '''convert''' , help='''Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.''' , ) train_parser.add_argument( '''--tfds_path''' , type=_UpperCAmelCase , required=_UpperCAmelCase , help='''Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.''' , ) train_parser.add_argument( '''--datasets_directory''' , type=_UpperCAmelCase , required=_UpperCAmelCase , help='''Path to the HuggingFace Datasets folder.''' ) train_parser.set_defaults(func=_UpperCAmelCase ) def __init__( self , _UpperCAmelCase , _UpperCAmelCase , *_UpperCAmelCase ): lowercase__: List[str] = get_logger('''datasets-cli/converting''' ) lowercase__: Optional[Any] = tfds_path lowercase__: Dict = datasets_directory def _snake_case ( self ): if os.path.isdir(self._tfds_path ): lowercase__: Optional[Any] = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): lowercase__: Optional[int] = os.path.dirname(self._tfds_path ) else: raise ValueError('''--tfds_path is neither a directory nor a file. Please check path.''' ) lowercase__: int = os.path.abspath(self._datasets_directory ) self._logger.info(F"""Converting datasets from {abs_tfds_path} to {abs_datasets_path}""" ) lowercase__: Tuple = [] lowercase__: Dict = [] lowercase__: Any = {} if os.path.isdir(self._tfds_path ): lowercase__: Dict = os.listdir(_UpperCAmelCase ) else: lowercase__: Dict = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(F"""Looking at file {f_name}""" ) lowercase__: Tuple = os.path.join(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Optional[int] = os.path.join(_UpperCAmelCase , _UpperCAmelCase ) if not os.path.isfile(_UpperCAmelCase ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info('''Skipping file''' ) continue with open(_UpperCAmelCase , encoding='''utf-8''' ) as f: lowercase__: Tuple = f.readlines() lowercase__: Optional[Any] = [] lowercase__: Dict = False lowercase__: List[str] = False lowercase__: List[Any] = [] for line in lines: lowercase__: List[str] = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: lowercase__: Optional[int] = '''import datasets\n''' elif "import tensorflow" in out_line: # order is important here lowercase__: Dict = '''''' continue elif "from absl import logging" in out_line: lowercase__: Tuple = '''from datasets import logging\n''' elif "getLogger" in out_line: lowercase__: Optional[Any] = out_line.replace('''getLogger''' , '''get_logger''' ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): lowercase__: Any = True lowercase__: str = list(filter(lambda _UpperCAmelCase : e in out_line , _UpperCAmelCase ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(_UpperCAmelCase ) + '''\n''' ) out_lines.append(_UpperCAmelCase ) out_lines.append(_UpperCAmelCase ) continue else: for pattern, replacement in TO_CONVERT: lowercase__: List[Any] = re.sub(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: lowercase__: Any = re.match(r'''from\stensorflow_datasets.*import\s([^\.\r\n]+)''' , _UpperCAmelCase ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split(''',''' ) ) lowercase__: List[str] = '''from . import ''' + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(F"""Error converting {out_line.strip()}""" ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: lowercase__: Optional[Any] = True out_lines.append(_UpperCAmelCase ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset lowercase__: Dict = f_name.replace('''.py''' , '''''' ) lowercase__: Dict = os.path.join(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Optional[Any] = os.path.join(_UpperCAmelCase , _UpperCAmelCase ) os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) self._logger.info(F"""Adding directory {output_dir}""" ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(_UpperCAmelCase ) if needs_manual_update: with_manual_update.append(_UpperCAmelCase ) with open(_UpperCAmelCase , '''w''' , encoding='''utf-8''' ) as f: f.writelines(_UpperCAmelCase ) self._logger.info(F"""Converted in {output_file}""" ) for utils_file in utils_files: try: lowercase__: str = os.path.basename(_UpperCAmelCase ) lowercase__: Union[str, Any] = imports_to_builder_map[f_name.replace('''.py''' , '''''' )] self._logger.info(F"""Moving {dest_folder} to {utils_file}""" ) shutil.copy(_UpperCAmelCase , _UpperCAmelCase ) except KeyError: self._logger.error(F"""Cannot find destination folder for {utils_file}. Please copy manually.""" ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( F"""You need to manually update file {file_path} to remove configurations using 'TextEncoderConfig'.""" )
2
1
"""simple docstring""" import inspect import unittest from transformers import DPTConfig from transformers.file_utils import is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MODEL_MAPPING, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel from transformers.models.dpt.modeling_dpt import DPT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class UpperCAmelCase_ : def __init__( self : str , __UpperCamelCase : Dict , __UpperCamelCase : List[Any]=2 , __UpperCamelCase : int=32 , __UpperCamelCase : int=16 , __UpperCamelCase : List[Any]=3 , __UpperCamelCase : int=True , __UpperCamelCase : str=True , __UpperCamelCase : Optional[int]=32 , __UpperCamelCase : Optional[int]=4 , __UpperCamelCase : List[Any]=[0, 1, 2, 3] , __UpperCamelCase : str=4 , __UpperCamelCase : int=37 , __UpperCamelCase : Union[str, Any]="gelu" , __UpperCamelCase : Any=0.1 , __UpperCamelCase : Optional[int]=0.1 , __UpperCamelCase : Any=0.0_2 , __UpperCamelCase : int=3 , __UpperCamelCase : List[str]=[1, 384, 24, 24] , __UpperCamelCase : List[Any]=True , __UpperCamelCase : str=None , ) -> Any: _UpperCamelCase = parent _UpperCamelCase = batch_size _UpperCamelCase = image_size _UpperCamelCase = patch_size _UpperCamelCase = num_channels _UpperCamelCase = is_training _UpperCamelCase = use_labels _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = backbone_out_indices _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = initializer_range _UpperCamelCase = num_labels _UpperCamelCase = backbone_featmap_shape _UpperCamelCase = scope _UpperCamelCase = is_hybrid # sequence length of DPT = num_patches + 1 (we add 1 for the [CLS] token) _UpperCamelCase = (image_size // patch_size) ** 2 _UpperCamelCase = num_patches + 1 def _UpperCamelCase ( self : int ) -> Tuple: _UpperCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCamelCase = None if self.use_labels: _UpperCamelCase = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) _UpperCamelCase = self.get_config() return config, pixel_values, labels def _UpperCamelCase ( self : str ) -> Optional[int]: _UpperCamelCase = { '''global_padding''': '''same''', '''layer_type''': '''bottleneck''', '''depths''': [3, 4, 9], '''out_features''': ['''stage1''', '''stage2''', '''stage3'''], '''embedding_dynamic_padding''': True, '''hidden_sizes''': [96, 192, 384, 768], '''num_groups''': 2, } return DPTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , backbone_out_indices=self.backbone_out_indices , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__UpperCamelCase , initializer_range=self.initializer_range , is_hybrid=self.is_hybrid , backbone_config=__UpperCamelCase , backbone_featmap_shape=self.backbone_featmap_shape , ) def _UpperCamelCase ( self : Dict , __UpperCamelCase : Dict , __UpperCamelCase : Any , __UpperCamelCase : int ) -> Tuple: _UpperCamelCase = DPTModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() _UpperCamelCase = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Optional[int] , __UpperCamelCase : Union[str, Any] , __UpperCamelCase : List[Any] , __UpperCamelCase : Tuple ) -> Union[str, Any]: _UpperCamelCase = self.num_labels _UpperCamelCase = DPTForDepthEstimation(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() _UpperCamelCase = model(__UpperCamelCase ) self.parent.assertEqual(result.predicted_depth.shape , (self.batch_size, self.image_size, self.image_size) ) def _UpperCamelCase ( self : Tuple , __UpperCamelCase : str , __UpperCamelCase : Optional[int] , __UpperCamelCase : List[Any] ) -> Union[str, Any]: _UpperCamelCase = self.num_labels _UpperCamelCase = DPTForSemanticSegmentation(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() _UpperCamelCase = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def _UpperCamelCase ( self : Optional[int] ) -> List[Any]: _UpperCamelCase = self.prepare_config_and_inputs() _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = config_and_inputs _UpperCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class UpperCAmelCase_ ( _lowercase , _lowercase , unittest.TestCase): snake_case__ = (DPTModel, DPTForDepthEstimation, DPTForSemanticSegmentation) if is_torch_available() else () snake_case__ = ( { '''depth-estimation''': DPTForDepthEstimation, '''feature-extraction''': DPTModel, '''image-segmentation''': DPTForSemanticSegmentation, } if is_torch_available() else {} ) snake_case__ = False snake_case__ = False snake_case__ = False def _UpperCamelCase ( self : List[str] ) -> List[str]: _UpperCamelCase = DPTModelTester(self ) _UpperCamelCase = ConfigTester(self , config_class=__UpperCamelCase , has_text_modality=__UpperCamelCase , hidden_size=37 ) def _UpperCamelCase ( self : int ) -> Optional[int]: self.config_tester.run_common_tests() @unittest.skip(reason='''DPT does not use inputs_embeds''' ) def _UpperCamelCase ( self : List[str] ) -> int: pass def _UpperCamelCase ( self : Dict ) -> Tuple: _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _UpperCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCamelCase , nn.Linear ) ) def _UpperCamelCase ( self : int ) -> Any: _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase = model_class(__UpperCamelCase ) _UpperCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase = [*signature.parameters.keys()] _UpperCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def _UpperCamelCase ( self : Dict ) -> Optional[int]: _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def _UpperCamelCase ( self : Tuple ) -> List[Any]: _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_depth_estimation(*__UpperCamelCase ) def _UpperCamelCase ( self : Union[str, Any] ) -> Optional[int]: _UpperCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*__UpperCamelCase ) def _UpperCamelCase ( self : Union[str, Any] ) -> List[Any]: for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = True if model_class in get_values(__UpperCamelCase ): continue _UpperCamelCase = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.train() _UpperCamelCase = self._prepare_for_class(__UpperCamelCase , __UpperCamelCase , return_labels=__UpperCamelCase ) _UpperCamelCase = model(**__UpperCamelCase ).loss loss.backward() def _UpperCamelCase ( self : str ) -> Any: for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = False _UpperCamelCase = True if model_class in get_values(__UpperCamelCase ) or not model_class.supports_gradient_checkpointing: continue _UpperCamelCase = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.gradient_checkpointing_enable() model.train() _UpperCamelCase = self._prepare_for_class(__UpperCamelCase , __UpperCamelCase , return_labels=__UpperCamelCase ) _UpperCamelCase = model(**__UpperCamelCase ).loss loss.backward() def _UpperCamelCase ( self : int ) -> int: _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = _config_zero_init(__UpperCamelCase ) for model_class in self.all_model_classes: _UpperCamelCase = model_class(config=__UpperCamelCase ) # Skip the check for the backbone _UpperCamelCase = [] for name, module in model.named_modules(): if module.__class__.__name__ == "DPTViTHybridEmbeddings": _UpperCamelCase = [F'''{name}.{key}''' for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def _UpperCamelCase ( self : Optional[Any] ) -> str: pass @slow def _UpperCamelCase ( self : Tuple ) -> Any: for model_name in DPT_PRETRAINED_MODEL_ARCHIVE_LIST[1:]: _UpperCamelCase = DPTModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def _UpperCamelCase ( self : Dict ) -> Dict: # We do this test only for DPTForDepthEstimation since it is the only model that uses readout_type _UpperCamelCase , _UpperCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase = '''add''' with self.assertRaises(__UpperCamelCase ): _UpperCamelCase = DPTForDepthEstimation(__UpperCamelCase ) def lowercase ( ) -> Dict: _UpperCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision @slow class UpperCAmelCase_ ( unittest.TestCase): def _UpperCamelCase ( self : List[str] ) -> List[Any]: _UpperCamelCase = DPTImageProcessor.from_pretrained('''Intel/dpt-hybrid-midas''' ) _UpperCamelCase = DPTForDepthEstimation.from_pretrained('''Intel/dpt-hybrid-midas''' ).to(__UpperCamelCase ) _UpperCamelCase = prepare_img() _UpperCamelCase = image_processor(images=__UpperCamelCase , return_tensors='''pt''' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): _UpperCamelCase = model(**__UpperCamelCase ) _UpperCamelCase = outputs.predicted_depth # verify the predicted depth _UpperCamelCase = torch.Size((1, 384, 384) ) self.assertEqual(predicted_depth.shape , __UpperCamelCase ) _UpperCamelCase = torch.tensor( [[[5.6_4_3_7, 5.6_1_4_6, 5.6_5_1_1], [5.4_3_7_1, 5.5_6_4_9, 5.5_9_5_8], [5.5_2_1_5, 5.5_1_8_4, 5.5_2_9_3]]] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.predicted_depth[:3, :3, :3] / 100 , __UpperCamelCase , atol=1E-4 ) )
256
"""simple docstring""" from typing import List import numpy as np def lowercase ( a__ : dict ) -> int: _UpperCamelCase = {key: len(a__ ) for key, value in gen_kwargs.items() if isinstance(a__ , a__ )} if len(set(lists_lengths.values() ) ) > 1: raise RuntimeError( ( '''Sharding is ambiguous for this dataset: ''' + '''we found several data sources lists of different lengths, and we don\'t know over which list we should parallelize:\n''' + '''\n'''.join(F'''\t- key {key} has length {length}''' for key, length in lists_lengths.items() ) + '''\nTo fix this, check the \'gen_kwargs\' and make sure to use lists only for data sources, ''' + '''and use tuples otherwise. In the end there should only be one single list, or several lists with the same length.''' ) ) _UpperCamelCase = max(lists_lengths.values() , default=0 ) return max(1 , a__ ) def lowercase ( a__ : int , a__ : int ) -> List[range]: _UpperCamelCase = [] for group_idx in range(a__ ): _UpperCamelCase = num_shards // max_num_jobs + (group_idx < (num_shards % max_num_jobs)) if num_shards_to_add == 0: break _UpperCamelCase = shards_indices_per_group[-1].stop if shards_indices_per_group else 0 _UpperCamelCase = range(a__ , start + num_shards_to_add ) shards_indices_per_group.append(a__ ) return shards_indices_per_group def lowercase ( a__ : dict , a__ : int ) -> List[dict]: _UpperCamelCase = _number_of_shards_in_gen_kwargs(a__ ) if num_shards == 1: return [dict(a__ )] else: _UpperCamelCase = _distribute_shards(num_shards=a__ , max_num_jobs=a__ ) return [ { key: [value[shard_idx] for shard_idx in shard_indices_per_group[group_idx]] if isinstance(a__ , a__ ) else value for key, value in gen_kwargs.items() } for group_idx in range(len(a__ ) ) ] def lowercase ( a__ : List[dict] ) -> dict: return { key: [value for gen_kwargs in gen_kwargs_list for value in gen_kwargs[key]] if isinstance(gen_kwargs_list[0][key] , a__ ) else gen_kwargs_list[0][key] for key in gen_kwargs_list[0] } def lowercase ( a__ : np.random.Generator , a__ : dict ) -> dict: _UpperCamelCase = {len(a__ ) for value in gen_kwargs.values() if isinstance(a__ , a__ )} _UpperCamelCase = {} for size in list_sizes: _UpperCamelCase = list(range(a__ ) ) rng.shuffle(indices_per_size[size] ) # Now let's copy the gen_kwargs and shuffle the lists based on their sizes _UpperCamelCase = dict(a__ ) for key, value in shuffled_kwargs.items(): if isinstance(a__ , a__ ): _UpperCamelCase = [value[i] for i in indices_per_size[len(a__ )]] return shuffled_kwargs
256
1
import re def __UpperCamelCase (_SCREAMING_SNAKE_CASE ) -> bool: lowercase__ = re.compile( R'^(?:0|94|\+94|0{2}94)' R'7(0|1|2|4|5|6|7|8)' R'(-| |)' R'\d{7}$' ) return bool(re.search(__UpperCAmelCase , __UpperCAmelCase ) ) if __name__ == "__main__": lowercase_ = '0094702343221' print(is_sri_lankan_phone_number(phone))
356
from typing import TYPE_CHECKING from ..utils import _LazyModule lowercase_ = { """config""": [ """EXTERNAL_DATA_FORMAT_SIZE_LIMIT""", """OnnxConfig""", """OnnxConfigWithPast""", """OnnxSeq2SeqConfigWithPast""", """PatchingSpec""", ], """convert""": ["""export""", """validate_model_outputs"""], """features""": ["""FeaturesManager"""], """utils""": ["""ParameterFormat""", """compute_serialized_parameters_size"""], } if TYPE_CHECKING: from .config import ( EXTERNAL_DATA_FORMAT_SIZE_LIMIT, OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast, PatchingSpec, ) from .convert import export, validate_model_outputs from .features import FeaturesManager from .utils import ParameterFormat, compute_serialized_parameters_size else: import sys lowercase_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
269
0
import darl # noqa import gym import tqdm from diffusers.experimental import ValueGuidedRLPipeline __snake_case :List[str] = { '''n_samples''': 64, '''horizon''': 32, '''num_inference_steps''': 20, '''n_guide_steps''': 2, # can set to 0 for faster sampling, does not use value network '''scale_grad_by_std''': True, '''scale''': 0.1, '''eta''': 0.0, '''t_grad_cutoff''': 2, '''device''': '''cpu''', } if __name__ == "__main__": __snake_case :Union[str, Any] = '''hopper-medium-v2''' __snake_case :Optional[int] = gym.make(env_name) __snake_case :Optional[Any] = ValueGuidedRLPipeline.from_pretrained( '''bglick13/hopper-medium-v2-value-function-hor32''', env=env, ) env.seed(0) __snake_case :Any = env.reset() __snake_case :Tuple = 0 __snake_case :Union[str, Any] = 0 __snake_case :Any = 1000 __snake_case :List[Any] = [obs.copy()] try: for t in tqdm.tqdm(range(T)): # call the policy __snake_case :Tuple = pipeline(obs, planning_horizon=32) # execute action in environment __snake_case ,__snake_case ,__snake_case ,__snake_case :List[Any] = env.step(denorm_actions) __snake_case :Dict = env.get_normalized_score(total_reward) # update return total_reward += reward total_score += score print( f'Step: {t}, Reward: {reward}, Total Reward: {total_reward}, Score: {score}, Total Score:' f' {total_score}' ) # save observations for rendering rollout.append(next_observation.copy()) __snake_case :Tuple = next_observation except KeyboardInterrupt: pass print(f'Total reward: {total_reward}')
49
'''simple docstring''' import argparse import os import shutil from pathlib import Path import onnx import torch from packaging import version from torch.onnx import export from diffusers import OnnxRuntimeModel, OnnxStableDiffusionPipeline, StableDiffusionPipeline _UpperCamelCase = version.parse(version.parse(torch.__version__).base_version) < version.parse('''1.11''') def lowercase_ ( lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : tuple , lowerCAmelCase__ : Path , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Optional[int]=False , ): """simple docstring""" output_path.parent.mkdir(parents=lowerCAmelCase__ , exist_ok=lowerCAmelCase__ ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( lowerCAmelCase__ , lowerCAmelCase__ , f=output_path.as_posix() , input_names=lowerCAmelCase__ , output_names=lowerCAmelCase__ , dynamic_axes=lowerCAmelCase__ , do_constant_folding=lowerCAmelCase__ , use_external_data_format=lowerCAmelCase__ , enable_onnx_checker=lowerCAmelCase__ , opset_version=lowerCAmelCase__ , ) else: export( lowerCAmelCase__ , lowerCAmelCase__ , f=output_path.as_posix() , input_names=lowerCAmelCase__ , output_names=lowerCAmelCase__ , dynamic_axes=lowerCAmelCase__ , do_constant_folding=lowerCAmelCase__ , opset_version=lowerCAmelCase__ , ) @torch.no_grad() def lowercase_ ( lowerCAmelCase__ : str , lowerCAmelCase__ : str , lowerCAmelCase__ : int , lowerCAmelCase__ : bool = False ): """simple docstring""" __UpperCAmelCase : Tuple = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): __UpperCAmelCase : Optional[int] = """cuda""" elif fpaa and not torch.cuda.is_available(): raise ValueError("""`float16` model export is only supported on GPUs with CUDA""" ) else: __UpperCAmelCase : Dict = """cpu""" __UpperCAmelCase : Optional[Any] = StableDiffusionPipeline.from_pretrained(lowerCAmelCase__ , torch_dtype=lowerCAmelCase__ ).to(lowerCAmelCase__ ) __UpperCAmelCase : List[str] = Path(lowerCAmelCase__ ) # TEXT ENCODER __UpperCAmelCase : Any = pipeline.text_encoder.config.max_position_embeddings __UpperCAmelCase : str = pipeline.text_encoder.config.hidden_size __UpperCAmelCase : Optional[Any] = pipeline.tokenizer( """A sample prompt""" , padding="""max_length""" , max_length=pipeline.tokenizer.model_max_length , truncation=lowerCAmelCase__ , return_tensors="""pt""" , ) onnx_export( pipeline.text_encoder , model_args=(text_input.input_ids.to(device=lowerCAmelCase__ , dtype=torch.intaa )) , output_path=output_path / """text_encoder""" / """model.onnx""" , ordered_input_names=["""input_ids"""] , output_names=["""last_hidden_state""", """pooler_output"""] , dynamic_axes={ """input_ids""": {0: """batch""", 1: """sequence"""}, } , opset=lowerCAmelCase__ , ) del pipeline.text_encoder # UNET __UpperCAmelCase : Optional[int] = pipeline.unet.config.in_channels __UpperCAmelCase : Tuple = pipeline.unet.config.sample_size __UpperCAmelCase : Dict = output_path / """unet""" / """model.onnx""" onnx_export( pipeline.unet , model_args=( torch.randn(2 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), torch.randn(2 ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), torch.randn(2 , lowerCAmelCase__ , lowerCAmelCase__ ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), False, ) , output_path=lowerCAmelCase__ , ordered_input_names=["""sample""", """timestep""", """encoder_hidden_states""", """return_dict"""] , output_names=["""out_sample"""] , dynamic_axes={ """sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, """timestep""": {0: """batch"""}, """encoder_hidden_states""": {0: """batch""", 1: """sequence"""}, } , opset=lowerCAmelCase__ , use_external_data_format=lowerCAmelCase__ , ) __UpperCAmelCase : Any = str(unet_path.absolute().as_posix() ) __UpperCAmelCase : int = os.path.dirname(lowerCAmelCase__ ) __UpperCAmelCase : Tuple = onnx.load(lowerCAmelCase__ ) # clean up existing tensor files shutil.rmtree(lowerCAmelCase__ ) os.mkdir(lowerCAmelCase__ ) # collate external tensor files into one onnx.save_model( lowerCAmelCase__ , lowerCAmelCase__ , save_as_external_data=lowerCAmelCase__ , all_tensors_to_one_file=lowerCAmelCase__ , location="""weights.pb""" , convert_attribute=lowerCAmelCase__ , ) del pipeline.unet # VAE ENCODER __UpperCAmelCase : Union[str, Any] = pipeline.vae __UpperCAmelCase : str = vae_encoder.config.in_channels __UpperCAmelCase : Any = vae_encoder.config.sample_size # need to get the raw tensor output (sample) from the encoder __UpperCAmelCase : str = lambda lowerCAmelCase__ , lowerCAmelCase__ : vae_encoder.encode(lowerCAmelCase__ , lowerCAmelCase__ )[0].sample() onnx_export( lowerCAmelCase__ , model_args=( torch.randn(1 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), False, ) , output_path=output_path / """vae_encoder""" / """model.onnx""" , ordered_input_names=["""sample""", """return_dict"""] , output_names=["""latent_sample"""] , dynamic_axes={ """sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, } , opset=lowerCAmelCase__ , ) # VAE DECODER __UpperCAmelCase : Optional[Any] = pipeline.vae __UpperCAmelCase : Optional[int] = vae_decoder.config.latent_channels __UpperCAmelCase : Dict = vae_decoder.config.out_channels # forward only through the decoder part __UpperCAmelCase : List[Any] = vae_encoder.decode onnx_export( lowerCAmelCase__ , model_args=( torch.randn(1 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), False, ) , output_path=output_path / """vae_decoder""" / """model.onnx""" , ordered_input_names=["""latent_sample""", """return_dict"""] , output_names=["""sample"""] , dynamic_axes={ """latent_sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, } , opset=lowerCAmelCase__ , ) del pipeline.vae # SAFETY CHECKER if pipeline.safety_checker is not None: __UpperCAmelCase : Tuple = pipeline.safety_checker __UpperCAmelCase : Union[str, Any] = safety_checker.config.vision_config.num_channels __UpperCAmelCase : Any = safety_checker.config.vision_config.image_size __UpperCAmelCase : Optional[int] = safety_checker.forward_onnx onnx_export( pipeline.safety_checker , model_args=( torch.randn( 1 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), torch.randn(1 , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ).to(device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ), ) , output_path=output_path / """safety_checker""" / """model.onnx""" , ordered_input_names=["""clip_input""", """images"""] , output_names=["""out_images""", """has_nsfw_concepts"""] , dynamic_axes={ """clip_input""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, """images""": {0: """batch""", 1: """height""", 2: """width""", 3: """channels"""}, } , opset=lowerCAmelCase__ , ) del pipeline.safety_checker __UpperCAmelCase : Optional[Any] = OnnxRuntimeModel.from_pretrained(output_path / """safety_checker""" ) __UpperCAmelCase : Any = pipeline.feature_extractor else: __UpperCAmelCase : List[str] = None __UpperCAmelCase : Any = None __UpperCAmelCase : Tuple = OnnxStableDiffusionPipeline( vae_encoder=OnnxRuntimeModel.from_pretrained(output_path / """vae_encoder""" ) , vae_decoder=OnnxRuntimeModel.from_pretrained(output_path / """vae_decoder""" ) , text_encoder=OnnxRuntimeModel.from_pretrained(output_path / """text_encoder""" ) , tokenizer=pipeline.tokenizer , unet=OnnxRuntimeModel.from_pretrained(output_path / """unet""" ) , scheduler=pipeline.scheduler , safety_checker=lowerCAmelCase__ , feature_extractor=lowerCAmelCase__ , requires_safety_checker=safety_checker is not None , ) onnx_pipeline.save_pretrained(lowerCAmelCase__ ) print("""ONNX pipeline saved to""" , lowerCAmelCase__ ) del pipeline del onnx_pipeline __UpperCAmelCase : Tuple = OnnxStableDiffusionPipeline.from_pretrained(lowerCAmelCase__ , provider="""CPUExecutionProvider""" ) print("""ONNX pipeline is loadable""" ) if __name__ == "__main__": _UpperCamelCase = argparse.ArgumentParser() parser.add_argument( '''--model_path''', type=str, required=True, help='''Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).''', ) parser.add_argument('''--output_path''', type=str, required=True, help='''Path to the output model.''') parser.add_argument( '''--opset''', default=14, type=int, help='''The version of the ONNX operator set to use.''', ) parser.add_argument('''--fp16''', action='''store_true''', default=False, help='''Export the models in `float16` mode''') _UpperCamelCase = parser.parse_args() convert_models(args.model_path, args.output_path, args.opset, args.fpaa)
254
0
import math from ...configuration_utils import PretrainedConfig from ...utils import logging A : Any = logging.get_logger(__name__) A : Optional[Any] = { 'facebook/data2vec-base-960h': 'https://huggingface.co/facebook/data2vec-audio-base-960h/resolve/main/config.json', # See all Data2VecAudio models at https://huggingface.co/models?filter=data2vec-audio } class A ( UpperCAmelCase__ ): '''simple docstring''' A__ = '''data2vec-audio''' def __init__(self : Dict , _UpperCAmelCase : List[Any]=32 , _UpperCAmelCase : List[str]=768 , _UpperCAmelCase : str=12 , _UpperCAmelCase : Union[str, Any]=12 , _UpperCAmelCase : str=3072 , _UpperCAmelCase : Tuple="gelu" , _UpperCAmelCase : Optional[Any]=0.1 , _UpperCAmelCase : Any=0.1 , _UpperCAmelCase : Optional[int]=0.1 , _UpperCAmelCase : Optional[int]=0.0 , _UpperCAmelCase : List[Any]=0.1 , _UpperCAmelCase : Optional[Any]=0.1 , _UpperCAmelCase : Dict=0.02 , _UpperCAmelCase : str=1E-5 , _UpperCAmelCase : Dict="gelu" , _UpperCAmelCase : List[str]=(512, 512, 512, 512, 512, 512, 512) , _UpperCAmelCase : Any=(5, 2, 2, 2, 2, 2, 2) , _UpperCAmelCase : List[Any]=(10, 3, 3, 3, 3, 2, 2) , _UpperCAmelCase : Any=False , _UpperCAmelCase : Tuple=16 , _UpperCAmelCase : Any=19 , _UpperCAmelCase : str=5 , _UpperCAmelCase : List[str]=0.05 , _UpperCAmelCase : Any=10 , _UpperCAmelCase : str=2 , _UpperCAmelCase : str=0.0 , _UpperCAmelCase : Optional[int]=10 , _UpperCAmelCase : List[str]=0 , _UpperCAmelCase : Tuple="sum" , _UpperCAmelCase : Union[str, Any]=False , _UpperCAmelCase : Union[str, Any]=False , _UpperCAmelCase : Any=256 , _UpperCAmelCase : str=(512, 512, 512, 512, 1500) , _UpperCAmelCase : Dict=(5, 3, 3, 1, 1) , _UpperCAmelCase : Optional[Any]=(1, 2, 3, 1, 1) , _UpperCAmelCase : List[str]=512 , _UpperCAmelCase : Any=0 , _UpperCAmelCase : Optional[Any]=1 , _UpperCAmelCase : Union[str, Any]=2 , _UpperCAmelCase : Union[str, Any]=False , _UpperCAmelCase : Optional[int]=3 , _UpperCAmelCase : Tuple=2 , _UpperCAmelCase : Dict=3 , _UpperCAmelCase : Optional[Any]=None , **_UpperCAmelCase : Optional[int] , ) -> List[str]: """simple docstring""" super().__init__(**_UpperCAmelCase , pad_token_id=_UpperCAmelCase , bos_token_id=_UpperCAmelCase , eos_token_id=_UpperCAmelCase ) lowercase__ = hidden_size lowercase__ = feat_extract_activation lowercase__ = list(_UpperCAmelCase ) lowercase__ = list(_UpperCAmelCase ) lowercase__ = list(_UpperCAmelCase ) lowercase__ = conv_bias lowercase__ = num_conv_pos_embeddings lowercase__ = num_conv_pos_embedding_groups lowercase__ = conv_pos_kernel_size lowercase__ = len(self.conv_dim ) lowercase__ = num_hidden_layers lowercase__ = intermediate_size lowercase__ = hidden_act lowercase__ = num_attention_heads lowercase__ = hidden_dropout lowercase__ = attention_dropout lowercase__ = activation_dropout lowercase__ = feat_proj_dropout lowercase__ = final_dropout lowercase__ = layerdrop lowercase__ = layer_norm_eps lowercase__ = initializer_range lowercase__ = vocab_size lowercase__ = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" f''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' f''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 lowercase__ = mask_time_prob lowercase__ = mask_time_length lowercase__ = mask_time_min_masks lowercase__ = mask_feature_prob lowercase__ = mask_feature_length lowercase__ = mask_feature_min_masks # ctc loss lowercase__ = ctc_loss_reduction lowercase__ = ctc_zero_infinity # adapter lowercase__ = add_adapter lowercase__ = adapter_kernel_size lowercase__ = adapter_stride lowercase__ = num_adapter_layers lowercase__ = output_hidden_size or hidden_size # SequenceClassification-specific parameter. Feel free to ignore for other classes. lowercase__ = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. lowercase__ = list(_UpperCAmelCase ) lowercase__ = list(_UpperCAmelCase ) lowercase__ = list(_UpperCAmelCase ) lowercase__ = xvector_output_dim @property def lowerCamelCase__ (self : Optional[int] ) -> Tuple: """simple docstring""" return math.prod(self.conv_stride )
351
import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor A : List[Any] = logging.get_logger(__name__) class A ( UpperCAmelCase__ ): '''simple docstring''' def __init__(self : List[Any] , *_UpperCAmelCase : int , **_UpperCAmelCase : List[str] ) -> None: """simple docstring""" warnings.warn( """The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use CLIPImageProcessor instead.""" , _UpperCAmelCase , ) super().__init__(*_UpperCAmelCase , **_UpperCAmelCase )
146
0
import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() a_ :Tuple = logging.get_logger(__name__) a_ :List[Any] = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } a_ :Optional[int] = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", ] def lowercase_ (A : Union[str, Any] , A : str , A : Dict , A : Optional[Any] , A : Optional[Any] ): for attribute in key.split('.' ): snake_case__ : Any = getattr(A , A ) if weight_type is not None: snake_case__ : Optional[Any] = getattr(A , A ).shape else: snake_case__ : Optional[int] = hf_pointer.shape assert hf_shape == value.shape, ( F'''Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": snake_case__ : Tuple = value elif weight_type == "weight_g": snake_case__ : Tuple = value elif weight_type == "weight_v": snake_case__ : List[Any] = value elif weight_type == "bias": snake_case__ : List[Any] = value else: snake_case__ : Optional[Any] = value logger.info(F'''{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.''' ) def lowercase_ (A : str , A : Any ): snake_case__ : Union[str, Any] = [] snake_case__ : Union[str, Any] = fairseq_model.state_dict() snake_case__ : Union[str, Any] = hf_model.feature_extractor snake_case__ : Any = hf_model.adapter for name, value in fairseq_dict.items(): snake_case__ : Any = False if "conv_layers" in name: load_conv_layer( A , A , A , A , hf_model.config.feat_extract_norm == 'group' , ) snake_case__ : List[Any] = True elif any(x in name for x in ['adaptor', 'w2v_encoder.proj.', 'w2v_proj_ln.'] ): load_adapter(A , A , A , A ) snake_case__ : Optional[Any] = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: snake_case__ : Tuple = True if "*" in mapped_key: snake_case__ : List[Any] = name.split(A )[0].split('.' )[-2] snake_case__ : Optional[int] = mapped_key.replace('*' , A ) if "weight_g" in name: snake_case__ : Optional[int] = 'weight_g' elif "weight_v" in name: snake_case__ : Optional[Any] = 'weight_v' elif "bias" in name: snake_case__ : Union[str, Any] = 'bias' elif "weight" in name: snake_case__ : Optional[int] = 'weight' else: snake_case__ : Tuple = None set_recursively(A , A , A , A , A ) continue if not is_used: unused_weights.append(A ) logger.warning(F'''Unused weights: {unused_weights}''' ) def lowercase_ (A : Union[str, Any] , A : Any , A : str , A : str , A : int ): snake_case__ : str = full_name.split('conv_layers.' )[-1] snake_case__ : Optional[int] = name.split('.' ) snake_case__ : Tuple = int(items[0] ) snake_case__ : Any = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) snake_case__ : Union[str, Any] = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) snake_case__ : Union[str, Any] = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'''{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was''' " found." ) snake_case__ : Optional[int] = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) snake_case__ : Optional[Any] = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(A ) def lowercase_ (A : Optional[Any] , A : Any , A : Tuple , A : Any ): snake_case__ : List[str] = full_name.split('adaptor.' )[-1] snake_case__ : Tuple = name.split('.' ) if items[1].isdigit(): snake_case__ : Optional[int] = int(items[1] ) else: snake_case__ : Any = None if "adaptor" not in full_name: if "proj_ln" in full_name: # has to be layer norm if "bias" in name: assert ( value.shape == adapter.proj_layer_norm.bias.data.shape ), F'''{full_name} has size {value.shape}, but {adapter.proj_layer_norm.bias.data.shape} was found.''' snake_case__ : List[Any] = value logger.info(F'''Adapter proj layer norm bias was initialized from {full_name}.''' ) if "weight" in name: assert ( value.shape == adapter.proj_layer_norm.weight.data.shape ), F'''{full_name} has size {value.shape}, but {adapter.proj_layer_norm.weight.data.shape} was found.''' snake_case__ : int = value else: # has to be projection layer if "bias" in name: assert ( value.shape == adapter.proj.bias.data.shape ), F'''{full_name} has size {value.shape}, but {adapter.proj.bias.data.shape} was found.''' snake_case__ : str = value logger.info(F'''Adapter proj layer bias was initialized from {full_name}.''' ) if "weight" in name: assert ( value.shape == adapter.proj.weight.data.shape ), F'''{full_name} has size {value.shape}, but {adapter.proj.weight.data.shape} was found.''' snake_case__ : Dict = value logger.info(F'''Adapter proj layer weight was initialized from {full_name}.''' ) elif isinstance(A , A ): if "bias" in name: assert ( value.shape == adapter.layers[layer_id].conv.bias.data.shape ), F'''{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.bias.data.shape} was found.''' snake_case__ : List[str] = value logger.info(F'''Adapter layer {layer_id} bias was initialized from {full_name}.''' ) elif "weight" in name: assert ( value.shape == adapter.layers[layer_id].conv.weight.data.shape ), F'''{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.weight.data.shape} was found.''' snake_case__ : List[str] = value logger.info(F'''Adapter layer {layer_id} bias was initialized from {full_name}.''' ) else: unused_weights.append(A ) def lowercase_ (A : int ): snake_case__ , snake_case__ : Union[str, Any] = emb.weight.shape snake_case__ : int = nn.Linear(A , A , bias=A ) snake_case__ : Optional[Any] = emb.weight.data return lin_layer @torch.no_grad() def lowercase_ (A : Tuple , A : Tuple , A : Any , A : Optional[Any] , A : int , A : Optional[Any] , A : Union[str, Any] , A : Union[str, Any] , A : Optional[Any] , A : List[Any] , A : Union[str, Any] , ): snake_case__ : Optional[Any] = WavaVecaConfig.from_pretrained( A , add_adapter=A , adapter_stride=A , adapter_kernel_size=A , use_auth_token=A , output_hidden_size=A , ) snake_case__ : Dict = MBartConfig.from_pretrained(A ) # load model snake_case__ , snake_case__ , snake_case__ : Any = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={ 'config_yaml': config_yaml_path, 'data': '/'.join(dict_path.split('/' )[:-1] ), 'w2v_path': checkpoint_path, 'load_pretrained_decoder_from': None, } , ) snake_case__ : List[Any] = model[0].eval() # load feature extractor snake_case__ : str = WavaVecaFeatureExtractor.from_pretrained(A , use_auth_token=A ) # set weights for wav2vec2 encoder snake_case__ : List[str] = WavaVecaModel(A ) recursively_load_weights_wavaveca(model.encoder , A ) # load decoder weights snake_case__ : Any = MBartForCausalLM(A ) snake_case__ , snake_case__ : int = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=A ) logger.warning(F'''The following keys are missing when loading the decoder weights: {missing_keys}''' ) logger.warning(F'''The following keys are unexpected when loading the decoder weights: {unexpected_keys}''' ) snake_case__ : Union[str, Any] = SpeechEncoderDecoderModel(encoder=A , decoder=A ) snake_case__ : str = False snake_case__ : int = MBartaaTokenizer(A ) tokenizer.save_pretrained(A ) snake_case__ : Any = hf_wavavec.config.to_dict() snake_case__ : Tuple = tokenizer.pad_token_id snake_case__ : Union[str, Any] = tokenizer.bos_token_id snake_case__ : Dict = tokenizer.eos_token_id snake_case__ : Optional[int] = 'mbart50' snake_case__ : Union[str, Any] = 'wav2vec2' snake_case__ : List[str] = tokenizer.eos_token_id snake_case__ : Union[str, Any] = 2_5_0_0_0_4 snake_case__ : int = tokenizer.eos_token_id snake_case__ : Union[str, Any] = SpeechEncoderDecoderConfig.from_dict(A ) hf_wavavec.save_pretrained(A ) feature_extractor.save_pretrained(A ) if __name__ == "__main__": a_ :str = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_yaml_path", default=None, type=str, help="Path to yaml file of fine-tuned model") parser.add_argument( "--encoder_config_path", default="facebook/wav2vec2-xls-r-1b", type=str, help="Path to hf encoder wav2vec2 checkpoint config", ) parser.add_argument( "--decoder_config_path", default="facebook/mbart-large-50-one-to-many-mmt", type=str, help="Path to hf decoder checkpoint config", ) parser.add_argument("--add_adapter", default=True, type=bool, help="whethere to add model adapter layers") parser.add_argument("--adapter_stride", default=2, type=int, help="stride of adapter layers") parser.add_argument("--adapter_kernel_size", default=3, type=int, help="kernel size of adapter layers") parser.add_argument("--encoder_output_dim", default=1_024, type=int, help="encoder output dim") parser.add_argument("--start_token_id", default=250_004, type=int, help="`decoder_start_token_id` of model config") a_ :Union[str, Any] = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, args.config_yaml_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, add_adapter=args.add_adapter, adapter_kernel_size=args.adapter_kernel_size, adapter_stride=args.adapter_stride, decoder_start_token_id=args.start_token_id, encoder_output_dim=args.encoder_output_dim, )
277
import argparse import json import pickle from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import MaskFormerConfig, MaskFormerForInstanceSegmentation, MaskFormerImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() a_ :str = logging.get_logger(__name__) def lowercase_ (A : str ): snake_case__ : Tuple = SwinConfig.from_pretrained( 'microsoft/swin-tiny-patch4-window7-224' , out_features=['stage1', 'stage2', 'stage3', 'stage4'] ) snake_case__ : List[Any] = MaskFormerConfig(backbone_config=A ) snake_case__ : Union[str, Any] = 'huggingface/label-files' if "ade20k-full" in model_name: # this should be ok snake_case__ : Dict = 8_4_7 snake_case__ : List[str] = 'maskformer-ade20k-full-id2label.json' elif "ade" in model_name: # this should be ok snake_case__ : Union[str, Any] = 1_5_0 snake_case__ : Any = 'ade20k-id2label.json' elif "coco-stuff" in model_name: # this should be ok snake_case__ : List[str] = 1_7_1 snake_case__ : Union[str, Any] = 'maskformer-coco-stuff-id2label.json' elif "coco" in model_name: # TODO snake_case__ : Dict = 1_3_3 snake_case__ : str = 'coco-panoptic-id2label.json' elif "cityscapes" in model_name: # this should be ok snake_case__ : List[str] = 1_9 snake_case__ : Union[str, Any] = 'cityscapes-id2label.json' elif "vistas" in model_name: # this should be ok snake_case__ : Tuple = 6_5 snake_case__ : List[str] = 'mapillary-vistas-id2label.json' snake_case__ : Dict = json.load(open(hf_hub_download(A , A , repo_type='dataset' ) , 'r' ) ) snake_case__ : List[str] = {int(A ): v for k, v in idalabel.items()} return config def lowercase_ (A : Any ): snake_case__ : Optional[int] = [] # stem # fmt: off rename_keys.append(('backbone.patch_embed.proj.weight', 'model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.weight') ) rename_keys.append(('backbone.patch_embed.proj.bias', 'model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.bias') ) rename_keys.append(('backbone.patch_embed.norm.weight', 'model.pixel_level_module.encoder.model.embeddings.norm.weight') ) rename_keys.append(('backbone.patch_embed.norm.bias', 'model.pixel_level_module.encoder.model.embeddings.norm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.norm1.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.norm1.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.attn.relative_position_bias_table''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.attn.relative_position_index''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.attn.proj.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.attn.proj.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.norm2.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.norm2.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.mlp.fc1.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.mlp.fc1.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.mlp.fc2.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.mlp.fc2.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.bias''') ) if i < 3: rename_keys.append((F'''backbone.layers.{i}.downsample.reduction.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.reduction.weight''') ) rename_keys.append((F'''backbone.layers.{i}.downsample.norm.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.weight''') ) rename_keys.append((F'''backbone.layers.{i}.downsample.norm.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.bias''') ) rename_keys.append((F'''backbone.norm{i}.weight''', F'''model.pixel_level_module.encoder.hidden_states_norms.{i}.weight''') ) rename_keys.append((F'''backbone.norm{i}.bias''', F'''model.pixel_level_module.encoder.hidden_states_norms.{i}.bias''') ) # FPN rename_keys.append(('sem_seg_head.layer_4.weight', 'model.pixel_level_module.decoder.fpn.stem.0.weight') ) rename_keys.append(('sem_seg_head.layer_4.norm.weight', 'model.pixel_level_module.decoder.fpn.stem.1.weight') ) rename_keys.append(('sem_seg_head.layer_4.norm.bias', 'model.pixel_level_module.decoder.fpn.stem.1.bias') ) for source_index, target_index in zip(range(3 , 0 , -1 ) , range(0 , 3 ) ): rename_keys.append((F'''sem_seg_head.adapter_{source_index}.weight''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.0.weight''') ) rename_keys.append((F'''sem_seg_head.adapter_{source_index}.norm.weight''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.weight''') ) rename_keys.append((F'''sem_seg_head.adapter_{source_index}.norm.bias''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.bias''') ) rename_keys.append((F'''sem_seg_head.layer_{source_index}.weight''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.0.weight''') ) rename_keys.append((F'''sem_seg_head.layer_{source_index}.norm.weight''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.weight''') ) rename_keys.append((F'''sem_seg_head.layer_{source_index}.norm.bias''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.bias''') ) rename_keys.append(('sem_seg_head.mask_features.weight', 'model.pixel_level_module.decoder.mask_projection.weight') ) rename_keys.append(('sem_seg_head.mask_features.bias', 'model.pixel_level_module.decoder.mask_projection.bias') ) # Transformer decoder for idx in range(config.decoder_config.decoder_layers ): # self-attention out projection rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.weight''', F'''model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.bias''', F'''model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.bias''') ) # cross-attention out projection rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.weight''', F'''model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.bias''', F'''model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.bias''') ) # MLP 1 rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.weight''', F'''model.transformer_module.decoder.layers.{idx}.fc1.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.bias''', F'''model.transformer_module.decoder.layers.{idx}.fc1.bias''') ) # MLP 2 rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.weight''', F'''model.transformer_module.decoder.layers.{idx}.fc2.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.bias''', F'''model.transformer_module.decoder.layers.{idx}.fc2.bias''') ) # layernorm 1 (self-attention layernorm) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.weight''', F'''model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.bias''', F'''model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.bias''') ) # layernorm 2 (cross-attention layernorm) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.weight''', F'''model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.bias''', F'''model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.bias''') ) # layernorm 3 (final layernorm) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.weight''', F'''model.transformer_module.decoder.layers.{idx}.final_layer_norm.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.bias''', F'''model.transformer_module.decoder.layers.{idx}.final_layer_norm.bias''') ) rename_keys.append(('sem_seg_head.predictor.transformer.decoder.norm.weight', 'model.transformer_module.decoder.layernorm.weight') ) rename_keys.append(('sem_seg_head.predictor.transformer.decoder.norm.bias', 'model.transformer_module.decoder.layernorm.bias') ) # heads on top rename_keys.append(('sem_seg_head.predictor.query_embed.weight', 'model.transformer_module.queries_embedder.weight') ) rename_keys.append(('sem_seg_head.predictor.input_proj.weight', 'model.transformer_module.input_projection.weight') ) rename_keys.append(('sem_seg_head.predictor.input_proj.bias', 'model.transformer_module.input_projection.bias') ) rename_keys.append(('sem_seg_head.predictor.class_embed.weight', 'class_predictor.weight') ) rename_keys.append(('sem_seg_head.predictor.class_embed.bias', 'class_predictor.bias') ) for i in range(3 ): rename_keys.append((F'''sem_seg_head.predictor.mask_embed.layers.{i}.weight''', F'''mask_embedder.{i}.0.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.mask_embed.layers.{i}.bias''', F'''mask_embedder.{i}.0.bias''') ) # fmt: on return rename_keys def lowercase_ (A : Tuple , A : Tuple , A : Optional[Any] ): snake_case__ : Optional[int] = dct.pop(A ) snake_case__ : Union[str, Any] = val def lowercase_ (A : Optional[Any] , A : Tuple ): snake_case__ : Optional[int] = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): snake_case__ : Optional[int] = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) snake_case__ : int = state_dict.pop(F'''backbone.layers.{i}.blocks.{j}.attn.qkv.weight''' ) snake_case__ : Tuple = state_dict.pop(F'''backbone.layers.{i}.blocks.{j}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict snake_case__ : str = in_proj_weight[:dim, :] snake_case__ : int = in_proj_bias[: dim] snake_case__ : List[Any] = in_proj_weight[ dim : dim * 2, : ] snake_case__ : List[str] = in_proj_bias[ dim : dim * 2 ] snake_case__ : List[Any] = in_proj_weight[ -dim :, : ] snake_case__ : Dict = in_proj_bias[-dim :] # fmt: on def lowercase_ (A : List[str] , A : List[Any] ): # fmt: off snake_case__ : str = config.decoder_config.hidden_size for idx in range(config.decoder_config.decoder_layers ): # read in weights + bias of self-attention input projection layer (in the original implementation, this is a single matrix + bias) snake_case__ : List[Any] = state_dict.pop(F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_weight''' ) snake_case__ : int = state_dict.pop(F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict snake_case__ : Any = in_proj_weight[: hidden_size, :] snake_case__ : Tuple = in_proj_bias[:config.hidden_size] snake_case__ : List[str] = in_proj_weight[hidden_size : hidden_size * 2, :] snake_case__ : Dict = in_proj_bias[hidden_size : hidden_size * 2] snake_case__ : Any = in_proj_weight[-hidden_size :, :] snake_case__ : int = in_proj_bias[-hidden_size :] # read in weights + bias of cross-attention input projection layer (in the original implementation, this is a single matrix + bias) snake_case__ : List[Any] = state_dict.pop(F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_weight''' ) snake_case__ : List[str] = state_dict.pop(F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict snake_case__ : Optional[int] = in_proj_weight[: hidden_size, :] snake_case__ : Optional[Any] = in_proj_bias[:config.hidden_size] snake_case__ : int = in_proj_weight[hidden_size : hidden_size * 2, :] snake_case__ : List[str] = in_proj_bias[hidden_size : hidden_size * 2] snake_case__ : List[str] = in_proj_weight[-hidden_size :, :] snake_case__ : str = in_proj_bias[-hidden_size :] # fmt: on def lowercase_ (): snake_case__ : Any = 'http://images.cocodataset.org/val2017/000000039769.jpg' snake_case__ : int = Image.open(requests.get(A , stream=A ).raw ) return im @torch.no_grad() def lowercase_ (A : str , A : str , A : str , A : bool = False ): snake_case__ : Optional[int] = get_maskformer_config(A ) # load original state_dict with open(A , 'rb' ) as f: snake_case__ : List[Any] = pickle.load(A ) snake_case__ : Optional[int] = data['model'] # for name, param in state_dict.items(): # print(name, param.shape) # rename keys snake_case__ : List[str] = create_rename_keys(A ) for src, dest in rename_keys: rename_key(A , A , A ) read_in_swin_q_k_v(A , config.backbone_config ) read_in_decoder_q_k_v(A , A ) # update to torch tensors for key, value in state_dict.items(): snake_case__ : int = torch.from_numpy(A ) # load 🤗 model snake_case__ : str = MaskFormerForInstanceSegmentation(A ) model.eval() for name, param in model.named_parameters(): print(A , param.shape ) snake_case__ , snake_case__ : Union[str, Any] = model.load_state_dict(A , strict=A ) assert missing_keys == [ "model.pixel_level_module.encoder.model.layernorm.weight", "model.pixel_level_module.encoder.model.layernorm.bias", ] assert len(A ) == 0, F'''Unexpected keys: {unexpected_keys}''' # verify results snake_case__ : Optional[Any] = prepare_img() if "vistas" in model_name: snake_case__ : int = 6_5 elif "cityscapes" in model_name: snake_case__ : Dict = 6_5_5_3_5 else: snake_case__ : Tuple = 2_5_5 snake_case__ : Optional[int] = True if 'ade' in model_name else False snake_case__ : Dict = MaskFormerImageProcessor(ignore_index=A , reduce_labels=A ) snake_case__ : Any = image_processor(A , return_tensors='pt' ) snake_case__ : Any = model(**A ) print('Logits:' , outputs.class_queries_logits[0, :3, :3] ) if model_name == "maskformer-swin-tiny-ade": snake_case__ : Tuple = torch.tensor( [[3.6353, -4.4770, -2.6065], [0.5081, -4.2394, -3.5343], [2.1909, -5.0353, -1.9323]] ) assert torch.allclose(outputs.class_queries_logits[0, :3, :3] , A , atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(F'''Saving model and image processor to {pytorch_dump_folder_path}''' ) Path(A ).mkdir(exist_ok=A ) model.save_pretrained(A ) image_processor.save_pretrained(A ) if push_to_hub: print('Pushing model and image processor to the hub...' ) model.push_to_hub(F'''nielsr/{model_name}''' ) image_processor.push_to_hub(F'''nielsr/{model_name}''' ) if __name__ == "__main__": a_ :Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="maskformer-swin-tiny-ade", type=str, help=("Name of the MaskFormer model you'd like to convert",), ) parser.add_argument( "--checkpoint_path", default="/Users/nielsrogge/Documents/MaskFormer_checkpoints/MaskFormer-Swin-tiny-ADE20k/model.pkl", type=str, help="Path to the original state dict (.pth file).", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) a_ :Dict = parser.parse_args() convert_maskformer_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
277
1
from __future__ import annotations UpperCamelCase_ = [-10, -5, 0, 5, 5.1, 11, 13, 21, 3, 4, -21, -10, -5, -1, 0] UpperCamelCase_ = [-5, 0, 5, 5.1, 11, 13, 21, -1, 4, -1, -10, -5, -1, 0, -1] def A ( __UpperCAmelCase ) -> list[float]: '''simple docstring''' UpperCAmelCase_ = [] UpperCAmelCase_ = len(_snake_case ) for i in range(_snake_case ): UpperCAmelCase_ = -1 for j in range(i + 1 , _snake_case ): if arr[i] < arr[j]: UpperCAmelCase_ = arr[j] break result.append(_snake_case ) return result def A ( __UpperCAmelCase ) -> list[float]: '''simple docstring''' UpperCAmelCase_ = [] for i, outer in enumerate(_snake_case ): UpperCAmelCase_ = -1 for inner in arr[i + 1 :]: if outer < inner: UpperCAmelCase_ = inner break result.append(_snake_case ) return result def A ( __UpperCAmelCase ) -> list[float]: '''simple docstring''' UpperCAmelCase_ = len(_snake_case ) UpperCAmelCase_ = [] UpperCAmelCase_ = [-1] * arr_size for index in reversed(range(_snake_case ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: UpperCAmelCase_ = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) UpperCamelCase_ = ( """from __main__ import arr, next_greatest_element_slow, """ """next_greatest_element_fast, next_greatest_element""" ) print( "next_greatest_element_slow():", timeit("next_greatest_element_slow(arr)", setup=setup), ) print( "next_greatest_element_fast():", timeit("next_greatest_element_fast(arr)", setup=setup), ) print( " next_greatest_element():", timeit("next_greatest_element(arr)", setup=setup), )
352
import os from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen, xsplitext from ..table import array_cast from ..utils.py_utils import no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: from .features import FeatureType UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = False, False, False @dataclass class a_ : UpperCamelCase__ : Optional[int] =None UpperCamelCase__ : bool =True UpperCamelCase__ : bool =True UpperCamelCase__ : Optional[str] =None # Automatically constructed UpperCamelCase__ : ClassVar[str] ="dict" UpperCamelCase__ : ClassVar[Any] =pa.struct({"bytes": pa.binary(), "path": pa.string()} ) UpperCamelCase__ : str =field(default="Audio" , init=_snake_case , repr=_snake_case ) def __call__( self :List[Any]) -> List[Any]: return self.pa_type def __a ( self :Any , _lowercase :Union[str, bytes, dict]) -> dict: try: import soundfile as sf # soundfile is a dependency of librosa, needed to decode audio files. except ImportError as err: raise ImportError('''To support encoding audio data, please install \'soundfile\'.''') from err if isinstance(_lowercase , _lowercase): return {"bytes": None, "path": value} elif isinstance(_lowercase , _lowercase): return {"bytes": value, "path": None} elif "array" in value: # convert the audio array to wav bytes UpperCAmelCase_ = BytesIO() sf.write(_lowercase , value['''array'''] , value['''sampling_rate'''] , format='''wav''') return {"bytes": buffer.getvalue(), "path": None} elif value.get('''path''') is not None and os.path.isfile(value['''path''']): # we set "bytes": None to not duplicate the data if they're already available locally if value["path"].endswith('''pcm'''): # "PCM" only has raw audio bytes if value.get('''sampling_rate''') is None: # At least, If you want to convert "PCM-byte" to "WAV-byte", you have to know sampling rate raise KeyError('''To use PCM files, please specify a \'sampling_rate\' in Audio object''') if value.get('''bytes'''): # If we already had PCM-byte, we don`t have to make "read file, make bytes" (just use it!) UpperCAmelCase_ = np.frombuffer(value['''bytes'''] , dtype=np.intaa).astype(np.floataa) / 32767 else: UpperCAmelCase_ = np.memmap(value['''path'''] , dtype='''h''' , mode='''r''').astype(np.floataa) / 32767 UpperCAmelCase_ = BytesIO(bytes()) sf.write(_lowercase , _lowercase , value['''sampling_rate'''] , format='''wav''') return {"bytes": buffer.getvalue(), "path": None} else: return {"bytes": None, "path": value.get('''path''')} elif value.get('''bytes''') is not None or value.get('''path''') is not None: # store the audio bytes, and path is used to infer the audio format using the file extension return {"bytes": value.get('''bytes'''), "path": value.get('''path''')} else: raise ValueError( f"An audio sample should have one of 'path' or 'bytes' but they are missing or None in {value}.") def __a ( self :Dict , _lowercase :dict , _lowercase :Optional[Dict[str, Union[str, bool, None]]] = None) -> dict: if not self.decode: raise RuntimeError('''Decoding is disabled for this feature. Please use Audio(decode=True) instead.''') UpperCAmelCase_ , UpperCAmelCase_ = (value['''path'''], BytesIO(value['''bytes'''])) if value['''bytes'''] is not None else (value['''path'''], None) if path is None and file is None: raise ValueError(f"An audio sample should have one of 'path' or 'bytes' but both are None in {value}.") try: import librosa import soundfile as sf except ImportError as err: raise ImportError('''To support decoding audio files, please install \'librosa\' and \'soundfile\'.''') from err UpperCAmelCase_ = xsplitext(_lowercase)[1][1:].lower() if path is not None else None if not config.IS_OPUS_SUPPORTED and audio_format == "opus": raise RuntimeError( '''Decoding \'opus\' files requires system library \'libsndfile\'>=1.0.31, ''' '''You can try to update `soundfile` python library: `pip install "soundfile>=0.12.1"`. ''') elif not config.IS_MP3_SUPPORTED and audio_format == "mp3": raise RuntimeError( '''Decoding \'mp3\' files requires system library \'libsndfile\'>=1.1.0, ''' '''You can try to update `soundfile` python library: `pip install "soundfile>=0.12.1"`. ''') if file is None: UpperCAmelCase_ = token_per_repo_id or {} UpperCAmelCase_ = path.split('''::''')[-1] try: UpperCAmelCase_ = string_to_dict(_lowercase , config.HUB_DATASETS_URL)['''repo_id'''] UpperCAmelCase_ = token_per_repo_id[repo_id] except (ValueError, KeyError): UpperCAmelCase_ = None with xopen(_lowercase , '''rb''' , use_auth_token=_lowercase) as f: UpperCAmelCase_ , UpperCAmelCase_ = sf.read(_lowercase) else: UpperCAmelCase_ , UpperCAmelCase_ = sf.read(_lowercase) UpperCAmelCase_ = array.T if self.mono: UpperCAmelCase_ = librosa.to_mono(_lowercase) if self.sampling_rate and self.sampling_rate != sampling_rate: UpperCAmelCase_ = librosa.resample(_lowercase , orig_sr=_lowercase , target_sr=self.sampling_rate) UpperCAmelCase_ = self.sampling_rate return {"path": path, "array": array, "sampling_rate": sampling_rate} def __a ( self :Union[str, Any]) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value if self.decode: raise ValueError('''Cannot flatten a decoded Audio feature.''') return { "bytes": Value('''binary'''), "path": Value('''string'''), } def __a ( self :int , _lowercase :Union[pa.StringArray, pa.StructArray]) -> pa.StructArray: if pa.types.is_string(storage.type): UpperCAmelCase_ = pa.array([None] * len(_lowercase) , type=pa.binary()) UpperCAmelCase_ = pa.StructArray.from_arrays([bytes_array, storage] , ['''bytes''', '''path'''] , mask=storage.is_null()) elif pa.types.is_binary(storage.type): UpperCAmelCase_ = pa.array([None] * len(_lowercase) , type=pa.string()) UpperCAmelCase_ = pa.StructArray.from_arrays([storage, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null()) elif pa.types.is_struct(storage.type) and storage.type.get_all_field_indices('''array'''): UpperCAmelCase_ = pa.array([Audio().encode_example(_lowercase) if x is not None else None for x in storage.to_pylist()]) elif pa.types.is_struct(storage.type): if storage.type.get_field_index('''bytes''') >= 0: UpperCAmelCase_ = storage.field('''bytes''') else: UpperCAmelCase_ = pa.array([None] * len(_lowercase) , type=pa.binary()) if storage.type.get_field_index('''path''') >= 0: UpperCAmelCase_ = storage.field('''path''') else: UpperCAmelCase_ = pa.array([None] * len(_lowercase) , type=pa.string()) UpperCAmelCase_ = pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null()) return array_cast(_lowercase , self.pa_type) def __a ( self :Any , _lowercase :pa.StructArray) -> pa.StructArray: @no_op_if_value_is_null def path_to_bytes(_lowercase :Tuple): with xopen(_lowercase , '''rb''') as f: UpperCAmelCase_ = f.read() return bytes_ UpperCAmelCase_ = pa.array( [ (path_to_bytes(x['''path''']) if x['''bytes'''] is None else x['''bytes''']) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) UpperCAmelCase_ = pa.array( [os.path.basename(_lowercase) if path is not None else None for path in storage.field('''path''').to_pylist()] , type=pa.string() , ) UpperCAmelCase_ = pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=bytes_array.is_null()) return array_cast(_lowercase , self.pa_type)
344
0
"""simple docstring""" import argparse import os import re __snake_case : Optional[int] = 'src/diffusers' # Pattern that looks at the indentation in a line. __snake_case : List[Any] = re.compile(R'^(\s*)\S') # Pattern that matches `"key":" and puts `key` in group 0. __snake_case : List[str] = re.compile(R'^\s*\"([^\"]+)\":') # Pattern that matches `_import_structure["key"]` and puts `key` in group 0. __snake_case : Optional[Any] = re.compile(R'^\s*_import_structure\[\"([^\"]+)\"\]') # Pattern that matches `"key",` and puts `key` in group 0. __snake_case : List[str] = re.compile(R'^\s*\"([^\"]+)\",\s*$') # Pattern that matches any `[stuff]` and puts `stuff` in group 0. __snake_case : str = re.compile(R'\[([^\]]+)\]') def _lowercase ( __snake_case ) -> Tuple: __lowerCAmelCase : List[str] = _re_indent.search(__snake_case ) return "" if search is None else search.groups()[0] def _lowercase ( __snake_case ,__snake_case="" ,__snake_case=None ,__snake_case=None ) -> Dict: __lowerCAmelCase : Dict = 0 __lowerCAmelCase : int = code.split("\n" ) if start_prompt is not None: while not lines[index].startswith(__snake_case ): index += 1 __lowerCAmelCase : Tuple = ["\n".join(lines[:index] )] else: __lowerCAmelCase : Optional[int] = [] # We split into blocks until we get to the `end_prompt` (or the end of the block). __lowerCAmelCase : int = [lines[index]] index += 1 while index < len(__snake_case ) and (end_prompt is None or not lines[index].startswith(__snake_case )): if len(lines[index] ) > 0 and get_indent(lines[index] ) == indent_level: if len(__snake_case ) > 0 and get_indent(current_block[-1] ).startswith(indent_level + " " ): current_block.append(lines[index] ) blocks.append("\n".join(__snake_case ) ) if index < len(__snake_case ) - 1: __lowerCAmelCase : Optional[Any] = [lines[index + 1]] index += 1 else: __lowerCAmelCase : List[Any] = [] else: blocks.append("\n".join(__snake_case ) ) __lowerCAmelCase : Union[str, Any] = [lines[index]] else: current_block.append(lines[index] ) index += 1 # Adds current block if it's nonempty. if len(__snake_case ) > 0: blocks.append("\n".join(__snake_case ) ) # Add final block after end_prompt if provided. if end_prompt is not None and index < len(__snake_case ): blocks.append("\n".join(lines[index:] ) ) return blocks def _lowercase ( __snake_case ) -> List[Any]: def _inner(__snake_case ): return key(__snake_case ).lower().replace("_" ,"" ) return _inner def _lowercase ( __snake_case ,__snake_case=None ) -> Union[str, Any]: def noop(__snake_case ): return x if key is None: __lowerCAmelCase : Any = noop # Constants are all uppercase, they go first. __lowerCAmelCase : int = [obj for obj in objects if key(__snake_case ).isupper()] # Classes are not all uppercase but start with a capital, they go second. __lowerCAmelCase : Union[str, Any] = [obj for obj in objects if key(__snake_case )[0].isupper() and not key(__snake_case ).isupper()] # Functions begin with a lowercase, they go last. __lowerCAmelCase : List[str] = [obj for obj in objects if not key(__snake_case )[0].isupper()] __lowerCAmelCase : str = ignore_underscore(__snake_case ) return sorted(__snake_case ,key=__snake_case ) + sorted(__snake_case ,key=__snake_case ) + sorted(__snake_case ,key=__snake_case ) def _lowercase ( __snake_case ) -> Optional[Any]: def _replace(__snake_case ): __lowerCAmelCase : str = match.groups()[0] if "," not in imports: return F"""[{imports}]""" __lowerCAmelCase : Tuple = [part.strip().replace("\"" ,"" ) for part in imports.split("," )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: __lowerCAmelCase : Tuple = keys[:-1] return "[" + ", ".join([F"""\"{k}\"""" for k in sort_objects(__snake_case )] ) + "]" __lowerCAmelCase : List[str] = import_statement.split("\n" ) if len(__snake_case ) > 3: # Here we have to sort internal imports that are on several lines (one per name): # key: [ # "object1", # "object2", # ... # ] # We may have to ignore one or two lines on each side. __lowerCAmelCase : Union[str, Any] = 2 if lines[1].strip() == "[" else 1 __lowerCAmelCase : Optional[int] = [(i, _re_strip_line.search(__snake_case ).groups()[0]) for i, line in enumerate(lines[idx:-idx] )] __lowerCAmelCase : str = sort_objects(__snake_case ,key=lambda __snake_case : x[1] ) __lowerCAmelCase : Tuple = [lines[x[0] + idx] for x in sorted_indices] return "\n".join(lines[:idx] + sorted_lines + lines[-idx:] ) elif len(__snake_case ) == 3: # Here we have to sort internal imports that are on one separate line: # key: [ # "object1", "object2", ... # ] if _re_bracket_content.search(lines[1] ) is not None: __lowerCAmelCase : int = _re_bracket_content.sub(_replace ,lines[1] ) else: __lowerCAmelCase : List[Any] = [part.strip().replace("\"" ,"" ) for part in lines[1].split("," )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: __lowerCAmelCase : int = keys[:-1] __lowerCAmelCase : List[Any] = get_indent(lines[1] ) + ", ".join([F"""\"{k}\"""" for k in sort_objects(__snake_case )] ) return "\n".join(__snake_case ) else: # Finally we have to deal with imports fitting on one line __lowerCAmelCase : List[Any] = _re_bracket_content.sub(_replace ,__snake_case ) return import_statement def _lowercase ( __snake_case ,__snake_case=True ) -> Any: with open(__snake_case ,"r" ) as f: __lowerCAmelCase : List[str] = f.read() if "_import_structure" not in code: return # Blocks of indent level 0 __lowerCAmelCase : Optional[Any] = split_code_in_indented_blocks( __snake_case ,start_prompt="_import_structure = {" ,end_prompt="if TYPE_CHECKING:" ) # We ignore block 0 (everything until start_prompt) and the last block (everything after end_prompt). for block_idx in range(1 ,len(__snake_case ) - 1 ): # Check if the block contains some `_import_structure`s thingy to sort. __lowerCAmelCase : Any = main_blocks[block_idx] __lowerCAmelCase : Union[str, Any] = block.split("\n" ) # Get to the start of the imports. __lowerCAmelCase : Tuple = 0 while line_idx < len(__snake_case ) and "_import_structure" not in block_lines[line_idx]: # Skip dummy import blocks if "import dummy" in block_lines[line_idx]: __lowerCAmelCase : List[str] = len(__snake_case ) else: line_idx += 1 if line_idx >= len(__snake_case ): continue # Ignore beginning and last line: they don't contain anything. __lowerCAmelCase : Optional[int] = "\n".join(block_lines[line_idx:-1] ) __lowerCAmelCase : Union[str, Any] = get_indent(block_lines[1] ) # Slit the internal block into blocks of indent level 1. __lowerCAmelCase : Optional[int] = split_code_in_indented_blocks(__snake_case ,indent_level=__snake_case ) # We have two categories of import key: list or _import_structure[key].append/extend __lowerCAmelCase : str = _re_direct_key if "_import_structure" in block_lines[0] else _re_indirect_key # Grab the keys, but there is a trap: some lines are empty or just comments. __lowerCAmelCase : str = [(pattern.search(__snake_case ).groups()[0] if pattern.search(__snake_case ) is not None else None) for b in internal_blocks] # We only sort the lines with a key. __lowerCAmelCase : Tuple = [(i, key) for i, key in enumerate(__snake_case ) if key is not None] __lowerCAmelCase : Any = [x[0] for x in sorted(__snake_case ,key=lambda __snake_case : x[1] )] # We reorder the blocks by leaving empty lines/comments as they were and reorder the rest. __lowerCAmelCase : Union[str, Any] = 0 __lowerCAmelCase : Dict = [] for i in range(len(__snake_case ) ): if keys[i] is None: reordered_blocks.append(internal_blocks[i] ) else: __lowerCAmelCase : List[str] = sort_objects_in_import(internal_blocks[sorted_indices[count]] ) reordered_blocks.append(__snake_case ) count += 1 # And we put our main block back together with its first and last line. __lowerCAmelCase : Union[str, Any] = "\n".join(block_lines[:line_idx] + reordered_blocks + [block_lines[-1]] ) if code != "\n".join(__snake_case ): if check_only: return True else: print(F"""Overwriting {file}.""" ) with open(__snake_case ,"w" ) as f: f.write("\n".join(__snake_case ) ) def _lowercase ( __snake_case=True ) -> List[str]: __lowerCAmelCase : List[str] = [] for root, _, files in os.walk(__snake_case ): if "__init__.py" in files: __lowerCAmelCase : List[Any] = sort_imports(os.path.join(__snake_case ,"__init__.py" ) ,check_only=__snake_case ) if result: __lowerCAmelCase : Union[str, Any] = [os.path.join(__snake_case ,"__init__.py" )] if len(__snake_case ) > 0: raise ValueError(F"""Would overwrite {len(__snake_case )} files, run `make style`.""" ) if __name__ == "__main__": __snake_case : Any = argparse.ArgumentParser() parser.add_argument('--check_only', action='store_true', help='Whether to only check or fix style.') __snake_case : str = parser.parse_args() sort_imports_in_all_inits(check_only=args.check_only)
269
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { "camembert-base": "https://huggingface.co/camembert-base/resolve/main/config.json", "umberto-commoncrawl-cased-v1": ( "https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json" ), "umberto-wikipedia-uncased-v1": ( "https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json" ), } class __A ( lowerCAmelCase ): '''simple docstring''' lowerCAmelCase_ = """camembert""" def __init__( self , __lowerCAmelCase=3_0_5_2_2 , __lowerCAmelCase=7_6_8 , __lowerCAmelCase=1_2 , __lowerCAmelCase=1_2 , __lowerCAmelCase=3_0_7_2 , __lowerCAmelCase="gelu" , __lowerCAmelCase=0.1 , __lowerCAmelCase=0.1 , __lowerCAmelCase=5_1_2 , __lowerCAmelCase=2 , __lowerCAmelCase=0.02 , __lowerCAmelCase=1E-12 , __lowerCAmelCase=1 , __lowerCAmelCase=0 , __lowerCAmelCase=2 , __lowerCAmelCase="absolute" , __lowerCAmelCase=True , __lowerCAmelCase=None , **__lowerCAmelCase , ): '''simple docstring''' super().__init__(pad_token_id=__lowerCAmelCase , bos_token_id=__lowerCAmelCase , eos_token_id=__lowerCAmelCase , **__lowerCAmelCase ) lowerCamelCase__ = vocab_size lowerCamelCase__ = hidden_size lowerCamelCase__ = num_hidden_layers lowerCamelCase__ = num_attention_heads lowerCamelCase__ = hidden_act lowerCamelCase__ = intermediate_size lowerCamelCase__ = hidden_dropout_prob lowerCamelCase__ = attention_probs_dropout_prob lowerCamelCase__ = max_position_embeddings lowerCamelCase__ = type_vocab_size lowerCamelCase__ = initializer_range lowerCamelCase__ = layer_norm_eps lowerCamelCase__ = position_embedding_type lowerCamelCase__ = use_cache lowerCamelCase__ = classifier_dropout class __A ( lowerCAmelCase ): '''simple docstring''' @property def __lowerCamelCase ( self ): '''simple docstring''' if self.task == "multiple-choice": lowerCamelCase__ = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowerCamelCase__ = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
209
0
from __future__ import annotations import math import numpy as np from numpy.linalg import norm def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase ) -> float: """simple docstring""" return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(lowercase__ , lowercase__ ) ) ) def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase ) -> list[list[list[float] | float]]: """simple docstring""" if dataset.ndim != value_array.ndim: snake_case__ : Dict = ( """Wrong input data's dimensions... """ f"""dataset : {dataset.ndim}, value_array : {value_array.ndim}""" ) raise ValueError(lowercase__ ) try: if dataset.shape[1] != value_array.shape[1]: snake_case__ : Union[str, Any] = ( """Wrong input data's shape... """ f"""dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}""" ) raise ValueError(lowercase__ ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError('''Wrong shape''' ) if dataset.dtype != value_array.dtype: snake_case__ : str = ( """Input data have different datatype... """ f"""dataset : {dataset.dtype}, value_array : {value_array.dtype}""" ) raise TypeError(lowercase__ ) snake_case__ : str = [] for value in value_array: snake_case__ : int = euclidean(lowercase__ , dataset[0] ) snake_case__ : Tuple = dataset[0].tolist() for dataset_value in dataset[1:]: snake_case__ : Any = euclidean(lowercase__ , lowercase__ ) if dist > temp_dist: snake_case__ : Any = temp_dist snake_case__ : List[Any] = dataset_value.tolist() answer.append([vector, dist] ) return answer def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase ) -> float: """simple docstring""" return np.dot(lowercase__ , lowercase__ ) / (norm(lowercase__ ) * norm(lowercase__ )) if __name__ == "__main__": import doctest doctest.testmod()
356
A__ = 0 # The first color of the flag. A__ = 1 # The second color of the flag. A__ = 2 # The third color of the flag. A__ = (red, white, blue) def _lowerCAmelCase ( __lowerCAmelCase ) -> list: """simple docstring""" if not sequence: return [] if len(__lowerCAmelCase ) == 1: return list(__lowerCAmelCase ) snake_case__ : List[Any] = 0 snake_case__ : str = len(__lowerCAmelCase ) - 1 snake_case__ : List[Any] = 0 while mid <= high: if sequence[mid] == colors[0]: snake_case__ , snake_case__ : List[Any] = sequence[mid], sequence[low] low += 1 mid += 1 elif sequence[mid] == colors[1]: mid += 1 elif sequence[mid] == colors[2]: snake_case__ , snake_case__ : int = sequence[high], sequence[mid] high -= 1 else: snake_case__ : List[Any] = f"""The elements inside the sequence must contains only {colors} values""" raise ValueError(__lowerCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod() A__ = input('''Enter numbers separated by commas:\n''').strip() A__ = [int(item.strip()) for item in user_input.split(''',''')] print(f"""{dutch_national_flag_sort(unsorted)}""")
44
0
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __A : List[str] = logging.get_logger(__name__) __A : int = { '''ut/deta''': '''https://huggingface.co/ut/deta/resolve/main/config.json''', } class _UpperCAmelCase ( _A ): SCREAMING_SNAKE_CASE_ : int = "deta" SCREAMING_SNAKE_CASE_ : List[str] = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self : Union[str, Any] , A : Optional[int]=None , A : Union[str, Any]=9_00 , A : Tuple=20_48 , A : int=6 , A : str=20_48 , A : Any=8 , A : Optional[int]=6 , A : Dict=10_24 , A : str=8 , A : Dict=0.0 , A : Union[str, Any]=True , A : List[Any]="relu" , A : Tuple=2_56 , A : Optional[int]=0.1 , A : int=0.0 , A : str=0.0 , A : List[Any]=0.02 , A : Union[str, Any]=1.0 , A : str=True , A : str=False , A : Optional[int]="sine" , A : Optional[Any]=5 , A : str=4 , A : Union[str, Any]=4 , A : Tuple=True , A : Union[str, Any]=3_00 , A : Optional[Any]=True , A : int=True , A : Dict=1 , A : Tuple=5 , A : Optional[Any]=2 , A : Optional[Any]=1 , A : Any=1 , A : int=5 , A : Optional[Any]=2 , A : List[str]=0.1 , A : Dict=0.25 , **A : Tuple , ) -> Dict: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) lowercase_ : Optional[int] = CONFIG_MAPPING['''resnet'''](out_features=['''stage2''', '''stage3''', '''stage4'''] ) else: if isinstance(A , A ): lowercase_ : List[str] = backbone_config.pop('''model_type''' ) lowercase_ : List[str] = CONFIG_MAPPING[backbone_model_type] lowercase_ : Union[str, Any] = config_class.from_dict(A ) lowercase_ : List[str] = backbone_config lowercase_ : Optional[int] = num_queries lowercase_ : str = max_position_embeddings lowercase_ : Any = d_model lowercase_ : Optional[Any] = encoder_ffn_dim lowercase_ : List[str] = encoder_layers lowercase_ : Dict = encoder_attention_heads lowercase_ : int = decoder_ffn_dim lowercase_ : List[Any] = decoder_layers lowercase_ : int = decoder_attention_heads lowercase_ : Optional[Any] = dropout lowercase_ : Tuple = attention_dropout lowercase_ : str = activation_dropout lowercase_ : List[str] = activation_function lowercase_ : int = init_std lowercase_ : Dict = init_xavier_std lowercase_ : List[Any] = encoder_layerdrop lowercase_ : str = auxiliary_loss lowercase_ : Dict = position_embedding_type # deformable attributes lowercase_ : Union[str, Any] = num_feature_levels lowercase_ : Optional[int] = encoder_n_points lowercase_ : Dict = decoder_n_points lowercase_ : Tuple = two_stage lowercase_ : Union[str, Any] = two_stage_num_proposals lowercase_ : Tuple = with_box_refine lowercase_ : Optional[int] = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError('''If two_stage is True, with_box_refine must be True.''' ) # Hungarian matcher lowercase_ : Optional[Any] = class_cost lowercase_ : Dict = bbox_cost lowercase_ : Optional[int] = giou_cost # Loss coefficients lowercase_ : Optional[int] = mask_loss_coefficient lowercase_ : Optional[Any] = dice_loss_coefficient lowercase_ : Dict = bbox_loss_coefficient lowercase_ : int = giou_loss_coefficient lowercase_ : Union[str, Any] = eos_coefficient lowercase_ : Dict = focal_alpha super().__init__(is_encoder_decoder=A , **A ) @property def A ( self : Any ) -> int: return self.encoder_attention_heads @property def A ( self : Optional[int] ) -> int: return self.d_model def A ( self : List[Any] ) -> Dict: lowercase_ : str = copy.deepcopy(self.__dict__ ) lowercase_ : Union[str, Any] = self.backbone_config.to_dict() lowercase_ : List[Any] = self.__class__.model_type return output
33
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import StableDiffusionKDiffusionPipeline from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() @slow @require_torch_gpu class UpperCAmelCase_ ( unittest.TestCase ): def __UpperCAmelCase ( self : Optional[int] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCAmelCase ( self : Tuple ) -> Any: lowerCAmelCase = StableDiffusionKDiffusionPipeline.from_pretrained('CompVis/stable-diffusion-v1-4' ) lowerCAmelCase = sd_pipe.to(UpperCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) sd_pipe.set_scheduler('sample_euler' ) lowerCAmelCase = 'A painting of a squirrel eating a burger' lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = sd_pipe([prompt] , generator=UpperCAmelCase__ , guidance_scale=9.0 , num_inference_steps=2_0 , output_type='np' ) lowerCAmelCase = output.images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) lowerCAmelCase = np.array([0.0_447, 0.0_492, 0.0_468, 0.0_408, 0.0_383, 0.0_408, 0.0_354, 0.0_380, 0.0_339] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __UpperCAmelCase ( self : List[str] ) -> Dict: lowerCAmelCase = StableDiffusionKDiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base' ) lowerCAmelCase = sd_pipe.to(UpperCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) sd_pipe.set_scheduler('sample_euler' ) lowerCAmelCase = 'A painting of a squirrel eating a burger' lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = sd_pipe([prompt] , generator=UpperCAmelCase__ , guidance_scale=9.0 , num_inference_steps=2_0 , output_type='np' ) lowerCAmelCase = output.images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) lowerCAmelCase = np.array([0.1_237, 0.1_320, 0.1_438, 0.1_359, 0.1_390, 0.1_132, 0.1_277, 0.1_175, 0.1_112] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-1 def __UpperCAmelCase ( self : Optional[Any] ) -> List[str]: lowerCAmelCase = StableDiffusionKDiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base' ) lowerCAmelCase = sd_pipe.to(UpperCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=UpperCAmelCase__ ) sd_pipe.set_scheduler('sample_dpmpp_2m' ) lowerCAmelCase = 'A painting of a squirrel eating a burger' lowerCAmelCase = torch.manual_seed(0 ) lowerCAmelCase = sd_pipe( [prompt] , generator=UpperCAmelCase__ , guidance_scale=7.5 , num_inference_steps=1_5 , output_type='np' , use_karras_sigmas=UpperCAmelCase__ , ) lowerCAmelCase = output.images lowerCAmelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) lowerCAmelCase = np.array( [0.11_381_689, 0.12_112_921, 0.1_389_457, 0.12_549_606, 0.1_244_964, 0.10_831_517, 0.11_562_866, 0.10_867_816, 0.10_499_048] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
4
0
import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class __a ( UpperCAmelCase , unittest.TestCase ): # FIXME: add fast tests pass @nightly @require_onnxruntime @require_torch_gpu class __a ( unittest.TestCase ): @property def UpperCAmelCase__ ( self ) -> int: """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCAmelCase__ ( self ) -> Tuple: """simple docstring""" _UpperCAmelCase = ort.SessionOptions() _UpperCAmelCase = False return options def UpperCAmelCase__ ( self ) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo.png' ) _UpperCAmelCase = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo_mask.png' ) _UpperCAmelCase = OnnxStableDiffusionInpaintPipeline.from_pretrained( 'runwayml/stable-diffusion-inpainting' , revision='onnx' , safety_checker=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = 'A red cat sitting on a park bench' _UpperCAmelCase = np.random.RandomState(0 ) _UpperCAmelCase = pipe( prompt=_SCREAMING_SNAKE_CASE , image=_SCREAMING_SNAKE_CASE , mask_image=_SCREAMING_SNAKE_CASE , guidance_scale=7.5 , num_inference_steps=10 , generator=_SCREAMING_SNAKE_CASE , output_type='np' , ) _UpperCAmelCase = output.images _UpperCAmelCase = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) _UpperCAmelCase = np.array([0.2514, 0.3007, 0.3517, 0.1790, 0.2382, 0.3167, 0.1944, 0.2273, 0.2464] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def UpperCAmelCase__ ( self ) -> str: """simple docstring""" _UpperCAmelCase = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo.png' ) _UpperCAmelCase = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo_mask.png' ) _UpperCAmelCase = LMSDiscreteScheduler.from_pretrained( 'runwayml/stable-diffusion-inpainting' , subfolder='scheduler' , revision='onnx' ) _UpperCAmelCase = OnnxStableDiffusionInpaintPipeline.from_pretrained( 'runwayml/stable-diffusion-inpainting' , revision='onnx' , scheduler=_SCREAMING_SNAKE_CASE , safety_checker=_SCREAMING_SNAKE_CASE , feature_extractor=_SCREAMING_SNAKE_CASE , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = 'A red cat sitting on a park bench' _UpperCAmelCase = np.random.RandomState(0 ) _UpperCAmelCase = pipe( prompt=_SCREAMING_SNAKE_CASE , image=_SCREAMING_SNAKE_CASE , mask_image=_SCREAMING_SNAKE_CASE , guidance_scale=7.5 , num_inference_steps=20 , generator=_SCREAMING_SNAKE_CASE , output_type='np' , ) _UpperCAmelCase = output.images _UpperCAmelCase = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) _UpperCAmelCase = np.array([0.0086, 0.0077, 0.0083, 0.0093, 0.0107, 0.0139, 0.0094, 0.0097, 0.0125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3
185
import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class __a : def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=99 , _SCREAMING_SNAKE_CASE=13 , _SCREAMING_SNAKE_CASE=7 , _SCREAMING_SNAKE_CASE=9 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=5 , _SCREAMING_SNAKE_CASE=4 , _SCREAMING_SNAKE_CASE=37 , _SCREAMING_SNAKE_CASE=8 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.002 , _SCREAMING_SNAKE_CASE=1 , _SCREAMING_SNAKE_CASE=0 , _SCREAMING_SNAKE_CASE=0 , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , ) -> Tuple: """simple docstring""" _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = encoder_seq_length _UpperCAmelCase = decoder_seq_length # For common tests _UpperCAmelCase = self.decoder_seq_length _UpperCAmelCase = is_training _UpperCAmelCase = use_attention_mask _UpperCAmelCase = use_labels _UpperCAmelCase = vocab_size _UpperCAmelCase = hidden_size _UpperCAmelCase = num_hidden_layers _UpperCAmelCase = num_attention_heads _UpperCAmelCase = d_ff _UpperCAmelCase = relative_attention_num_buckets _UpperCAmelCase = dropout_rate _UpperCAmelCase = initializer_factor _UpperCAmelCase = eos_token_id _UpperCAmelCase = pad_token_id _UpperCAmelCase = decoder_start_token_id _UpperCAmelCase = None _UpperCAmelCase = decoder_layers def UpperCAmelCase__ ( self ) -> List[str]: """simple docstring""" return TaConfig.from_pretrained('google/umt5-base' ) def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , ) -> List[str]: """simple docstring""" if attention_mask is None: _UpperCAmelCase = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: _UpperCAmelCase = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: _UpperCAmelCase = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=_SCREAMING_SNAKE_CASE ) if decoder_head_mask is None: _UpperCAmelCase = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=_SCREAMING_SNAKE_CASE ) if cross_attn_head_mask is None: _UpperCAmelCase = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=_SCREAMING_SNAKE_CASE ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def UpperCAmelCase__ ( self ) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) _UpperCAmelCase = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input _UpperCAmelCase = input_ids.clamp(self.pad_token_id + 1 ) _UpperCAmelCase = decoder_input_ids.clamp(self.pad_token_id + 1 ) _UpperCAmelCase = self.get_config() _UpperCAmelCase = config.num_attention_heads _UpperCAmelCase = self.prepare_inputs_dict(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) return config, input_dict def UpperCAmelCase__ ( self ) -> Any: """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.prepare_config_and_inputs() return config, inputs_dict def UpperCAmelCase__ ( self ) -> Dict: """simple docstring""" return TaConfig( vocab_size=166 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def UpperCAmelCase__ ( self ) -> List[Any]: """simple docstring""" return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) -> str: """simple docstring""" _UpperCAmelCase = UMTaModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() _UpperCAmelCase = model( input_ids=_SCREAMING_SNAKE_CASE , decoder_input_ids=_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , decoder_attention_mask=_SCREAMING_SNAKE_CASE , ) _UpperCAmelCase = model(input_ids=_SCREAMING_SNAKE_CASE , decoder_input_ids=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = result.last_hidden_state _UpperCAmelCase = result.past_key_values _UpperCAmelCase = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(_SCREAMING_SNAKE_CASE ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) -> Optional[int]: """simple docstring""" _UpperCAmelCase = UMTaModel(config=_SCREAMING_SNAKE_CASE ).get_decoder().to(_SCREAMING_SNAKE_CASE ).eval() # first forward pass _UpperCAmelCase = model(_SCREAMING_SNAKE_CASE , use_cache=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = model(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = model(_SCREAMING_SNAKE_CASE , use_cache=_SCREAMING_SNAKE_CASE ) self.parent.assertTrue(len(_SCREAMING_SNAKE_CASE ) == len(_SCREAMING_SNAKE_CASE ) ) self.parent.assertTrue(len(_SCREAMING_SNAKE_CASE ) == len(_SCREAMING_SNAKE_CASE ) + 1 ) _UpperCAmelCase , _UpperCAmelCase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids _UpperCAmelCase = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and _UpperCAmelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _UpperCAmelCase = model(_SCREAMING_SNAKE_CASE )['last_hidden_state'] _UpperCAmelCase = model(_SCREAMING_SNAKE_CASE , past_key_values=_SCREAMING_SNAKE_CASE )['last_hidden_state'] # select random slice _UpperCAmelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _UpperCAmelCase = output_from_no_past[:, -1, random_slice_idx].detach() _UpperCAmelCase = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1e-3 ) ) def UpperCAmelCase__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) -> List[str]: """simple docstring""" _UpperCAmelCase = UMTaModel(config=_SCREAMING_SNAKE_CASE ).to(_SCREAMING_SNAKE_CASE ).half().eval() _UpperCAmelCase = model(**_SCREAMING_SNAKE_CASE )['last_hidden_state'] self.parent.assertFalse(torch.isnan(_SCREAMING_SNAKE_CASE ).any().item() ) @require_torch class __a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , unittest.TestCase ): _a : Union[str, Any] = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) _a : List[Any] = (UMTaForConditionalGeneration,) if is_torch_available() else () _a : Tuple = ( { 'conversational': UMTaForConditionalGeneration, 'feature-extraction': UMTaModel, 'summarization': UMTaForConditionalGeneration, 'text2text-generation': UMTaForConditionalGeneration, 'translation': UMTaForConditionalGeneration, 'question-answering': UMTaForQuestionAnswering, } if is_torch_available() else {} ) _a : List[str] = True _a : List[Any] = False _a : Tuple = False _a : List[Any] = True _a : str = True # The small UMT5 model needs higher percentages for CPU/MP tests _a : Tuple = [0.8, 0.9] def UpperCAmelCase__ ( self ) -> List[str]: """simple docstring""" _UpperCAmelCase = UMTaModelTester(self ) @unittest.skip('Test has a segmentation fault on torch 1.8.0' ) def UpperCAmelCase__ ( self ) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() _UpperCAmelCase = UMTaModel(config_and_inputs[0] ).to(_SCREAMING_SNAKE_CASE ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( _SCREAMING_SNAKE_CASE , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , f'''{tmpdirname}/t5_test.onnx''' , export_params=_SCREAMING_SNAKE_CASE , opset_version=9 , input_names=['input_ids', 'decoder_input_ids'] , ) @unittest.skipIf(torch_device == 'cpu' , 'Cant do half precision' ) def UpperCAmelCase__ ( self ) -> Tuple: """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__ ( self ) -> int: """simple docstring""" _UpperCAmelCase = ['encoder_attentions', 'decoder_attentions', 'cross_attentions'] _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() _UpperCAmelCase = config_and_inputs[0] _UpperCAmelCase = UMTaForConditionalGeneration(_SCREAMING_SNAKE_CASE ).eval() model.to(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = { 'head_mask': torch.zeros(config.num_layers , config.num_heads , device=_SCREAMING_SNAKE_CASE ), 'decoder_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=_SCREAMING_SNAKE_CASE ), 'cross_attn_head_mask': torch.zeros(config.num_decoder_layers , config.num_heads , device=_SCREAMING_SNAKE_CASE ), } for attn_name, (name, mask) in zip(_SCREAMING_SNAKE_CASE , head_masking.items() ): _UpperCAmelCase = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": _UpperCAmelCase = torch.ones( config.num_decoder_layers , config.num_heads , device=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = model.generate( config_and_inputs[1]['input_ids'] , num_beams=1 , max_length=3 , output_attentions=_SCREAMING_SNAKE_CASE , return_dict_in_generate=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE , ) # We check the state of decoder_attentions and cross_attentions just from the last step _UpperCAmelCase = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip('Does not work on the tiny model as we keep hitting edge cases.' ) def UpperCAmelCase__ ( self ) -> int: """simple docstring""" pass @require_torch @require_sentencepiece @require_tokenizers class __a ( unittest.TestCase ): @slow @unittest.skip( 'Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged' ) def UpperCAmelCase__ ( self ) -> List[str]: """simple docstring""" _UpperCAmelCase = UMTaForConditionalGeneration.from_pretrained('google/umt5-small' , return_dict=_SCREAMING_SNAKE_CASE ).to(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = AutoTokenizer.from_pretrained('google/umt5-small' , use_fast=_SCREAMING_SNAKE_CASE , legacy=_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = [ 'Bonjour monsieur <extra_id_0> bien <extra_id_1>.', 'No se como puedo <extra_id_0>.', 'This is the reason why we <extra_id_0> them.', 'The <extra_id_0> walks in <extra_id_1>, seats', 'A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.', ] _UpperCAmelCase = tokenizer(_SCREAMING_SNAKE_CASE , return_tensors='pt' , padding=_SCREAMING_SNAKE_CASE ).input_ids # fmt: off _UpperCAmelCase = torch.tensor( [ [ 38530, 210703, 256299, 1410, 256298, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 826, 321, 671, 25922, 256299, 274, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 1460, 339, 312, 19014, 10620, 758, 256299, 2355,274, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 517, 256299, 14869, 281, 301, 256298, 275, 119983,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 320, 256299, 14869, 281, 2234, 289, 2275, 333,61391, 289, 256298, 543, 256297, 168714, 329, 256296,274, 1], ] ) # fmt: on torch.testing.assert_allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) _UpperCAmelCase = model.generate(input_ids.to(_SCREAMING_SNAKE_CASE ) ) _UpperCAmelCase = [ '<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>', '<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', '<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>', ] _UpperCAmelCase = tokenizer.batch_decode(_SCREAMING_SNAKE_CASE ) self.assertEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )
185
1
"""simple docstring""" import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": A: Tuple = argparse.ArgumentParser() parser.add_argument( "--checkpoint_path", default=None, type=str, required=True, help="Path to the checkpoint to convert." ) parser.add_argument( "--original_config_file", type=str, required=True, help="The YAML config file corresponding to the original architecture.", ) parser.add_argument( "--num_in_channels", default=None, type=int, help="The number of input channels. If `None` number of input channels will be automatically inferred.", ) parser.add_argument( "--image_size", default=5_1_2, type=int, help=( "The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2" " Base. Use 768 for Stable Diffusion v2." ), ) parser.add_argument( "--extract_ema", action="store_true", help=( "Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights" " or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield" " higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning." ), ) parser.add_argument( "--upcast_attention", action="store_true", help=( "Whether the attention computation should always be upcasted. This is necessary when running stable" " diffusion 2.1." ), ) parser.add_argument( "--from_safetensors", action="store_true", help="If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.", ) parser.add_argument( "--to_safetensors", action="store_true", help="Whether to store pipeline in safetensors format or not.", ) parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the output model.") parser.add_argument("--device", type=str, help="Device to use (e.g. cpu, cuda:0, cuda:1, etc.)") def _snake_case ( UpperCamelCase : int ): if string == "True": return True elif string == "False": return False else: raise ValueError(F"could not parse string as bool {string}" ) parser.add_argument( "--use_linear_projection", help="Override for use linear projection", required=False, type=parse_bool ) parser.add_argument("--cross_attention_dim", help="Override for cross attention_dim", required=False, type=int) A: Union[str, Any] = parser.parse_args() A: Tuple = download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
109
"""simple docstring""" import argparse import re from typing import Dict import torch from datasets import Audio, Dataset, load_dataset, load_metric from transformers import AutoFeatureExtractor, pipeline def _snake_case ( UpperCamelCase : Dataset , UpperCamelCase : Dict[str, str] ): UpperCAmelCase : Any = args.log_outputs UpperCAmelCase : Any = """_""".join(args.dataset.split("""/""" ) + [args.config, args.split] ) # load metric UpperCAmelCase : List[Any] = load_metric("""wer""" ) UpperCAmelCase : Any = load_metric("""cer""" ) # compute metrics UpperCAmelCase : int = wer.compute(references=result["""target"""] , predictions=result["""prediction"""] ) UpperCAmelCase : str = cer.compute(references=result["""target"""] , predictions=result["""prediction"""] ) # print & log results UpperCAmelCase : Tuple = F"WER: {wer_result}\nCER: {cer_result}" print(UpperCamelCase ) with open(F"{dataset_id}_eval_results.txt" , """w""" ) as f: f.write(UpperCamelCase ) # log all results in text file. Possibly interesting for analysis if log_outputs is not None: UpperCAmelCase : str = F"log_{dataset_id}_predictions.txt" UpperCAmelCase : Tuple = F"log_{dataset_id}_targets.txt" with open(UpperCamelCase , """w""" ) as p, open(UpperCamelCase , """w""" ) as t: # mapping function to write output def write_to_file(UpperCamelCase : List[Any] , UpperCamelCase : List[Any] ): p.write(F"{i}" + """\n""" ) p.write(batch["""prediction"""] + """\n""" ) t.write(F"{i}" + """\n""" ) t.write(batch["""target"""] + """\n""" ) result.map(UpperCamelCase , with_indices=UpperCamelCase ) def _snake_case ( UpperCamelCase : str ): UpperCAmelCase : List[str] = """[,?.!\-\;\:\"“%‘”�—’…–]""" # noqa: W605 IMPORTANT: this should correspond to the chars that were ignored during training UpperCAmelCase : Dict = re.sub(UpperCamelCase , """""" , text.lower() ) # In addition, we can normalize the target text, e.g. removing new lines characters etc... # note that order is important here! UpperCAmelCase : List[str] = ["""\n\n""", """\n""", """ """, """ """] for t in token_sequences_to_ignore: UpperCAmelCase : Optional[Any] = """ """.join(text.split(UpperCamelCase ) ) return text def _snake_case ( UpperCamelCase : Tuple ): # load dataset UpperCAmelCase : Union[str, Any] = load_dataset(args.dataset , args.config , split=args.split , use_auth_token=UpperCamelCase ) # for testing: only process the first two examples as a test # dataset = dataset.select(range(10)) # load processor UpperCAmelCase : Optional[int] = AutoFeatureExtractor.from_pretrained(args.model_id ) UpperCAmelCase : Any = feature_extractor.sampling_rate # resample audio UpperCAmelCase : List[str] = dataset.cast_column("""audio""" , Audio(sampling_rate=UpperCamelCase ) ) # load eval pipeline if args.device is None: UpperCAmelCase : Optional[int] = 0 if torch.cuda.is_available() else -1 UpperCAmelCase : Tuple = pipeline("""automatic-speech-recognition""" , model=args.model_id , device=args.device ) # map function to decode audio def map_to_pred(UpperCamelCase : Any ): UpperCAmelCase : Any = asr( batch["""audio"""]["""array"""] , chunk_length_s=args.chunk_length_s , stride_length_s=args.stride_length_s ) UpperCAmelCase : Tuple = prediction["""text"""] UpperCAmelCase : List[str] = normalize_text(batch["""sentence"""] ) return batch # run inference on all examples UpperCAmelCase : int = dataset.map(UpperCamelCase , remove_columns=dataset.column_names ) # compute and log_results # do not change function below log_results(UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": A: List[Any] = argparse.ArgumentParser() parser.add_argument( "--model_id", type=str, required=True, help="Model identifier. Should be loadable with 🤗 Transformers" ) parser.add_argument( "--dataset", type=str, required=True, help="Dataset name to evaluate the `model_id`. Should be loadable with 🤗 Datasets", ) parser.add_argument( "--config", type=str, required=True, help="Config of the dataset. *E.g.* `'en'` for Common Voice" ) parser.add_argument("--split", type=str, required=True, help="Split of the dataset. *E.g.* `'test'`") parser.add_argument( "--chunk_length_s", type=float, default=None, help="Chunk length in seconds. Defaults to 5 seconds." ) parser.add_argument( "--stride_length_s", type=float, default=None, help="Stride of the audio chunks. Defaults to 1 second." ) parser.add_argument( "--log_outputs", action="store_true", help="If defined, write outputs to log file for analysis." ) parser.add_argument( "--device", type=int, default=None, help="The device to run the pipeline on. -1 for CPU (default), 0 for the first GPU and so on.", ) A: Union[str, Any] = parser.parse_args() main(args)
109
1
def UpperCAmelCase ( _lowerCamelCase , _lowerCamelCase ): return int((input_a, input_a).count(1 ) != 0 ) def UpperCAmelCase ( ): assert or_gate(0 , 0 ) == 0 assert or_gate(0 , 1 ) == 1 assert or_gate(1 , 0 ) == 1 assert or_gate(1 , 1 ) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
256
import os import random import sys from . import cryptomath_module as cryptomath from . import rabin_miller __SCREAMING_SNAKE_CASE = 3 def UpperCAmelCase ( _lowerCamelCase ): print("Generating primitive root of p" ) while True: A : str = random.randrange(3 , _lowerCamelCase ) if pow(_lowerCamelCase , 2 , _lowerCamelCase ) == 1: continue if pow(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) == 1: continue return g def UpperCAmelCase ( _lowerCamelCase ): print("Generating prime p..." ) A : int = rabin_miller.generate_large_prime(_lowerCamelCase ) # select large prime number. A : List[str] = primitive_root(_lowerCamelCase ) # one primitive root on modulo p. A : int = random.randrange(3 , _lowerCamelCase ) # private_key -> have to be greater than 2 for safety. A : Tuple = cryptomath.find_mod_inverse(pow(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) , _lowerCamelCase ) A : int = (key_size, e_a, e_a, p) A : str = (key_size, d) return public_key, private_key def UpperCAmelCase ( _lowerCamelCase , _lowerCamelCase ): if os.path.exists(f"""{name}_pubkey.txt""" ) or os.path.exists(f"""{name}_privkey.txt""" ): print("\nWARNING:" ) print( f"""\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n""" "Use a different name or delete these files and re-run this program." ) sys.exit() A , A : Any = generate_key(_lowerCamelCase ) print(f"""\nWriting public key to file {name}_pubkey.txt...""" ) with open(f"""{name}_pubkey.txt""" , "w" ) as fo: fo.write(f"""{public_key[0]},{public_key[1]},{public_key[2]},{public_key[3]}""" ) print(f"""Writing private key to file {name}_privkey.txt...""" ) with open(f"""{name}_privkey.txt""" , "w" ) as fo: fo.write(f"""{private_key[0]},{private_key[1]}""" ) def UpperCAmelCase ( ): print("Making key files..." ) make_key_files("elgamal" , 2048 ) print("Key files generation successful" ) if __name__ == "__main__": main()
256
1
import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def lowerCamelCase_ ( UpperCamelCase__ : int , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Tuple , UpperCamelCase__ : Tuple , UpperCamelCase__ : Union[str, Any] ) -> Dict: """simple docstring""" __lowerCamelCase = TapasConfig.from_json_file(UpperCamelCase__ ) # set absolute/relative position embeddings parameter __lowerCamelCase = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": __lowerCamelCase = TapasForQuestionAnswering(config=UpperCamelCase__ ) elif task == "WTQ": # run_task_main.py hparams __lowerCamelCase = 4 __lowerCamelCase = True # hparam_utils.py hparams __lowerCamelCase = 0.66_46_94 __lowerCamelCase = 0.20_79_51 __lowerCamelCase = 0.12_11_94 __lowerCamelCase = True __lowerCamelCase = True __lowerCamelCase = False __lowerCamelCase = 0.0_35_25_13 __lowerCamelCase = TapasForQuestionAnswering(config=UpperCamelCase__ ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams __lowerCamelCase = 4 __lowerCamelCase = False # hparam_utils.py hparams __lowerCamelCase = 36.45_19 __lowerCamelCase = 0.90_34_21 __lowerCamelCase = 2_22.0_88 __lowerCamelCase = True __lowerCamelCase = True __lowerCamelCase = True __lowerCamelCase = 0.76_31_41 __lowerCamelCase = TapasForQuestionAnswering(config=UpperCamelCase__ ) elif task == "TABFACT": __lowerCamelCase = TapasForSequenceClassification(config=UpperCamelCase__ ) elif task == "MLM": __lowerCamelCase = TapasForMaskedLM(config=UpperCamelCase__ ) elif task == "INTERMEDIATE_PRETRAINING": __lowerCamelCase = TapasModel(config=UpperCamelCase__ ) else: raise ValueError(F"""Task {task} not supported.""" ) print(F"""Building PyTorch model from configuration: {config}""" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # Save pytorch-model (weights and configuration) print(F"""Save PyTorch model to {pytorch_dump_path}""" ) model.save_pretrained(UpperCamelCase__ ) # Save tokenizer files print(F"""Save tokenizer files to {pytorch_dump_path}""" ) __lowerCamelCase = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + 'vocab.txt' , model_max_length=512 ) tokenizer.save_pretrained(UpperCamelCase__ ) print('Used relative position embeddings:' , model.config.reset_position_index_per_cell ) if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument( "--task", default="SQA", type=str, help="Model task for which to convert a checkpoint. Defaults to SQA." ) parser.add_argument( "--reset_position_index_per_cell", default=False, action="store_true", help="Whether to use relative position embeddings or not. Defaults to True.", ) parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--tapas_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained TAPAS model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) __A = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
90
'''simple docstring''' def lowerCAmelCase_ ( snake_case__ = 10 ): '''simple docstring''' if not isinstance(snake_case__ , snake_case__ ) or n < 0: raise ValueError('''Invalid input''' ) A : List[str] = 10**n A : Tuple = 2_8433 * (pow(2 , 783_0457 , snake_case__ )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(f'''{solution(10) = }''')
3
0
def __UpperCAmelCase ( __a : int ,__a : float ,__a : float ) -> float: """simple docstring""" return round(float(moles / volume ) * nfactor ) def __UpperCAmelCase ( __a : float ,__a : float ,__a : float ) -> float: """simple docstring""" return round(float((moles * 0.08_21 * temperature) / (volume) ) ) def __UpperCAmelCase ( __a : float ,__a : float ,__a : float ) -> float: """simple docstring""" return round(float((moles * 0.08_21 * temperature) / (pressure) ) ) def __UpperCAmelCase ( __a : float ,__a : float ,__a : float ) -> float: """simple docstring""" return round(float((pressure * volume) / (0.08_21 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
15
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( HubertConfig, HubertForCTC, HubertModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() a__ = logging.get_logger(__name__) a__ = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } def __UpperCAmelCase ( __a : List[Any] ,__a : Optional[int] ,__a : Optional[int] ,__a : List[str] ,__a : List[Any] ) -> Union[str, Any]: """simple docstring""" for attribute in key.split('''.''' ): _a : Optional[Any] = getattr(__a ,__a ) if weight_type is not None: _a : Dict = getattr(__a ,__a ).shape else: _a : Optional[int] = hf_pointer.shape assert hf_shape == value.shape, ( F"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": _a : List[Any] = value elif weight_type == "weight_g": _a : Any = value elif weight_type == "weight_v": _a : Union[str, Any] = value elif weight_type == "bias": _a : Optional[int] = value else: _a : List[Any] = value logger.info(F"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def __UpperCAmelCase ( __a : Any ,__a : Union[str, Any] ,__a : Union[str, Any] ) -> int: """simple docstring""" _a : Union[str, Any] = [] _a : Union[str, Any] = fairseq_model.state_dict() _a : Union[str, Any] = hf_model.hubert.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): _a : int = False if "conv_layers" in name: load_conv_layer( __a ,__a ,__a ,__a ,hf_model.config.feat_extract_norm == '''group''' ,) _a : Optional[Any] = True else: for key, mapped_key in MAPPING.items(): _a : Union[str, Any] = '''hubert.''' + mapped_key if (is_finetuned and mapped_key != '''lm_head''') else mapped_key if key in name or (key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0] and not is_finetuned): _a : Any = True if "*" in mapped_key: _a : Optional[int] = name.split(__a )[0].split('''.''' )[-2] _a : Any = mapped_key.replace('''*''' ,__a ) if "weight_g" in name: _a : List[Any] = '''weight_g''' elif "weight_v" in name: _a : List[str] = '''weight_v''' elif "weight" in name: _a : Any = '''weight''' elif "bias" in name: _a : str = '''bias''' else: _a : Any = None set_recursively(__a ,__a ,__a ,__a ,__a ) continue if not is_used: unused_weights.append(__a ) logger.warning(F"""Unused weights: {unused_weights}""" ) def __UpperCAmelCase ( __a : int ,__a : Optional[Any] ,__a : Dict ,__a : List[str] ,__a : Any ) -> Tuple: """simple docstring""" _a : int = full_name.split('''conv_layers.''' )[-1] _a : Any = name.split('''.''' ) _a : List[Any] = int(items[0] ) _a : Optional[int] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) _a : Optional[int] = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) _a : Optional[Any] = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) _a : int = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) _a : Any = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(__a ) @torch.no_grad() def __UpperCAmelCase ( __a : Dict ,__a : List[Any] ,__a : List[str]=None ,__a : Optional[int]=None ,__a : int=True ) -> List[Any]: """simple docstring""" if config_path is not None: _a : Tuple = HubertConfig.from_pretrained(__a ) else: _a : Any = HubertConfig() if is_finetuned: if dict_path: _a : Tuple = Dictionary.load(__a ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _a : Any = target_dict.pad_index _a : Tuple = target_dict.bos_index _a : Optional[int] = target_dict.eos_index _a : Optional[Any] = len(target_dict.symbols ) _a : Tuple = os.path.join(__a ,'''vocab.json''' ) if not os.path.isdir(__a ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(__a ) ) return os.makedirs(__a ,exist_ok=__a ) with open(__a ,'''w''' ,encoding='''utf-8''' ) as vocab_handle: json.dump(target_dict.indices ,__a ) _a : Tuple = WavaVecaCTCTokenizer( __a ,unk_token=target_dict.unk_word ,pad_token=target_dict.pad_word ,bos_token=target_dict.bos_word ,eos_token=target_dict.eos_word ,word_delimiter_token='''|''' ,do_lower_case=__a ,) _a : Tuple = True if config.feat_extract_norm == '''layer''' else False _a : List[Any] = WavaVecaFeatureExtractor( feature_size=1 ,sampling_rate=16_000 ,padding_value=0 ,do_normalize=__a ,return_attention_mask=__a ,) _a : List[Any] = WavaVecaProcessor(feature_extractor=__a ,tokenizer=__a ) processor.save_pretrained(__a ) _a : Tuple = HubertForCTC(__a ) else: _a : Tuple = HubertModel(__a ) if is_finetuned: _a , _a , _a : int = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] ,arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: _a , _a , _a : str = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) _a : Any = model[0].eval() recursively_load_weights(__a ,__a ,__a ) hf_wavavec.save_pretrained(__a ) if __name__ == "__main__": a__ = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) a__ = parser.parse_args() convert_hubert_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
15
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : Any = logging.get_logger(__name__) lowerCamelCase : Tuple = { 'facebook/vit-mae-base': 'https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json', # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class __lowerCAmelCase (lowercase_ ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = """vit_mae""" def __init__(self : Any , UpperCamelCase : List[str]=768 , UpperCamelCase : Dict=12 , UpperCamelCase : Tuple=12 , UpperCamelCase : Any=3072 , UpperCamelCase : Dict="gelu" , UpperCamelCase : Tuple=0.0 , UpperCamelCase : Union[str, Any]=0.0 , UpperCamelCase : List[str]=0.02 , UpperCamelCase : str=1E-12 , UpperCamelCase : Optional[Any]=224 , UpperCamelCase : Tuple=16 , UpperCamelCase : str=3 , UpperCamelCase : int=True , UpperCamelCase : List[Any]=16 , UpperCamelCase : str=512 , UpperCamelCase : int=8 , UpperCamelCase : Optional[int]=2048 , UpperCamelCase : List[str]=0.75 , UpperCamelCase : Union[str, Any]=False , **UpperCamelCase : int , ): '''simple docstring''' super().__init__(**UpperCamelCase ) lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = intermediate_size lowercase__ = hidden_act lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = image_size lowercase__ = patch_size lowercase__ = num_channels lowercase__ = qkv_bias lowercase__ = decoder_num_attention_heads lowercase__ = decoder_hidden_size lowercase__ = decoder_num_hidden_layers lowercase__ = decoder_intermediate_size lowercase__ = mask_ratio lowercase__ = norm_pix_loss
2
'''simple docstring''' import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class __lowerCAmelCase : '''simple docstring''' def __init__(self : str , UpperCamelCase : Tuple , UpperCamelCase : Optional[int]=99 , UpperCamelCase : Optional[int]=13 , UpperCamelCase : Tuple=16 , UpperCamelCase : Union[str, Any]=7 , UpperCamelCase : List[Any]=True , UpperCamelCase : List[str]=True , UpperCamelCase : str=True , UpperCamelCase : Tuple=False , UpperCamelCase : str=True , UpperCamelCase : Tuple=2 , UpperCamelCase : Optional[int]=32 , UpperCamelCase : Any=4 , UpperCamelCase : Optional[int]=4 , UpperCamelCase : Tuple=30 , UpperCamelCase : str=0 , UpperCamelCase : Tuple=1 , UpperCamelCase : List[Any]=2 , UpperCamelCase : str=None , ): '''simple docstring''' lowercase__ = parent lowercase__ = batch_size lowercase__ = decoder_seq_length # For common tests lowercase__ = self.decoder_seq_length lowercase__ = is_training lowercase__ = use_attention_mask lowercase__ = use_labels lowercase__ = vocab_size lowercase__ = d_model lowercase__ = d_model lowercase__ = decoder_layers lowercase__ = decoder_layers lowercase__ = decoder_ffn_dim lowercase__ = decoder_attention_heads lowercase__ = decoder_attention_heads lowercase__ = eos_token_id lowercase__ = bos_token_id lowercase__ = pad_token_id lowercase__ = decoder_start_token_id lowercase__ = use_cache lowercase__ = max_position_embeddings lowercase__ = None lowercase__ = decoder_seq_length lowercase__ = 2 lowercase__ = 1 def UpperCamelCase__ (self : str ): '''simple docstring''' lowercase__ = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) lowercase__ = None if self.use_attention_mask: lowercase__ = ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) lowercase__ = None if self.use_labels: lowercase__ = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) lowercase__ = TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def UpperCamelCase__ (self : Tuple , UpperCamelCase : List[Any] , UpperCamelCase : Optional[Any] , UpperCamelCase : Tuple , UpperCamelCase : List[str] , ): '''simple docstring''' lowercase__ = True lowercase__ = TrOCRDecoder(config=UpperCamelCase ).to(UpperCamelCase ).eval() lowercase__ = input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass lowercase__ = model(UpperCamelCase , use_cache=UpperCamelCase ) lowercase__ = model(UpperCamelCase ) lowercase__ = model(UpperCamelCase , use_cache=UpperCamelCase ) self.parent.assertTrue(len(UpperCamelCase ) == len(UpperCamelCase ) ) self.parent.assertTrue(len(UpperCamelCase ) == len(UpperCamelCase ) + 1 ) lowercase__ = outputs['''past_key_values'''] # create hypothetical next token and extent to next_input_ids lowercase__ = ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and lowercase__ = torch.cat([input_ids, next_tokens] , dim=-1 ) lowercase__ = model(UpperCamelCase )['''last_hidden_state'''] lowercase__ = model(UpperCamelCase , past_key_values=UpperCamelCase )['''last_hidden_state'''] # select random slice lowercase__ = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowercase__ = output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() lowercase__ = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(UpperCamelCase , UpperCamelCase , atol=1E-3 ) def UpperCamelCase__ (self : Optional[Any] ): '''simple docstring''' lowercase__ = self.prepare_config_and_inputs() lowercase__ ,lowercase__ ,lowercase__ ,lowercase__ = config_and_inputs lowercase__ = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_torch class __lowerCAmelCase (lowercase_ , lowercase_ , lowercase_ , unittest.TestCase ): '''simple docstring''' lowerCAmelCase__ : List[str] = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () lowerCAmelCase__ : List[Any] = (TrOCRForCausalLM,) if is_torch_available() else () lowerCAmelCase__ : Optional[Any] = {"""text-generation""": TrOCRForCausalLM} if is_torch_available() else {} lowerCAmelCase__ : Optional[Any] = True lowerCAmelCase__ : List[str] = False def UpperCamelCase__ (self : Any ): '''simple docstring''' lowercase__ = TrOCRStandaloneDecoderModelTester(self , is_training=UpperCamelCase ) lowercase__ = ConfigTester(self , config_class=UpperCamelCase ) def UpperCamelCase__ (self : List[str] ): '''simple docstring''' pass def UpperCamelCase__ (self : Optional[int] ): '''simple docstring''' pass def UpperCamelCase__ (self : Any ): '''simple docstring''' pass def UpperCamelCase__ (self : Any ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase__ (self : Union[str, Any] ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*UpperCamelCase ) def UpperCamelCase__ (self : Optional[int] ): '''simple docstring''' return @unittest.skip('''The model doesn\'t support left padding''' ) # and it's not used enough to be worth fixing :) def UpperCamelCase__ (self : List[str] ): '''simple docstring''' pass
2
1
"""simple docstring""" import itertools import string from collections.abc import Generator, Iterable def _lowerCamelCase( a , a ): __a = iter(a ) while True: __a = tuple(itertools.islice(a , a ) ) if not chunk: return yield chunk def _lowerCamelCase( a ): __a = "".join([c.upper() for c in dirty if c in string.ascii_letters] ) __a = "" if len(a ) < 2: return dirty for i in range(len(a ) - 1 ): clean += dirty[i] if dirty[i] == dirty[i + 1]: clean += "X" clean += dirty[-1] if len(a ) & 1: clean += "X" return clean def _lowerCamelCase( a ): # I and J are used interchangeably to allow # us to use a 5x5 table (25 letters) __a = "ABCDEFGHIKLMNOPQRSTUVWXYZ" # we're using a list instead of a '2d' array because it makes the math # for setting up the table and doing the actual encoding/decoding simpler __a = [] # copy key chars into the table if they are in `alphabet` ignoring duplicates for char in key.upper(): if char not in table and char in alphabet: table.append(a ) # fill the rest of the table in with the remaining alphabet chars for char in alphabet: if char not in table: table.append(a ) return table def _lowerCamelCase( a , a ): __a = generate_table(a ) __a = prepare_input(a ) __a = "" # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(a , 2 ): __a , __a = divmod(table.index(a ) , 5 ) __a , __a = divmod(table.index(a ) , 5 ) if rowa == rowa: ciphertext += table[rowa * 5 + (cola + 1) % 5] ciphertext += table[rowa * 5 + (cola + 1) % 5] elif cola == cola: ciphertext += table[((rowa + 1) % 5) * 5 + cola] ciphertext += table[((rowa + 1) % 5) * 5 + cola] else: # rectangle ciphertext += table[rowa * 5 + cola] ciphertext += table[rowa * 5 + cola] return ciphertext def _lowerCamelCase( a , a ): __a = generate_table(a ) __a = "" # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(a , 2 ): __a , __a = divmod(table.index(a ) , 5 ) __a , __a = divmod(table.index(a ) , 5 ) if rowa == rowa: plaintext += table[rowa * 5 + (cola - 1) % 5] plaintext += table[rowa * 5 + (cola - 1) % 5] elif cola == cola: plaintext += table[((rowa - 1) % 5) * 5 + cola] plaintext += table[((rowa - 1) % 5) * 5 + cola] else: # rectangle plaintext += table[rowa * 5 + cola] plaintext += table[rowa * 5 + cola] return plaintext
268
"""simple docstring""" SCREAMING_SNAKE_CASE__:Any = { """A""": """.-""", """B""": """-...""", """C""": """-.-.""", """D""": """-..""", """E""": """.""", """F""": """..-.""", """G""": """--.""", """H""": """....""", """I""": """..""", """J""": """.---""", """K""": """-.-""", """L""": """.-..""", """M""": """--""", """N""": """-.""", """O""": """---""", """P""": """.--.""", """Q""": """--.-""", """R""": """.-.""", """S""": """...""", """T""": """-""", """U""": """..-""", """V""": """...-""", """W""": """.--""", """X""": """-..-""", """Y""": """-.--""", """Z""": """--..""", """1""": """.----""", """2""": """..---""", """3""": """...--""", """4""": """....-""", """5""": """.....""", """6""": """-....""", """7""": """--...""", """8""": """---..""", """9""": """----.""", """0""": """-----""", """&""": """.-...""", """@""": """.--.-.""", """:""": """---...""", """,""": """--..--""", """.""": """.-.-.-""", """'""": """.----.""", """\"""": """.-..-.""", """?""": """..--..""", """/""": """-..-.""", """=""": """-...-""", """+""": """.-.-.""", """-""": """-....-""", """(""": """-.--.""", """)""": """-.--.-""", """!""": """-.-.--""", """ """: """/""" } # Exclamation mark is not in ITU-R recommendation # fmt: on SCREAMING_SNAKE_CASE__:Any = {value: key for key, value in MORSE_CODE_DICT.items()} def _lowerCamelCase( a ): return " ".join(MORSE_CODE_DICT[char] for char in message.upper() ) def _lowerCamelCase( a ): return "".join(REVERSE_DICT[char] for char in message.split() ) def _lowerCamelCase( ): __a = "Morse code here!" print(a ) __a = encrypt(a ) print(a ) __a = decrypt(a ) print(a ) if __name__ == "__main__": main()
268
1
'''simple docstring''' import math def __lowercase ( __lowercase = 100 ) -> int: '''simple docstring''' _A = sum(i * i for i in range(1 , n + 1 ) ) _A = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(F"""{solution() = }""")
79
import torch from torch import nn class snake_case ( nn.Module ): '''simple docstring''' def __init__( self : int , lowerCAmelCase : Tuple , lowerCAmelCase : int , lowerCAmelCase : Any , lowerCAmelCase : Tuple , lowerCAmelCase : int=1 , lowerCAmelCase : List[Any]=False) -> str: """simple docstring""" super().__init__() _snake_case : List[str] = n_token _snake_case : Any = d_embed _snake_case : List[str] = d_proj _snake_case : Optional[int] = cutoffs + [n_token] _snake_case : Dict = [0] + self.cutoffs _snake_case : Optional[Any] = div_val _snake_case : Tuple = self.cutoffs[0] _snake_case : List[str] = len(self.cutoffs) - 1 _snake_case : str = self.shortlist_size + self.n_clusters if self.n_clusters > 0: _snake_case : int = nn.Parameter(torch.zeros(self.n_clusters , self.d_embed)) _snake_case : Any = nn.Parameter(torch.zeros(self.n_clusters)) _snake_case : Tuple = nn.ModuleList() _snake_case : int = nn.ParameterList() if div_val == 1: for i in range(len(self.cutoffs)): if d_proj != d_embed: self.out_projs.append(nn.Parameter(torch.FloatTensor(lowerCAmelCase , lowerCAmelCase))) else: self.out_projs.append(lowerCAmelCase) self.out_layers.append(nn.Linear(lowerCAmelCase , lowerCAmelCase)) else: for i in range(len(self.cutoffs)): _snake_case , _snake_case : Any = self.cutoff_ends[i], self.cutoff_ends[i + 1] _snake_case : Dict = d_embed // (div_val**i) self.out_projs.append(nn.Parameter(torch.FloatTensor(lowerCAmelCase , lowerCAmelCase))) self.out_layers.append(nn.Linear(lowerCAmelCase , r_idx - l_idx)) _snake_case : Tuple = keep_order def UpperCamelCase_ ( self : List[str] , lowerCAmelCase : Any , lowerCAmelCase : Any , lowerCAmelCase : Dict , lowerCAmelCase : Optional[int]) -> List[str]: """simple docstring""" if proj is None: _snake_case : List[Any] = nn.functional.linear(lowerCAmelCase , lowerCAmelCase , bias=lowerCAmelCase) else: # if CUDA_MAJOR <= 9 and CUDA_MINOR <= 1: _snake_case : List[str] = nn.functional.linear(lowerCAmelCase , proj.t().contiguous()) _snake_case : Optional[int] = nn.functional.linear(lowerCAmelCase , lowerCAmelCase , bias=lowerCAmelCase) # else: # logit = torch.einsum('bd,de,ev->bv', (hidden, proj, weight.t())) # if bias is not None: # logit = logit + bias return logit def UpperCamelCase_ ( self : Optional[Any] , lowerCAmelCase : Dict , lowerCAmelCase : Optional[Any]=None , lowerCAmelCase : int=False) -> Tuple: """simple docstring""" if labels is not None: # Shift so that tokens < n predict n _snake_case : List[str] = hidden[..., :-1, :].contiguous() _snake_case : int = labels[..., 1:].contiguous() _snake_case : int = hidden.view(-1 , hidden.size(-1)) _snake_case : str = labels.view(-1) if hidden.size(0) != labels.size(0): raise RuntimeError("""Input and labels should have the same size in the batch dimension.""") else: _snake_case : List[Any] = hidden.view(-1 , hidden.size(-1)) if self.n_clusters == 0: _snake_case : int = self._compute_logit(lowerCAmelCase , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0]) if labels is not None: _snake_case : Optional[int] = labels != -100 _snake_case : Union[str, Any] = torch.zeros_like(lowerCAmelCase , dtype=hidden.dtype , device=hidden.device) _snake_case : Union[str, Any] = ( -nn.functional.log_softmax(lowerCAmelCase , dim=-1)[mask].gather(1 , labels[mask].unsqueeze(1)).squeeze(1) ) else: _snake_case : Optional[int] = nn.functional.log_softmax(lowerCAmelCase , dim=-1) else: # construct weights and biases _snake_case , _snake_case : Optional[int] = [], [] for i in range(len(self.cutoffs)): if self.div_val == 1: _snake_case , _snake_case : Any = self.cutoff_ends[i], self.cutoff_ends[i + 1] _snake_case : Dict = self.out_layers[0].weight[l_idx:r_idx] _snake_case : Tuple = self.out_layers[0].bias[l_idx:r_idx] else: _snake_case : Any = self.out_layers[i].weight _snake_case : Optional[int] = self.out_layers[i].bias if i == 0: _snake_case : Dict = torch.cat([weight_i, self.cluster_weight] , dim=0) _snake_case : List[str] = torch.cat([bias_i, self.cluster_bias] , dim=0) weights.append(lowerCAmelCase) biases.append(lowerCAmelCase) _snake_case , _snake_case , _snake_case : List[Any] = weights[0], biases[0], self.out_projs[0] _snake_case : List[str] = self._compute_logit(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase) _snake_case : Dict = nn.functional.log_softmax(lowerCAmelCase , dim=1) if labels is None: _snake_case : List[Any] = hidden.new_empty((head_logit.size(0), self.n_token)) else: _snake_case : Optional[Any] = torch.zeros_like(lowerCAmelCase , dtype=hidden.dtype , device=hidden.device) _snake_case : Optional[int] = 0 _snake_case : Union[str, Any] = [0] + self.cutoffs for i in range(len(lowerCAmelCase) - 1): _snake_case , _snake_case : Any = cutoff_values[i], cutoff_values[i + 1] if labels is not None: _snake_case : Optional[int] = (labels >= l_idx) & (labels < r_idx) _snake_case : Dict = mask_i.nonzero().squeeze() if indices_i.numel() == 0: continue _snake_case : Dict = labels.index_select(0 , lowerCAmelCase) - l_idx _snake_case : List[Any] = head_logprob.index_select(0 , lowerCAmelCase) _snake_case : Dict = hidden.index_select(0 , lowerCAmelCase) else: _snake_case : Optional[Any] = hidden if i == 0: if labels is not None: _snake_case : str = head_logprob_i.gather(1 , target_i[:, None]).squeeze(1) else: _snake_case : int = head_logprob[:, : self.cutoffs[0]] else: _snake_case , _snake_case , _snake_case : Dict = weights[i], biases[i], self.out_projs[i] _snake_case : int = self._compute_logit(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase) _snake_case : List[str] = nn.functional.log_softmax(lowerCAmelCase , dim=1) _snake_case : str = self.cutoffs[0] + i - 1 # No probability for the head cluster if labels is not None: _snake_case : Dict = head_logprob_i[:, cluster_prob_idx] + tail_logprob_i.gather( 1 , target_i[:, None]).squeeze(1) else: _snake_case : Tuple = head_logprob[:, cluster_prob_idx, None] + tail_logprob_i _snake_case : int = logprob_i if labels is not None: if (hasattr(self , """keep_order""") and self.keep_order) or keep_order: out.index_copy_(0 , lowerCAmelCase , -logprob_i) else: out[offset : offset + logprob_i.size(0)].copy_(-logprob_i) offset += logprob_i.size(0) return out def UpperCamelCase_ ( self : Union[str, Any] , lowerCAmelCase : Optional[int]) -> Tuple: """simple docstring""" if self.n_clusters == 0: _snake_case : Optional[Any] = self._compute_logit(lowerCAmelCase , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0]) return nn.functional.log_softmax(lowerCAmelCase , dim=-1) else: # construct weights and biases _snake_case , _snake_case : Optional[int] = [], [] for i in range(len(self.cutoffs)): if self.div_val == 1: _snake_case , _snake_case : Optional[Any] = self.cutoff_ends[i], self.cutoff_ends[i + 1] _snake_case : Optional[Any] = self.out_layers[0].weight[l_idx:r_idx] _snake_case : Union[str, Any] = self.out_layers[0].bias[l_idx:r_idx] else: _snake_case : Tuple = self.out_layers[i].weight _snake_case : Any = self.out_layers[i].bias if i == 0: _snake_case : Tuple = torch.cat([weight_i, self.cluster_weight] , dim=0) _snake_case : Optional[Any] = torch.cat([bias_i, self.cluster_bias] , dim=0) weights.append(lowerCAmelCase) biases.append(lowerCAmelCase) _snake_case , _snake_case , _snake_case : int = weights[0], biases[0], self.out_projs[0] _snake_case : Union[str, Any] = self._compute_logit(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase) _snake_case : Any = hidden.new_empty((head_logit.size(0), self.n_token)) _snake_case : Optional[Any] = nn.functional.log_softmax(lowerCAmelCase , dim=1) _snake_case : List[Any] = [0] + self.cutoffs for i in range(len(lowerCAmelCase) - 1): _snake_case , _snake_case : Any = cutoff_values[i], cutoff_values[i + 1] if i == 0: _snake_case : Union[str, Any] = head_logprob[:, : self.cutoffs[0]] else: _snake_case , _snake_case , _snake_case : str = weights[i], biases[i], self.out_projs[i] _snake_case : List[str] = self._compute_logit(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase) _snake_case : str = nn.functional.log_softmax(lowerCAmelCase , dim=1) _snake_case : Dict = head_logprob[:, -i] + tail_logprob_i _snake_case : Any = logprob_i return out
317
0
"""simple docstring""" from __future__ import annotations import math _lowerCAmelCase :Tuple = '2020.9.26' _lowerCAmelCase :int = 'xcodz-dot, cclaus, dhruvmanila' def lowerCamelCase_ (UpperCamelCase__ : Dict , UpperCamelCase__ : List[Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Union[str, Any] ): if not all(isinstance(_lowerCAmelCase , (float, int) ) for val in locals().values() ): _UpperCAmelCase : Tuple = F'Input values must either be float or int: {list(locals().values() )}' raise TypeError(_lowerCAmelCase ) _UpperCAmelCase : List[str] = ((x * distance) / (z + distance)) * scale _UpperCAmelCase : List[Any] = ((y * distance) / (z + distance)) * scale return projected_x, projected_y def lowerCamelCase_ (UpperCamelCase__ : List[Any] , UpperCamelCase__ : str , UpperCamelCase__ : str , UpperCamelCase__ : Any , UpperCamelCase__ : List[str] ): if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise TypeError('''Axis must be a str''' ) _UpperCAmelCase : List[str] = locals() del input_variables["axis"] if not all(isinstance(_lowerCAmelCase , (float, int) ) for val in input_variables.values() ): _UpperCAmelCase : Union[str, Any] = ( """Input values except axis must either be float or int: """ F'{list(input_variables.values() )}' ) raise TypeError(_lowerCAmelCase ) _UpperCAmelCase : Optional[int] = (angle % 360) / 450 * 180 / math.pi if axis == "z": _UpperCAmelCase : Dict = x * math.cos(_lowerCAmelCase ) - y * math.sin(_lowerCAmelCase ) _UpperCAmelCase : Any = y * math.cos(_lowerCAmelCase ) + x * math.sin(_lowerCAmelCase ) _UpperCAmelCase : Union[str, Any] = z elif axis == "x": _UpperCAmelCase : Tuple = y * math.cos(_lowerCAmelCase ) - z * math.sin(_lowerCAmelCase ) _UpperCAmelCase : Any = z * math.cos(_lowerCAmelCase ) + y * math.sin(_lowerCAmelCase ) _UpperCAmelCase : Optional[int] = x elif axis == "y": _UpperCAmelCase : Tuple = x * math.cos(_lowerCAmelCase ) - z * math.sin(_lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = z * math.cos(_lowerCAmelCase ) + x * math.sin(_lowerCAmelCase ) _UpperCAmelCase : List[str] = y else: raise ValueError('''not a valid axis, choose one of \'x\', \'y\', \'z\'''' ) return new_x, new_y, new_z if __name__ == "__main__": import doctest doctest.testmod() print(f"{convert_to_ad(1.0, 2.0, 3.0, 10.0, 10.0) = }") print(f"{rotate(1.0, 2.0, 3.0, 'y', 90.0) = }")
371
"""simple docstring""" from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def lowerCamelCase_ (): _UpperCAmelCase : int = HfArgumentParser(UpperCamelCase__ ) _UpperCAmelCase : Optional[Any] = parser.parse_args_into_dataclasses()[0] _UpperCAmelCase : Union[str, Any] = TensorFlowBenchmark(args=UpperCamelCase__ ) try: _UpperCAmelCase : Optional[int] = parser.parse_args_into_dataclasses()[0] except ValueError as e: _UpperCAmelCase : List[Any] = '''Arg --no_{0} is no longer used, please use --no-{0} instead.''' _UpperCAmelCase : Tuple = ''' '''.join(str(UpperCamelCase__ ).split(''' ''' )[:-1] ) _UpperCAmelCase : int = '''''' _UpperCAmelCase : List[Any] = eval(str(UpperCamelCase__ ).split(''' ''' )[-1] ) _UpperCAmelCase : int = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(UpperCamelCase__ ) if len(UpperCamelCase__ ) > 0: _UpperCAmelCase : Union[str, Any] = full_error_msg + begin_error_msg + str(UpperCamelCase__ ) raise ValueError(UpperCamelCase__ ) benchmark.run() if __name__ == "__main__": main()
68
0
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ChineseCLIPImageProcessor class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" def __init__( self , snake_case__ , snake_case__=7 , snake_case__=3 , snake_case__=18 , snake_case__=30 , snake_case__=400 , snake_case__=True , snake_case__=None , snake_case__=True , snake_case__=None , snake_case__=True , snake_case__=[0.48145466, 0.4578275, 0.40821073] , snake_case__=[0.26862954, 0.26130258, 0.27577711] , snake_case__=True , ): """simple docstring""" lowerCAmelCase : str = size if size is not None else {"height": 224, "width": 224} lowerCAmelCase : str = crop_size if crop_size is not None else {"height": 18, "width": 18} lowerCAmelCase : Union[str, Any] = parent lowerCAmelCase : List[Any] = batch_size lowerCAmelCase : Union[str, Any] = num_channels lowerCAmelCase : Tuple = image_size lowerCAmelCase : Dict = min_resolution lowerCAmelCase : Optional[Any] = max_resolution lowerCAmelCase : int = do_resize lowerCAmelCase : Any = size lowerCAmelCase : Any = do_center_crop lowerCAmelCase : Tuple = crop_size lowerCAmelCase : List[str] = do_normalize lowerCAmelCase : Any = image_mean lowerCAmelCase : Any = image_std lowerCAmelCase : Tuple = do_convert_rgb def lowercase__ ( self ): """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_convert_rgb": self.do_convert_rgb, } def lowercase__ ( self , snake_case__=False , snake_case__=False , snake_case__=False ): """simple docstring""" assert not (numpify and torchify), "You cannot specify both numpy and PyTorch tensors at the same time" if equal_resolution: lowerCAmelCase : Any = [] for i in range(self.batch_size ): image_inputs.append( np.random.randint( 255 , size=(self.num_channels, self.max_resolution, self.max_resolution) , dtype=np.uinta ) ) else: lowerCAmelCase : List[str] = [] for i in range(self.batch_size ): lowerCAmelCase , lowerCAmelCase : Optional[Any] = np.random.choice(np.arange(self.min_resolution , self.max_resolution ) , 2 ) image_inputs.append(np.random.randint(255 , size=(self.num_channels, width, height) , dtype=np.uinta ) ) if not numpify and not torchify: # PIL expects the channel dimension as last dimension lowerCAmelCase : int = [Image.fromarray(np.moveaxis(snake_case__ , 0 , -1 ) ) for x in image_inputs] if torchify: lowerCAmelCase : Optional[int] = [torch.from_numpy(snake_case__ ) for x in image_inputs] return image_inputs @require_torch @require_vision class SCREAMING_SNAKE_CASE__ ( lowercase , unittest.TestCase ): """simple docstring""" a : int =ChineseCLIPImageProcessor if is_vision_available() else None def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : int = ChineseCLIPImageProcessingTester(self , do_center_crop=snake_case__ ) @property def lowercase__ ( self ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case__ , "do_resize" ) ) self.assertTrue(hasattr(snake_case__ , "size" ) ) self.assertTrue(hasattr(snake_case__ , "do_center_crop" ) ) self.assertTrue(hasattr(snake_case__ , "center_crop" ) ) self.assertTrue(hasattr(snake_case__ , "do_normalize" ) ) self.assertTrue(hasattr(snake_case__ , "image_mean" ) ) self.assertTrue(hasattr(snake_case__ , "image_std" ) ) self.assertTrue(hasattr(snake_case__ , "do_convert_rgb" ) ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Tuple = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"height": 224, "width": 224} ) self.assertEqual(image_processor.crop_size , {"height": 18, "width": 18} ) lowerCAmelCase : str = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84} ) def lowercase__ ( self ): """simple docstring""" pass def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase : Tuple = self.image_processor_tester.prepare_inputs(equal_resolution=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , Image.Image ) # Test not batched input lowerCAmelCase : Optional[int] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched lowerCAmelCase : Tuple = image_processing(snake_case__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase : Any = self.image_processor_tester.prepare_inputs(equal_resolution=snake_case__ , numpify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , np.ndarray ) # Test not batched input lowerCAmelCase : Optional[int] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched lowerCAmelCase : Tuple = image_processing(snake_case__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase : int = self.image_processor_tester.prepare_inputs(equal_resolution=snake_case__ , torchify=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , torch.Tensor ) # Test not batched input lowerCAmelCase : Any = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched lowerCAmelCase : List[Any] = image_processing(snake_case__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) @require_torch @require_vision class SCREAMING_SNAKE_CASE__ ( lowercase , unittest.TestCase ): """simple docstring""" a : Optional[int] =ChineseCLIPImageProcessor if is_vision_available() else None def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : List[Any] = ChineseCLIPImageProcessingTester(self , num_channels=4 , do_center_crop=snake_case__ ) lowerCAmelCase : Optional[int] = 3 @property def lowercase__ ( self ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : Any = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case__ , "do_resize" ) ) self.assertTrue(hasattr(snake_case__ , "size" ) ) self.assertTrue(hasattr(snake_case__ , "do_center_crop" ) ) self.assertTrue(hasattr(snake_case__ , "center_crop" ) ) self.assertTrue(hasattr(snake_case__ , "do_normalize" ) ) self.assertTrue(hasattr(snake_case__ , "image_mean" ) ) self.assertTrue(hasattr(snake_case__ , "image_std" ) ) self.assertTrue(hasattr(snake_case__ , "do_convert_rgb" ) ) def lowercase__ ( self ): """simple docstring""" pass def lowercase__ ( self ): """simple docstring""" lowerCAmelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase : Any = self.image_processor_tester.prepare_inputs(equal_resolution=snake_case__ ) for image in image_inputs: self.assertIsInstance(snake_case__ , Image.Image ) # Test not batched input lowerCAmelCase : Tuple = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched lowerCAmelCase : Dict = image_processing(snake_case__ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
108
import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class __magic_name__ ( unittest.TestCase): def UpperCAmelCase__ ( self : List[Any] ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ : Optional[int] = '''ZinengTang/tvlt-base''' UpperCamelCase__ : int = tempfile.mkdtemp() def UpperCAmelCase__ ( self : int , **lowerCamelCase__ : List[str] ) -> List[Any]: '''simple docstring''' return TvltImageProcessor.from_pretrained(self.checkpoint , **lowerCamelCase__ ) def UpperCAmelCase__ ( self : Optional[Any] , **lowerCamelCase__ : Tuple ) -> List[Any]: '''simple docstring''' return TvltFeatureExtractor.from_pretrained(self.checkpoint , **lowerCamelCase__ ) def UpperCAmelCase__ ( self : str ) -> Tuple: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def UpperCAmelCase__ ( self : Any ) -> int: '''simple docstring''' UpperCamelCase__ : int = self.get_image_processor() UpperCamelCase__ : Union[str, Any] = self.get_feature_extractor() UpperCamelCase__ : List[str] = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) processor.save_pretrained(self.tmpdirname ) UpperCamelCase__ : Optional[int] = TvltProcessor.from_pretrained(self.tmpdirname ) self.assertIsInstance(processor.feature_extractor , lowerCamelCase__ ) self.assertIsInstance(processor.image_processor , lowerCamelCase__ ) def UpperCAmelCase__ ( self : List[Any] ) -> Tuple: '''simple docstring''' UpperCamelCase__ : str = self.get_image_processor() UpperCamelCase__ : List[Any] = self.get_feature_extractor() UpperCamelCase__ : Dict = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) UpperCamelCase__ : Any = np.ones([12000] ) UpperCamelCase__ : Union[str, Any] = feature_extractor(lowerCamelCase__ , return_tensors='''np''' ) UpperCamelCase__ : Any = processor(audio=lowerCamelCase__ , return_tensors='''np''' ) for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCAmelCase__ ( self : List[Any] ) -> List[Any]: '''simple docstring''' UpperCamelCase__ : List[Any] = self.get_image_processor() UpperCamelCase__ : Any = self.get_feature_extractor() UpperCamelCase__ : int = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) UpperCamelCase__ : int = np.ones([3, 224, 224] ) UpperCamelCase__ : List[str] = image_processor(lowerCamelCase__ , return_tensors='''np''' ) UpperCamelCase__ : str = processor(images=lowerCamelCase__ , return_tensors='''np''' ) for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1E-2 ) def UpperCAmelCase__ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ : Any = self.get_image_processor() UpperCamelCase__ : Dict = self.get_feature_extractor() UpperCamelCase__ : Union[str, Any] = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) UpperCamelCase__ : List[str] = np.ones([12000] ) UpperCamelCase__ : Tuple = np.ones([3, 224, 224] ) UpperCamelCase__ : Optional[Any] = processor(audio=lowerCamelCase__ , images=lowerCamelCase__ ) self.assertListEqual(list(inputs.keys() ) , ['''audio_values''', '''audio_mask''', '''pixel_values''', '''pixel_mask'''] ) # test if it raises when no input is passed with pytest.raises(lowerCamelCase__ ): processor() def UpperCAmelCase__ ( self : Dict ) -> int: '''simple docstring''' UpperCamelCase__ : List[str] = self.get_image_processor() UpperCamelCase__ : str = self.get_feature_extractor() UpperCamelCase__ : Tuple = TvltProcessor(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg='''`processor` and `image_processor`+`feature_extractor` model input names do not match''' , )
146
0
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class UpperCamelCase__( lowerCAmelCase , unittest.TestCase ): __magic_name__ : Optional[int] = TextToVideoSDPipeline __magic_name__ : List[Any] = TEXT_TO_IMAGE_PARAMS __magic_name__ : List[str] = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. __magic_name__ : List[Any] = frozenset( [ "num_inference_steps", "generator", "latents", "return_dict", "callback", "callback_steps", ] ) def a__( self : str )-> Union[str, Any]: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''DownBlock3D''') , up_block_types=('''UpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''') , cross_attention_dim=32 , attention_head_dim=4 , ) UpperCAmelCase = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=lowerCAmelCase , set_alpha_to_one=lowerCAmelCase , ) torch.manual_seed(0 ) UpperCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) UpperCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='''gelu''' , projection_dim=512 , ) UpperCAmelCase = CLIPTextModel(lowerCAmelCase ) UpperCAmelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) UpperCAmelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def a__( self : str , lowerCAmelCase : int , lowerCAmelCase : Optional[Any]=0 )-> Dict: """simple docstring""" if str(lowerCAmelCase ).startswith('''mps''' ): UpperCAmelCase = torch.manual_seed(lowerCAmelCase ) else: UpperCAmelCase = torch.Generator(device=lowerCAmelCase ).manual_seed(lowerCAmelCase ) UpperCAmelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def a__( self : Optional[Any] )-> List[str]: """simple docstring""" UpperCAmelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = TextToVideoSDPipeline(**lowerCAmelCase ) UpperCAmelCase = sd_pipe.to(lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase ) UpperCAmelCase = self.get_dummy_inputs(lowerCAmelCase ) UpperCAmelCase = '''np''' UpperCAmelCase = sd_pipe(**lowerCAmelCase ).frames UpperCAmelCase = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) UpperCAmelCase = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def a__( self : List[str] )-> List[str]: """simple docstring""" self._test_attention_slicing_forward_pass(test_mean_pixel_difference=lowerCAmelCase , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def a__( self : List[str] )-> str: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCAmelCase , expected_max_diff=1E-2 ) @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''' ) def a__( self : Union[str, Any] )-> str: """simple docstring""" pass @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''' ) def a__( self : List[Any] )-> int: """simple docstring""" pass @unittest.skip(reason='''`num_images_per_prompt` argument is not supported for this pipeline.''' ) def a__( self : int )-> Tuple: """simple docstring""" pass def a__( self : Dict )-> Optional[Any]: """simple docstring""" return super().test_progress_bar() @slow @skip_mps class UpperCamelCase__( unittest.TestCase ): def a__( self : str )-> List[Any]: """simple docstring""" UpperCAmelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy''' ) UpperCAmelCase = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''' ) UpperCAmelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) UpperCAmelCase = pipe.to('''cuda''' ) UpperCAmelCase = '''Spiderman is surfing''' UpperCAmelCase = torch.Generator(device='''cpu''' ).manual_seed(0 ) UpperCAmelCase = pipe(lowerCAmelCase , generator=lowerCAmelCase , num_inference_steps=25 , output_type='''pt''' ).frames UpperCAmelCase = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def a__( self : int )-> str: """simple docstring""" UpperCAmelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy''' ) UpperCAmelCase = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''' ) UpperCAmelCase = pipe.to('''cuda''' ) UpperCAmelCase = '''Spiderman is surfing''' UpperCAmelCase = torch.Generator(device='''cpu''' ).manual_seed(0 ) UpperCAmelCase = pipe(lowerCAmelCase , generator=lowerCAmelCase , num_inference_steps=2 , output_type='''pt''' ).frames UpperCAmelCase = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
91
'''simple docstring''' import numpy as np from cva import COLOR_BGR2GRAY, CV_8UC3, cvtColor, filteraD, imread, imshow, waitKey def lowerCamelCase__ ( A : int , A : int , A : int , A : int , A : int , A : int ): '''simple docstring''' if (ksize % 2) == 0: UpperCAmelCase = ksize + 1 UpperCAmelCase = np.zeros((ksize, ksize) , dtype=np.floataa ) # each value for y in range(A ): for x in range(A ): # distance from center UpperCAmelCase = x - ksize // 2 UpperCAmelCase = y - ksize // 2 # degree to radiant UpperCAmelCase = theta / 1_80 * np.pi UpperCAmelCase = np.cos(_theta ) UpperCAmelCase = np.sin(_theta ) # get kernel x UpperCAmelCase = cos_theta * px + sin_theta * py # get kernel y UpperCAmelCase = -sin_theta * px + cos_theta * py # fill kernel UpperCAmelCase = np.exp( -(_x**2 + gamma**2 * _y**2) / (2 * sigma**2) ) * np.cos(2 * np.pi * _x / lambd + psi ) return gabor if __name__ == "__main__": import doctest doctest.testmod() # read original image _lowercase : Tuple = imread("""../image_data/lena.jpg""") # turn image in gray scale value _lowercase : int = cvtColor(img, COLOR_BGR2GRAY) # Apply multiple Kernel to detect edges _lowercase : List[str] = np.zeros(gray.shape[:2]) for theta in [0, 30, 60, 90, 120, 150]: _lowercase : List[Any] = gabor_filter_kernel(10, 8, theta, 10, 0, 0) out += filteraD(gray, CV_8UC3, kernel_aa) _lowercase : Optional[int] = out / out.max() * 255 _lowercase : Optional[int] = out.astype(np.uinta) imshow("""Original""", gray) imshow("""Gabor filter with 20x20 mask and 6 directions""", out) waitKey(0)
91
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = {'vocab_file': 'sentencepiece.bpe.model'} lowerCAmelCase_ = { 'vocab_file': { 'moussaKam/mbarthez': 'https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model', 'moussaKam/barthez': 'https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model', 'moussaKam/barthez-orangesum-title': ( 'https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model' ), }, } lowerCAmelCase_ = { 'moussaKam/mbarthez': 10_24, 'moussaKam/barthez': 10_24, 'moussaKam/barthez-orangesum-title': 10_24, } lowerCAmelCase_ = '▁' class _A ( _lowerCamelCase ): _UpperCamelCase : Union[str, Any] = VOCAB_FILES_NAMES _UpperCamelCase : List[str] = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCamelCase : int = ['''input_ids''', '''attention_mask'''] def __init__( self : Optional[Any] , _A : Optional[int] , _A : List[Any]="<s>" , _A : List[str]="</s>" , _A : int="</s>" , _A : Any="<s>" , _A : Optional[int]="<unk>" , _A : Tuple="<pad>" , _A : Tuple="<mask>" , _A : Optional[Dict[str, Any]] = None , **_A : int , ) -> None: """simple docstring""" lowercase : Union[str, Any] = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else mask_token lowercase : List[str] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_A , eos_token=_A , unk_token=_A , sep_token=_A , cls_token=_A , pad_token=_A , mask_token=_A , sp_model_kwargs=self.sp_model_kwargs , **_A , ) lowercase : int = vocab_file lowercase : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_A ) ) lowercase : Optional[int] = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} lowercase : Optional[int] = len(self.sp_model ) - 1 lowercase : Tuple = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __a ( self : Any , _A : List[int] , _A : Optional[List[int]] = None ) -> List[int]: """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase : Optional[Any] = [self.cls_token_id] lowercase : Dict = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __a ( self : int , _A : List[int] , _A : Optional[List[int]] = None , _A : bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_A , token_ids_a=_A , already_has_special_tokens=_A ) if token_ids_a is None: return [1] + ([0] * len(_A )) + [1] return [1] + ([0] * len(_A )) + [1, 1] + ([0] * len(_A )) + [1] def __a ( self : Any , _A : List[int] , _A : Optional[List[int]] = None ) -> List[int]: """simple docstring""" lowercase : str = [self.sep_token_id] lowercase : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __a ( self : int ) -> Tuple: """simple docstring""" return len(self.sp_model ) def __a ( self : List[str] ) -> Optional[Any]: """simple docstring""" lowercase : List[Any] = {self.convert_ids_to_tokens(_A ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __a ( self : Any , _A : str ) -> List[str]: """simple docstring""" return self.sp_model.encode(_A , out_type=_A ) def __a ( self : Optional[int] , _A : Optional[Any] ) -> Tuple: """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] lowercase : Any = self.sp_model.PieceToId(_A ) return spm_id if spm_id else self.unk_token_id def __a ( self : Optional[Any] , _A : Optional[int] ) -> Optional[Any]: """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(_A ) def __a ( self : List[str] , _A : str ) -> Dict: """simple docstring""" lowercase : Union[str, Any] = [] lowercase : Any = '''''' lowercase : Optional[Any] = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(_A ) + token lowercase : str = True lowercase : int = [] else: current_sub_tokens.append(_A ) lowercase : List[str] = False out_string += self.sp_model.decode(_A ) return out_string.strip() def __getstate__( self : Union[str, Any] ) -> List[Any]: """simple docstring""" lowercase : Any = self.__dict__.copy() lowercase : Union[str, Any] = None return state def __setstate__( self : List[Any] , _A : Dict ) -> int: """simple docstring""" lowercase : str = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): lowercase : List[str] = {} lowercase : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __a ( self : int , _A : str , _A : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(_A ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowercase : Optional[Any] = os.path.join( _A , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_A ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _A ) elif not os.path.isfile(self.vocab_file ): with open(_A , '''wb''' ) as fi: lowercase : List[Any] = self.sp_model.serialized_model_proto() fi.write(_A ) return (out_vocab_file,)
308
def snake_case( __magic_name__ , __magic_name__ ) -> float: '''simple docstring''' return price * (1 + tax_rate) if __name__ == "__main__": print(f'''{price_plus_tax(1_00, 0.2_5) = }''') print(f'''{price_plus_tax(1_2_5.5_0, 0.0_5) = }''')
308
1
'''simple docstring''' from __future__ import annotations def _a ( _lowerCamelCase : List[str] ) -> list: """simple docstring""" if len(_lowerCamelCase ) == 0: return [] __snake_case : Tuple = min(_lowerCamelCase ), max(_lowerCamelCase ) __snake_case : List[Any] = int(max_value - min_value ) + 1 __snake_case : list[list] = [[] for _ in range(_lowerCamelCase )] for i in my_list: buckets[int(i - min_value )].append(_lowerCamelCase ) return [v for bucket in buckets for v in sorted(_lowerCamelCase )] if __name__ == "__main__": from doctest import testmod testmod() assert bucket_sort([4, 5, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bucket_sort([0, 1, -10, 15, 2, -2]) == [-10, -2, 0, 1, 2, 15]
358
'''simple docstring''' from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = { "Salesforce/codegen-350M-nl": "https://huggingface.co/Salesforce/codegen-350M-nl/resolve/main/config.json", "Salesforce/codegen-350M-multi": "https://huggingface.co/Salesforce/codegen-350M-multi/resolve/main/config.json", "Salesforce/codegen-350M-mono": "https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/config.json", "Salesforce/codegen-2B-nl": "https://huggingface.co/Salesforce/codegen-2B-nl/resolve/main/config.json", "Salesforce/codegen-2B-multi": "https://huggingface.co/Salesforce/codegen-2B-multi/resolve/main/config.json", "Salesforce/codegen-2B-mono": "https://huggingface.co/Salesforce/codegen-2B-mono/resolve/main/config.json", "Salesforce/codegen-6B-nl": "https://huggingface.co/Salesforce/codegen-6B-nl/resolve/main/config.json", "Salesforce/codegen-6B-multi": "https://huggingface.co/Salesforce/codegen-6B-multi/resolve/main/config.json", "Salesforce/codegen-6B-mono": "https://huggingface.co/Salesforce/codegen-6B-mono/resolve/main/config.json", "Salesforce/codegen-16B-nl": "https://huggingface.co/Salesforce/codegen-16B-nl/resolve/main/config.json", "Salesforce/codegen-16B-multi": "https://huggingface.co/Salesforce/codegen-16B-multi/resolve/main/config.json", "Salesforce/codegen-16B-mono": "https://huggingface.co/Salesforce/codegen-16B-mono/resolve/main/config.json", } class _A ( __lowercase ): lowercase__: str = '''codegen''' lowercase__: Optional[int] = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : Union[str, Any] , __magic_name__ : Optional[Any]=5_04_00 , __magic_name__ : Any=20_48 , __magic_name__ : List[str]=20_48 , __magic_name__ : Union[str, Any]=40_96 , __magic_name__ : Tuple=28 , __magic_name__ : Dict=16 , __magic_name__ : List[str]=64 , __magic_name__ : str=None , __magic_name__ : Tuple="gelu_new" , __magic_name__ : Tuple=0.0 , __magic_name__ : Tuple=0.0 , __magic_name__ : Dict=0.0 , __magic_name__ : Optional[Any]=1E-5 , __magic_name__ : int=0.02 , __magic_name__ : List[Any]=True , __magic_name__ : int=5_02_56 , __magic_name__ : int=5_02_56 , __magic_name__ : Any=False , **__magic_name__ : Optional[int] , ) -> int: """simple docstring""" __snake_case : List[str] = vocab_size __snake_case : Union[str, Any] = n_ctx __snake_case : int = n_positions __snake_case : str = n_embd __snake_case : Dict = n_layer __snake_case : List[Any] = n_head __snake_case : Any = n_inner __snake_case : str = rotary_dim __snake_case : List[str] = activation_function __snake_case : Tuple = resid_pdrop __snake_case : Dict = embd_pdrop __snake_case : int = attn_pdrop __snake_case : Tuple = layer_norm_epsilon __snake_case : Union[str, Any] = initializer_range __snake_case : Optional[Any] = use_cache __snake_case : Dict = bos_token_id __snake_case : Union[str, Any] = eos_token_id super().__init__( bos_token_id=__magic_name__ , eos_token_id=__magic_name__ , tie_word_embeddings=__magic_name__ , **__magic_name__ ) class _A ( __lowercase ): def __init__( self : int , __magic_name__ : PretrainedConfig , __magic_name__ : str = "default" , __magic_name__ : List[PatchingSpec] = None , __magic_name__ : bool = False , ) -> Tuple: """simple docstring""" super().__init__(__magic_name__ , task=__magic_name__ , patching_specs=__magic_name__ , use_past=__magic_name__ ) if not getattr(self._config , """pad_token_id""" , __magic_name__ ): # TODO: how to do that better? __snake_case : List[str] = 0 @property def lowercase__ ( self : Tuple ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" __snake_case : Dict = OrderedDict({"""input_ids""": {0: """batch""", 1: """sequence"""}} ) if self.use_past: self.fill_with_past_key_values_(__magic_name__ , direction="""inputs""" ) __snake_case : Optional[Any] = {0: """batch""", 1: """past_sequence + sequence"""} else: __snake_case : Union[str, Any] = {0: """batch""", 1: """sequence"""} return common_inputs @property def lowercase__ ( self : Tuple ) -> int: """simple docstring""" return self._config.n_layer @property def lowercase__ ( self : Union[str, Any] ) -> int: """simple docstring""" return self._config.n_head def lowercase__ ( self : Dict , __magic_name__ : PreTrainedTokenizer , __magic_name__ : int = -1 , __magic_name__ : int = -1 , __magic_name__ : bool = False , __magic_name__ : Optional[TensorType] = None , ) -> Mapping[str, Any]: """simple docstring""" __snake_case : Tuple = super(__magic_name__ , self ).generate_dummy_inputs( __magic_name__ , batch_size=__magic_name__ , seq_length=__magic_name__ , is_pair=__magic_name__ , framework=__magic_name__ ) # We need to order the input in the way they appears in the forward() __snake_case : Union[str, Any] = OrderedDict({"""input_ids""": common_inputs["""input_ids"""]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch __snake_case , __snake_case : str = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values __snake_case : Tuple = seqlen + 2 __snake_case : Union[str, Any] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) __snake_case : List[str] = [ (torch.zeros(__magic_name__ ), torch.zeros(__magic_name__ )) for _ in range(self.num_layers ) ] __snake_case : Optional[int] = common_inputs["""attention_mask"""] if self.use_past: __snake_case : Union[str, Any] = ordered_inputs["""attention_mask"""].dtype __snake_case : Optional[Any] = torch.cat( [ordered_inputs["""attention_mask"""], torch.ones(__magic_name__ , __magic_name__ , dtype=__magic_name__ )] , dim=1 ) return ordered_inputs @property def lowercase__ ( self : Union[str, Any] ) -> int: """simple docstring""" return 13
13
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer import diffusers from diffusers import ( AutoencoderKL, EulerDiscreteScheduler, StableDiffusionLatentUpscalePipeline, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.schedulers import KarrasDiffusionSchedulers from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() def lowerCamelCase (_SCREAMING_SNAKE_CASE : Optional[int] ): __a : Optional[Any] = [tensor.shape for tensor in tensor_list] return all(shape == shapes[0] for shape in shapes[1:] ) class __UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): A_ = StableDiffusionLatentUpscalePipeline A_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { "height", "width", "cross_attention_kwargs", "negative_prompt_embeds", "prompt_embeds", } A_ = PipelineTesterMixin.required_optional_params - {"num_images_per_prompt"} A_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS A_ = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess A_ = frozenset([] ) A_ = True @property def __UpperCAmelCase ( self ): '''simple docstring''' __a : int = 1 __a : Any = 4 __a : List[str] = (16, 16) __a : List[Any] = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(__a ) return image def __UpperCAmelCase ( self ): '''simple docstring''' torch.manual_seed(0 ) __a : List[Any] = UNetaDConditionModel( act_fn='gelu' , attention_head_dim=8 , norm_num_groups=__a , block_out_channels=[32, 32, 64, 64] , time_cond_proj_dim=160 , conv_in_kernel=1 , conv_out_kernel=1 , cross_attention_dim=32 , down_block_types=( 'KDownBlock2D', 'KCrossAttnDownBlock2D', 'KCrossAttnDownBlock2D', 'KCrossAttnDownBlock2D', ) , in_channels=8 , mid_block_type=__a , only_cross_attention=__a , out_channels=5 , resnet_time_scale_shift='scale_shift' , time_embedding_type='fourier' , timestep_post_act='gelu' , up_block_types=('KCrossAttnUpBlock2D', 'KCrossAttnUpBlock2D', 'KCrossAttnUpBlock2D', 'KUpBlock2D') , ) __a : Dict = AutoencoderKL( block_out_channels=[32, 32, 64, 64] , in_channels=3 , out_channels=3 , down_block_types=[ 'DownEncoderBlock2D', 'DownEncoderBlock2D', 'DownEncoderBlock2D', 'DownEncoderBlock2D', ] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D', 'UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) __a : str = EulerDiscreteScheduler(prediction_type='sample' ) __a : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='quick_gelu' , projection_dim=512 , ) __a : Optional[Any] = CLIPTextModel(__a ) __a : Any = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __a : Any = { 'unet': model.eval(), 'vae': vae.eval(), 'scheduler': scheduler, 'text_encoder': text_encoder, 'tokenizer': tokenizer, } return components def __UpperCAmelCase ( self , __a , __a=0 ): '''simple docstring''' if str(__a ).startswith('mps' ): __a : str = torch.manual_seed(__a ) else: __a : Tuple = torch.Generator(device=__a ).manual_seed(__a ) __a : Optional[int] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': self.dummy_image.cpu(), 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[Any] = 'cpu' __a : List[Any] = self.get_dummy_components() __a : Optional[int] = self.pipeline_class(**__a ) pipe.to(__a ) pipe.set_progress_bar_config(disable=__a ) __a : Dict = self.get_dummy_inputs(__a ) __a : Tuple = pipe(**__a ).images __a : List[Any] = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 256, 256, 3) ) __a : List[str] = np.array( [0.47222412, 0.41921633, 0.44717434, 0.46874192, 0.42588258, 0.46150726, 0.4677534, 0.45583832, 0.48579055] ) __a : Tuple = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(__a , 1E-3 ) def __UpperCAmelCase ( self ): '''simple docstring''' super().test_attention_slicing_forward_pass(expected_max_diff=7E-3 ) def __UpperCAmelCase ( self ): '''simple docstring''' super().test_cpu_offload_forward_pass(expected_max_diff=3E-3 ) def __UpperCAmelCase ( self ): '''simple docstring''' super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def __UpperCAmelCase ( self ): '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=7E-3 ) def __UpperCAmelCase ( self ): '''simple docstring''' super().test_pt_np_pil_outputs_equivalent(expected_max_diff=3E-3 ) def __UpperCAmelCase ( self ): '''simple docstring''' super().test_save_load_local(expected_max_difference=3E-3 ) def __UpperCAmelCase ( self ): '''simple docstring''' super().test_save_load_optional_components(expected_max_difference=3E-3 ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : List[Any] = [ 'DDIMScheduler', 'DDPMScheduler', 'PNDMScheduler', 'HeunDiscreteScheduler', 'EulerAncestralDiscreteScheduler', 'KDPM2DiscreteScheduler', 'KDPM2AncestralDiscreteScheduler', 'DPMSolverSDEScheduler', ] __a : Tuple = self.get_dummy_components() __a : Tuple = self.pipeline_class(**__a ) # make sure that PNDM does not need warm-up pipe.scheduler.register_to_config(skip_prk_steps=__a ) pipe.to(__a ) pipe.set_progress_bar_config(disable=__a ) __a : List[str] = self.get_dummy_inputs(__a ) __a : Any = 2 __a : Tuple = [] for scheduler_enum in KarrasDiffusionSchedulers: if scheduler_enum.name in skip_schedulers: # no sigma schedulers are not supported # no schedulers continue __a : Tuple = getattr(__a , scheduler_enum.name ) __a : Optional[Any] = scheduler_cls.from_config(pipe.scheduler.config ) __a : int = pipe(**__a )[0] outputs.append(__a ) assert check_same_shape(__a ) @require_torch_gpu @slow class __UpperCamelCase ( unittest.TestCase ): def __UpperCAmelCase ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCAmelCase ( self ): '''simple docstring''' __a : Dict = torch.manual_seed(33 ) __a : str = StableDiffusionPipeline.from_pretrained('CompVis/stable-diffusion-v1-4' , torch_dtype=torch.floataa ) pipe.to('cuda' ) __a : str = StableDiffusionLatentUpscalePipeline.from_pretrained( 'stabilityai/sd-x2-latent-upscaler' , torch_dtype=torch.floataa ) upscaler.to('cuda' ) __a : Union[str, Any] = 'a photo of an astronaut high resolution, unreal engine, ultra realistic' __a : int = pipe(__a , generator=__a , output_type='latent' ).images __a : Union[str, Any] = upscaler( prompt=__a , image=__a , num_inference_steps=20 , guidance_scale=0 , generator=__a , output_type='np' , ).images[0] __a : Optional[Any] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/astronaut_1024.npy' ) assert np.abs((expected_image - image).mean() ) < 5E-2 def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[Any] = torch.manual_seed(33 ) __a : Union[str, Any] = StableDiffusionLatentUpscalePipeline.from_pretrained( 'stabilityai/sd-x2-latent-upscaler' , torch_dtype=torch.floataa ) upscaler.to('cuda' ) __a : Optional[int] = 'the temple of fire by Ross Tran and Gerardo Dottori, oil on canvas' __a : List[str] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_512.png' ) __a : List[str] = upscaler( prompt=__a , image=__a , num_inference_steps=20 , guidance_scale=0 , generator=__a , output_type='np' , ).images[0] __a : Tuple = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_1024.npy' ) assert np.abs((expected_image - image).max() ) < 5E-2
27
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XGLMTokenizer, XGLMTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin _a : Optional[int] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class __A ( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): _UpperCamelCase : List[Any] = XGLMTokenizer _UpperCamelCase : List[Any] = XGLMTokenizerFast _UpperCamelCase : Dict = True _UpperCamelCase : Tuple = True def __A ( self ): super().setUp() # We have a SentencePiece fixture for testing _lowerCAmelCase : List[Any] = XGLMTokenizer(a__ , keep_accents=a__ ) tokenizer.save_pretrained(self.tmpdirname ) def __A ( self ): _lowerCAmelCase : List[str] = """<pad>""" _lowerCAmelCase : List[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a__ ) , a__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a__ ) , a__ ) def __A ( self ): _lowerCAmelCase : Optional[int] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(len(a__ ) , 1008 ) def __A ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1008 ) def __A ( self ): _lowerCAmelCase : List[Any] = XGLMTokenizer(a__ , keep_accents=a__ ) _lowerCAmelCase : Dict = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(a__ , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(a__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) _lowerCAmelCase : Any = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( a__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) _lowerCAmelCase : List[str] = tokenizer.convert_tokens_to_ids(a__ ) self.assertListEqual( a__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) _lowerCAmelCase : Optional[int] = tokenizer.convert_ids_to_tokens(a__ ) self.assertListEqual( a__ , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) @cached_property def __A ( self ): return XGLMTokenizer.from_pretrained("""facebook/xglm-564M""" ) def __A ( self ): with tempfile.NamedTemporaryFile() as f: shutil.copyfile(a__ , f.name ) _lowerCAmelCase : Union[str, Any] = XGLMTokenizer(f.name , keep_accents=a__ ) _lowerCAmelCase : List[str] = pickle.dumps(a__ ) pickle.loads(a__ ) def __A ( self ): if not self.test_rust_tokenizer: return _lowerCAmelCase : List[str] = self.get_tokenizer() _lowerCAmelCase : Optional[Any] = self.get_rust_tokenizer() _lowerCAmelCase : Tuple = """I was born in 92000, and this is falsé.""" _lowerCAmelCase : List[Any] = tokenizer.tokenize(a__ ) _lowerCAmelCase : Tuple = rust_tokenizer.tokenize(a__ ) self.assertListEqual(a__ , a__ ) _lowerCAmelCase : Union[str, Any] = tokenizer.encode(a__ , add_special_tokens=a__ ) _lowerCAmelCase : str = rust_tokenizer.encode(a__ , add_special_tokens=a__ ) self.assertListEqual(a__ , a__ ) _lowerCAmelCase : int = self.get_rust_tokenizer() _lowerCAmelCase : Dict = tokenizer.encode(a__ ) _lowerCAmelCase : List[Any] = rust_tokenizer.encode(a__ ) self.assertListEqual(a__ , a__ ) @slow def __A ( self ): _lowerCAmelCase : int = """Hello World!""" _lowerCAmelCase : Optional[int] = [2, 31227, 4447, 35] self.assertListEqual(a__ , self.big_tokenizer.encode(a__ ) ) @slow def __A ( self ): _lowerCAmelCase : Any = ( """This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will""" """ add words that should not exsist and be tokenized to unk, such as saoneuhaoesuth""" ) # fmt: off _lowerCAmelCase : List[str] = [2, 1018, 67, 11, 1988, 2617, 5631, 278, 11, 3407, 48, 71630, 28085, 4, 3234, 157, 13, 6, 5, 6, 4, 3526, 768, 15, 659, 57, 298, 3983, 864, 129, 21, 6, 5, 13675, 377, 652, 7580, 10341, 155, 2817, 422, 1666, 7, 1674, 53, 113, 202277, 17892, 33, 60, 87, 4, 3234, 157, 61, 2667, 52376, 19, 88, 23, 735] # fmt: on self.assertListEqual(a__ , self.big_tokenizer.encode(a__ ) ) @slow def __A ( self ): # fmt: off _lowerCAmelCase : List[str] = { """input_ids""": [[2, 108825, 1163, 15, 88010, 473, 15898, 157, 13672, 1857, 312, 8, 238021, 1163, 53, 13672, 1857, 312, 8, 53283, 182396, 8, 18566, 16, 36733, 4101, 8, 230, 244017, 122553, 7, 15, 132597, 4, 293, 12511, 7610, 4, 3414, 132597, 9, 4, 32361, 362, 4, 734, 28512, 32569, 18, 4, 32361, 26096, 14982, 73, 18715, 21433, 235261, 15, 492, 12427, 16, 53, 18715, 21433, 65454, 15, 23659, 563, 16, 278, 597, 2843, 595, 7931, 182396, 64186, 22, 886, 595, 132981, 53, 25540, 3449, 43982, 39901, 5951, 878, 330, 4, 27694, 80269, 312, 53, 6517, 11780, 611, 20408, 5], [2, 6, 132597, 67, 42897, 33, 592, 8, 163729, 25540, 361, 136997, 109514, 173230, 7, 501, 60, 102913, 196, 5631, 235, 63243, 473, 6, 231757, 74, 5277, 7905, 53, 3095, 37317, 22, 454, 183874, 5], [2, 268, 31298, 46530, 6, 132935, 43831, 7, 597, 32, 24, 3688, 9865, 5]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] } # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=a__ , model_name="""facebook/xglm-564M""" , padding=a__ , )
44
0
from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
367
def __lowerCAmelCase (SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE )-> list[int]: """simple docstring""" snake_case_ = int(SCREAMING_SNAKE_CASE ) # Initialize Result snake_case_ = [] # Traverse through all denomination for denomination in reversed(SCREAMING_SNAKE_CASE ): # Find denominations while int(SCREAMING_SNAKE_CASE ) >= int(SCREAMING_SNAKE_CASE ): total_value -= int(SCREAMING_SNAKE_CASE ) answer.append(SCREAMING_SNAKE_CASE ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": UpperCAmelCase = [] UpperCAmelCase = """0""" if ( input("""Do you want to enter your denominations ? (yY/n): """).strip().lower() == "y" ): UpperCAmelCase = int(input("""Enter the number of denominations you want to add: """).strip()) for i in range(0, n): denominations.append(int(input(f'''Denomination {i}: ''').strip())) UpperCAmelCase = input("""Enter the change you want to make in Indian Currency: """).strip() else: # All denominations of Indian Currency if user does not enter UpperCAmelCase = [1, 2, 5, 10, 20, 50, 100, 500, 2000] UpperCAmelCase = input("""Enter the change you want to make: """).strip() if int(value) == 0 or int(value) < 0: print("""The total value cannot be zero or negative.""") else: print(f'''Following is minimal change for {value}: ''') UpperCAmelCase = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=""" """)
267
0
'''simple docstring''' # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def UpperCAmelCase__ ( UpperCAmelCase_ : Dict ) -> Any: return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def UpperCAmelCase__ ( UpperCAmelCase_ : str ) -> Optional[Any]: __lowerCamelCase : Optional[int] = create_tensor(UpperCAmelCase_ ) __lowerCamelCase : int = gather(UpperCAmelCase_ ) assert gathered_tensor.tolist() == list(range(1 , state.num_processes**2 + 1 ) ) def UpperCAmelCase__ ( UpperCAmelCase_ : Optional[int] ) -> Any: __lowerCamelCase : str = [state.process_index] __lowerCamelCase : List[str] = gather_object(UpperCAmelCase_ ) assert len(UpperCAmelCase_ ) == state.num_processes, F'{gathered_obj}, {len(UpperCAmelCase_ )} != {state.num_processes}' assert gathered_obj == list(range(state.num_processes ) ), F'{gathered_obj} != {list(range(state.num_processes ) )}' def UpperCAmelCase__ ( UpperCAmelCase_ : Tuple ) -> List[Any]: __lowerCamelCase : Union[str, Any] = create_tensor(UpperCAmelCase_ ) __lowerCamelCase : Union[str, Any] = broadcast(UpperCAmelCase_ ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1 , state.num_processes + 1 ) ) def UpperCAmelCase__ ( UpperCAmelCase_ : int ) -> Dict: # We need to pad the tensor with one more element if we are the main process # to ensure that we can pad if state.is_main_process: __lowerCamelCase : Dict = torch.arange(state.num_processes + 1 ).to(state.device ) else: __lowerCamelCase : Union[str, Any] = torch.arange(state.num_processes ).to(state.device ) __lowerCamelCase : int = pad_across_processes(UpperCAmelCase_ ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0 , state.num_processes ) ) + [0] def UpperCAmelCase__ ( UpperCAmelCase_ : str ) -> int: # For now runs on only two processes if state.num_processes != 2: return __lowerCamelCase : Dict = create_tensor(UpperCAmelCase_ ) __lowerCamelCase : Tuple = reduce(UpperCAmelCase_ , 'sum' ) __lowerCamelCase : Union[str, Any] = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(UpperCAmelCase_ , UpperCAmelCase_ ), F'{reduced_tensor} != {truth_tensor}' def UpperCAmelCase__ ( UpperCAmelCase_ : List[str] ) -> Optional[int]: # For now runs on only two processes if state.num_processes != 2: return __lowerCamelCase : Dict = create_tensor(UpperCAmelCase_ ) __lowerCamelCase : Optional[Any] = reduce(UpperCAmelCase_ , 'mean' ) __lowerCamelCase : List[Any] = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(UpperCAmelCase_ , UpperCAmelCase_ ), F'{reduced_tensor} != {truth_tensor}' def UpperCAmelCase__ ( UpperCAmelCase_ : int ) -> Tuple: # For xla_spawn (TPUs) main() def UpperCAmelCase__ ( ) -> List[Any]: __lowerCamelCase : Dict = PartialState() state.print(F'State: {state}' ) state.print('testing gather' ) test_gather(UpperCAmelCase_ ) state.print('testing gather_object' ) test_gather_object(UpperCAmelCase_ ) state.print('testing broadcast' ) test_broadcast(UpperCAmelCase_ ) state.print('testing pad_across_processes' ) test_pad_across_processes(UpperCAmelCase_ ) state.print('testing reduce_sum' ) test_reduce_sum(UpperCAmelCase_ ) state.print('testing reduce_mean' ) test_reduce_mean(UpperCAmelCase_ ) if __name__ == "__main__": main()
185
'''simple docstring''' A__ : Optional[int] = """ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/""" def UpperCAmelCase__ ( UpperCAmelCase_ : bytes ) -> bytes: # Make sure the supplied data is a bytes-like object if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): __lowerCamelCase : Dict = F'a bytes-like object is required, not \'{data.__class__.__name__}\'' raise TypeError(UpperCAmelCase_ ) __lowerCamelCase : Optional[int] = ''.join(bin(UpperCAmelCase_ )[2:].zfill(8 ) for byte in data ) __lowerCamelCase : Any = len(UpperCAmelCase_ ) % 6 != 0 if padding_needed: # The padding that will be added later __lowerCamelCase : int = B'=' * ((6 - len(UpperCAmelCase_ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(UpperCAmelCase_ ) % 6) else: __lowerCamelCase : str = B'' # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(UpperCAmelCase_ ) , 6 ) ).encode() + padding ) def UpperCAmelCase__ ( UpperCAmelCase_ : str ) -> bytes: # Make sure encoded_data is either a string or a bytes-like object if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): __lowerCamelCase : Dict = ( 'argument should be a bytes-like object or ASCII string, ' F'not \'{encoded_data.__class__.__name__}\'' ) raise TypeError(UpperCAmelCase_ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): try: __lowerCamelCase : int = encoded_data.decode('utf-8' ) except UnicodeDecodeError: raise ValueError('base64 encoded data should only contain ASCII characters' ) __lowerCamelCase : Union[str, Any] = encoded_data.count('=' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(UpperCAmelCase_ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one __lowerCamelCase : Any = encoded_data[:-padding] __lowerCamelCase : Optional[Any] = ''.join( bin(B64_CHARSET.index(UpperCAmelCase_ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: __lowerCamelCase : Any = ''.join( bin(B64_CHARSET.index(UpperCAmelCase_ ) )[2:].zfill(6 ) for char in encoded_data ) __lowerCamelCase : str = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(UpperCAmelCase_ ) , 8 ) ] return bytes(UpperCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
185
1
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class lowercase ( UpperCamelCase__,unittest.TestCase ): _a = BertJapaneseTokenizer _a = False _a = True def a__ ( self ) -> Any: super().setUp() _A : Dict = [ """[UNK]""", """[CLS]""", """[SEP]""", """こんにちは""", """こん""", """にちは""", """ばんは""", """##こん""", """##にちは""", """##ばんは""", """世界""", """##世界""", """、""", """##、""", """。""", """##。""", ] _A : int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def a__ ( self , _a ) -> Dict: _A : Optional[Any] = """こんにちは、世界。 \nこんばんは、世界。""" _A : Optional[Any] = """こんにちは 、 世界 。 こんばんは 、 世界 。""" return input_text, output_text def a__ ( self , _a ) -> Union[str, Any]: _A , _A : List[str] = self.get_input_output_texts(_a ) _A : Optional[Any] = tokenizer.encode(_a , add_special_tokens=_a ) _A : int = tokenizer.decode(_a , clean_up_tokenization_spaces=_a ) return text, ids def a__ ( self ) -> int: pass # TODO add if relevant def a__ ( self ) -> List[str]: pass # TODO add if relevant def a__ ( self ) -> Optional[Any]: pass # TODO add if relevant def a__ ( self ) -> Optional[int]: _A : int = self.tokenizer_class(self.vocab_file ) _A : List[str] = tokenizer.tokenize("""こんにちは、世界。\nこんばんは、世界。""" ) self.assertListEqual(_a , ["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) def a__ ( self ) -> Optional[Any]: _A : List[Any] = self.tokenizer_class(self.vocab_file , word_tokenizer_type="""mecab""" ) self.assertIsNotNone(_a ) _A : Tuple = """こんにちは、世界。\nこんばんは、世界。""" _A : Union[str, Any] = tokenizer.tokenize(_a ) self.assertListEqual(_a , ["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) _A : str = os.path.join(self.tmpdirname , """tokenizer.bin""" ) with open(_a , """wb""" ) as handle: pickle.dump(_a , _a ) with open(_a , """rb""" ) as handle: _A : List[str] = pickle.load(_a ) _A : str = tokenizer_new.tokenize(_a ) self.assertListEqual(_a , _a ) def a__ ( self ) -> List[Any]: _A : str = MecabTokenizer(mecab_dic="""ipadic""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップルストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) def a__ ( self ) -> Tuple: try: _A : str = MecabTokenizer(mecab_dic="""unidic_lite""" ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) def a__ ( self ) -> List[Any]: try: _A : Any = MecabTokenizer(mecab_dic="""unidic""" ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) def a__ ( self ) -> Optional[int]: _A : List[Any] = MecabTokenizer(do_lower_case=_a , mecab_dic="""ipadic""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップルストア""", """で""", """iphone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) def a__ ( self ) -> Optional[Any]: try: _A : List[Any] = MecabTokenizer( do_lower_case=_a , normalize_text=_a , mecab_option="""-d /usr/local/lib/mecab/dic/jumandic""" ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップルストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れた""", """\u3000""", """。"""] , ) def a__ ( self ) -> Optional[Any]: _A : Dict = MecabTokenizer(normalize_text=_a , mecab_dic="""ipadic""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップルストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """ """, """。"""] , ) @require_sudachi def a__ ( self ) -> List[str]: _A : Optional[Any] = self.tokenizer_class(self.vocab_file , word_tokenizer_type="""sudachi""" ) self.assertIsNotNone(_a ) _A : List[str] = """こんにちは、世界。\nこんばんは、世界。""" _A : Dict = tokenizer.tokenize(_a ) self.assertListEqual(_a , ["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) _A : List[str] = os.path.join(self.tmpdirname , """tokenizer.bin""" ) with open(_a , """wb""" ) as handle: pickle.dump(_a , _a ) with open(_a , """rb""" ) as handle: _A : Any = pickle.load(_a ) _A : int = tokenizer_new.tokenize(_a ) self.assertListEqual(_a , _a ) @require_sudachi def a__ ( self ) -> Optional[int]: _A : Optional[int] = SudachiTokenizer(sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , [""" """, """\t""", """アップル""", """ストア""", """で""", """iPhone""", """8""", """ """, """が""", """ """, """ """, """\n """, """発売""", """さ""", """れ""", """た""", """ """, """。""", """ """, """ """] , ) @require_sudachi def a__ ( self ) -> List[Any]: _A : int = SudachiTokenizer(sudachi_dict_type="""core""" , sudachi_split_mode="""A""" ) self.assertListEqual(tokenizer.tokenize("""外国人参政権""" ) , ["""外国""", """人""", """参政""", """権"""] ) @require_sudachi def a__ ( self ) -> Optional[Any]: _A : Optional[int] = SudachiTokenizer(sudachi_dict_type="""core""" , sudachi_split_mode="""B""" ) self.assertListEqual(tokenizer.tokenize("""外国人参政権""" ) , ["""外国人""", """参政権"""] ) @require_sudachi def a__ ( self ) -> Tuple: _A : str = SudachiTokenizer(sudachi_dict_type="""core""" , sudachi_split_mode="""C""" ) self.assertListEqual(tokenizer.tokenize("""外国人参政権""" ) , ["""外国人参政権"""] ) @require_sudachi def a__ ( self ) -> Optional[Any]: _A : str = SudachiTokenizer(do_lower_case=_a , sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , [""" """, """\t""", """アップル""", """ストア""", """で""", """iphone""", """8""", """ """, """が""", """ """, """ """, """\n """, """発売""", """さ""", """れ""", """た""", """ """, """。""", """ """, """ """] , ) @require_sudachi def a__ ( self ) -> Optional[int]: _A : Tuple = SudachiTokenizer(normalize_text=_a , sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , [""" """, """\t""", """アップル""", """ストア""", """で""", """iPhone""", """8""", """ """, """が""", """ """, """ """, """\n """, """発売""", """さ""", """れ""", """た""", """\u3000""", """。""", """ """, """ """] , ) @require_sudachi def a__ ( self ) -> List[str]: _A : List[Any] = SudachiTokenizer(trim_whitespace=_a , sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) @require_jumanpp def a__ ( self ) -> Tuple: _A : Dict = self.tokenizer_class(self.vocab_file , word_tokenizer_type="""jumanpp""" ) self.assertIsNotNone(_a ) _A : List[Any] = """こんにちは、世界。\nこんばんは、世界。""" _A : str = tokenizer.tokenize(_a ) self.assertListEqual(_a , ["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) _A : Optional[Any] = os.path.join(self.tmpdirname , """tokenizer.bin""" ) with open(_a , """wb""" ) as handle: pickle.dump(_a , _a ) with open(_a , """rb""" ) as handle: _A : Any = pickle.load(_a ) _A : Dict = tokenizer_new.tokenize(_a ) self.assertListEqual(_a , _a ) @require_jumanpp def a__ ( self ) -> Optional[Any]: _A : List[str] = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """\u3000""", """が""", """\u3000""", """\u3000""", """\u3000""", """発売""", """さ""", """れた""", """\u3000""", """。"""] , ) @require_jumanpp def a__ ( self ) -> Optional[int]: _A : Union[str, Any] = JumanppTokenizer(do_lower_case=_a ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iphone""", """8""", """\u3000""", """が""", """\u3000""", """\u3000""", """\u3000""", """発売""", """さ""", """れた""", """\u3000""", """。"""] , ) @require_jumanpp def a__ ( self ) -> List[str]: _A : str = JumanppTokenizer(normalize_text=_a ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""ア""", """ッ""", """フ""", """゚""", """ル""", """ストア""", """で""", """iPhone""", """8""", """\u3000""", """が""", """\u3000""", """\u3000""", """\u3000""", """発売""", """さ""", """れた""", """\u3000""", """。"""] , ) @require_jumanpp def a__ ( self ) -> Optional[Any]: _A : Union[str, Any] = JumanppTokenizer(trim_whitespace=_a ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れた""", """。"""] , ) @require_jumanpp def a__ ( self ) -> str: _A : Tuple = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize("""ありがとうございますm(_ _)m見つけるのが大変です。""" ) , ["""ありがとう""", """ございます""", """m(_ _)m""", """見つける""", """の""", """が""", """大変です""", """。"""] , ) def a__ ( self ) -> Optional[Any]: _A : Dict = ["""[UNK]""", """[CLS]""", """[SEP]""", """こんにちは""", """こん""", """にちは""", """ばんは""", """##こん""", """##にちは""", """##ばんは"""] _A : int = {} for i, token in enumerate(_a ): _A : List[Any] = i _A : Any = WordpieceTokenizer(vocab=_a , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""こんにちは""" ) , ["""こんにちは"""] ) self.assertListEqual(tokenizer.tokenize("""こんばんは""" ) , ["""こん""", """##ばんは"""] ) self.assertListEqual(tokenizer.tokenize("""こんばんは こんばんにちは こんにちは""" ) , ["""こん""", """##ばんは""", """[UNK]""", """こんにちは"""] ) def a__ ( self ) -> List[Any]: _A : Tuple = BertJapaneseTokenizer.from_pretrained("""nlp-waseda/roberta-base-japanese-with-auto-jumanpp""" ) _A : str = tokenizer.subword_tokenizer _A : Any = subword_tokenizer.tokenize("""国境 の 長い トンネル を 抜ける と 雪国 であった 。""" ) self.assertListEqual(_a , ["""▁国境""", """▁の""", """▁長い""", """▁トンネル""", """▁を""", """▁抜ける""", """▁と""", """▁雪""", """国""", """▁であった""", """▁。"""] ) _A : Dict = subword_tokenizer.tokenize("""こんばんは こんばん にち は こんにちは""" ) self.assertListEqual(_a , ["""▁こん""", """ばん""", """は""", """▁こん""", """ばん""", """▁に""", """ち""", """▁は""", """▁こんにちは"""] ) def a__ ( self ) -> Dict: _A : str = self.tokenizer_class.from_pretrained("""cl-tohoku/bert-base-japanese""" ) _A : Optional[Any] = tokenizer.encode("""ありがとう。""" , add_special_tokens=_a ) _A : str = tokenizer.encode("""どういたしまして。""" , add_special_tokens=_a ) _A : List[Any] = tokenizer.build_inputs_with_special_tokens(_a ) _A : Tuple = tokenizer.build_inputs_with_special_tokens(_a , _a ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class lowercase ( UpperCamelCase__,unittest.TestCase ): _a = BertJapaneseTokenizer _a = False def a__ ( self ) -> Any: super().setUp() _A : List[Any] = ["""[UNK]""", """[CLS]""", """[SEP]""", """こ""", """ん""", """に""", """ち""", """は""", """ば""", """世""", """界""", """、""", """。"""] _A : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def a__ ( self , **_a ) -> Optional[int]: return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type="""character""" , **_a ) def a__ ( self , _a ) -> int: _A : str = """こんにちは、世界。 \nこんばんは、世界。""" _A : int = """こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。""" return input_text, output_text def a__ ( self ) -> Tuple: pass # TODO add if relevant def a__ ( self ) -> Tuple: pass # TODO add if relevant def a__ ( self ) -> str: pass # TODO add if relevant def a__ ( self ) -> Any: _A : List[str] = self.tokenizer_class(self.vocab_file , subword_tokenizer_type="""character""" ) _A : List[str] = tokenizer.tokenize("""こんにちは、世界。 \nこんばんは、世界。""" ) self.assertListEqual( _a , ["""こ""", """ん""", """に""", """ち""", """は""", """、""", """世""", """界""", """。""", """こ""", """ん""", """ば""", """ん""", """は""", """、""", """世""", """界""", """。"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_a ) , [3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def a__ ( self ) -> Tuple: _A : List[Any] = ["""[UNK]""", """[CLS]""", """[SEP]""", """こ""", """ん""", """に""", """ち""", """は""", """ば""", """世""", """界""", """、""", """。"""] _A : Dict = {} for i, token in enumerate(_a ): _A : Tuple = i _A : int = CharacterTokenizer(vocab=_a , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""こんにちは""" ) , ["""こ""", """ん""", """に""", """ち""", """は"""] ) self.assertListEqual(tokenizer.tokenize("""こんにちほ""" ) , ["""こ""", """ん""", """に""", """ち""", """[UNK]"""] ) def a__ ( self ) -> Dict: _A : int = self.tokenizer_class.from_pretrained("""cl-tohoku/bert-base-japanese-char""" ) _A : int = tokenizer.encode("""ありがとう。""" , add_special_tokens=_a ) _A : Optional[Any] = tokenizer.encode("""どういたしまして。""" , add_special_tokens=_a ) _A : Optional[Any] = tokenizer.build_inputs_with_special_tokens(_a ) _A : int = tokenizer.build_inputs_with_special_tokens(_a , _a ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class lowercase ( unittest.TestCase ): def a__ ( self ) -> Union[str, Any]: _A : List[Any] = """cl-tohoku/bert-base-japanese""" _A : List[str] = AutoTokenizer.from_pretrained(_a ) self.assertIsInstance(_a , _a ) class lowercase ( unittest.TestCase ): def a__ ( self ) -> Optional[int]: _A : Dict = """cl-tohoku/bert-base-japanese""" with self.assertLogs("""transformers""" , level="""WARNING""" ) as cm: BertTokenizer.from_pretrained(_a ) self.assertTrue( cm.records[0].message.startswith( """The tokenizer class you load from this checkpoint is not the same type as the class this function""" """ is called from.""" ) ) _A : str = """bert-base-cased""" with self.assertLogs("""transformers""" , level="""WARNING""" ) as cm: BertJapaneseTokenizer.from_pretrained(_a ) self.assertTrue( cm.records[0].message.startswith( """The tokenizer class you load from this checkpoint is not the same type as the class this function""" """ is called from.""" ) )
343
import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) _snake_case = logging.getLogger() def lowerCAmelCase_ ( ): _A : Optional[Any] = argparse.ArgumentParser() parser.add_argument("""-f""" ) _A : Optional[Any] = parser.parse_args() return args.f class lowercase ( UpperCamelCase__ ): def a__ ( self ) -> None: _A : List[Any] = logging.StreamHandler(sys.stdout ) logger.addHandler(_a ) def a__ ( self , _a ) -> Dict: _A : Tuple = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , """run_glue_deebert.py""" ) with patch.object(_a , """argv""" , _a ): _A : Optional[Any] = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(_a , 0.666 ) @slow @require_torch_non_multi_gpu def a__ ( self ) -> Optional[int]: _A : Tuple = """ --model_type roberta --model_name_or_path roberta-base --task_name MRPC --do_train --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --max_seq_length 128 --per_gpu_eval_batch_size=1 --per_gpu_train_batch_size=8 --learning_rate 2e-4 --num_train_epochs 3 --overwrite_output_dir --seed 42 --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --save_steps 0 --overwrite_cache --eval_after_first_stage """.split() self.run_and_check(_a ) _A : Optional[Any] = """ --model_type roberta --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --task_name MRPC --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --max_seq_length 128 --eval_each_highway --eval_highway --overwrite_cache --per_gpu_eval_batch_size=1 """.split() self.run_and_check(_a ) _A : List[str] = """ --model_type roberta --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --task_name MRPC --do_eval --do_lower_case --data_dir ./tests/fixtures/tests_samples/MRPC/ --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage --plot_data_dir ./examples/deebert/results/ --max_seq_length 128 --early_exit_entropy 0.1 --eval_highway --overwrite_cache --per_gpu_eval_batch_size=1 """.split() self.run_and_check(_a )
343
1
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase = {"""configuration_focalnet""": ["""FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """FocalNetConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ """FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """FocalNetForImageClassification""", """FocalNetForMaskedImageModeling""", """FocalNetBackbone""", """FocalNetModel""", """FocalNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_focalnet import FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FocalNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_focalnet import ( FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST, FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, FocalNetPreTrainedModel, ) else: import sys UpperCAmelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
256
"""simple docstring""" import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder UpperCAmelCase = """base_with_context""" def lowercase ( a__ : Optional[Any] , a__ : Optional[int] ) -> int: _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''token_embedder''']['''embedding'''] ) ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=a__ ) for lyr_num, lyr in enumerate(model.encoders ): _UpperCamelCase = weights[F'''layers_{lyr_num}'''] _UpperCamelCase = nn.Parameter( torch.FloatTensor(ly_weight['''pre_attention_layer_norm''']['''scale'''] ) ) _UpperCamelCase = ly_weight['''attention'''] _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''encoder_norm''']['''scale'''] ) ) return model def lowercase ( a__ : List[Any] , a__ : Dict ) -> Optional[Any]: _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''input_proj''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=a__ ) for lyr_num, lyr in enumerate(model.encoders ): _UpperCamelCase = weights[F'''layers_{lyr_num}'''] _UpperCamelCase = ly_weight['''attention'''] _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(ly_weight['''pre_attention_layer_norm''']['''scale'''] ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''encoder_norm''']['''scale'''] ) ) return model def lowercase ( a__ : List[Any] , a__ : Union[str, Any] ) -> str: _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''time_emb_dense0''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''time_emb_dense1''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=a__ ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(weights['''continuous_inputs_projection''']['''kernel'''].T ) ) for lyr_num, lyr in enumerate(model.decoders ): _UpperCamelCase = weights[F'''layers_{lyr_num}'''] _UpperCamelCase = nn.Parameter( torch.FloatTensor(ly_weight['''pre_self_attention_layer_norm''']['''scale'''] ) ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(ly_weight['''FiLMLayer_0''']['''DenseGeneral_0''']['''kernel'''].T ) ) _UpperCamelCase = ly_weight['''self_attention'''] _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) _UpperCamelCase = ly_weight['''MultiHeadDotProductAttention_0'''] _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(ly_weight['''pre_cross_attention_layer_norm''']['''scale'''] ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) _UpperCamelCase = nn.Parameter( torch.FloatTensor(ly_weight['''FiLMLayer_1''']['''DenseGeneral_0''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''decoder_norm''']['''scale'''] ) ) _UpperCamelCase = nn.Parameter(torch.FloatTensor(weights['''spec_out_dense''']['''kernel'''].T ) ) return model def lowercase ( a__ : Union[str, Any] ) -> int: _UpperCamelCase = checkpoints.load_tax_checkpoint(args.checkpoint_path ) _UpperCamelCase = jnp.tree_util.tree_map(onp.array , a__ ) _UpperCamelCase = [ '''from __gin__ import dynamic_registration''', '''from music_spectrogram_diffusion.models.diffusion import diffusion_utils''', '''diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0''', '''diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()''', ] _UpperCamelCase = os.path.join(args.checkpoint_path , '''..''' , '''config.gin''' ) _UpperCamelCase = inference.parse_training_gin_file(a__ , a__ ) _UpperCamelCase = inference.InferenceModel(args.checkpoint_path , a__ ) _UpperCamelCase = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''' , variance_type='''fixed_large''' ) _UpperCamelCase = SpectrogramNotesEncoder( max_length=synth_model.sequence_length['''inputs'''] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='''gated-gelu''' , ) _UpperCamelCase = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length['''targets_context'''] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='''gated-gelu''' , ) _UpperCamelCase = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length['''targets_context'''] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) _UpperCamelCase = load_notes_encoder(ta_checkpoint['''target''']['''token_encoder'''] , a__ ) _UpperCamelCase = load_continuous_encoder(ta_checkpoint['''target''']['''continuous_encoder'''] , a__ ) _UpperCamelCase = load_decoder(ta_checkpoint['''target''']['''decoder'''] , a__ ) _UpperCamelCase = OnnxRuntimeModel.from_pretrained('''kashif/soundstream_mel_decoder''' ) _UpperCamelCase = SpectrogramDiffusionPipeline( notes_encoder=a__ , continuous_encoder=a__ , decoder=a__ , scheduler=a__ , melgan=a__ , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() parser.add_argument("""--output_path""", default=None, type=str, required=True, help="""Path to the converted model.""") parser.add_argument( """--save""", default=True, type=bool, required=False, help="""Whether to save the converted model or not.""" ) parser.add_argument( """--checkpoint_path""", default=F'''{MODEL}/checkpoint_500000''', type=str, required=False, help="""Path to the original jax model checkpoint.""", ) UpperCAmelCase = parser.parse_args() main(args)
256
1
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def lowerCamelCase ( ) -> tuple[list[int], int]: lowerCAmelCase_ = [randint(-1_000 , 1_000 ) for i in range(10 )] lowerCAmelCase_ = randint(-5_000 , 5_000 ) return (arr, r) lowerCamelCase_ = make_dataset() def lowerCamelCase ( a_ , a_ ) -> tuple[int, ...]: for triplet in permutations(a_ , 3 ): if sum(a_ ) == target: return tuple(sorted(a_ ) ) return (0, 0, 0) def lowerCamelCase ( a_ , a_ ) -> tuple[int, int, int]: arr.sort() lowerCAmelCase_ = len(a_ ) for i in range(n - 1 ): lowerCAmelCase_ , lowerCAmelCase_ = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def lowerCamelCase ( ) -> tuple[float, float]: lowerCAmelCase_ = '\nfrom __main__ import dataset, triplet_sum1, triplet_sum2\n' lowerCAmelCase_ = '\ntriplet_sum1(*dataset)\n' lowerCAmelCase_ = '\ntriplet_sum2(*dataset)\n' lowerCAmelCase_ = repeat(setup=a_ , stmt=a_ , repeat=5 , number=10_000 ) lowerCAmelCase_ = repeat(setup=a_ , stmt=a_ , repeat=5 , number=10_000 ) return (min(a_ ), min(a_ )) if __name__ == "__main__": from doctest import testmod testmod() lowerCamelCase_ = solution_times() print(f'''The time for naive implementation is {times[0]}.''') print(f'''The time for optimized implementation is {times[1]}.''')
14
def lowerCamelCase ( a_ , a_ ) -> List[Any]: lowerCAmelCase_ = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def lowerCamelCase ( a_ , a_ , a_ ) -> Union[str, Any]: lowerCAmelCase_ = 0 while b > 0: if b & 1: lowerCAmelCase_ = ((res % c) + (a % c)) % c a += a b >>= 1 return res
14
1
def UpperCAmelCase ( a_ , a_ , a_ ) -> float: """simple docstring""" return round(float(moles / volume ) * nfactor ) def UpperCAmelCase ( a_ , a_ , a_ ) -> float: """simple docstring""" return round(float((moles * 0.0_821 * temperature) / (volume) ) ) def UpperCAmelCase ( a_ , a_ , a_ ) -> float: """simple docstring""" return round(float((moles * 0.0_821 * temperature) / (pressure) ) ) def UpperCAmelCase ( a_ , a_ , a_ ) -> float: """simple docstring""" return round(float((pressure * volume) / (0.0_821 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
15
import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse('9.1.0'): SCREAMING_SNAKE_CASE :Any = { 'linear': PIL.Image.Resampling.BILINEAR, 'bilinear': PIL.Image.Resampling.BILINEAR, 'bicubic': PIL.Image.Resampling.BICUBIC, 'lanczos': PIL.Image.Resampling.LANCZOS, 'nearest': PIL.Image.Resampling.NEAREST, } else: SCREAMING_SNAKE_CASE :int = { 'linear': PIL.Image.LINEAR, 'bilinear': PIL.Image.BILINEAR, 'bicubic': PIL.Image.BICUBIC, 'lanczos': PIL.Image.LANCZOS, 'nearest': PIL.Image.NEAREST, } def UpperCAmelCase ( a_ ) -> Optional[Any]: """simple docstring""" __A = (images / 2 + 0.5).clamp(0 , 1 ) __A = images.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() __A = numpy_to_pil(a_ ) return images def UpperCAmelCase ( a_ ) -> int: """simple docstring""" if images.ndim == 3: __A = images[None, ...] __A = (images * 2_5_5).round().astype("uint8" ) if images.shape[-1] == 1: # special case for grayscale (single channel) images __A = [Image.fromarray(image.squeeze() , mode="L" ) for image in images] else: __A = [Image.fromarray(a_ ) for image in images] return pil_images
15
1
"""simple docstring""" import gc import random import unittest import torch from diffusers import ( IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ) from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import floats_tensor, load_numpy, require_torch_gpu, skip_mps, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference from . import IFPipelineTesterMixin @skip_mps class lowerCAmelCase__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): SCREAMING_SNAKE_CASE_ =IFPipeline SCREAMING_SNAKE_CASE_ =TEXT_TO_IMAGE_PARAMS - {'''width''', '''height''', '''latents'''} SCREAMING_SNAKE_CASE_ =TEXT_TO_IMAGE_BATCH_PARAMS SCREAMING_SNAKE_CASE_ =PipelineTesterMixin.required_optional_params - {'''latents'''} def __a ( self : Dict ): '''simple docstring''' return self._get_dummy_components() def __a ( self : Any , snake_case__ : Dict , snake_case__ : Optional[Any]=0 ): '''simple docstring''' if str(snake_case__ ).startswith("mps" ): UpperCAmelCase__ : str = torch.manual_seed(snake_case__ ) else: UpperCAmelCase__ : Optional[int] = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) UpperCAmelCase__ : Tuple = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def __a ( self : Tuple ): '''simple docstring''' self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def __a ( self : Tuple ): '''simple docstring''' # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1 ) def __a ( self : Dict ): '''simple docstring''' self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def __a ( self : int ): '''simple docstring''' self._test_save_load_local() def __a ( self : Any ): '''simple docstring''' self._test_inference_batch_single_identical( expected_max_diff=1e-2 , ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __a ( self : Optional[Any] ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): def __a ( self : str ): '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __a ( self : Tuple ): '''simple docstring''' # if UpperCAmelCase__ : Any = IFPipeline.from_pretrained("DeepFloyd/IF-I-XL-v1.0" , variant="fp16" , torch_dtype=torch.floataa ) UpperCAmelCase__ : Union[str, Any] = IFSuperResolutionPipeline.from_pretrained( "DeepFloyd/IF-II-L-v1.0" , variant="fp16" , torch_dtype=torch.floataa , text_encoder=snake_case__ , tokenizer=snake_case__ ) # pre compute text embeddings and remove T5 to save memory pipe_a.text_encoder.to("cuda" ) UpperCAmelCase__ , UpperCAmelCase__ : Any = pipe_a.encode_prompt("anime turtle" , device="cuda" ) del pipe_a.tokenizer del pipe_a.text_encoder gc.collect() UpperCAmelCase__ : Tuple = None UpperCAmelCase__ : List[Any] = None pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # img2img UpperCAmelCase__ : List[str] = IFImgaImgPipeline(**pipe_a.components ) UpperCAmelCase__ : List[str] = IFImgaImgSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_imgaimg(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) pipe_a.remove_all_hooks() pipe_a.remove_all_hooks() # inpainting UpperCAmelCase__ : List[str] = IFInpaintingPipeline(**pipe_a.components ) UpperCAmelCase__ : List[str] = IFInpaintingSuperResolutionPipeline(**pipe_a.components ) pipe_a.enable_model_cpu_offload() pipe_a.enable_model_cpu_offload() pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) pipe_a.unet.set_attn_processor(AttnAddedKVProcessor() ) self._test_if_inpainting(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) def __a ( self : List[str] , snake_case__ : Tuple , snake_case__ : List[Any] , snake_case__ : List[str] , snake_case__ : List[Any] ): '''simple docstring''' # pipeline 1 _start_torch_memory_measurement() UpperCAmelCase__ : List[str] = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase__ : Dict = pipe_a( prompt_embeds=snake_case__ , negative_prompt_embeds=snake_case__ , num_inference_steps=2 , generator=snake_case__ , output_type="np" , ) UpperCAmelCase__ : List[Any] = output.images[0] assert image.shape == (6_4, 6_4, 3) UpperCAmelCase__ : Optional[Any] = torch.cuda.max_memory_allocated() assert mem_bytes < 1_3 * 1_0**9 UpperCAmelCase__ : str = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if.npy" ) assert_mean_pixel_difference(snake_case__ , snake_case__ ) # pipeline 2 _start_torch_memory_measurement() UpperCAmelCase__ : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase__ : Dict = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(0 ) ).to(snake_case__ ) UpperCAmelCase__ : str = pipe_a( prompt_embeds=snake_case__ , negative_prompt_embeds=snake_case__ , image=snake_case__ , generator=snake_case__ , num_inference_steps=2 , output_type="np" , ) UpperCAmelCase__ : Union[str, Any] = output.images[0] assert image.shape == (2_5_6, 2_5_6, 3) UpperCAmelCase__ : List[str] = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 1_0**9 UpperCAmelCase__ : Dict = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_superresolution_stage_II.npy" ) assert_mean_pixel_difference(snake_case__ , snake_case__ ) def __a ( self : Optional[Any] , snake_case__ : Optional[int] , snake_case__ : Any , snake_case__ : Optional[Any] , snake_case__ : List[str] ): '''simple docstring''' # pipeline 1 _start_torch_memory_measurement() UpperCAmelCase__ : List[str] = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(0 ) ).to(snake_case__ ) UpperCAmelCase__ : int = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase__ : Tuple = pipe_a( prompt_embeds=snake_case__ , negative_prompt_embeds=snake_case__ , image=snake_case__ , num_inference_steps=2 , generator=snake_case__ , output_type="np" , ) UpperCAmelCase__ : str = output.images[0] assert image.shape == (6_4, 6_4, 3) UpperCAmelCase__ : Optional[Any] = torch.cuda.max_memory_allocated() assert mem_bytes < 1_0 * 1_0**9 UpperCAmelCase__ : List[str] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img.npy" ) assert_mean_pixel_difference(snake_case__ , snake_case__ ) # pipeline 2 _start_torch_memory_measurement() UpperCAmelCase__ : int = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase__ : Optional[int] = floats_tensor((1, 3, 2_5_6, 2_5_6) , rng=random.Random(0 ) ).to(snake_case__ ) UpperCAmelCase__ : Tuple = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(0 ) ).to(snake_case__ ) UpperCAmelCase__ : Dict = pipe_a( prompt_embeds=snake_case__ , negative_prompt_embeds=snake_case__ , image=snake_case__ , original_image=snake_case__ , generator=snake_case__ , num_inference_steps=2 , output_type="np" , ) UpperCAmelCase__ : Optional[Any] = output.images[0] assert image.shape == (2_5_6, 2_5_6, 3) UpperCAmelCase__ : Dict = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 1_0**9 UpperCAmelCase__ : str = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_img2img_superresolution_stage_II.npy" ) assert_mean_pixel_difference(snake_case__ , snake_case__ ) def __a ( self : Union[str, Any] , snake_case__ : List[str] , snake_case__ : int , snake_case__ : int , snake_case__ : Optional[int] ): '''simple docstring''' # pipeline 1 _start_torch_memory_measurement() UpperCAmelCase__ : str = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(0 ) ).to(snake_case__ ) UpperCAmelCase__ : Dict = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(1 ) ).to(snake_case__ ) UpperCAmelCase__ : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase__ : int = pipe_a( prompt_embeds=snake_case__ , negative_prompt_embeds=snake_case__ , image=snake_case__ , mask_image=snake_case__ , num_inference_steps=2 , generator=snake_case__ , output_type="np" , ) UpperCAmelCase__ : int = output.images[0] assert image.shape == (6_4, 6_4, 3) UpperCAmelCase__ : Union[str, Any] = torch.cuda.max_memory_allocated() assert mem_bytes < 1_0 * 1_0**9 UpperCAmelCase__ : int = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting.npy" ) assert_mean_pixel_difference(snake_case__ , snake_case__ ) # pipeline 2 _start_torch_memory_measurement() UpperCAmelCase__ : int = torch.Generator(device="cpu" ).manual_seed(0 ) UpperCAmelCase__ : Optional[Any] = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(0 ) ).to(snake_case__ ) UpperCAmelCase__ : Optional[int] = floats_tensor((1, 3, 2_5_6, 2_5_6) , rng=random.Random(0 ) ).to(snake_case__ ) UpperCAmelCase__ : List[Any] = floats_tensor((1, 3, 2_5_6, 2_5_6) , rng=random.Random(1 ) ).to(snake_case__ ) UpperCAmelCase__ : Union[str, Any] = pipe_a( prompt_embeds=snake_case__ , negative_prompt_embeds=snake_case__ , image=snake_case__ , mask_image=snake_case__ , original_image=snake_case__ , generator=snake_case__ , num_inference_steps=2 , output_type="np" , ) UpperCAmelCase__ : Tuple = output.images[0] assert image.shape == (2_5_6, 2_5_6, 3) UpperCAmelCase__ : List[str] = torch.cuda.max_memory_allocated() assert mem_bytes < 4 * 1_0**9 UpperCAmelCase__ : List[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/if/test_if_inpainting_superresolution_stage_II.npy" ) assert_mean_pixel_difference(snake_case__ , snake_case__ ) def SCREAMING_SNAKE_CASE__ ( )-> Any: '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats()
298
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import tensorflow as tf from transformers import AutoTokenizer, TFAutoModelForSeqaSeqLM @require_tf @require_sentencepiece @require_tokenizers class lowerCAmelCase__ ( unittest.TestCase ): @slow def __a ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[Any] = TFAutoModelForSeqaSeqLM.from_pretrained("google/mt5-small" ) UpperCAmelCase__ : int = AutoTokenizer.from_pretrained("google/mt5-small" ) UpperCAmelCase__ : Dict = tokenizer("Hello there" , return_tensors="tf" ).input_ids UpperCAmelCase__ : Union[str, Any] = tokenizer("Hi I am" , return_tensors="tf" ).input_ids UpperCAmelCase__ : Dict = model(snake_case__ , labels=snake_case__ ).loss UpperCAmelCase__ : Optional[Any] = -tf.math.reduce_mean(snake_case__ ).numpy() UpperCAmelCase__ : List[Any] = -21.22_8168 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 2e-4 )
298
1
"""simple docstring""" from collections import namedtuple lowerCamelCase_ = namedtuple('''from_to''', '''from_ to''') lowerCamelCase_ = { '''cubicmeter''': from_to(1, 1), '''litre''': from_to(0.001, 1000), '''kilolitre''': from_to(1, 1), '''gallon''': from_to(0.00454, 264.172), '''cubicyard''': from_to(0.76455, 1.30795), '''cubicfoot''': from_to(0.028, 35.3147), '''cup''': from_to(0.000236588, 4226.75), } def snake_case ( A__ ,A__ ,A__ ): if from_type not in METRIC_CONVERSION: raise ValueError( F"""Invalid 'from_type' value: {from_type!r} Supported values are:\n""" + ", ".join(A__ ) ) if to_type not in METRIC_CONVERSION: raise ValueError( F"""Invalid 'to_type' value: {to_type!r}. Supported values are:\n""" + ", ".join(A__ ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
268
"""simple docstring""" from torch import nn def snake_case ( A__ ): if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(F"""Unsupported activation function: {act_fn}""" )
268
1
"""simple docstring""" import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowerCamelCase : def __init__( self : Dict , UpperCamelCase : Optional[Any] , UpperCamelCase : str=13 , UpperCamelCase : Union[str, Any]=30 , UpperCamelCase : Optional[Any]=2 , UpperCamelCase : Union[str, Any]=3 , UpperCamelCase : Optional[int]=True , UpperCamelCase : Optional[Any]=True , UpperCamelCase : Optional[Any]=32 , UpperCamelCase : int=5 , UpperCamelCase : Optional[Any]=4 , UpperCamelCase : int=37 , UpperCamelCase : Optional[Any]="gelu" , UpperCamelCase : Tuple=0.1 , UpperCamelCase : Any=0.1 , UpperCamelCase : str=10 , UpperCamelCase : str=0.02 , UpperCamelCase : str=None , UpperCamelCase : Optional[int]=2 , ) -> Optional[Any]: """simple docstring""" lowerCAmelCase__ : Union[str, Any] = parent lowerCAmelCase__ : Optional[int] = batch_size lowerCAmelCase__ : int = image_size lowerCAmelCase__ : Any = patch_size lowerCAmelCase__ : Union[str, Any] = num_channels lowerCAmelCase__ : str = is_training lowerCAmelCase__ : Any = use_labels lowerCAmelCase__ : Any = hidden_size lowerCAmelCase__ : Union[str, Any] = num_hidden_layers lowerCAmelCase__ : List[Any] = num_attention_heads lowerCAmelCase__ : Optional[int] = intermediate_size lowerCAmelCase__ : Optional[Any] = hidden_act lowerCAmelCase__ : Any = hidden_dropout_prob lowerCAmelCase__ : Union[str, Any] = attention_probs_dropout_prob lowerCAmelCase__ : Optional[int] = type_sequence_label_size lowerCAmelCase__ : Union[str, Any] = initializer_range lowerCAmelCase__ : Optional[int] = scope lowerCAmelCase__ : str = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCAmelCase__ : int = (image_size // patch_size) ** 2 lowerCAmelCase__ : int = num_patches + 1 def _lowerCAmelCase ( self : Any ) -> Dict: """simple docstring""" lowerCAmelCase__ : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : Optional[int] = None if self.use_labels: lowerCAmelCase__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : Dict = self.get_config() return config, pixel_values, labels def _lowerCAmelCase ( self : Tuple ) -> List[Any]: """simple docstring""" return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=UpperCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowerCAmelCase ( self : Tuple , UpperCamelCase : Tuple , UpperCamelCase : Tuple , UpperCamelCase : Any ) -> List[Any]: """simple docstring""" lowerCAmelCase__ : Dict = ViTModel(config=UpperCamelCase ) model.to(UpperCamelCase ) model.eval() lowerCAmelCase__ : Optional[Any] = model(UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCAmelCase ( self : Optional[Any] , UpperCamelCase : str , UpperCamelCase : Union[str, Any] , UpperCamelCase : str ) -> Optional[int]: """simple docstring""" lowerCAmelCase__ : Any = ViTForMaskedImageModeling(config=UpperCamelCase ) model.to(UpperCamelCase ) model.eval() lowerCAmelCase__ : Optional[Any] = model(UpperCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase__ : Tuple = 1 lowerCAmelCase__ : Optional[Any] = ViTForMaskedImageModeling(UpperCamelCase ) model.to(UpperCamelCase ) model.eval() lowerCAmelCase__ : int = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : List[Any] = model(UpperCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowerCAmelCase ( self : List[Any] , UpperCamelCase : str , UpperCamelCase : Any , UpperCamelCase : str ) -> List[Any]: """simple docstring""" lowerCAmelCase__ : Optional[int] = self.type_sequence_label_size lowerCAmelCase__ : Tuple = ViTForImageClassification(UpperCamelCase ) model.to(UpperCamelCase ) model.eval() lowerCAmelCase__ : Any = model(UpperCamelCase , labels=UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCAmelCase__ : Any = 1 lowerCAmelCase__ : Union[str, Any] = ViTForImageClassification(UpperCamelCase ) model.to(UpperCamelCase ) model.eval() lowerCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : Dict = model(UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCAmelCase ( self : Any ) -> Optional[int]: """simple docstring""" lowerCAmelCase__ : Optional[Any] = self.prepare_config_and_inputs() ( ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ) : str = config_and_inputs lowerCAmelCase__ : Dict = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowerCamelCase ( a_ , a_ , unittest.TestCase ): _lowerCamelCase :Optional[int] = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) _lowerCamelCase :Any = ( {"feature-extraction": ViTModel, "image-classification": ViTForImageClassification} if is_torch_available() else {} ) _lowerCamelCase :Tuple = True _lowerCamelCase :Union[str, Any] = False _lowerCamelCase :Optional[int] = False _lowerCamelCase :Union[str, Any] = False def _lowerCAmelCase ( self : Optional[int] ) -> List[Any]: """simple docstring""" lowerCAmelCase__ : Union[str, Any] = ViTModelTester(self ) lowerCAmelCase__ : Dict = ConfigTester(self , config_class=UpperCamelCase , has_text_modality=UpperCamelCase , hidden_size=37 ) def _lowerCAmelCase ( self : Tuple ) -> Union[str, Any]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""ViT does not use inputs_embeds""" ) def _lowerCAmelCase ( self : Dict ) -> Dict: """simple docstring""" pass def _lowerCAmelCase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Optional[Any] = model_class(UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCamelCase , nn.Linear ) ) def _lowerCAmelCase ( self : Optional[Any] ) -> Dict: """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : int = model_class(UpperCamelCase ) lowerCAmelCase__ : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : List[Any] = [*signature.parameters.keys()] lowerCAmelCase__ : List[Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase ) def _lowerCAmelCase ( self : str ) -> Optional[int]: """simple docstring""" lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase ) def _lowerCAmelCase ( self : List[str] ) -> str: """simple docstring""" lowerCAmelCase__ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*UpperCamelCase ) def _lowerCAmelCase ( self : Optional[Any] ) -> Any: """simple docstring""" lowerCAmelCase__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase ) @slow def _lowerCAmelCase ( self : Dict ) -> Dict: """simple docstring""" for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : str = ViTModel.from_pretrained(UpperCamelCase ) self.assertIsNotNone(UpperCamelCase ) def lowercase_ ( ) -> Optional[Any]: lowerCAmelCase__ : str = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowerCamelCase ( unittest.TestCase ): @cached_property def _lowerCAmelCase ( self : List[Any] ) -> Dict: """simple docstring""" return ViTImageProcessor.from_pretrained("""google/vit-base-patch16-224""" ) if is_vision_available() else None @slow def _lowerCAmelCase ( self : Optional[Any] ) -> Dict: """simple docstring""" lowerCAmelCase__ : str = ViTForImageClassification.from_pretrained("""google/vit-base-patch16-224""" ).to(UpperCamelCase ) lowerCAmelCase__ : Optional[int] = self.default_image_processor lowerCAmelCase__ : str = prepare_img() lowerCAmelCase__ : Tuple = image_processor(images=UpperCamelCase , return_tensors="""pt""" ).to(UpperCamelCase ) # forward pass with torch.no_grad(): lowerCAmelCase__ : Dict = model(**UpperCamelCase ) # verify the logits lowerCAmelCase__ : Union[str, Any] = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , UpperCamelCase ) lowerCAmelCase__ : Optional[int] = torch.tensor([-0.2744, 0.8215, -0.0836] ).to(UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCamelCase , atol=1E-4 ) ) @slow def _lowerCAmelCase ( self : Tuple ) -> List[str]: """simple docstring""" # ViT models have an `interpolate_pos_encoding` argument in their forward method, # allowing to interpolate the pre-trained position embeddings in order to use # the model on higher resolutions. The DINO model by Facebook AI leverages this # to visualize self-attention on higher resolution images. lowerCAmelCase__ : Dict = ViTModel.from_pretrained("""facebook/dino-vits8""" ).to(UpperCamelCase ) lowerCAmelCase__ : Optional[Any] = ViTImageProcessor.from_pretrained("""facebook/dino-vits8""" , size=4_80 ) lowerCAmelCase__ : Union[str, Any] = prepare_img() lowerCAmelCase__ : Tuple = image_processor(images=UpperCamelCase , return_tensors="""pt""" ) lowerCAmelCase__ : Any = inputs.pixel_values.to(UpperCamelCase ) # forward pass with torch.no_grad(): lowerCAmelCase__ : List[Any] = model(UpperCamelCase , interpolate_pos_encoding=UpperCamelCase ) # verify the logits lowerCAmelCase__ : Tuple = torch.Size((1, 36_01, 3_84) ) self.assertEqual(outputs.last_hidden_state.shape , UpperCamelCase ) lowerCAmelCase__ : Dict = torch.tensor( [[4.2340, 4.3906, -6.6692], [4.5463, 1.8928, -6.7257], [4.4429, 0.8496, -5.8585]] ).to(UpperCamelCase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , UpperCamelCase , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def _lowerCAmelCase ( self : int ) -> Tuple: """simple docstring""" lowerCAmelCase__ : Tuple = ViTModel.from_pretrained("""facebook/dino-vits8""" , torch_dtype=torch.floataa , device_map="""auto""" ) lowerCAmelCase__ : Dict = self.default_image_processor lowerCAmelCase__ : List[str] = prepare_img() lowerCAmelCase__ : Dict = image_processor(images=UpperCamelCase , return_tensors="""pt""" ) lowerCAmelCase__ : Optional[Any] = inputs.pixel_values.to(UpperCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): lowerCAmelCase__ : str = model(UpperCamelCase )
212
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _A = logging.get_logger(__name__) _A = {"""vocab_file""": """vocab.json"""} _A = { """vocab_file""": { """mgp-str""": """https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json""", } } _A = {"""mgp-str""": 2_7} class _lowerCamelCase ( a_ ): _lowerCamelCase :Dict = VOCAB_FILES_NAMES _lowerCamelCase :Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP _lowerCamelCase :Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Optional[int] , UpperCamelCase : int , UpperCamelCase : Union[str, Any]="[GO]" , UpperCamelCase : Any="[GO]" , UpperCamelCase : Tuple="[s]" , UpperCamelCase : List[Any]="[GO]" , **UpperCamelCase : Dict ) -> Union[str, Any]: """simple docstring""" super().__init__( unk_token=UpperCamelCase , bos_token=UpperCamelCase , eos_token=UpperCamelCase , pad_token=UpperCamelCase , **UpperCamelCase , ) with open(UpperCamelCase , encoding="""utf-8""" ) as vocab_handle: lowerCAmelCase__ : Optional[Any] = json.load(UpperCamelCase ) lowerCAmelCase__ : Union[str, Any] = {v: k for k, v in self.vocab.items()} @property def _lowerCAmelCase ( self : int ) -> Dict: """simple docstring""" return len(self.vocab ) def _lowerCAmelCase ( self : Dict ) -> Any: """simple docstring""" return dict(self.vocab , **self.added_tokens_encoder ) def _lowerCAmelCase ( self : List[Any] , UpperCamelCase : Tuple ) -> Optional[Any]: """simple docstring""" lowerCAmelCase__ : Optional[Any] = [] for s in text: char_tokens.extend(UpperCamelCase ) return char_tokens def _lowerCAmelCase ( self : List[Any] , UpperCamelCase : str ) -> Tuple: """simple docstring""" return self.vocab.get(UpperCamelCase , self.vocab.get(self.unk_token ) ) def _lowerCAmelCase ( self : List[str] , UpperCamelCase : int ) -> Any: """simple docstring""" return self.decoder.get(UpperCamelCase ) def _lowerCAmelCase ( self : Dict , UpperCamelCase : str , UpperCamelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(UpperCamelCase ): logger.error("""Vocabulary path ({}) should be a directory""".format(UpperCamelCase ) ) return lowerCAmelCase__ : Union[str, Any] = os.path.join( UpperCamelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) with open(UpperCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=UpperCamelCase , ensure_ascii=UpperCamelCase ) + """\n""" ) return (vocab_file,)
212
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _A : str =logging.get_logger(__name__) _A : List[str] ={ '''microsoft/focalnet-tiny''': '''https://huggingface.co/microsoft/focalnet-tiny/resolve/main/config.json''', } class _lowercase ( _lowercase , _lowercase ): a = """focalnet""" def __init__( self: Dict , UpperCamelCase__: List[str]=224 , UpperCamelCase__: str=4 , UpperCamelCase__: List[Any]=3 , UpperCamelCase__: List[Any]=96 , UpperCamelCase__: Dict=False , UpperCamelCase__: Union[str, Any]=[192, 384, 768, 768] , UpperCamelCase__: Union[str, Any]=[2, 2, 6, 2] , UpperCamelCase__: Any=[2, 2, 2, 2] , UpperCamelCase__: Union[str, Any]=[3, 3, 3, 3] , UpperCamelCase__: Dict="gelu" , UpperCamelCase__: Dict=4.0 , UpperCamelCase__: Optional[int]=0.0 , UpperCamelCase__: str=0.1 , UpperCamelCase__: Union[str, Any]=False , UpperCamelCase__: str=1e-4 , UpperCamelCase__: List[str]=False , UpperCamelCase__: str=False , UpperCamelCase__: List[str]=False , UpperCamelCase__: Optional[Any]=0.02 , UpperCamelCase__: str=1e-5 , UpperCamelCase__: Union[str, Any]=32 , UpperCamelCase__: int=None , UpperCamelCase__: Optional[int]=None , **UpperCamelCase__: int , ): super().__init__(**UpperCamelCase__ ) lowerCamelCase__ : List[str] = image_size lowerCamelCase__ : Dict = patch_size lowerCamelCase__ : int = num_channels lowerCamelCase__ : str = embed_dim lowerCamelCase__ : List[str] = use_conv_embed lowerCamelCase__ : Dict = hidden_sizes lowerCamelCase__ : List[Any] = depths lowerCamelCase__ : int = focal_levels lowerCamelCase__ : Dict = focal_windows lowerCamelCase__ : str = hidden_act lowerCamelCase__ : Optional[Any] = mlp_ratio lowerCamelCase__ : int = hidden_dropout_prob lowerCamelCase__ : Any = drop_path_rate lowerCamelCase__ : Dict = use_layerscale lowerCamelCase__ : Union[str, Any] = layerscale_value lowerCamelCase__ : Optional[Any] = use_post_layernorm lowerCamelCase__ : Tuple = use_post_layernorm_in_modulation lowerCamelCase__ : Optional[int] = normalize_modulator lowerCamelCase__ : Dict = initializer_range lowerCamelCase__ : Any = layer_norm_eps lowerCamelCase__ : Union[str, Any] = encoder_stride lowerCamelCase__ : Union[str, Any] = ["""stem"""] + [F'''stage{idx}''' for idx in range(1 , len(self.depths ) + 1 )] lowerCamelCase__ , lowerCamelCase__ : str = get_aligned_output_features_output_indices( out_features=UpperCamelCase__ , out_indices=UpperCamelCase__ , stage_names=self.stage_names )
41
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a__ ( snake_case , snake_case , unittest.TestCase ): """simple docstring""" __lowerCamelCase = AutoencoderKL __lowerCamelCase = 'sample' __lowerCamelCase = 1e-2 @property def UpperCamelCase ( self ) -> Optional[Any]: '''simple docstring''' A__ = 4 A__ = 3 A__ = (32, 32) A__ = floats_tensor((batch_size, num_channels) + sizes ).to(lowercase ) return {"sample": image} @property def UpperCamelCase ( self ) -> Optional[Any]: '''simple docstring''' return (3, 32, 32) @property def UpperCamelCase ( self ) -> Optional[Any]: '''simple docstring''' return (3, 32, 32) def UpperCamelCase ( self ) -> Optional[Any]: '''simple docstring''' A__ = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } A__ = self.dummy_input return init_dict, inputs_dict def UpperCamelCase ( self ) -> Tuple: '''simple docstring''' pass def UpperCamelCase ( self ) -> Any: '''simple docstring''' pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def UpperCamelCase ( self ) -> Tuple: '''simple docstring''' A__ , A__ = self.prepare_init_args_and_inputs_for_common() A__ = self.model_class(**lowercase ) model.to(lowercase ) assert not model.is_gradient_checkpointing and model.training A__ = model(**lowercase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() A__ = torch.randn_like(lowercase ) A__ = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing A__ = self.model_class(**lowercase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowercase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training A__ = model_a(**lowercase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() A__ = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1e-5 ) A__ = dict(model.named_parameters() ) A__ = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5e-5 ) ) def UpperCamelCase ( self ) -> Optional[int]: '''simple docstring''' A__ , A__ = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowercase ) self.assertIsNotNone(lowercase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowercase ) A__ = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def UpperCamelCase ( self ) -> Any: '''simple docstring''' A__ = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) A__ = model.to(lowercase ) model.eval() if torch_device == "mps": A__ = torch.manual_seed(0 ) else: A__ = torch.Generator(device=lowercase ).manual_seed(0 ) A__ = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) A__ = image.to(lowercase ) with torch.no_grad(): A__ = model(lowercase , sample_posterior=lowercase , generator=lowercase ).sample A__ = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": A__ = torch.tensor( [ -4.00_78e-01, -3.83_23e-04, -1.26_81e-01, -1.14_62e-01, 2.00_95e-01, 1.08_93e-01, -8.82_47e-02, -3.03_61e-01, -9.86_44e-03, ] ) elif torch_device == "cpu": A__ = torch.tensor( [-0.1352, 0.0878, 0.0419, -0.0818, -0.1069, 0.0688, -0.1458, -0.4446, -0.0026] ) else: A__ = torch.tensor( [-0.2421, 0.4642, 0.2507, -0.0438, 0.0682, 0.3160, -0.2018, -0.0727, 0.2485] ) self.assertTrue(torch_all_close(lowercase , lowercase , rtol=1e-2 ) ) @slow class a__ ( unittest.TestCase ): """simple docstring""" def UpperCamelCase ( self , lowercase , lowercase ) -> str: '''simple docstring''' return F'gaussian_noise_s={seed}_shape={"_".join([str(lowercase ) for s in shape] )}.npy' def UpperCamelCase ( self ) -> Optional[int]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self , lowercase=0 , lowercase=(4, 3, 512, 512) , lowercase=False ) -> Optional[int]: '''simple docstring''' A__ = torch.floataa if fpaa else torch.floataa A__ = torch.from_numpy(load_hf_numpy(self.get_file_format(lowercase , lowercase ) ) ).to(lowercase ).to(lowercase ) return image def UpperCamelCase ( self , lowercase="CompVis/stable-diffusion-v1-4" , lowercase=False ) -> Any: '''simple docstring''' A__ = "fp16" if fpaa else None A__ = torch.floataa if fpaa else torch.floataa A__ = AutoencoderKL.from_pretrained( lowercase , subfolder="vae" , torch_dtype=lowercase , revision=lowercase , ) model.to(lowercase ).eval() return model def UpperCamelCase ( self , lowercase=0 ) -> List[str]: '''simple docstring''' if torch_device == "mps": return torch.manual_seed(lowercase ) return torch.Generator(device=lowercase ).manual_seed(lowercase ) @parameterized.expand( [ # fmt: off [33, [-0.1603, 0.9878, -0.0495, -0.0790, -0.2709, 0.8375, -0.2060, -0.0824], [-0.2395, 0.0098, 0.0102, -0.0709, -0.2840, -0.0274, -0.0718, -0.1824]], [47, [-0.2376, 0.1168, 0.1332, -0.4840, -0.2508, -0.0791, -0.0493, -0.4089], [0.0350, 0.0847, 0.0467, 0.0344, -0.0842, -0.0547, -0.0633, -0.1131]], # fmt: on ] ) def UpperCamelCase ( self , lowercase , lowercase , lowercase ) -> int: '''simple docstring''' A__ = self.get_sd_vae_model() A__ = self.get_sd_image(lowercase ) A__ = self.get_generator(lowercase ) with torch.no_grad(): A__ = model(lowercase , generator=lowercase , sample_posterior=lowercase ).sample assert sample.shape == image.shape A__ = sample[-1, -2:, -2:, :2].flatten().float().cpu() A__ = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowercase , lowercase , atol=3e-3 ) @parameterized.expand( [ # fmt: off [33, [-0.0513, 0.0289, 1.3799, 0.2166, -0.2573, -0.0871, 0.5103, -0.0999]], [47, [-0.4128, -0.1320, -0.3704, 0.1965, -0.4116, -0.2332, -0.3340, 0.2247]], # fmt: on ] ) @require_torch_gpu def UpperCamelCase ( self , lowercase , lowercase ) -> List[Any]: '''simple docstring''' A__ = self.get_sd_vae_model(fpaa=lowercase ) A__ = self.get_sd_image(lowercase , fpaa=lowercase ) A__ = self.get_generator(lowercase ) with torch.no_grad(): A__ = model(lowercase , generator=lowercase , sample_posterior=lowercase ).sample assert sample.shape == image.shape A__ = sample[-1, -2:, :2, -2:].flatten().float().cpu() A__ = torch.tensor(lowercase ) assert torch_all_close(lowercase , lowercase , atol=1e-2 ) @parameterized.expand( [ # fmt: off [33, [-0.1609, 0.9866, -0.0487, -0.0777, -0.2716, 0.8368, -0.2055, -0.0814], [-0.2395, 0.0098, 0.0102, -0.0709, -0.2840, -0.0274, -0.0718, -0.1824]], [47, [-0.2377, 0.1147, 0.1333, -0.4841, -0.2506, -0.0805, -0.0491, -0.4085], [0.0350, 0.0847, 0.0467, 0.0344, -0.0842, -0.0547, -0.0633, -0.1131]], # fmt: on ] ) def UpperCamelCase ( self , lowercase , lowercase , lowercase ) -> Dict: '''simple docstring''' A__ = self.get_sd_vae_model() A__ = self.get_sd_image(lowercase ) with torch.no_grad(): A__ = model(lowercase ).sample assert sample.shape == image.shape A__ = sample[-1, -2:, -2:, :2].flatten().float().cpu() A__ = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowercase , lowercase , atol=3e-3 ) @parameterized.expand( [ # fmt: off [13, [-0.2051, -0.1803, -0.2311, -0.2114, -0.3292, -0.3574, -0.2953, -0.3323]], [37, [-0.2632, -0.2625, -0.2199, -0.2741, -0.4539, -0.4990, -0.3720, -0.4925]], # fmt: on ] ) @require_torch_gpu def UpperCamelCase ( self , lowercase , lowercase ) -> Tuple: '''simple docstring''' A__ = self.get_sd_vae_model() A__ = self.get_sd_image(lowercase , shape=(3, 4, 64, 64) ) with torch.no_grad(): A__ = model.decode(lowercase ).sample assert list(sample.shape ) == [3, 3, 512, 512] A__ = sample[-1, -2:, :2, -2:].flatten().cpu() A__ = torch.tensor(lowercase ) assert torch_all_close(lowercase , lowercase , atol=1e-3 ) @parameterized.expand( [ # fmt: off [27, [-0.0369, 0.0207, -0.0776, -0.0682, -0.1747, -0.1930, -0.1465, -0.2039]], [16, [-0.1628, -0.2134, -0.2747, -0.2642, -0.3774, -0.4404, -0.3687, -0.4277]], # fmt: on ] ) @require_torch_gpu def UpperCamelCase ( self , lowercase , lowercase ) -> Union[str, Any]: '''simple docstring''' A__ = self.get_sd_vae_model(fpaa=lowercase ) A__ = self.get_sd_image(lowercase , shape=(3, 4, 64, 64) , fpaa=lowercase ) with torch.no_grad(): A__ = model.decode(lowercase ).sample assert list(sample.shape ) == [3, 3, 512, 512] A__ = sample[-1, -2:, :2, -2:].flatten().float().cpu() A__ = torch.tensor(lowercase ) assert torch_all_close(lowercase , lowercase , atol=5e-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def UpperCamelCase ( self , lowercase ) -> Optional[Any]: '''simple docstring''' A__ = self.get_sd_vae_model(fpaa=lowercase ) A__ = self.get_sd_image(lowercase , shape=(3, 4, 64, 64) , fpaa=lowercase ) with torch.no_grad(): A__ = model.decode(lowercase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): A__ = model.decode(lowercase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowercase , lowercase , atol=1e-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def UpperCamelCase ( self , lowercase ) -> List[str]: '''simple docstring''' A__ = self.get_sd_vae_model() A__ = self.get_sd_image(lowercase , shape=(3, 4, 64, 64) ) with torch.no_grad(): A__ = model.decode(lowercase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): A__ = model.decode(lowercase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowercase , lowercase , atol=1e-2 ) @parameterized.expand( [ # fmt: off [33, [-0.3001, 0.0918, -2.6984, -3.9720, -3.2099, -5.0353, 1.7338, -0.2065, 3.4267]], [47, [-1.5030, -4.3871, -6.0355, -9.1157, -1.6661, -2.7853, 2.1607, -5.0823, 2.5633]], # fmt: on ] ) def UpperCamelCase ( self , lowercase , lowercase ) -> str: '''simple docstring''' A__ = self.get_sd_vae_model() A__ = self.get_sd_image(lowercase ) A__ = self.get_generator(lowercase ) with torch.no_grad(): A__ = model.encode(lowercase ).latent_dist A__ = dist.sample(generator=lowercase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] A__ = sample[0, -1, -3:, -3:].flatten().cpu() A__ = torch.tensor(lowercase ) A__ = 3e-3 if torch_device != "mps" else 1e-2 assert torch_all_close(lowercase , lowercase , atol=lowercase )
68
0
"""simple docstring""" import argparse import logging import pickle from collections import Counter logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) UpperCamelCase_ = logging.getLogger(__name__) if __name__ == "__main__": UpperCamelCase_ = argparse.ArgumentParser( description='Token Counts for smoothing the masking probabilities in MLM (cf XLM/word2vec)' ) parser.add_argument( '--data_file', type=str, default='data/dump.bert-base-uncased.pickle', help='The binarized dataset.' ) parser.add_argument( '--token_counts_dump', type=str, default='data/token_counts.bert-base-uncased.pickle', help='The dump file.' ) parser.add_argument('--vocab_size', default=30522, type=int) UpperCamelCase_ = parser.parse_args() logger.info(F"""Loading data from {args.data_file}""") with open(args.data_file, 'rb') as fp: UpperCamelCase_ = pickle.load(fp) logger.info('Counting occurrences for MLM.') UpperCamelCase_ = Counter() for tk_ids in data: counter.update(tk_ids) UpperCamelCase_ = [0] * args.vocab_size for k, v in counter.items(): UpperCamelCase_ = v logger.info(F"""Dump to {args.token_counts_dump}""") with open(args.token_counts_dump, 'wb') as handle: pickle.dump(counts, handle, protocol=pickle.HIGHEST_PROTOCOL)
366
"""simple docstring""" import unittest from transformers import LiltConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, ) from transformers.models.lilt.modeling_lilt import LILT_PRETRAINED_MODEL_ARCHIVE_LIST class snake_case : def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=99 , __UpperCAmelCase=24 , __UpperCAmelCase=2 , __UpperCAmelCase=6 , __UpperCAmelCase=37 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=5_12 , __UpperCAmelCase=16 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=3 , __UpperCAmelCase=None , __UpperCAmelCase=10_00 , ) ->List[str]: a_ = parent a_ = batch_size a_ = seq_length a_ = is_training a_ = use_input_mask a_ = use_token_type_ids a_ = use_labels a_ = vocab_size a_ = hidden_size a_ = num_hidden_layers a_ = num_attention_heads a_ = intermediate_size a_ = hidden_act a_ = hidden_dropout_prob a_ = attention_probs_dropout_prob a_ = max_position_embeddings a_ = type_vocab_size a_ = type_sequence_label_size a_ = initializer_range a_ = num_labels a_ = scope a_ = range_bbox def UpperCAmelCase__ ( self) ->int: a_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) a_ = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox) # Ensure that bbox is legal for i in range(bbox.shape[0]): for j in range(bbox.shape[1]): if bbox[i, j, 3] < bbox[i, j, 1]: a_ = bbox[i, j, 3] a_ = bbox[i, j, 1] a_ = t if bbox[i, j, 2] < bbox[i, j, 0]: a_ = bbox[i, j, 2] a_ = bbox[i, j, 0] a_ = t a_ = None if self.use_input_mask: a_ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2) a_ = None if self.use_token_type_ids: a_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) a_ = None a_ = None if self.use_labels: a_ = ids_tensor([self.batch_size] , self.type_sequence_label_size) a_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) a_ = self.get_config() return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels def UpperCAmelCase__ ( self) ->List[str]: return LiltConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ) ->Any: a_ = LiltModel(config=__UpperCAmelCase) model.to(__UpperCAmelCase) model.eval() a_ = model(__UpperCAmelCase , bbox=__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase) a_ = model(__UpperCAmelCase , bbox=__UpperCAmelCase , token_type_ids=__UpperCAmelCase) a_ = model(__UpperCAmelCase , bbox=__UpperCAmelCase) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size)) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ) ->Union[str, Any]: a_ = self.num_labels a_ = LiltForTokenClassification(config=__UpperCAmelCase) model.to(__UpperCAmelCase) model.eval() a_ = model( __UpperCAmelCase , bbox=__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ) ->Dict: a_ = LiltForQuestionAnswering(config=__UpperCAmelCase) model.to(__UpperCAmelCase) model.eval() a_ = model( __UpperCAmelCase , bbox=__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , start_positions=__UpperCAmelCase , end_positions=__UpperCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def UpperCAmelCase__ ( self) ->str: a_ = self.prepare_config_and_inputs() ( ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ) = config_and_inputs a_ = { "input_ids": input_ids, "bbox": bbox, "token_type_ids": token_type_ids, "attention_mask": input_mask, } return config, inputs_dict @require_torch class snake_case ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , unittest.TestCase ): a_ : List[Any] = ( ( LiltModel, LiltForSequenceClassification, LiltForTokenClassification, LiltForQuestionAnswering, ) if is_torch_available() else () ) a_ : List[str] = ( { """feature-extraction""": LiltModel, """question-answering""": LiltForQuestionAnswering, """text-classification""": LiltForSequenceClassification, """token-classification""": LiltForTokenClassification, """zero-shot""": LiltForSequenceClassification, } if is_torch_available() else {} ) a_ : Any = False a_ : Dict = False def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) ->int: return True def UpperCAmelCase__ ( self) ->str: a_ = LiltModelTester(self) a_ = ConfigTester(self , config_class=__UpperCAmelCase , hidden_size=37) def UpperCAmelCase__ ( self) ->List[Any]: self.config_tester.run_common_tests() def UpperCAmelCase__ ( self) ->Tuple: a_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase) def UpperCAmelCase__ ( self) ->Dict: a_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: a_ = type self.model_tester.create_and_check_model(*__UpperCAmelCase) def UpperCAmelCase__ ( self) ->List[str]: a_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCAmelCase) def UpperCAmelCase__ ( self) ->str: a_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCAmelCase) @slow def UpperCAmelCase__ ( self) ->List[Any]: for model_name in LILT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a_ = LiltModel.from_pretrained(__UpperCAmelCase) self.assertIsNotNone(__UpperCAmelCase) @require_torch @slow class snake_case ( unittest.TestCase ): def UpperCAmelCase__ ( self) ->List[Any]: a_ = LiltModel.from_pretrained("SCUT-DLVCLab/lilt-roberta-en-base").to(__UpperCAmelCase) a_ = torch.tensor([[1, 2]] , device=__UpperCAmelCase) a_ = torch.tensor([[[1, 2, 3, 4], [5, 6, 7, 8]]] , device=__UpperCAmelCase) # forward pass with torch.no_grad(): a_ = model(input_ids=__UpperCAmelCase , bbox=__UpperCAmelCase) a_ = torch.Size([1, 2, 7_68]) a_ = torch.tensor( [[-0.0_653, 0.0_950, -0.0_061], [-0.0_545, 0.0_926, -0.0_324]] , device=__UpperCAmelCase , ) self.assertTrue(outputs.last_hidden_state.shape , __UpperCAmelCase) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :, :3] , __UpperCAmelCase , atol=1E-3))
303
0
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: UpperCAmelCase_ : Union[str, Any] = None UpperCAmelCase_ : List[str] = logging.get_logger(__name__) UpperCAmelCase_ : Union[str, Any] = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} UpperCAmelCase_ : Optional[Any] = { """vocab_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/spiece.model""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/spiece.model""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/spiece.model""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/spiece.model""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model""", }, """tokenizer_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json""", }, } UpperCAmelCase_ : Any = { """albert-base-v1""": 512, """albert-large-v1""": 512, """albert-xlarge-v1""": 512, """albert-xxlarge-v1""": 512, """albert-base-v2""": 512, """albert-large-v2""": 512, """albert-xlarge-v2""": 512, """albert-xxlarge-v2""": 512, } UpperCAmelCase_ : Union[str, Any] = """▁""" class lowerCAmelCase__ ( UpperCAmelCase__ ): '''simple docstring''' __UpperCamelCase = VOCAB_FILES_NAMES __UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase = AlbertTokenizer def __init__( self : List[Any] , lowercase_ : List[Any]=None , lowercase_ : Optional[Any]=None , lowercase_ : Dict=True , lowercase_ : Optional[Any]=True , lowercase_ : Optional[int]=False , lowercase_ : List[str]="[CLS]" , lowercase_ : str="[SEP]" , lowercase_ : Dict="<unk>" , lowercase_ : int="[SEP]" , lowercase_ : List[str]="<pad>" , lowercase_ : Optional[Any]="[CLS]" , lowercase_ : Optional[int]="[MASK]" , **lowercase_ : List[str] , ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = ( AddedToken(lowercase_ , lstrip=lowercase_ , rstrip=lowercase_ , normalized=lowercase_) if isinstance(lowercase_ , lowercase_) else mask_token ) super().__init__( lowercase_ , tokenizer_file=lowercase_ , do_lower_case=lowercase_ , remove_space=lowercase_ , keep_accents=lowercase_ , bos_token=lowercase_ , eos_token=lowercase_ , unk_token=lowercase_ , sep_token=lowercase_ , pad_token=lowercase_ , cls_token=lowercase_ , mask_token=lowercase_ , **lowercase_ , ) SCREAMING_SNAKE_CASE_ : Optional[int] = do_lower_case SCREAMING_SNAKE_CASE_ : Optional[Any] = remove_space SCREAMING_SNAKE_CASE_ : List[str] = keep_accents SCREAMING_SNAKE_CASE_ : List[Any] = vocab_file SCREAMING_SNAKE_CASE_ : List[str] = False if not self.vocab_file else True def _SCREAMING_SNAKE_CASE ( self : List[str] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = [self.sep_token_id] SCREAMING_SNAKE_CASE_ : str = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[str] = [self.sep_token_id] SCREAMING_SNAKE_CASE_ : Dict = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def _SCREAMING_SNAKE_CASE ( self : List[str] , lowercase_ : str , lowercase_ : Optional[str] = None): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''') if not os.path.isdir(lowercase_): logger.error(F'Vocabulary path ({save_directory}) should be a directory') return SCREAMING_SNAKE_CASE_ : List[Any] = os.path.join( lowercase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowercase_): copyfile(self.vocab_file , lowercase_) return (out_vocab_file,)
91
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ : Optional[int] = logging.get_logger(__name__) UpperCAmelCase_ : Any = {"""openai-gpt""": """https://huggingface.co/openai-gpt/resolve/main/config.json"""} class lowerCAmelCase__ ( UpperCAmelCase__ ): '''simple docstring''' __UpperCamelCase = "openai-gpt" __UpperCamelCase = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : List[str] , lowercase_ : List[str]=40478 , lowercase_ : List[str]=512 , lowercase_ : Optional[Any]=768 , lowercase_ : Tuple=12 , lowercase_ : Tuple=12 , lowercase_ : Union[str, Any]="gelu" , lowercase_ : List[Any]=0.1 , lowercase_ : Tuple=0.1 , lowercase_ : List[Any]=0.1 , lowercase_ : List[Any]=1e-5 , lowercase_ : int=0.02 , lowercase_ : Optional[int]="cls_index" , lowercase_ : Any=True , lowercase_ : List[Any]=None , lowercase_ : List[str]=True , lowercase_ : Optional[Any]=0.1 , **lowercase_ : List[str] , ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Optional[int] = vocab_size SCREAMING_SNAKE_CASE_ : Tuple = n_positions SCREAMING_SNAKE_CASE_ : Optional[int] = n_embd SCREAMING_SNAKE_CASE_ : Dict = n_layer SCREAMING_SNAKE_CASE_ : Any = n_head SCREAMING_SNAKE_CASE_ : Union[str, Any] = afn SCREAMING_SNAKE_CASE_ : int = resid_pdrop SCREAMING_SNAKE_CASE_ : List[str] = embd_pdrop SCREAMING_SNAKE_CASE_ : Union[str, Any] = attn_pdrop SCREAMING_SNAKE_CASE_ : Union[str, Any] = layer_norm_epsilon SCREAMING_SNAKE_CASE_ : List[Any] = initializer_range SCREAMING_SNAKE_CASE_ : List[str] = summary_type SCREAMING_SNAKE_CASE_ : Tuple = summary_use_proj SCREAMING_SNAKE_CASE_ : Union[str, Any] = summary_activation SCREAMING_SNAKE_CASE_ : Any = summary_first_dropout SCREAMING_SNAKE_CASE_ : List[str] = summary_proj_to_labels super().__init__(**lowercase_)
91
1
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _a ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): UpperCamelCase = AltDiffusionPipeline UpperCamelCase = TEXT_TO_IMAGE_PARAMS UpperCamelCase = TEXT_TO_IMAGE_BATCH_PARAMS UpperCamelCase = TEXT_TO_IMAGE_IMAGE_PARAMS UpperCamelCase = TEXT_TO_IMAGE_IMAGE_PARAMS def snake_case ( self : int ) -> int: '''simple docstring''' torch.manual_seed(0 ) _UpperCamelCase : Dict = UNetaDConditionModel( block_out_channels=(3_2, 6_4), layers_per_block=2, sample_size=3_2, in_channels=4, out_channels=4, down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D'''), up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D'''), cross_attention_dim=3_2, ) _UpperCamelCase : Union[str, Any] = DDIMScheduler( beta_start=0.00_085, beta_end=0.012, beta_schedule='''scaled_linear''', clip_sample=lowerCAmelCase__, set_alpha_to_one=lowerCAmelCase__, ) torch.manual_seed(0 ) _UpperCamelCase : List[str] = AutoencoderKL( block_out_channels=[3_2, 6_4], in_channels=3, out_channels=3, down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''], up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''], latent_channels=4, ) # TODO: address the non-deterministic text encoder (fails for save-load tests) # torch.manual_seed(0) # text_encoder_config = RobertaSeriesConfig( # hidden_size=32, # project_dim=32, # intermediate_size=37, # layer_norm_eps=1e-05, # num_attention_heads=4, # num_hidden_layers=5, # vocab_size=5002, # ) # text_encoder = RobertaSeriesModelWithTransformation(text_encoder_config) torch.manual_seed(0 ) _UpperCamelCase : str = CLIPTextConfig( bos_token_id=0, eos_token_id=2, hidden_size=3_2, projection_dim=3_2, intermediate_size=3_7, layer_norm_eps=1e-0_5, num_attention_heads=4, num_hidden_layers=5, pad_token_id=1, vocab_size=5_0_0_2, ) _UpperCamelCase : List[Any] = CLIPTextModel(lowerCAmelCase__ ) _UpperCamelCase : Dict = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) _UpperCamelCase : str = 7_7 _UpperCamelCase : int = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def snake_case ( self : Dict, lowerCAmelCase__ : Any, lowerCAmelCase__ : int=0 ) -> Optional[int]: '''simple docstring''' if str(lowerCAmelCase__ ).startswith('''mps''' ): _UpperCamelCase : Any = torch.manual_seed(lowerCAmelCase__ ) else: _UpperCamelCase : Dict = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) _UpperCamelCase : str = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def snake_case ( self : List[Any] ) -> List[str]: '''simple docstring''' super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def snake_case ( self : List[Any] ) -> Tuple: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def snake_case ( self : List[str] ) -> List[Any]: '''simple docstring''' _UpperCamelCase : Any = '''cpu''' # ensure determinism for the device-dependent torch.Generator _UpperCamelCase : int = self.get_dummy_components() torch.manual_seed(0 ) _UpperCamelCase : Any = RobertaSeriesConfig( hidden_size=3_2, project_dim=3_2, intermediate_size=3_7, layer_norm_eps=1e-0_5, num_attention_heads=4, num_hidden_layers=5, vocab_size=5_0_0_2, ) # TODO: remove after fixing the non-deterministic text encoder _UpperCamelCase : Tuple = RobertaSeriesModelWithTransformation(lowerCAmelCase__ ) _UpperCamelCase : str = text_encoder _UpperCamelCase : List[Any] = AltDiffusionPipeline(**lowerCAmelCase__ ) _UpperCamelCase : List[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) _UpperCamelCase : Optional[int] = self.get_dummy_inputs(lowerCAmelCase__ ) _UpperCamelCase : Optional[Any] = '''A photo of an astronaut''' _UpperCamelCase : Any = alt_pipe(**lowerCAmelCase__ ) _UpperCamelCase : Any = output.images _UpperCamelCase : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) _UpperCamelCase : List[Any] = np.array( [0.5_748_162, 0.60_447_145, 0.48_821_217, 0.50_100_636, 0.5_431_185, 0.45_763_683, 0.49_657_696, 0.48_132_733, 0.47_573_093] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def snake_case ( self : int ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase : List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator _UpperCamelCase : Optional[Any] = self.get_dummy_components() _UpperCamelCase : str = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) torch.manual_seed(0 ) _UpperCamelCase : int = RobertaSeriesConfig( hidden_size=3_2, project_dim=3_2, intermediate_size=3_7, layer_norm_eps=1e-0_5, num_attention_heads=4, num_hidden_layers=5, vocab_size=5_0_0_2, ) # TODO: remove after fixing the non-deterministic text encoder _UpperCamelCase : Tuple = RobertaSeriesModelWithTransformation(lowerCAmelCase__ ) _UpperCamelCase : int = text_encoder _UpperCamelCase : str = AltDiffusionPipeline(**lowerCAmelCase__ ) _UpperCamelCase : List[str] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) _UpperCamelCase : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase__ ) _UpperCamelCase : Optional[int] = alt_pipe(**lowerCAmelCase__ ) _UpperCamelCase : List[str] = output.images _UpperCamelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) _UpperCamelCase : List[Any] = np.array( [0.51_605_093, 0.5_707_241, 0.47_365_507, 0.50_578_886, 0.5_633_877, 0.4_642_503, 0.5_182_081, 0.48_763_484, 0.49_084_237] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class _a ( unittest.TestCase ): def snake_case ( self : List[str] ) -> Any: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case ( self : Union[str, Any] ) -> Any: '''simple docstring''' _UpperCamelCase : str = AltDiffusionPipeline.from_pretrained('''BAAI/AltDiffusion''', safety_checker=lowerCAmelCase__ ) _UpperCamelCase : List[Any] = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) _UpperCamelCase : int = '''A painting of a squirrel eating a burger''' _UpperCamelCase : int = torch.manual_seed(0 ) _UpperCamelCase : Dict = alt_pipe([prompt], generator=lowerCAmelCase__, guidance_scale=6.0, num_inference_steps=2_0, output_type='''np''' ) _UpperCamelCase : Optional[int] = output.images _UpperCamelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) _UpperCamelCase : List[str] = np.array([0.1_010, 0.0_800, 0.0_794, 0.0_885, 0.0_843, 0.0_762, 0.0_769, 0.0_729, 0.0_586] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def snake_case ( self : str ) -> str: '''simple docstring''' _UpperCamelCase : Any = DDIMScheduler.from_pretrained('''BAAI/AltDiffusion''', subfolder='''scheduler''' ) _UpperCamelCase : Dict = AltDiffusionPipeline.from_pretrained('''BAAI/AltDiffusion''', scheduler=lowerCAmelCase__, safety_checker=lowerCAmelCase__ ) _UpperCamelCase : Dict = alt_pipe.to(lowerCAmelCase__ ) alt_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) _UpperCamelCase : Optional[Any] = '''A painting of a squirrel eating a burger''' _UpperCamelCase : Optional[Any] = torch.manual_seed(0 ) _UpperCamelCase : Union[str, Any] = alt_pipe([prompt], generator=lowerCAmelCase__, num_inference_steps=2, output_type='''numpy''' ) _UpperCamelCase : Tuple = output.images _UpperCamelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) _UpperCamelCase : str = np.array([0.4_019, 0.4_052, 0.3_810, 0.4_119, 0.3_916, 0.3_982, 0.4_651, 0.4_195, 0.5_323] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
352
"""simple docstring""" import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import FEATURE_EXTRACTOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import ChineseCLIPImageProcessor, ChineseCLIPProcessor @require_vision class _a ( unittest.TestCase ): def snake_case ( self : Tuple ) -> Dict: '''simple docstring''' _UpperCamelCase : int = tempfile.mkdtemp() _UpperCamelCase : List[str] = [ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''的''', '''价''', '''格''', '''是''', '''15''', '''便''', '''alex''', '''##andra''', ''',''', '''。''', '''-''', '''t''', '''shirt''', ] _UpperCamelCase : Optional[Any] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file, '''w''', encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) _UpperCamelCase : Dict = { '''do_resize''': True, '''size''': {'''height''': 2_2_4, '''width''': 2_2_4}, '''do_center_crop''': True, '''crop_size''': {'''height''': 1_8, '''width''': 1_8}, '''do_normalize''': True, '''image_mean''': [0.48_145_466, 0.4_578_275, 0.40_821_073], '''image_std''': [0.26_862_954, 0.26_130_258, 0.27_577_711], '''do_convert_rgb''': True, } _UpperCamelCase : Optional[Any] = os.path.join(self.tmpdirname, lowerCAmelCase__ ) with open(self.image_processor_file, '''w''', encoding='''utf-8''' ) as fp: json.dump(lowerCAmelCase__, lowerCAmelCase__ ) def snake_case ( self : str, **lowerCAmelCase__ : List[Any] ) -> Optional[int]: '''simple docstring''' return BertTokenizer.from_pretrained(self.tmpdirname, **lowerCAmelCase__ ) def snake_case ( self : Union[str, Any], **lowerCAmelCase__ : Tuple ) -> str: '''simple docstring''' return BertTokenizerFast.from_pretrained(self.tmpdirname, **lowerCAmelCase__ ) def snake_case ( self : Any, **lowerCAmelCase__ : Optional[int] ) -> Optional[Any]: '''simple docstring''' return ChineseCLIPImageProcessor.from_pretrained(self.tmpdirname, **lowerCAmelCase__ ) def snake_case ( self : str ) -> Optional[int]: '''simple docstring''' shutil.rmtree(self.tmpdirname ) def snake_case ( self : Any ) -> int: '''simple docstring''' _UpperCamelCase : List[str] = [np.random.randint(2_5_5, size=(3, 3_0, 4_0_0), dtype=np.uinta )] _UpperCamelCase : List[Any] = [Image.fromarray(np.moveaxis(lowerCAmelCase__, 0, -1 ) ) for x in image_inputs] return image_inputs def snake_case ( self : str ) -> Any: '''simple docstring''' _UpperCamelCase : Any = self.get_tokenizer() _UpperCamelCase : int = self.get_rust_tokenizer() _UpperCamelCase : int = self.get_image_processor() _UpperCamelCase : Tuple = ChineseCLIPProcessor(tokenizer=lowerCAmelCase__, image_processor=lowerCAmelCase__ ) processor_slow.save_pretrained(self.tmpdirname ) _UpperCamelCase : List[Any] = ChineseCLIPProcessor.from_pretrained(self.tmpdirname, use_fast=lowerCAmelCase__ ) _UpperCamelCase : List[Any] = ChineseCLIPProcessor(tokenizer=lowerCAmelCase__, image_processor=lowerCAmelCase__ ) processor_fast.save_pretrained(self.tmpdirname ) _UpperCamelCase : List[Any] = ChineseCLIPProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab(), tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab(), tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab(), tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer, lowerCAmelCase__ ) self.assertIsInstance(processor_fast.tokenizer, lowerCAmelCase__ ) self.assertEqual(processor_slow.image_processor.to_json_string(), image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string(), image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor, lowerCAmelCase__ ) self.assertIsInstance(processor_fast.image_processor, lowerCAmelCase__ ) def snake_case ( self : int ) -> Tuple: '''simple docstring''' _UpperCamelCase : List[Any] = ChineseCLIPProcessor(tokenizer=self.get_tokenizer(), image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) _UpperCamelCase : Dict = self.get_tokenizer(cls_token='''(CLS)''', sep_token='''(SEP)''' ) _UpperCamelCase : List[str] = self.get_image_processor(do_normalize=lowerCAmelCase__ ) _UpperCamelCase : Optional[Any] = ChineseCLIPProcessor.from_pretrained( self.tmpdirname, cls_token='''(CLS)''', sep_token='''(SEP)''', do_normalize=lowerCAmelCase__ ) self.assertEqual(processor.tokenizer.get_vocab(), tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer, lowerCAmelCase__ ) self.assertEqual(processor.image_processor.to_json_string(), image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor, lowerCAmelCase__ ) def snake_case ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase : List[str] = self.get_image_processor() _UpperCamelCase : str = self.get_tokenizer() _UpperCamelCase : Optional[Any] = ChineseCLIPProcessor(tokenizer=lowerCAmelCase__, image_processor=lowerCAmelCase__ ) _UpperCamelCase : List[str] = self.prepare_image_inputs() _UpperCamelCase : Any = image_processor(lowerCAmelCase__, return_tensors='''np''' ) _UpperCamelCase : Any = processor(images=lowerCAmelCase__, return_tensors='''np''' ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum(), input_processor[key].sum(), delta=1e-2 ) def snake_case ( self : Optional[Any] ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase : Tuple = self.get_image_processor() _UpperCamelCase : Optional[Any] = self.get_tokenizer() _UpperCamelCase : Any = ChineseCLIPProcessor(tokenizer=lowerCAmelCase__, image_processor=lowerCAmelCase__ ) _UpperCamelCase : Tuple = '''Alexandra,T-shirt的价格是15便士。''' _UpperCamelCase : List[str] = processor(text=lowerCAmelCase__ ) _UpperCamelCase : Any = tokenizer(lowerCAmelCase__ ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key], encoded_processor[key] ) def snake_case ( self : Dict ) -> Tuple: '''simple docstring''' _UpperCamelCase : Tuple = self.get_image_processor() _UpperCamelCase : Optional[Any] = self.get_tokenizer() _UpperCamelCase : Dict = ChineseCLIPProcessor(tokenizer=lowerCAmelCase__, image_processor=lowerCAmelCase__ ) _UpperCamelCase : Any = '''Alexandra,T-shirt的价格是15便士。''' _UpperCamelCase : Union[str, Any] = self.prepare_image_inputs() _UpperCamelCase : str = processor(text=lowerCAmelCase__, images=lowerCAmelCase__ ) self.assertListEqual(list(inputs.keys() ), ['''input_ids''', '''token_type_ids''', '''attention_mask''', '''pixel_values'''] ) # test if it raises when no input is passed with pytest.raises(lowerCAmelCase__ ): processor() def snake_case ( self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase : int = self.get_image_processor() _UpperCamelCase : int = self.get_tokenizer() _UpperCamelCase : Optional[Any] = ChineseCLIPProcessor(tokenizer=lowerCAmelCase__, image_processor=lowerCAmelCase__ ) _UpperCamelCase : List[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _UpperCamelCase : List[Any] = processor.batch_decode(lowerCAmelCase__ ) _UpperCamelCase : Dict = tokenizer.batch_decode(lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__, lowerCAmelCase__ ) def snake_case ( self : Union[str, Any] ) -> Dict: '''simple docstring''' _UpperCamelCase : Any = self.get_image_processor() _UpperCamelCase : Optional[int] = self.get_tokenizer() _UpperCamelCase : Optional[Any] = ChineseCLIPProcessor(tokenizer=lowerCAmelCase__, image_processor=lowerCAmelCase__ ) _UpperCamelCase : Any = '''Alexandra,T-shirt的价格是15便士。''' _UpperCamelCase : int = self.prepare_image_inputs() _UpperCamelCase : Dict = processor(text=lowerCAmelCase__, images=lowerCAmelCase__ ) self.assertListEqual(list(inputs.keys() ), processor.model_input_names )
128
0
'''simple docstring''' import pytest import datasets # Import fixture modules as plugins lowerCamelCase = ["""tests.fixtures.files""", """tests.fixtures.hub""", """tests.fixtures.fsspec"""] def _A ( _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" for item in items: if any(marker in item.keywords for marker in ['integration', 'unit'] ): continue item.add_marker(pytest.mark.unit ) def _A ( _lowerCAmelCase ): """simple docstring""" config.addinivalue_line('markers' , 'torchaudio_latest: mark test to run with torchaudio>=0.12' ) @pytest.fixture(autouse=_UpperCAmelCase ) def _A ( _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =tmp_path_factory.getbasetemp() / "cache" __lowercase =test_hf_cache_home / "datasets" __lowercase =test_hf_cache_home / "metrics" __lowercase =test_hf_cache_home / "modules" monkeypatch.setattr('datasets.config.HF_DATASETS_CACHE' , str(_UpperCAmelCase ) ) monkeypatch.setattr('datasets.config.HF_METRICS_CACHE' , str(_UpperCAmelCase ) ) monkeypatch.setattr('datasets.config.HF_MODULES_CACHE' , str(_UpperCAmelCase ) ) __lowercase =test_hf_datasets_cache / "downloads" monkeypatch.setattr('datasets.config.DOWNLOADED_DATASETS_PATH' , str(_UpperCAmelCase ) ) __lowercase =test_hf_datasets_cache / "downloads" / "extracted" monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_PATH' , str(_UpperCAmelCase ) ) @pytest.fixture(autouse=_UpperCAmelCase , scope='session' ) def _A ( ): """simple docstring""" datasets.disable_progress_bar() @pytest.fixture(autouse=_UpperCAmelCase ) def _A ( _lowerCAmelCase ): """simple docstring""" monkeypatch.setattr('datasets.config.HF_UPDATE_DOWNLOAD_COUNTS' , _UpperCAmelCase ) @pytest.fixture def _A ( _lowerCAmelCase ): """simple docstring""" monkeypatch.setattr('sqlalchemy.util.deprecations.SILENCE_UBER_WARNING' , _UpperCAmelCase )
166
def A_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ): SCREAMING_SNAKE_CASE_: Optional[int] = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError("All input parameters must be positive" ) if any(p > 1 for p in parameters[1:4] ): raise ValueError("Relative densities cannot be greater than one" ) else: SCREAMING_SNAKE_CASE_: int = 1 - (matter_density + radiation_density + dark_energy) SCREAMING_SNAKE_CASE_: Dict = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) SCREAMING_SNAKE_CASE_: Any = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation lowerCAmelCase : List[Any] = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1E-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
13
0
import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): __lowerCamelCase = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right __lowerCamelCase = 12_80_22 __lowerCamelCase = 12_80_28 @require_sentencepiece class UpperCAmelCase ( A_ ,unittest.TestCase ): A__ : Tuple = MaMaaaTokenizer A__ : Optional[Any] = False A__ : Optional[int] = False A__ : Union[str, Any] = True def _SCREAMING_SNAKE_CASE (self : List[str] ) -> List[str]: '''simple docstring''' super().setUp() snake_case : int = ["</s>", "<unk>", "▁This", "▁is", "▁a", "▁t", "est", "\u0120", "<pad>"] snake_case : List[Any] = dict(zip(snake_case__ , range(len(snake_case__ ) ) ) ) snake_case : int = Path(self.tmpdirname ) save_json(snake_case__ , save_dir / VOCAB_FILES_NAMES["vocab_file"] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(snake_case__ , save_dir / VOCAB_FILES_NAMES["spm_file"] ) snake_case : str = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def _SCREAMING_SNAKE_CASE (self : int , **snake_case__ : int ) -> Tuple: '''simple docstring''' return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **snake_case__ ) def _SCREAMING_SNAKE_CASE (self : Any , snake_case__ : Tuple ) -> Optional[int]: '''simple docstring''' return ( "This is a test", "This is a test", ) def _SCREAMING_SNAKE_CASE (self : Optional[Any] ) -> int: '''simple docstring''' snake_case : Tuple = "</s>" snake_case : Optional[Any] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case__ ) , snake_case__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case__ ) , snake_case__ ) def _SCREAMING_SNAKE_CASE (self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' snake_case : List[Any] = self.get_tokenizer() snake_case : Tuple = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "</s>" ) self.assertEqual(vocab_keys[1] , "<unk>" ) self.assertEqual(vocab_keys[-1] , "<s>" ) self.assertEqual(len(snake_case__ ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip("Skip this test while all models are still to be uploaded." ) def _SCREAMING_SNAKE_CASE (self : str ) -> Optional[Any]: '''simple docstring''' pass def _SCREAMING_SNAKE_CASE (self : Optional[int] ) -> int: '''simple docstring''' snake_case : str = self.get_tokenizer() snake_case : Union[str, Any] = tokenizer.tokenize("This is a test" ) self.assertListEqual(snake_case__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(snake_case__ ) , [2, 3, 4, 5, 6] , ) snake_case : Union[str, Any] = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(snake_case__ , ["▁This", "▁is", "▁a", "▁t", "est"] ) snake_case : int = tokenizer.convert_tokens_to_string(snake_case__ ) self.assertEqual(snake_case__ , "This is a test" ) @slow def _SCREAMING_SNAKE_CASE (self : Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case : Dict = {"input_ids": [[12_80_22, 11_01_08, 3_97, 11, 3_82_72, 22_47, 12_48_11, 2_85, 1_81_05, 15_86, 2_07, 7, 3_95_34, 44_28, 3_97, 10_19, 1_81_05, 15_86, 2_07, 7, 4_13_37, 1_67_86, 2_41, 7, 2_02_14, 17, 12_56_90, 1_03_98, 7, 4_43_78, 5_80_69, 6_83_42, 77_98, 73_43, 11, 2_99, 3_33_10, 4, 1_58, 3_73_50, 9_40_77, 45_69, 2_99, 3_33_10, 90, 4, 5_28_40, 2_90, 4, 3_12_70, 1_12, 2_99, 6_82, 4, 5_28_40, 3_99_53, 1_40_79, 1_93, 5_25_19, 9_08_94, 1_78_94, 12_06_97, 11, 4_04_45, 5_51, 17, 10_19, 5_25_19, 9_08_94, 1_77_56, 9_63, 11, 4_04_45, 4_80, 17, 97_92, 11_20, 51_73, 13_93, 62_40, 1_67_86, 2_41, 12_09_96, 28, 12_45, 13_93, 11_82_40, 1_11_23, 10_19, 9_36_12, 26_91, 1_06_18, 9_80_58, 12_04_09, 19_28, 2_79, 4, 4_06_83, 3_67, 1_78, 2_07, 10_19, 1_03, 10_31_21, 5_06, 6_52_96, 5, 2], [12_80_22, 2_12_17, 3_67, 1_17, 12_54_50, 1_28, 7_19, 7, 73_08, 40, 9_36_12, 1_26_69, 11_16, 1_67_04, 71, 1_77_85, 36_99, 1_55_92, 35, 1_44, 95_84, 2_41, 1_19_43, 7_13, 9_50, 7_99, 22_47, 8_84_27, 1_50, 1_49, 11_88_13, 12_07_06, 10_19, 10_69_06, 8_15_18, 28, 12_24, 2_27_99, 3_97, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [12_80_22, 16_58, 12_33_11, 51_55, 55_78, 47_22, 2_79, 1_49_47, 23_66, 11_20, 11_97, 14, 13_48, 92_32, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=snake_case__ , model_name="facebook/m2m100_418M" , revision="c168bae485c864188cf9aa0e4108b0b6934dc91e" , ) @require_torch @require_sentencepiece @require_tokenizers class UpperCAmelCase ( unittest.TestCase ): A__ : int = "facebook/m2m100_418M" A__ : int = [ "In my opinion, there are two levels of response from the French government.", "NSA Affair Emphasizes Complete Lack of Debate on Intelligence", ] A__ : Union[str, Any] = [ "Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.", "L'affaire NSA souligne l'absence totale de débat sur le renseignement", ] # fmt: off A__ : Optional[Any] = [EN_CODE, 5_93, 19_49, 11_57_81, 4, 7_15_86, 42_34, 6_06_33, 12_62_33, 4_32, 12_38_08, 1_55_92, 11_97, 11_71_32, 12_06_18, 5, 2] @classmethod def _SCREAMING_SNAKE_CASE (cls : List[Any] ) -> Union[str, Any]: '''simple docstring''' snake_case : MaMaaaTokenizer = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en" , tgt_lang="fr" ) snake_case : Dict = 1 return cls def _SCREAMING_SNAKE_CASE (self : Optional[Any] ) -> str: '''simple docstring''' self.assertEqual(self.tokenizer.get_lang_id("ar" ) , 12_80_06 ) self.assertEqual(self.tokenizer.get_lang_id("en" ) , 12_80_22 ) self.assertEqual(self.tokenizer.get_lang_id("ro" ) , 12_80_76 ) self.assertEqual(self.tokenizer.get_lang_id("mr" ) , 12_80_63 ) def _SCREAMING_SNAKE_CASE (self : List[Any] ) -> Dict: '''simple docstring''' snake_case : int = self.tokenizer.get_vocab() self.assertEqual(len(snake_case__ ) , self.tokenizer.vocab_size ) self.assertEqual(vocab["<unk>"] , 3 ) self.assertIn(self.tokenizer.get_lang_token("en" ) , snake_case__ ) def _SCREAMING_SNAKE_CASE (self : List[str] ) -> Dict: '''simple docstring''' snake_case : Union[str, Any] = "en" snake_case : str = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , snake_case__ ) def _SCREAMING_SNAKE_CASE (self : Any ) -> Tuple: '''simple docstring''' self.assertIn(snake_case__ , self.tokenizer.all_special_ids ) # fmt: off snake_case : List[Any] = [FR_CODE, 53_64, 82, 86_42, 4, 2_94, 47, 8, 1_40_28, 1_36, 32_86, 97_06, 6, 9_07_97, 6, 14_40_12, 1_62, 8_81_28, 3_00_61, 5, 2] # fmt: on snake_case : Dict = self.tokenizer.decode(snake_case__ , skip_special_tokens=snake_case__ ) snake_case : List[Any] = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=snake_case__ ) self.assertEqual(snake_case__ , snake_case__ ) self.assertNotIn(self.tokenizer.eos_token , snake_case__ ) def _SCREAMING_SNAKE_CASE (self : Union[str, Any] ) -> List[str]: '''simple docstring''' snake_case : str = tempfile.mkdtemp() snake_case : Any = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(snake_case__ ) snake_case : Union[str, Any] = MaMaaaTokenizer.from_pretrained(snake_case__ ) self.assertDictEqual(new_tok.lang_token_to_id , snake_case__ ) @require_torch def _SCREAMING_SNAKE_CASE (self : int ) -> str: '''simple docstring''' snake_case : Any = "en" snake_case : Union[str, Any] = "fr" snake_case : Optional[int] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=snake_case__ , return_tensors="pt" ) snake_case : List[str] = shift_tokens_right( batch["labels"] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: snake_case : Dict = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def _SCREAMING_SNAKE_CASE (self : str ) -> Any: '''simple docstring''' snake_case : Dict = "mr" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) snake_case : Any = "zh" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def _SCREAMING_SNAKE_CASE (self : Tuple ) -> Dict: '''simple docstring''' snake_case : Optional[int] = "mr" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) snake_case : Any = "zh" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def _SCREAMING_SNAKE_CASE (self : int ) -> List[str]: '''simple docstring''' snake_case : int = self.tokenizer._build_translation_inputs("A test" , return_tensors="pt" , src_lang="en" , tgt_lang="ar" ) self.assertEqual( nested_simplify(snake_case__ ) , { # en_XX, A, test, EOS "input_ids": [[12_80_22, 58, 41_83, 2]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 12_80_06, } , )
10
from __future__ import annotations __lowerCamelCase = { """A""": ["""B""", """C""", """E"""], """B""": ["""A""", """D""", """E"""], """C""": ["""A""", """F""", """G"""], """D""": ["""B"""], """E""": ["""A""", """B""", """D"""], """F""": ["""C"""], """G""": ["""C"""], } class UpperCAmelCase : def __init__(self : Tuple , snake_case__ : dict[str, list[str]] , snake_case__ : str ) -> None: '''simple docstring''' snake_case : str = graph # mapping node to its parent in resulting breadth first tree snake_case : dict[str, str | None] = {} snake_case : Union[str, Any] = source_vertex def _SCREAMING_SNAKE_CASE (self : Union[str, Any] ) -> None: '''simple docstring''' snake_case : Any = {self.source_vertex} snake_case : str = None snake_case : List[str] = [self.source_vertex] # first in first out queue while queue: snake_case : List[Any] = queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(snake_case__ ) snake_case : Any = vertex queue.append(snake_case__ ) def _SCREAMING_SNAKE_CASE (self : str , snake_case__ : str ) -> str: '''simple docstring''' if target_vertex == self.source_vertex: return self.source_vertex snake_case : str = self.parent.get(snake_case__ ) if target_vertex_parent is None: snake_case : Optional[Any] = ( f"""No path from vertex: {self.source_vertex} to vertex: {target_vertex}""" ) raise ValueError(snake_case__ ) return self.shortest_path(snake_case__ ) + f"""->{target_vertex}""" if __name__ == "__main__": __lowerCamelCase = Graph(graph, """G""") g.breath_first_search() print(g.shortest_path("""D""")) print(g.shortest_path("""G""")) print(g.shortest_path("""Foo"""))
10
1
def __lowerCamelCase ( UpperCamelCase__ ): '''simple docstring''' return 10 - x * x def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' if equation(UpperCamelCase__ ) * equation(UpperCamelCase__ ) >= 0: raise ValueError('Wrong space!' ) snake_case_ = a while (b - a) >= 0.01: # Find middle point snake_case_ = (a + b) / 2 # Check if middle point is root if equation(UpperCamelCase__ ) == 0.0: break # Decide the side to repeat the steps if equation(UpperCamelCase__ ) * equation(UpperCamelCase__ ) < 0: snake_case_ = c else: snake_case_ = c return c if __name__ == "__main__": import doctest doctest.testmod() print(bisection(-2, 5)) print(bisection(0, 6))
285
def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' global f # a global dp table for knapsack if f[i][j] < 0: if j < wt[i - 1]: snake_case_ = mf_knapsack(i - 1 , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else: snake_case_ = max( mf_knapsack(i - 1 , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) , mf_knapsack(i - 1 , UpperCamelCase__ , UpperCamelCase__ , j - wt[i - 1] ) + val[i - 1] , ) snake_case_ = val return f[i][j] def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' snake_case_ = [[0] * (w + 1) for _ in range(n + 1 )] for i in range(1 , n + 1 ): for w_ in range(1 , w + 1 ): if wt[i - 1] <= w_: snake_case_ = max(val[i - 1] + dp[i - 1][w_ - wt[i - 1]] , dp[i - 1][w_] ) else: snake_case_ = dp[i - 1][w_] return dp[n][w_], dp def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' if not (isinstance(UpperCamelCase__ , (list, tuple) ) and isinstance(UpperCamelCase__ , (list, tuple) )): raise ValueError( 'Both the weights and values vectors must be either lists or tuples' ) snake_case_ = len(UpperCamelCase__ ) if num_items != len(UpperCamelCase__ ): snake_case_ = ( 'The number of weights must be the same as the number of values.\n' F'''But got {num_items} weights and {len(UpperCamelCase__ )} values''' ) raise ValueError(UpperCamelCase__ ) for i in range(UpperCamelCase__ ): if not isinstance(wt[i] , UpperCamelCase__ ): snake_case_ = ( 'All weights must be integers but got weight of ' F'''type {type(wt[i] )} at index {i}''' ) raise TypeError(UpperCamelCase__ ) snake_case_ , snake_case_ = knapsack(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) snake_case_ = set() _construct_solution(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) return optimal_val, example_optional_set def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' if i > 0 and j > 0: if dp[i - 1][j] == dp[i][j]: _construct_solution(UpperCamelCase__ , UpperCamelCase__ , i - 1 , UpperCamelCase__ , UpperCamelCase__ ) else: optimal_set.add(UpperCamelCase__ ) _construct_solution(UpperCamelCase__ , UpperCamelCase__ , i - 1 , j - wt[i - 1] , UpperCamelCase__ ) if __name__ == "__main__": _UpperCAmelCase : Tuple = [3, 2, 4, 4] _UpperCAmelCase : Optional[Any] = [4, 3, 2, 3] _UpperCAmelCase : List[str] = 4 _UpperCAmelCase : str = 6 _UpperCAmelCase : Tuple = [[0] * (w + 1)] + [[0] + [-1] * (w + 1) for _ in range(n + 1)] _UpperCAmelCase , _UpperCAmelCase : List[Any] = knapsack(w, wt, val, n) print(optimal_solution) print(mf_knapsack(n, wt, val, w)) # switched the n and w # testing the dynamic programming problem with example # the optimal subset for the above example are items 3 and 4 _UpperCAmelCase , _UpperCAmelCase : Any = knapsack_with_example_solution(w, wt, val) assert optimal_solution == 8 assert optimal_subset == {3, 4} print("""optimal_value = """, optimal_solution) print("""An optimal subset corresponding to the optimal value""", optimal_subset)
285
1
def UpperCamelCase( lowercase_ ) -> Dict: '''simple docstring''' stooge(lowercase_ , 0 , len(lowercase_ ) - 1 ) return arr def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ ) -> Any: '''simple docstring''' if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: snake_case_ , snake_case_ = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: snake_case_ = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(lowercase_ , lowercase_ , (h - t) ) # Recursively sort last 2/3 elements stooge(lowercase_ , i + t , (lowercase_) ) # Recursively sort first 2/3 elements stooge(lowercase_ , lowercase_ , (h - t) ) if __name__ == "__main__": lowerCamelCase_ = input('''Enter numbers separated by a comma:\n''').strip() lowerCamelCase_ = [int(item) for item in user_input.split(''',''')] print(stooge_sort(unsorted))
353
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { '''facebook/levit-128S''': '''https://huggingface.co/facebook/levit-128S/resolve/main/config.json''', # See all LeViT models at https://huggingface.co/models?filter=levit } class __lowerCamelCase ( __snake_case ): lowerCamelCase_ : Tuple = 'levit' def __init__( self , lowerCamelCase=224 , lowerCamelCase=3 , lowerCamelCase=3 , lowerCamelCase=2 , lowerCamelCase=1 , lowerCamelCase=16 , lowerCamelCase=[128, 256, 384] , lowerCamelCase=[4, 8, 12] , lowerCamelCase=[4, 4, 4] , lowerCamelCase=[16, 16, 16] , lowerCamelCase=0 , lowerCamelCase=[2, 2, 2] , lowerCamelCase=[2, 2, 2] , lowerCamelCase=0.02 , **lowerCamelCase , ) -> Tuple: super().__init__(**lowerCamelCase ) snake_case_ = image_size snake_case_ = num_channels snake_case_ = kernel_size snake_case_ = stride snake_case_ = padding snake_case_ = hidden_sizes snake_case_ = num_attention_heads snake_case_ = depths snake_case_ = key_dim snake_case_ = drop_path_rate snake_case_ = patch_size snake_case_ = attention_ratio snake_case_ = mlp_ratio snake_case_ = initializer_range snake_case_ = [ ["""Subsample""", key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ["""Subsample""", key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] class __lowerCamelCase ( __snake_case ): lowerCamelCase_ : Any = version.parse('1.11' ) @property def lowerCAmelCase_ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def lowerCAmelCase_ ( self ) -> float: return 1e-4
34
0
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase , unittest.TestCase ): __lowerCAmelCase = BertJapaneseTokenizer __lowerCAmelCase = False __lowerCAmelCase = True def lowerCamelCase_ ( self : str ): """simple docstring""" super().setUp() UpperCamelCase = [ """[UNK]""", """[CLS]""", """[SEP]""", """こんにちは""", """こん""", """にちは""", """ばんは""", """##こん""", """##にちは""", """##ばんは""", """世界""", """##世界""", """、""", """##、""", """。""", """##。""", ] UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def lowerCamelCase_ ( self : Union[str, Any] , lowerCamelCase_ : List[str] ): """simple docstring""" UpperCamelCase = """こんにちは、世界。 \nこんばんは、世界。""" UpperCamelCase = """こんにちは 、 世界 。 こんばんは 、 世界 。""" return input_text, output_text def lowerCamelCase_ ( self : Union[str, Any] , lowerCamelCase_ : Any ): """simple docstring""" UpperCamelCase , UpperCamelCase = self.get_input_output_texts(lowerCamelCase_ ) UpperCamelCase = tokenizer.encode(lowerCamelCase_ , add_special_tokens=lowerCamelCase_ ) UpperCamelCase = tokenizer.decode(lowerCamelCase_ , clean_up_tokenization_spaces=lowerCamelCase_ ) return text, ids def lowerCamelCase_ ( self : str ): """simple docstring""" pass # TODO add if relevant def lowerCamelCase_ ( self : Dict ): """simple docstring""" pass # TODO add if relevant def lowerCamelCase_ ( self : Tuple ): """simple docstring""" pass # TODO add if relevant def lowerCamelCase_ ( self : Tuple ): """simple docstring""" UpperCamelCase = self.tokenizer_class(self.vocab_file ) UpperCamelCase = tokenizer.tokenize("""こんにちは、世界。\nこんばんは、世界。""" ) self.assertListEqual(lowerCamelCase_ , ["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) def lowerCamelCase_ ( self : Tuple ): """simple docstring""" UpperCamelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type="""mecab""" ) self.assertIsNotNone(lowerCamelCase_ ) UpperCamelCase = """こんにちは、世界。\nこんばんは、世界。""" UpperCamelCase = tokenizer.tokenize(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , ["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) UpperCamelCase = os.path.join(self.tmpdirname , """tokenizer.bin""" ) with open(lowerCamelCase_ , """wb""" ) as handle: pickle.dump(lowerCamelCase_ , lowerCamelCase_ ) with open(lowerCamelCase_ , """rb""" ) as handle: UpperCamelCase = pickle.load(lowerCamelCase_ ) UpperCamelCase = tokenizer_new.tokenize(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) def lowerCamelCase_ ( self : int ): """simple docstring""" UpperCamelCase = MecabTokenizer(mecab_dic="""ipadic""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップルストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" try: UpperCamelCase = MecabTokenizer(mecab_dic="""unidic_lite""" ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" try: UpperCamelCase = MecabTokenizer(mecab_dic="""unidic""" ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) def lowerCamelCase_ ( self : int ): """simple docstring""" UpperCamelCase = MecabTokenizer(do_lower_case=lowerCamelCase_ , mecab_dic="""ipadic""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップルストア""", """で""", """iphone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) def lowerCamelCase_ ( self : Any ): """simple docstring""" try: UpperCamelCase = MecabTokenizer( do_lower_case=lowerCamelCase_ , normalize_text=lowerCamelCase_ , mecab_option="""-d /usr/local/lib/mecab/dic/jumandic""" ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップルストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れた""", """\u3000""", """。"""] , ) def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" UpperCamelCase = MecabTokenizer(normalize_text=lowerCamelCase_ , mecab_dic="""ipadic""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップルストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """ """, """。"""] , ) @require_sudachi def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" UpperCamelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type="""sudachi""" ) self.assertIsNotNone(lowerCamelCase_ ) UpperCamelCase = """こんにちは、世界。\nこんばんは、世界。""" UpperCamelCase = tokenizer.tokenize(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , ["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) UpperCamelCase = os.path.join(self.tmpdirname , """tokenizer.bin""" ) with open(lowerCamelCase_ , """wb""" ) as handle: pickle.dump(lowerCamelCase_ , lowerCamelCase_ ) with open(lowerCamelCase_ , """rb""" ) as handle: UpperCamelCase = pickle.load(lowerCamelCase_ ) UpperCamelCase = tokenizer_new.tokenize(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) @require_sudachi def lowerCamelCase_ ( self : Any ): """simple docstring""" UpperCamelCase = SudachiTokenizer(sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , [""" """, """\t""", """アップル""", """ストア""", """で""", """iPhone""", """8""", """ """, """が""", """ """, """ """, """\n """, """発売""", """さ""", """れ""", """た""", """ """, """。""", """ """, """ """] , ) @require_sudachi def lowerCamelCase_ ( self : Tuple ): """simple docstring""" UpperCamelCase = SudachiTokenizer(sudachi_dict_type="""core""" , sudachi_split_mode="""A""" ) self.assertListEqual(tokenizer.tokenize("""外国人参政権""" ) , ["""外国""", """人""", """参政""", """権"""] ) @require_sudachi def lowerCamelCase_ ( self : str ): """simple docstring""" UpperCamelCase = SudachiTokenizer(sudachi_dict_type="""core""" , sudachi_split_mode="""B""" ) self.assertListEqual(tokenizer.tokenize("""外国人参政権""" ) , ["""外国人""", """参政権"""] ) @require_sudachi def lowerCamelCase_ ( self : int ): """simple docstring""" UpperCamelCase = SudachiTokenizer(sudachi_dict_type="""core""" , sudachi_split_mode="""C""" ) self.assertListEqual(tokenizer.tokenize("""外国人参政権""" ) , ["""外国人参政権"""] ) @require_sudachi def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" UpperCamelCase = SudachiTokenizer(do_lower_case=lowerCamelCase_ , sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , [""" """, """\t""", """アップル""", """ストア""", """で""", """iphone""", """8""", """ """, """が""", """ """, """ """, """\n """, """発売""", """さ""", """れ""", """た""", """ """, """。""", """ """, """ """] , ) @require_sudachi def lowerCamelCase_ ( self : List[str] ): """simple docstring""" UpperCamelCase = SudachiTokenizer(normalize_text=lowerCamelCase_ , sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , [""" """, """\t""", """アップル""", """ストア""", """で""", """iPhone""", """8""", """ """, """が""", """ """, """ """, """\n """, """発売""", """さ""", """れ""", """た""", """\u3000""", """。""", """ """, """ """] , ) @require_sudachi def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" UpperCamelCase = SudachiTokenizer(trim_whitespace=lowerCamelCase_ , sudachi_dict_type="""core""" ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れ""", """た""", """。"""] , ) @require_jumanpp def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" UpperCamelCase = self.tokenizer_class(self.vocab_file , word_tokenizer_type="""jumanpp""" ) self.assertIsNotNone(lowerCamelCase_ ) UpperCamelCase = """こんにちは、世界。\nこんばんは、世界。""" UpperCamelCase = tokenizer.tokenize(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , ["""こんにちは""", """、""", """世界""", """。""", """こん""", """##ばんは""", """、""", """世界""", """。"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase_ ) , [3, 12, 10, 14, 4, 9, 12, 10, 14] ) UpperCamelCase = os.path.join(self.tmpdirname , """tokenizer.bin""" ) with open(lowerCamelCase_ , """wb""" ) as handle: pickle.dump(lowerCamelCase_ , lowerCamelCase_ ) with open(lowerCamelCase_ , """rb""" ) as handle: UpperCamelCase = pickle.load(lowerCamelCase_ ) UpperCamelCase = tokenizer_new.tokenize(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) @require_jumanpp def lowerCamelCase_ ( self : Dict ): """simple docstring""" UpperCamelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """\u3000""", """が""", """\u3000""", """\u3000""", """\u3000""", """発売""", """さ""", """れた""", """\u3000""", """。"""] , ) @require_jumanpp def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" UpperCamelCase = JumanppTokenizer(do_lower_case=lowerCamelCase_ ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iphone""", """8""", """\u3000""", """が""", """\u3000""", """\u3000""", """\u3000""", """発売""", """さ""", """れた""", """\u3000""", """。"""] , ) @require_jumanpp def lowerCamelCase_ ( self : Union[str, Any] ): """simple docstring""" UpperCamelCase = JumanppTokenizer(normalize_text=lowerCamelCase_ ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""ア""", """ッ""", """フ""", """゚""", """ル""", """ストア""", """で""", """iPhone""", """8""", """\u3000""", """が""", """\u3000""", """\u3000""", """\u3000""", """発売""", """さ""", """れた""", """\u3000""", """。"""] , ) @require_jumanpp def lowerCamelCase_ ( self : Dict ): """simple docstring""" UpperCamelCase = JumanppTokenizer(trim_whitespace=lowerCamelCase_ ) self.assertListEqual( tokenizer.tokenize(""" \tアップルストアでiPhone8 が \n 発売された 。 """ ) , ["""アップル""", """ストア""", """で""", """iPhone""", """8""", """が""", """発売""", """さ""", """れた""", """。"""] , ) @require_jumanpp def lowerCamelCase_ ( self : Union[str, Any] ): """simple docstring""" UpperCamelCase = JumanppTokenizer() self.assertListEqual( tokenizer.tokenize("""ありがとうございますm(_ _)m見つけるのが大変です。""" ) , ["""ありがとう""", """ございます""", """m(_ _)m""", """見つける""", """の""", """が""", """大変です""", """。"""] , ) def lowerCamelCase_ ( self : str ): """simple docstring""" UpperCamelCase = ["""[UNK]""", """[CLS]""", """[SEP]""", """こんにちは""", """こん""", """にちは""", """ばんは""", """##こん""", """##にちは""", """##ばんは"""] UpperCamelCase = {} for i, token in enumerate(lowerCamelCase_ ): UpperCamelCase = i UpperCamelCase = WordpieceTokenizer(vocab=lowerCamelCase_ , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""こんにちは""" ) , ["""こんにちは"""] ) self.assertListEqual(tokenizer.tokenize("""こんばんは""" ) , ["""こん""", """##ばんは"""] ) self.assertListEqual(tokenizer.tokenize("""こんばんは こんばんにちは こんにちは""" ) , ["""こん""", """##ばんは""", """[UNK]""", """こんにちは"""] ) def lowerCamelCase_ ( self : Tuple ): """simple docstring""" UpperCamelCase = BertJapaneseTokenizer.from_pretrained("""nlp-waseda/roberta-base-japanese-with-auto-jumanpp""" ) UpperCamelCase = tokenizer.subword_tokenizer UpperCamelCase = subword_tokenizer.tokenize("""国境 の 長い トンネル を 抜ける と 雪国 であった 。""" ) self.assertListEqual(lowerCamelCase_ , ["""▁国境""", """▁の""", """▁長い""", """▁トンネル""", """▁を""", """▁抜ける""", """▁と""", """▁雪""", """国""", """▁であった""", """▁。"""] ) UpperCamelCase = subword_tokenizer.tokenize("""こんばんは こんばん にち は こんにちは""" ) self.assertListEqual(lowerCamelCase_ , ["""▁こん""", """ばん""", """は""", """▁こん""", """ばん""", """▁に""", """ち""", """▁は""", """▁こんにちは"""] ) def lowerCamelCase_ ( self : Union[str, Any] ): """simple docstring""" UpperCamelCase = self.tokenizer_class.from_pretrained("""cl-tohoku/bert-base-japanese""" ) UpperCamelCase = tokenizer.encode("""ありがとう。""" , add_special_tokens=lowerCamelCase_ ) UpperCamelCase = tokenizer.encode("""どういたしまして。""" , add_special_tokens=lowerCamelCase_ ) UpperCamelCase = tokenizer.build_inputs_with_special_tokens(lowerCamelCase_ ) UpperCamelCase = tokenizer.build_inputs_with_special_tokens(lowerCamelCase_ , lowerCamelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase , unittest.TestCase ): __lowerCAmelCase = BertJapaneseTokenizer __lowerCAmelCase = False def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" super().setUp() UpperCamelCase = ["""[UNK]""", """[CLS]""", """[SEP]""", """こ""", """ん""", """に""", """ち""", """は""", """ば""", """世""", """界""", """、""", """。"""] UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def lowerCamelCase_ ( self : Tuple , **lowerCamelCase_ : Union[str, Any] ): """simple docstring""" return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type="""character""" , **lowerCamelCase_ ) def lowerCamelCase_ ( self : Union[str, Any] , lowerCamelCase_ : Any ): """simple docstring""" UpperCamelCase = """こんにちは、世界。 \nこんばんは、世界。""" UpperCamelCase = """こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。""" return input_text, output_text def lowerCamelCase_ ( self : List[str] ): """simple docstring""" pass # TODO add if relevant def lowerCamelCase_ ( self : Any ): """simple docstring""" pass # TODO add if relevant def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" pass # TODO add if relevant def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" UpperCamelCase = self.tokenizer_class(self.vocab_file , subword_tokenizer_type="""character""" ) UpperCamelCase = tokenizer.tokenize("""こんにちは、世界。 \nこんばんは、世界。""" ) self.assertListEqual( lowerCamelCase_ , ["""こ""", """ん""", """に""", """ち""", """は""", """、""", """世""", """界""", """。""", """こ""", """ん""", """ば""", """ん""", """は""", """、""", """世""", """界""", """。"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase_ ) , [3, 4, 5, 6, 7, 11, 9, 10, 12, 3, 4, 8, 4, 7, 11, 9, 10, 12] ) def lowerCamelCase_ ( self : int ): """simple docstring""" UpperCamelCase = ["""[UNK]""", """[CLS]""", """[SEP]""", """こ""", """ん""", """に""", """ち""", """は""", """ば""", """世""", """界""", """、""", """。"""] UpperCamelCase = {} for i, token in enumerate(lowerCamelCase_ ): UpperCamelCase = i UpperCamelCase = CharacterTokenizer(vocab=lowerCamelCase_ , unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ) , [] ) self.assertListEqual(tokenizer.tokenize("""こんにちは""" ) , ["""こ""", """ん""", """に""", """ち""", """は"""] ) self.assertListEqual(tokenizer.tokenize("""こんにちほ""" ) , ["""こ""", """ん""", """に""", """ち""", """[UNK]"""] ) def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" UpperCamelCase = self.tokenizer_class.from_pretrained("""cl-tohoku/bert-base-japanese-char""" ) UpperCamelCase = tokenizer.encode("""ありがとう。""" , add_special_tokens=lowerCamelCase_ ) UpperCamelCase = tokenizer.encode("""どういたしまして。""" , add_special_tokens=lowerCamelCase_ ) UpperCamelCase = tokenizer.build_inputs_with_special_tokens(lowerCamelCase_ ) UpperCamelCase = tokenizer.build_inputs_with_special_tokens(lowerCamelCase_ , lowerCamelCase_ ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): def lowerCamelCase_ ( self : int ): """simple docstring""" UpperCamelCase = """cl-tohoku/bert-base-japanese""" UpperCamelCase = AutoTokenizer.from_pretrained(lowerCamelCase_ ) self.assertIsInstance(lowerCamelCase_ , lowerCamelCase_ ) class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): def lowerCamelCase_ ( self : int ): """simple docstring""" UpperCamelCase = """cl-tohoku/bert-base-japanese""" with self.assertLogs("""transformers""" , level="""WARNING""" ) as cm: BertTokenizer.from_pretrained(lowerCamelCase_ ) self.assertTrue( cm.records[0].message.startswith( """The tokenizer class you load from this checkpoint is not the same type as the class this function""" """ is called from.""" ) ) UpperCamelCase = """bert-base-cased""" with self.assertLogs("""transformers""" , level="""WARNING""" ) as cm: BertJapaneseTokenizer.from_pretrained(lowerCamelCase_ ) self.assertTrue( cm.records[0].message.startswith( """The tokenizer class you load from this checkpoint is not the same type as the class this function""" """ is called from.""" ) )
343
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _SCREAMING_SNAKE_CASE = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.weight''', F'''encoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (F'''transformer.encoder.layers.{i}.self_attn.out_proj.bias''', F'''encoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.weight''', F'''encoder.layers.{i}.fc1.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear1.bias''', F'''encoder.layers.{i}.fc1.bias''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.weight''', F'''encoder.layers.{i}.fc2.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.linear2.bias''', F'''encoder.layers.{i}.fc2.bias''')) rename_keys.append( (F'''transformer.encoder.layers.{i}.norm1.weight''', F'''encoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''transformer.encoder.layers.{i}.norm1.bias''', F'''encoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.weight''', F'''encoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((F'''transformer.encoder.layers.{i}.norm2.bias''', F'''encoder.layers.{i}.final_layer_norm.bias''')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.weight''', F'''decoder.layers.{i}.self_attn.out_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.self_attn.out_proj.bias''', F'''decoder.layers.{i}.self_attn.out_proj.bias''') ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.cross_attn.out_proj.weight''', F'''decoder.layers.{i}.encoder_attn.out_proj.weight''', ) ) rename_keys.append( ( F'''transformer.decoder.layers.{i}.cross_attn.out_proj.bias''', F'''decoder.layers.{i}.encoder_attn.out_proj.bias''', ) ) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.weight''', F'''decoder.layers.{i}.fc1.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear1.bias''', F'''decoder.layers.{i}.fc1.bias''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.weight''', F'''decoder.layers.{i}.fc2.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.linear2.bias''', F'''decoder.layers.{i}.fc2.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm1.weight''', F'''decoder.layers.{i}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm1.bias''', F'''decoder.layers.{i}.self_attn_layer_norm.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.weight''', F'''decoder.layers.{i}.encoder_attn_layer_norm.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.norm2.bias''', F'''decoder.layers.{i}.encoder_attn_layer_norm.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.weight''', F'''decoder.layers.{i}.final_layer_norm.weight''')) rename_keys.append((F'''transformer.decoder.layers.{i}.norm3.bias''', F'''decoder.layers.{i}.final_layer_norm.bias''')) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (F'''transformer.decoder.layers.{i}.sa_qcontent_proj.weight''', F'''decoder.layers.{i}.sa_qcontent_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.sa_kcontent_proj.weight''', F'''decoder.layers.{i}.sa_kcontent_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.sa_qpos_proj.weight''', F'''decoder.layers.{i}.sa_qpos_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.sa_kpos_proj.weight''', F'''decoder.layers.{i}.sa_kpos_proj.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.sa_v_proj.weight''', F'''decoder.layers.{i}.sa_v_proj.weight''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.ca_qcontent_proj.weight''', F'''decoder.layers.{i}.ca_qcontent_proj.weight''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (F'''transformer.decoder.layers.{i}.ca_kcontent_proj.weight''', F'''decoder.layers.{i}.ca_kcontent_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.ca_kpos_proj.weight''', F'''decoder.layers.{i}.ca_kpos_proj.weight''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.ca_v_proj.weight''', F'''decoder.layers.{i}.ca_v_proj.weight''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight''', F'''decoder.layers.{i}.ca_qpos_sine_proj.weight''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.sa_qcontent_proj.bias''', F'''decoder.layers.{i}.sa_qcontent_proj.bias''') ) rename_keys.append( (F'''transformer.decoder.layers.{i}.sa_kcontent_proj.bias''', F'''decoder.layers.{i}.sa_kcontent_proj.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.sa_qpos_proj.bias''', F'''decoder.layers.{i}.sa_qpos_proj.bias''')) rename_keys.append((F'''transformer.decoder.layers.{i}.sa_kpos_proj.bias''', F'''decoder.layers.{i}.sa_kpos_proj.bias''')) rename_keys.append((F'''transformer.decoder.layers.{i}.sa_v_proj.bias''', F'''decoder.layers.{i}.sa_v_proj.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.ca_qcontent_proj.bias''', F'''decoder.layers.{i}.ca_qcontent_proj.bias''') ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (F'''transformer.decoder.layers.{i}.ca_kcontent_proj.bias''', F'''decoder.layers.{i}.ca_kcontent_proj.bias''') ) rename_keys.append((F'''transformer.decoder.layers.{i}.ca_kpos_proj.bias''', F'''decoder.layers.{i}.ca_kpos_proj.bias''')) rename_keys.append((F'''transformer.decoder.layers.{i}.ca_v_proj.bias''', F'''decoder.layers.{i}.ca_v_proj.bias''')) rename_keys.append( (F'''transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias''', F'''decoder.layers.{i}.ca_qpos_sine_proj.bias''') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ("""transformer.decoder.ref_point_head.layers.0.weight""", """decoder.ref_point_head.layers.0.weight"""), ("""transformer.decoder.ref_point_head.layers.0.bias""", """decoder.ref_point_head.layers.0.bias"""), ("""transformer.decoder.ref_point_head.layers.1.weight""", """decoder.ref_point_head.layers.1.weight"""), ("""transformer.decoder.ref_point_head.layers.1.bias""", """decoder.ref_point_head.layers.1.bias"""), ("""transformer.decoder.query_scale.layers.0.weight""", """decoder.query_scale.layers.0.weight"""), ("""transformer.decoder.query_scale.layers.0.bias""", """decoder.query_scale.layers.0.bias"""), ("""transformer.decoder.query_scale.layers.1.weight""", """decoder.query_scale.layers.1.weight"""), ("""transformer.decoder.query_scale.layers.1.bias""", """decoder.query_scale.layers.1.bias"""), ("""transformer.decoder.layers.0.ca_qpos_proj.weight""", """decoder.layers.0.ca_qpos_proj.weight"""), ("""transformer.decoder.layers.0.ca_qpos_proj.bias""", """decoder.layers.0.ca_qpos_proj.bias"""), ] ) def lowercase( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> Optional[Any]: '''simple docstring''' UpperCamelCase = state_dict.pop(UpperCamelCase_ ) UpperCamelCase = val def lowercase( UpperCamelCase_ ) -> Any: '''simple docstring''' UpperCamelCase = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: UpperCamelCase = key.replace("""backbone.0.body""" , """backbone.conv_encoder.model""" ) UpperCamelCase = value else: UpperCamelCase = value return new_state_dict def lowercase( UpperCamelCase_ , UpperCamelCase_=False ) -> Optional[int]: '''simple docstring''' UpperCamelCase = """""" if is_panoptic: UpperCamelCase = """conditional_detr.""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) UpperCamelCase = state_dict.pop(f"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight""" ) UpperCamelCase = state_dict.pop(f"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCamelCase = in_proj_weight[:256, :] UpperCamelCase = in_proj_bias[:256] UpperCamelCase = in_proj_weight[256:512, :] UpperCamelCase = in_proj_bias[256:512] UpperCamelCase = in_proj_weight[-256:, :] UpperCamelCase = in_proj_bias[-256:] def lowercase( ) -> Any: '''simple docstring''' UpperCamelCase = """http://images.cocodataset.org/val2017/000000039769.jpg""" UpperCamelCase = Image.open(requests.get(UpperCamelCase_ , stream=UpperCamelCase_ ).raw ) return im @torch.no_grad() def lowercase( UpperCamelCase_ , UpperCamelCase_ ) -> Any: '''simple docstring''' UpperCamelCase = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: UpperCamelCase = """resnet101""" if "dc5" in model_name: UpperCamelCase = True UpperCamelCase = """panoptic""" in model_name if is_panoptic: UpperCamelCase = 250 else: UpperCamelCase = 91 UpperCamelCase = """huggingface/label-files""" UpperCamelCase = """coco-detection-id2label.json""" UpperCamelCase = json.load(open(hf_hub_download(UpperCamelCase_ , UpperCamelCase_ , repo_type="""dataset""" ) , """r""" ) ) UpperCamelCase = {int(UpperCamelCase_ ): v for k, v in idalabel.items()} UpperCamelCase = idalabel UpperCamelCase = {v: k for k, v in idalabel.items()} # load image processor UpperCamelCase = """coco_panoptic""" if is_panoptic else """coco_detection""" UpperCamelCase = ConditionalDetrImageProcessor(format=UpperCamelCase_ ) # prepare image UpperCamelCase = prepare_img() UpperCamelCase = image_processor(images=UpperCamelCase_ , return_tensors="""pt""" ) UpperCamelCase = encoding["""pixel_values"""] logger.info(f"""Converting model {model_name}...""" ) # load original model from torch hub UpperCamelCase = torch.hub.load("""DeppMeng/ConditionalDETR""" , UpperCamelCase_ , pretrained=UpperCamelCase_ ).eval() UpperCamelCase = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: UpperCamelCase = """conditional_detr.""" + src rename_key(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) UpperCamelCase = rename_backbone_keys(UpperCamelCase_ ) # query, key and value matrices need special treatment read_in_q_k_v(UpperCamelCase_ , is_panoptic=UpperCamelCase_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them UpperCamelCase = """conditional_detr.model.""" if is_panoptic else """model.""" for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("""conditional_detr""" ) and not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ) ): UpperCamelCase = state_dict.pop(UpperCamelCase_ ) UpperCamelCase = val elif "class_labels_classifier" in key or "bbox_predictor" in key: UpperCamelCase = state_dict.pop(UpperCamelCase_ ) UpperCamelCase = val elif key.startswith("""bbox_attention""" ) or key.startswith("""mask_head""" ): continue else: UpperCamelCase = state_dict.pop(UpperCamelCase_ ) UpperCamelCase = val else: if not key.startswith("""class_labels_classifier""" ) and not key.startswith("""bbox_predictor""" ): UpperCamelCase = state_dict.pop(UpperCamelCase_ ) UpperCamelCase = val # finally, create HuggingFace model and load state dict UpperCamelCase = ConditionalDetrForSegmentation(UpperCamelCase_ ) if is_panoptic else ConditionalDetrForObjectDetection(UpperCamelCase_ ) model.load_state_dict(UpperCamelCase_ ) model.eval() model.push_to_hub(repo_id=UpperCamelCase_ , organization="""DepuMeng""" , commit_message="""Add model""" ) # verify our conversion UpperCamelCase = conditional_detr(UpperCamelCase_ ) UpperCamelCase = model(UpperCamelCase_ ) assert torch.allclose(outputs.logits , original_outputs["""pred_logits"""] , atol=1E-4 ) assert torch.allclose(outputs.pred_boxes , original_outputs["""pred_boxes"""] , atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["""pred_masks"""] , atol=1E-4 ) # Save model and image processor logger.info(f"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(UpperCamelCase_ ).mkdir(exist_ok=UpperCamelCase_ ) model.save_pretrained(UpperCamelCase_ ) image_processor.save_pretrained(UpperCamelCase_ ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument( """--model_name""", default="""conditional_detr_resnet50""", type=str, help="""Name of the CONDITIONAL_DETR model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) _SCREAMING_SNAKE_CASE = parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
343
1
from math import ceil, sqrt def __UpperCAmelCase ( __a : int = 1_000_000 ) -> int: """simple docstring""" _a : Any = 0 for outer_width in range(3 ,(limit // 4) + 2 ): if outer_width**2 > limit: _a : Any = max(ceil(sqrt(outer_width**2 - limit ) ) ,1 ) else: _a : List[str] = 1 if (outer_width - hole_width_lower_bound) % 2: hole_width_lower_bound += 1 answer += (outer_width - hole_width_lower_bound - 2) // 2 + 1 return answer if __name__ == "__main__": print(f'''{solution() = }''')
15
import argparse import os import torch from transformers.utils import WEIGHTS_NAME a__ = ['''small''', '''medium''', '''large'''] a__ = '''lm_head.decoder.weight''' a__ = '''lm_head.weight''' def __UpperCAmelCase ( __a : str ,__a : str ) -> List[str]: """simple docstring""" _a : Any = torch.load(__a ) _a : List[str] = d.pop(__a ) os.makedirs(__a ,exist_ok=__a ) torch.save(__a ,os.path.join(__a ,__a ) ) if __name__ == "__main__": a__ = argparse.ArgumentParser() parser.add_argument('''--dialogpt_path''', default='''.''', type=str) a__ = parser.parse_args() for MODEL in DIALOGPT_MODELS: a__ = os.path.join(args.dialogpt_path, f'''{MODEL}_ft.pkl''') a__ = f'''./DialoGPT-{MODEL}''' convert_dialogpt_checkpoint( checkpoint_path, pytorch_dump_folder_path, )
15
1
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def SCREAMING_SNAKE_CASE ( ) -> tuple[list[int], int]: """simple docstring""" A__ = [randint(-1_000 , 1_000 ) for i in range(10 )] A__ = randint(-5_000 , 5_000 ) return (arr, r) _lowerCamelCase : Tuple = make_dataset() def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ ) -> tuple[int, ...]: """simple docstring""" for triplet in permutations(lowercase_ , 3 ): if sum(lowercase_ ) == target: return tuple(sorted(lowercase_ ) ) return (0, 0, 0) def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ ) -> tuple[int, int, int]: """simple docstring""" arr.sort() A__ = len(lowercase_ ) for i in range(n - 1 ): A__ , A__ = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def SCREAMING_SNAKE_CASE ( ) -> tuple[float, float]: """simple docstring""" A__ = ''' from __main__ import dataset, triplet_sum1, triplet_sum2 ''' A__ = ''' triplet_sum1(*dataset) ''' A__ = ''' triplet_sum2(*dataset) ''' A__ = repeat(setup=lowercase_ , stmt=lowercase_ , repeat=5 , number=10_000 ) A__ = repeat(setup=lowercase_ , stmt=lowercase_ , repeat=5 , number=10_000 ) return (min(lowercase_ ), min(lowercase_ )) if __name__ == "__main__": from doctest import testmod testmod() _lowerCamelCase : Optional[Any] = solution_times() print(F'''The time for naive implementation is {times[0]}.''') print(F'''The time for optimized implementation is {times[1]}.''')
14
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCamelCase : List[str] = {"""configuration_vit_msn""": ["""VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ViTMSNConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : List[Any] = [ """VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST""", """ViTMSNModel""", """ViTMSNForImageClassification""", """ViTMSNPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys _lowerCamelCase : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
14
1
def __lowercase ( lowerCamelCase : int = 1000 ): return sum(e for e in range(3 , UpperCamelCase__ ) if e % 3 == 0 or e % 5 == 0 ) if __name__ == "__main__": print(F"""{solution() = }""")
363
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor a_ = random.Random() def __lowercase ( lowerCamelCase : Optional[int] , lowerCamelCase : int=1.0 , lowerCamelCase : Optional[int]=None , lowerCamelCase : Optional[int]=None ): if rng is None: UpperCamelCase_ : Union[str, Any] = global_rng UpperCamelCase_ : Union[str, Any] = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class _lowercase ( unittest.TestCase ): def __init__( self : Optional[Any] , snake_case : Tuple , snake_case : str=7 , snake_case : Tuple=4_0_0 , snake_case : List[Any]=2_0_0_0 , snake_case : Optional[Any]=2_4 , snake_case : Tuple=2_4 , snake_case : Dict=0.0 , snake_case : Any=1_6_0_0_0 , snake_case : Tuple=True , snake_case : List[str]=True , ) -> Optional[int]: """simple docstring""" UpperCamelCase_ : int = parent UpperCamelCase_ : int = batch_size UpperCamelCase_ : str = min_seq_length UpperCamelCase_ : str = max_seq_length UpperCamelCase_ : Optional[int] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) UpperCamelCase_ : int = feature_size UpperCamelCase_ : Optional[int] = num_mel_bins UpperCamelCase_ : str = padding_value UpperCamelCase_ : Union[str, Any] = sampling_rate UpperCamelCase_ : Tuple = return_attention_mask UpperCamelCase_ : List[str] = do_normalize def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def SCREAMING_SNAKE_CASE__ ( self : Any , snake_case : Dict=False , snake_case : List[str]=False ) -> int: """simple docstring""" def _flatten(snake_case : Optional[Any] ): return list(itertools.chain(*snake_case ) ) if equal_length: UpperCamelCase_ : Any = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size UpperCamelCase_ : Optional[Any] = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: UpperCamelCase_ : List[str] = [np.asarray(snake_case ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class _lowercase ( snake_case_ , unittest.TestCase ): lowercase = SpeechaTextFeatureExtractor if is_speech_available() else None def SCREAMING_SNAKE_CASE__ ( self : Dict ) -> Tuple: """simple docstring""" UpperCamelCase_ : List[str] = SpeechaTextFeatureExtractionTester(self ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , snake_case : str ) -> Tuple: """simple docstring""" self.assertTrue(np.all(np.mean(snake_case , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(snake_case , axis=0 ) - 1 ) < 1e-3 ) ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] ) -> List[str]: """simple docstring""" UpperCamelCase_ : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 UpperCamelCase_ : Union[str, Any] = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] UpperCamelCase_ : List[Any] = [np.asarray(snake_case ) for speech_input in speech_inputs] # Test feature size UpperCamelCase_ : Tuple = feature_extractor(snake_case , padding=snake_case , return_tensors='np' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input UpperCamelCase_ : int = feature_extractor(speech_inputs[0] , return_tensors='np' ).input_features UpperCamelCase_ : str = feature_extractor(np_speech_inputs[0] , return_tensors='np' ).input_features self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) # Test batched UpperCamelCase_ : Union[str, Any] = feature_extractor(snake_case , return_tensors='np' ).input_features UpperCamelCase_ : List[str] = feature_extractor(snake_case , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(snake_case , snake_case ): self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. UpperCamelCase_ : int = [floats_list((1, x) )[0] for x in (8_0_0, 8_0_0, 8_0_0)] UpperCamelCase_ : List[str] = np.asarray(snake_case ) UpperCamelCase_ : Any = feature_extractor(snake_case , return_tensors='np' ).input_features UpperCamelCase_ : str = feature_extractor(snake_case , return_tensors='np' ).input_features for enc_seq_a, enc_seq_a in zip(snake_case , snake_case ): self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] ) -> Tuple: """simple docstring""" UpperCamelCase_ : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase_ : Optional[int] = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] UpperCamelCase_ : Union[str, Any] = ['longest', 'max_length', 'do_not_pad'] UpperCamelCase_ : Tuple = [None, 1_6, None] for max_length, padding in zip(snake_case , snake_case ): UpperCamelCase_ : Optional[Any] = feature_extractor( snake_case , padding=snake_case , max_length=snake_case , return_attention_mask=snake_case ) UpperCamelCase_ : List[str] = inputs.input_features UpperCamelCase_ : List[str] = inputs.attention_mask UpperCamelCase_ : Optional[int] = [np.sum(snake_case ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> List[str]: """simple docstring""" UpperCamelCase_ : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase_ : Any = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] UpperCamelCase_ : List[str] = ['longest', 'max_length', 'do_not_pad'] UpperCamelCase_ : Optional[Any] = [None, 1_6, None] for max_length, padding in zip(snake_case , snake_case ): UpperCamelCase_ : Any = feature_extractor( snake_case , max_length=snake_case , padding=snake_case , return_tensors='np' , return_attention_mask=snake_case ) UpperCamelCase_ : int = inputs.input_features UpperCamelCase_ : Optional[int] = inputs.attention_mask UpperCamelCase_ : str = [np.sum(snake_case ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] ) -> List[str]: """simple docstring""" UpperCamelCase_ : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase_ : List[Any] = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] UpperCamelCase_ : str = feature_extractor( snake_case , padding='max_length' , max_length=4 , truncation=snake_case , return_tensors='np' , return_attention_mask=snake_case , ) UpperCamelCase_ : int = inputs.input_features UpperCamelCase_ : Union[str, Any] = inputs.attention_mask UpperCamelCase_ : Dict = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> str: """simple docstring""" UpperCamelCase_ : str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase_ : Optional[int] = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] UpperCamelCase_ : Any = feature_extractor( snake_case , padding='longest' , max_length=4 , truncation=snake_case , return_tensors='np' , return_attention_mask=snake_case , ) UpperCamelCase_ : Dict = inputs.input_features UpperCamelCase_ : List[Any] = inputs.attention_mask UpperCamelCase_ : Tuple = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 2_4) ) UpperCamelCase_ : Dict = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] UpperCamelCase_ : int = feature_extractor( snake_case , padding='longest' , max_length=1_6 , truncation=snake_case , return_tensors='np' , return_attention_mask=snake_case , ) UpperCamelCase_ : Dict = inputs.input_features UpperCamelCase_ : Union[str, Any] = inputs.attention_mask UpperCamelCase_ : Dict = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 2_4) ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ) -> Tuple: """simple docstring""" import torch UpperCamelCase_ : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase_ : Optional[Any] = np.random.rand(1_0_0 , 3_2 ).astype(np.floataa ) UpperCamelCase_ : int = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: UpperCamelCase_ : Tuple = feature_extractor.pad([{'input_features': inputs}] , return_tensors='np' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) UpperCamelCase_ : Tuple = feature_extractor.pad([{'input_features': inputs}] , return_tensors='pt' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def SCREAMING_SNAKE_CASE__ ( self : Tuple , snake_case : Tuple ) -> Dict: """simple docstring""" from datasets import load_dataset UpperCamelCase_ : Optional[int] = load_dataset('hf-internal-testing/librispeech_asr_dummy' , 'clean' , split='validation' ) # automatic decoding with librispeech UpperCamelCase_ : Optional[Any] = ds.sort('id' ).select(range(snake_case ) )[:num_samples]['audio'] return [x["array"] for x in speech_samples] def SCREAMING_SNAKE_CASE__ ( self : Dict ) -> str: """simple docstring""" UpperCamelCase_ : Union[str, Any] = np.array([ -1.5745, -1.7713, -1.7020, -1.6069, -1.2250, -1.1105, -0.9072, -0.8241, -1.2310, -0.8098, -0.3320, -0.4101, -0.7985, -0.4996, -0.8213, -0.9128, -1.0420, -1.1286, -1.0440, -0.7999, -0.8405, -1.2275, -1.5443, -1.4625, ] ) # fmt: on UpperCamelCase_ : str = self._load_datasamples(1 ) UpperCamelCase_ : str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) UpperCamelCase_ : str = feature_extractor(snake_case , return_tensors='pt' ).input_features self.assertEquals(input_features.shape , (1, 5_8_4, 2_4) ) self.assertTrue(np.allclose(input_features[0, 0, :3_0] , snake_case , atol=1e-4 ) )
50
0
'''simple docstring''' import copy from typing import Dict, List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING _lowerCAmelCase = { '''facebook/mask2former-swin-small-coco-instance''': ( '''https://huggingface.co/facebook/mask2former-swin-small-coco-instance/blob/main/config.json''' ) # See all Mask2Former models at https://huggingface.co/models?filter=mask2former } _lowerCAmelCase = logging.get_logger(__name__) class A ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' A = "mask2former" A = ["swin"] A = {"hidden_size": "hidden_dim"} def __init__(self , _UpperCAmelCase = None , _UpperCAmelCase = 2_5_6 , _UpperCAmelCase = 2_5_6 , _UpperCAmelCase = 2_5_6 , _UpperCAmelCase = 1_0_2_4 , _UpperCAmelCase = "relu" , _UpperCAmelCase = 6 , _UpperCAmelCase = 1_0 , _UpperCAmelCase = 8 , _UpperCAmelCase = 0.0 , _UpperCAmelCase = 2_0_4_8 , _UpperCAmelCase = False , _UpperCAmelCase = False , _UpperCAmelCase = 4 , _UpperCAmelCase = 2_5_5 , _UpperCAmelCase = 1_0_0 , _UpperCAmelCase = 0.1 , _UpperCAmelCase = 2.0 , _UpperCAmelCase = 5.0 , _UpperCAmelCase = 5.0 , _UpperCAmelCase = 1_2_5_4_4 , _UpperCAmelCase = 3.0 , _UpperCAmelCase = 0.75 , _UpperCAmelCase = 0.02 , _UpperCAmelCase = 1.0 , _UpperCAmelCase = True , _UpperCAmelCase = [4, 8, 1_6, 3_2] , _UpperCAmelCase = None , **_UpperCAmelCase , ) -> Any: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `Swin` backbone." ) __UpperCamelCase : Optional[int] = CONFIG_MAPPING["swin"]( image_size=2_2_4 , in_channels=3 , patch_size=4 , embed_dim=9_6 , depths=[2, 2, 1_8, 2] , num_heads=[3, 6, 1_2, 2_4] , window_size=7 , drop_path_rate=0.3 , use_absolute_embeddings=_UpperCAmelCase , out_features=["stage1", "stage2", "stage3", "stage4"] , ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ): __UpperCamelCase : Optional[int] = backbone_config.pop("model_type" ) __UpperCamelCase : List[Any] = CONFIG_MAPPING[backbone_model_type] __UpperCamelCase : Optional[int] = config_class.from_dict(_UpperCAmelCase ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( f"Backbone {backbone_config.model_type} is not a supported model and may not be compatible with Mask2Former. " f"Supported model types: {','.join(self.backbones_supported )}" ) __UpperCamelCase : Optional[Any] = backbone_config __UpperCamelCase : Union[str, Any] = feature_size __UpperCamelCase : List[str] = mask_feature_size __UpperCamelCase : List[str] = hidden_dim __UpperCamelCase : Dict = encoder_feedforward_dim __UpperCamelCase : str = activation_function __UpperCamelCase : Any = encoder_layers __UpperCamelCase : Dict = decoder_layers __UpperCamelCase : Dict = num_attention_heads __UpperCamelCase : Optional[Any] = dropout __UpperCamelCase : Any = dim_feedforward __UpperCamelCase : int = pre_norm __UpperCamelCase : int = enforce_input_projection __UpperCamelCase : List[str] = common_stride __UpperCamelCase : List[str] = ignore_value __UpperCamelCase : int = num_queries __UpperCamelCase : List[str] = no_object_weight __UpperCamelCase : Optional[int] = class_weight __UpperCamelCase : str = mask_weight __UpperCamelCase : Any = dice_weight __UpperCamelCase : str = train_num_points __UpperCamelCase : Optional[Any] = oversample_ratio __UpperCamelCase : List[Any] = importance_sample_ratio __UpperCamelCase : Optional[int] = init_std __UpperCamelCase : Optional[int] = init_xavier_std __UpperCamelCase : Dict = use_auxiliary_loss __UpperCamelCase : Union[str, Any] = feature_strides __UpperCamelCase : Any = output_auxiliary_logits __UpperCamelCase : Union[str, Any] = decoder_layers super().__init__(**_UpperCAmelCase ) @classmethod def a_ (cls , _UpperCAmelCase , **_UpperCAmelCase ) -> str: return cls( backbone_config=_UpperCAmelCase , **_UpperCAmelCase , ) def a_ (self ) -> Dict[str, any]: __UpperCamelCase : Union[str, Any] = copy.deepcopy(self.__dict__ ) __UpperCamelCase : str = self.backbone_config.to_dict() __UpperCamelCase : Any = self.__class__.model_type return output
298
'''simple docstring''' def __lowerCAmelCase ( snake_case__ ): return "".join([hex(snake_case__ )[2:].zfill(2 ).upper() for byte in list(snake_case__ )] ) def __lowerCAmelCase ( snake_case__ ): # Check data validity, following RFC3548 # https://www.ietf.org/rfc/rfc3548.txt if (len(snake_case__ ) % 2) != 0: raise ValueError( "Base16 encoded data is invalid:\nData does not have an even number of hex digits." ) # Check the character set - the standard base16 alphabet # is uppercase according to RFC3548 section 6 if not set(snake_case__ ) <= set("0123456789ABCDEF" ): raise ValueError( "Base16 encoded data is invalid:\nData is not uppercase hex or it contains invalid characters." ) # For every two hexadecimal digits (= a byte), turn it into an integer. # Then, string the result together into bytes, and return it. return bytes(int(data[i] + data[i + 1] , 16 ) for i in range(0 , len(snake_case__ ) , 2 ) ) if __name__ == "__main__": import doctest doctest.testmod()
298
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __a :List[Any] = logging.get_logger(__name__) __a :Union[str, Any] = { 'shi-labs/dinat-mini-in1k-224': 'https://huggingface.co/shi-labs/dinat-mini-in1k-224/resolve/main/config.json', # See all Dinat models at https://huggingface.co/models?filter=dinat } class _a ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): """simple docstring""" _lowerCamelCase : List[Any] = 'dinat' _lowerCamelCase : Dict = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self : str , UpperCAmelCase : str=4 , UpperCAmelCase : Any=3 , UpperCAmelCase : str=64 , UpperCAmelCase : int=[3, 4, 6, 5] , UpperCAmelCase : Union[str, Any]=[2, 4, 8, 16] , UpperCAmelCase : Optional[int]=7 , UpperCAmelCase : int=[[1, 8, 1], [1, 4, 1, 4], [1, 2, 1, 2, 1, 2], [1, 1, 1, 1, 1]] , UpperCAmelCase : Any=3.0 , UpperCAmelCase : int=True , UpperCAmelCase : List[str]=0.0 , UpperCAmelCase : Optional[int]=0.0 , UpperCAmelCase : int=0.1 , UpperCAmelCase : str="gelu" , UpperCAmelCase : Optional[Any]=0.02 , UpperCAmelCase : Dict=1E-5 , UpperCAmelCase : Union[str, Any]=0.0 , UpperCAmelCase : Dict=None , UpperCAmelCase : Optional[Any]=None , **UpperCAmelCase : int , ): super().__init__(**_SCREAMING_SNAKE_CASE ) A_ = patch_size A_ = num_channels A_ = embed_dim A_ = depths A_ = len(_SCREAMING_SNAKE_CASE ) A_ = num_heads A_ = kernel_size A_ = dilations A_ = mlp_ratio A_ = qkv_bias A_ = hidden_dropout_prob A_ = attention_probs_dropout_prob A_ = drop_path_rate A_ = hidden_act A_ = layer_norm_eps A_ = initializer_range # we set the hidden_size attribute in order to make Dinat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model A_ = int(embed_dim * 2 ** (len(_SCREAMING_SNAKE_CASE ) - 1) ) A_ = layer_scale_init_value A_ = ["stem"] + [f'''stage{idx}''' for idx in range(1 , len(_SCREAMING_SNAKE_CASE ) + 1 )] A_ = get_aligned_output_features_output_indices( out_features=_SCREAMING_SNAKE_CASE , out_indices=_SCREAMING_SNAKE_CASE , stage_names=self.stage_names )
364
from ..utils import DummyObject, requires_backends class _a ( metaclass=snake_case_ ): """simple docstring""" _lowerCamelCase : Union[str, Any] = ['torch', 'transformers', 'onnx'] def __init__( self : List[Any] , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : str ): requires_backends(self , ["torch", "transformers", "onnx"] ) @classmethod def __A ( cls : Tuple , *UpperCAmelCase : Tuple , **UpperCAmelCase : Union[str, Any] ): requires_backends(cls , ["torch", "transformers", "onnx"] ) @classmethod def __A ( cls : Dict , *UpperCAmelCase : List[Any] , **UpperCAmelCase : Tuple ): requires_backends(cls , ["torch", "transformers", "onnx"] ) class _a ( metaclass=snake_case_ ): """simple docstring""" _lowerCamelCase : Tuple = ['torch', 'transformers', 'onnx'] def __init__( self : Optional[Any] , *UpperCAmelCase : List[Any] , **UpperCAmelCase : List[Any] ): requires_backends(self , ["torch", "transformers", "onnx"] ) @classmethod def __A ( cls : List[Any] , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : str ): requires_backends(cls , ["torch", "transformers", "onnx"] ) @classmethod def __A ( cls : Tuple , *UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : int ): requires_backends(cls , ["torch", "transformers", "onnx"] ) class _a ( metaclass=snake_case_ ): """simple docstring""" _lowerCamelCase : Any = ['torch', 'transformers', 'onnx'] def __init__( self : Dict , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : Optional[int] ): requires_backends(self , ["torch", "transformers", "onnx"] ) @classmethod def __A ( cls : Union[str, Any] , *UpperCAmelCase : Tuple , **UpperCAmelCase : Optional[int] ): requires_backends(cls , ["torch", "transformers", "onnx"] ) @classmethod def __A ( cls : Tuple , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : int ): requires_backends(cls , ["torch", "transformers", "onnx"] ) class _a ( metaclass=snake_case_ ): """simple docstring""" _lowerCamelCase : List[str] = ['torch', 'transformers', 'onnx'] def __init__( self : List[Any] , *UpperCAmelCase : List[Any] , **UpperCAmelCase : int ): requires_backends(self , ["torch", "transformers", "onnx"] ) @classmethod def __A ( cls : Any , *UpperCAmelCase : List[Any] , **UpperCAmelCase : str ): requires_backends(cls , ["torch", "transformers", "onnx"] ) @classmethod def __A ( cls : Optional[int] , *UpperCAmelCase : str , **UpperCAmelCase : int ): requires_backends(cls , ["torch", "transformers", "onnx"] ) class _a ( metaclass=snake_case_ ): """simple docstring""" _lowerCamelCase : Dict = ['torch', 'transformers', 'onnx'] def __init__( self : str , *UpperCAmelCase : int , **UpperCAmelCase : Tuple ): requires_backends(self , ["torch", "transformers", "onnx"] ) @classmethod def __A ( cls : Optional[int] , *UpperCAmelCase : str , **UpperCAmelCase : Dict ): requires_backends(cls , ["torch", "transformers", "onnx"] ) @classmethod def __A ( cls : int , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : List[str] ): requires_backends(cls , ["torch", "transformers", "onnx"] ) class _a ( metaclass=snake_case_ ): """simple docstring""" _lowerCamelCase : List[Any] = ['torch', 'transformers', 'onnx'] def __init__( self : str , *UpperCAmelCase : str , **UpperCAmelCase : List[Any] ): requires_backends(self , ["torch", "transformers", "onnx"] ) @classmethod def __A ( cls : List[Any] , *UpperCAmelCase : Optional[Any] , **UpperCAmelCase : List[Any] ): requires_backends(cls , ["torch", "transformers", "onnx"] ) @classmethod def __A ( cls : Optional[int] , *UpperCAmelCase : List[str] , **UpperCAmelCase : int ): requires_backends(cls , ["torch", "transformers", "onnx"] )
329
0
import random def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: lowerCAmelCase__ : Optional[int] = a[left_index] lowerCAmelCase__ : Optional[int] = left_index + 1 for j in range(left_index + 1 , SCREAMING_SNAKE_CASE_ ): if a[j] < pivot: lowerCAmelCase__ , lowerCAmelCase__ : int = a[i], a[j] i += 1 lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = a[i - 1], a[left_index] return i - 1 def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: if left < right: lowerCAmelCase__ : str = random.randint(SCREAMING_SNAKE_CASE_ , right - 1 ) lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = ( a[left], a[pivot], ) # switches the pivot with the left most bound lowerCAmelCase__ : List[str] = partition(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) quick_sort_random( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # recursive quicksort to the left of the pivot point quick_sort_random( SCREAMING_SNAKE_CASE_ , pivot_index + 1 , SCREAMING_SNAKE_CASE_ ) # recursive quicksort to the right of the pivot point def lowerCAmelCase__ ( ) -> Union[str, Any]: lowerCAmelCase__ : List[Any] = input('Enter numbers separated by a comma:\n' ).strip() lowerCAmelCase__ : Optional[Any] = [int(SCREAMING_SNAKE_CASE_ ) for item in user_input.split(',' )] quick_sort_random(SCREAMING_SNAKE_CASE_ , 0 , len(SCREAMING_SNAKE_CASE_ ) ) print(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": main()
212
import re def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> str: if len(re.findall('[ATCG]' , SCREAMING_SNAKE_CASE_ ) ) != len(SCREAMING_SNAKE_CASE_ ): raise ValueError('Invalid Strand' ) return dna.translate(dna.maketrans('ATCG' , 'TAGC' ) ) if __name__ == "__main__": import doctest doctest.testmod()
212
1
"""simple docstring""" def UpperCamelCase_ ( lowerCAmelCase__ : str ) -> str: """simple docstring""" return " ".join( ''.join(word[::-1] ) if len(lowerCAmelCase__ ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words("""Hey wollef sroirraw"""))
289
"""simple docstring""" def UpperCamelCase_ ( lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : str ) -> List[Any]: """simple docstring""" return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def UpperCamelCase_ ( lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : str=0 ) -> Union[str, Any]: """simple docstring""" return sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__ : x[column] ) def UpperCamelCase_ ( lowerCAmelCase__ : Dict , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Optional[Any]=float('inf' ) ) -> Optional[int]: """simple docstring""" for i in range(points_counts - 1 ): for j in range(i + 1 , lowerCAmelCase__ ): lowerCAmelCase_ : Union[str, Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: lowerCAmelCase_ : Optional[int] = current_dis return min_dis def UpperCamelCase_ ( lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Union[str, Any]=float('inf' ) ) -> Dict: """simple docstring""" for i in range(min(6 , points_counts - 1 ) , lowerCAmelCase__ ): for j in range(max(0 , i - 6 ) , lowerCAmelCase__ ): lowerCAmelCase_ : Any = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: lowerCAmelCase_ : Union[str, Any] = current_dis return min_dis def UpperCamelCase_ ( lowerCAmelCase__ : Dict , lowerCAmelCase__ : int , lowerCAmelCase__ : Optional[int] ) -> Dict: """simple docstring""" if points_counts <= 3: return dis_between_closest_pair(lowerCAmelCase__ , lowerCAmelCase__ ) # recursion lowerCAmelCase_ : int = points_counts // 2 lowerCAmelCase_ : Optional[Any] = closest_pair_of_points_sqr( lowerCAmelCase__ , points_sorted_on_y[:mid] , lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = closest_pair_of_points_sqr( lowerCAmelCase__ , points_sorted_on_y[mid:] , points_counts - mid ) lowerCAmelCase_ : Any = min(lowerCAmelCase__ , lowerCAmelCase__ ) lowerCAmelCase_ : str = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = dis_between_closest_in_strip( lowerCAmelCase__ , len(lowerCAmelCase__ ) , lowerCAmelCase__ ) return min(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCamelCase_ ( lowerCAmelCase__ : str , lowerCAmelCase__ : Tuple ) -> List[Any]: """simple docstring""" lowerCAmelCase_ : List[str] = column_based_sort(lowerCAmelCase__ , column=0 ) lowerCAmelCase_ : Dict = column_based_sort(lowerCAmelCase__ , column=1 ) return ( closest_pair_of_points_sqr( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) ) ** 0.5 if __name__ == "__main__": lowercase__ : List[str] = [(2, 3), (1_2, 3_0), (4_0, 5_0), (5, 1), (1_2, 1_0), (3, 4)] print("""Distance:""", closest_pair_of_points(points, len(points)))
289
1
from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor _a = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def __A ( __lowerCAmelCase )-> str: """simple docstring""" if isinstance(__lowerCAmelCase , torch.Tensor ): return image elif isinstance(__lowerCAmelCase , PIL.Image.Image ): _UpperCAmelCase = [image] _UpperCAmelCase = [trans(img.convert('RGB' ) ) for img in image] _UpperCAmelCase = torch.stack(__lowerCAmelCase ) return image class __lowerCamelCase ( snake_case__): """simple docstring""" def __init__( self , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" super().__init__() # make sure scheduler can always be converted to DDIM _UpperCAmelCase = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=UpperCAmelCase , scheduler=UpperCAmelCase ) def UpperCamelCase ( self , UpperCAmelCase ): """simple docstring""" if strength < 0 or strength > 1: raise ValueError(F"""The value of strength should in [0.0, 1.0] but is {strength}""" ) def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = min(int(num_inference_steps * strength ) , UpperCAmelCase ) _UpperCAmelCase = max(num_inference_steps - init_timestep , 0 ) _UpperCAmelCase = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def UpperCamelCase ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=None ): """simple docstring""" if not isinstance(UpperCAmelCase , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( F"""`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(UpperCAmelCase )}""" ) _UpperCAmelCase = image.to(device=UpperCAmelCase , dtype=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) and len(UpperCAmelCase ) != batch_size: raise ValueError( F"""You have passed a list of generators of length {len(UpperCAmelCase )}, but requested an effective batch""" F""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) _UpperCAmelCase = init_latents.shape _UpperCAmelCase = randn_tensor(UpperCAmelCase , generator=UpperCAmelCase , device=UpperCAmelCase , dtype=UpperCAmelCase ) # get latents print('add noise to latents at timestep' , UpperCAmelCase ) _UpperCAmelCase = self.scheduler.add_noise(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) _UpperCAmelCase = init_latents return latents @torch.no_grad() def __call__( self , UpperCAmelCase = None , UpperCAmelCase = 0.8 , UpperCAmelCase = 1 , UpperCAmelCase = None , UpperCAmelCase = 0.0 , UpperCAmelCase = 50 , UpperCAmelCase = None , UpperCAmelCase = "pil" , UpperCAmelCase = True , ): """simple docstring""" self.check_inputs(UpperCAmelCase ) # 2. Preprocess image _UpperCAmelCase = preprocess(UpperCAmelCase ) # 3. set timesteps self.scheduler.set_timesteps(UpperCAmelCase , device=self.device ) _UpperCAmelCase , _UpperCAmelCase = self.get_timesteps(UpperCAmelCase , UpperCAmelCase , self.device ) _UpperCAmelCase = timesteps[:1].repeat(UpperCAmelCase ) # 4. Prepare latent variables _UpperCAmelCase = self.prepare_latents(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , self.unet.dtype , self.device , UpperCAmelCase ) _UpperCAmelCase = latents # 5. Denoising loop for t in self.progress_bar(UpperCAmelCase ): # 1. predict noise model_output _UpperCAmelCase = self.unet(UpperCAmelCase , UpperCAmelCase ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 _UpperCAmelCase = self.scheduler.step( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , eta=UpperCAmelCase , use_clipped_model_output=UpperCAmelCase , generator=UpperCAmelCase , ).prev_sample _UpperCAmelCase = (image / 2 + 0.5).clamp(0 , 1 ) _UpperCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _UpperCAmelCase = self.numpy_to_pil(UpperCAmelCase ) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=UpperCAmelCase )
39
import math import os import sys def a__ ( snake_case ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = '''''' try: with open(snake_case , '''rb''' ) as binary_file: __SCREAMING_SNAKE_CASE : int = binary_file.read() for dat in data: __SCREAMING_SNAKE_CASE : Optional[Any] = F'''{dat:08b}''' result += curr_byte return result except OSError: print('''File not accessible''' ) sys.exit() def a__ ( snake_case , snake_case , snake_case , snake_case ): """simple docstring""" lexicon.pop(snake_case ) __SCREAMING_SNAKE_CASE : List[str] = last_match_id if math.loga(snake_case ).is_integer(): for curr_key in lexicon: __SCREAMING_SNAKE_CASE : int = '''0''' + lexicon[curr_key] __SCREAMING_SNAKE_CASE : List[str] = bin(snake_case )[2:] def a__ ( snake_case ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = {'''0''': '''0''', '''1''': '''1'''} __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : Any = '''''', '''''' __SCREAMING_SNAKE_CASE : Optional[Any] = len(snake_case ) for i in range(len(snake_case ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue __SCREAMING_SNAKE_CASE : Any = lexicon[curr_string] result += last_match_id add_key_to_lexicon(snake_case , snake_case , snake_case , snake_case ) index += 1 __SCREAMING_SNAKE_CASE : Tuple = '''''' while curr_string != "" and curr_string not in lexicon: curr_string += "0" if curr_string != "": __SCREAMING_SNAKE_CASE : Dict = lexicon[curr_string] result += last_match_id return result def a__ ( snake_case , snake_case ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = os.path.getsize(snake_case ) __SCREAMING_SNAKE_CASE : Union[str, Any] = bin(snake_case )[2:] __SCREAMING_SNAKE_CASE : int = len(snake_case ) return "0" * (length_length - 1) + file_length_binary + compressed def a__ ( snake_case , snake_case ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = 8 try: with open(snake_case , '''wb''' ) as opened_file: __SCREAMING_SNAKE_CASE : Optional[int] = [ to_write[i : i + byte_length] for i in range(0 , len(snake_case ) , snake_case ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append('''10000000''' ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array: opened_file.write(int(snake_case , 2 ).to_bytes(1 , byteorder='''big''' ) ) except OSError: print('''File not accessible''' ) sys.exit() def a__ ( snake_case , snake_case ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = read_file_binary(snake_case ) __SCREAMING_SNAKE_CASE : Optional[int] = compress_data(snake_case ) __SCREAMING_SNAKE_CASE : Dict = add_file_length(snake_case , snake_case ) write_file_binary(snake_case , snake_case ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
303
0
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_xlnet import XLNetTokenizer else: __snake_case = None __snake_case = logging.get_logger(__name__) __snake_case = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} __snake_case = { '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/tokenizer.json''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/tokenizer.json''', }, } __snake_case = { '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } __snake_case = '''▁''' # Segments (not really needed) __snake_case = 0 __snake_case = 1 __snake_case = 2 __snake_case = 3 __snake_case = 4 class lowercase ( A__ ): """simple docstring""" _a = VOCAB_FILES_NAMES _a = PRETRAINED_VOCAB_FILES_MAP _a = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _a = 'left' _a = XLNetTokenizer def __init__( self , UpperCamelCase_=None , UpperCamelCase_=None , UpperCamelCase_=False , UpperCamelCase_=True , UpperCamelCase_=False , UpperCamelCase_="<s>" , UpperCamelCase_="</s>" , UpperCamelCase_="<unk>" , UpperCamelCase_="<sep>" , UpperCamelCase_="<pad>" , UpperCamelCase_="<cls>" , UpperCamelCase_="<mask>" , UpperCamelCase_=["<eop>", "<eod>"] , **UpperCamelCase_ , ): '''simple docstring''' UpperCamelCase__ :Union[str, Any] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else mask_token super().__init__( vocab_file=UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , do_lower_case=UpperCamelCase_ , remove_space=UpperCamelCase_ , keep_accents=UpperCamelCase_ , bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , cls_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , **UpperCamelCase_ , ) UpperCamelCase__ :Optional[Any] = 3 UpperCamelCase__ :Any = do_lower_case UpperCamelCase__ :str = remove_space UpperCamelCase__ :Optional[int] = keep_accents UpperCamelCase__ :List[str] = vocab_file UpperCamelCase__ :Dict = False if not self.vocab_file else True def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = None ): '''simple docstring''' UpperCamelCase__ :Tuple = [self.sep_token_id] UpperCamelCase__ :Any = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = None ): '''simple docstring''' UpperCamelCase__ :Optional[int] = [self.sep_token_id] UpperCamelCase__ :Optional[Any] = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(UpperCamelCase_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCamelCase__ :Optional[Any] = os.path.join( UpperCamelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase_ ): copyfile(self.vocab_file , UpperCamelCase_ ) return (out_vocab_file,)
219
'''simple docstring''' # Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING, Dict, Optional import numpy as np import pyarrow as pa from .. import config from ..utils.logging import get_logger from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import jax import jaxlib __snake_case = get_logger() __snake_case = None class lowercase ( TensorFormatter[Mapping, 'jax.Array', Mapping] ): """simple docstring""" def __init__( self , UpperCamelCase_=None , UpperCamelCase_=None , **UpperCamelCase_ ): '''simple docstring''' super().__init__(features=UpperCamelCase_ ) import jax from jaxlib.xla_client import Device if isinstance(UpperCamelCase_ , UpperCamelCase_ ): raise ValueError( F'''Expected {device} to be a `str` not {type(UpperCamelCase_ )}, as `jaxlib.xla_extension.Device` ''' '''is not serializable neither with `pickle` nor with `dill`. Instead you can surround ''' '''the device with `str()` to get its string identifier that will be internally mapped ''' '''to the actual `jaxlib.xla_extension.Device`.''' ) UpperCamelCase__ :Tuple = device if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else str(jax.devices()[0] ) # using global variable since `jaxlib.xla_extension.Device` is not serializable neither # with `pickle` nor with `dill`, so we need to use a global variable instead global DEVICE_MAPPING if DEVICE_MAPPING is None: UpperCamelCase__ :Optional[Any] = self._map_devices_to_str() if self.device not in list(DEVICE_MAPPING.keys() ): logger.warning( F'''Device with string identifier {self.device} not listed among the available ''' F'''devices: {list(DEVICE_MAPPING.keys() )}, so falling back to the default ''' F'''device: {str(jax.devices()[0] )}.''' ) UpperCamelCase__ :Optional[int] = str(jax.devices()[0] ) UpperCamelCase__ :Tuple = jnp_array_kwargs @staticmethod def lowerCAmelCase__ ( ): '''simple docstring''' import jax return {str(UpperCamelCase_ ): device for device in jax.devices()} def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' import jax import jax.numpy as jnp if isinstance(UpperCamelCase_ , UpperCamelCase_ ) and column: if all( isinstance(UpperCamelCase_ , jax.Array ) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column ): return jnp.stack(UpperCamelCase_ , axis=0 ) return column def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' import jax import jax.numpy as jnp if isinstance(UpperCamelCase_ , (str, bytes, type(UpperCamelCase_ )) ): return value elif isinstance(UpperCamelCase_ , (np.character, np.ndarray) ) and np.issubdtype(value.dtype , np.character ): return value.tolist() UpperCamelCase__ :Optional[int] = {} if isinstance(UpperCamelCase_ , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.integer ): # the default int precision depends on the jax config # see https://jax.readthedocs.io/en/latest/notebooks/Common_Gotchas_in_JAX.html#double-64bit-precision if jax.config.jax_enable_xaa: UpperCamelCase__ :List[str] = {'''dtype''': jnp.intaa} else: UpperCamelCase__ :Union[str, Any] = {'''dtype''': jnp.intaa} elif isinstance(UpperCamelCase_ , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.floating ): UpperCamelCase__ :Optional[Any] = {'''dtype''': jnp.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(UpperCamelCase_ , PIL.Image.Image ): UpperCamelCase__ :str = np.asarray(UpperCamelCase_ ) # using global variable since `jaxlib.xla_extension.Device` is not serializable neither # with `pickle` nor with `dill`, so we need to use a global variable instead global DEVICE_MAPPING if DEVICE_MAPPING is None: UpperCamelCase__ :Dict = self._map_devices_to_str() with jax.default_device(DEVICE_MAPPING[self.device] ): # calling jnp.array on a np.ndarray does copy the data # see https://github.com/google/jax/issues/4486 return jnp.array(UpperCamelCase_ , **{**default_dtype, **self.jnp_array_kwargs} ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' import jax # support for torch, tf, jax etc. if config.TORCH_AVAILABLE and "torch" in sys.modules: import torch if isinstance(UpperCamelCase_ , torch.Tensor ): return self._tensorize(data_struct.detach().cpu().numpy()[()] ) if hasattr(UpperCamelCase_ , '''__array__''' ) and not isinstance(UpperCamelCase_ , jax.Array ): UpperCamelCase__ :int = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(UpperCamelCase_ , np.ndarray ): if data_struct.dtype == object: # jax arrays cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(UpperCamelCase_ ) for substruct in data_struct] ) elif isinstance(UpperCamelCase_ , (list, tuple) ): return self._consolidate([self.recursive_tensorize(UpperCamelCase_ ) for substruct in data_struct] ) return self._tensorize(UpperCamelCase_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' return map_nested(self._recursive_tensorize , UpperCamelCase_ , map_list=UpperCamelCase_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :List[Any] = self.numpy_arrow_extractor().extract_row(UpperCamelCase_ ) UpperCamelCase__ :List[Any] = self.python_features_decoder.decode_row(UpperCamelCase_ ) return self.recursive_tensorize(UpperCamelCase_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :List[Any] = self.numpy_arrow_extractor().extract_column(UpperCamelCase_ ) UpperCamelCase__ :Tuple = self.python_features_decoder.decode_column(UpperCamelCase_ , pa_table.column_names[0] ) UpperCamelCase__ :Dict = self.recursive_tensorize(UpperCamelCase_ ) UpperCamelCase__ :str = self._consolidate(UpperCamelCase_ ) return column def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :int = self.numpy_arrow_extractor().extract_batch(UpperCamelCase_ ) UpperCamelCase__ :Union[str, Any] = self.python_features_decoder.decode_batch(UpperCamelCase_ ) UpperCamelCase__ :List[str] = self.recursive_tensorize(UpperCamelCase_ ) for column_name in batch: UpperCamelCase__ :Optional[int] = self._consolidate(batch[column_name] ) return batch
219
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase : Optional[int] = { """configuration_whisper""": ["""WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """WhisperConfig""", """WhisperOnnxConfig"""], """feature_extraction_whisper""": ["""WhisperFeatureExtractor"""], """processing_whisper""": ["""WhisperProcessor"""], """tokenization_whisper""": ["""WhisperTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : str = ["""WhisperTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : str = [ """WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST""", """WhisperForConditionalGeneration""", """WhisperModel""", """WhisperPreTrainedModel""", """WhisperForAudioClassification""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : str = [ """TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFWhisperForConditionalGeneration""", """TFWhisperModel""", """TFWhisperPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Optional[Any] = [ """FlaxWhisperForConditionalGeneration""", """FlaxWhisperModel""", """FlaxWhisperPreTrainedModel""", """FlaxWhisperForAudioClassification""", ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys lowerCAmelCase : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
291
import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import BertGenerationDecoder, BertGenerationEncoder class _lowercase : '''simple docstring''' def __init__( self , snake_case__ , snake_case__=13 , snake_case__=7 , snake_case__=True , snake_case__=True , snake_case__=99 , snake_case__=32 , snake_case__=5 , snake_case__=4 , snake_case__=37 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=50 , snake_case__=0.02 , snake_case__=True , snake_case__=None , ): '''simple docstring''' UpperCamelCase_ = parent UpperCamelCase_ = batch_size UpperCamelCase_ = seq_length UpperCamelCase_ = is_training UpperCamelCase_ = use_input_mask UpperCamelCase_ = vocab_size UpperCamelCase_ = hidden_size UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = num_attention_heads UpperCamelCase_ = intermediate_size UpperCamelCase_ = hidden_act UpperCamelCase_ = hidden_dropout_prob UpperCamelCase_ = attention_probs_dropout_prob UpperCamelCase_ = max_position_embeddings UpperCamelCase_ = initializer_range UpperCamelCase_ = use_labels UpperCamelCase_ = scope def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase_ = None if self.use_input_mask: UpperCamelCase_ = random_attention_mask([self.batch_size, self.seq_length] ) if self.use_labels: UpperCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase_ = self.get_config() return config, input_ids, input_mask, token_labels def _lowerCamelCase ( self ): '''simple docstring''' return BertGenerationConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) def _lowerCamelCase ( self ): '''simple docstring''' ( ( UpperCamelCase_ ) , ( UpperCamelCase_ ) , ( UpperCamelCase_ ) , ( UpperCamelCase_ ) , ) = self.prepare_config_and_inputs() UpperCamelCase_ = True UpperCamelCase_ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, token_labels, encoder_hidden_states, encoder_attention_mask, ) def _lowerCamelCase ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , **snake_case__ , ): '''simple docstring''' UpperCamelCase_ = BertGenerationEncoder(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCamelCase_ = model(snake_case__ , attention_mask=snake_case__ ) UpperCamelCase_ = model(snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , **snake_case__ , ): '''simple docstring''' UpperCamelCase_ = True UpperCamelCase_ = BertGenerationEncoder(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCamelCase_ = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , encoder_attention_mask=snake_case__ , ) UpperCamelCase_ = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , **snake_case__ , ): '''simple docstring''' UpperCamelCase_ = True UpperCamelCase_ = True UpperCamelCase_ = BertGenerationDecoder(config=snake_case__ ).to(snake_case__ ).eval() # first forward pass UpperCamelCase_ = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , encoder_attention_mask=snake_case__ , use_cache=snake_case__ , ) UpperCamelCase_ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids UpperCamelCase_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) UpperCamelCase_ = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and UpperCamelCase_ = torch.cat([input_ids, next_tokens] , dim=-1 ) UpperCamelCase_ = torch.cat([input_mask, next_mask] , dim=-1 ) UpperCamelCase_ = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , encoder_attention_mask=snake_case__ , output_hidden_states=snake_case__ , )["hidden_states"][0] UpperCamelCase_ = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , encoder_attention_mask=snake_case__ , past_key_values=snake_case__ , output_hidden_states=snake_case__ , )["hidden_states"][0] # select random slice UpperCamelCase_ = ids_tensor((1,) , output_from_past.shape[-1] ).item() UpperCamelCase_ = output_from_no_past[:, -3:, random_slice_idx].detach() UpperCamelCase_ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1e-3 ) ) def _lowerCamelCase ( self , snake_case__ , snake_case__ , snake_case__ , snake_case__ , *snake_case__ , ): '''simple docstring''' UpperCamelCase_ = BertGenerationDecoder(snake_case__ ) model.to(snake_case__ ) model.eval() UpperCamelCase_ = model(snake_case__ , attention_mask=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = self.prepare_config_and_inputs() UpperCamelCase_ = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class _lowercase (a_ , a_ , a_ , unittest.TestCase ): '''simple docstring''' lowercase__ = (BertGenerationEncoder, BertGenerationDecoder) if is_torch_available() else () lowercase__ = (BertGenerationDecoder,) if is_torch_available() else () lowercase__ = ( {"""feature-extraction""": BertGenerationEncoder, """text-generation""": BertGenerationDecoder} if is_torch_available() else {} ) def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = BertGenerationEncoderTester(self ) UpperCamelCase_ = ConfigTester(self , config_class=snake_case__ , hidden_size=37 ) def _lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = self.model_tester.prepare_config_and_inputs() UpperCamelCase_ = "bert" self.model_tester.create_and_check_model(snake_case__ , snake_case__ , snake_case__ , snake_case__ ) def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*snake_case__ ) def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_decoder_model_past_large_inputs(*snake_case__ ) def _lowerCamelCase ( self ): '''simple docstring''' ( ( UpperCamelCase_ ) , ( UpperCamelCase_ ) , ( UpperCamelCase_ ) , ( UpperCamelCase_ ) , ( UpperCamelCase_ ) , ( UpperCamelCase_ ) , ) = self.model_tester.prepare_config_and_inputs_for_decoder() UpperCamelCase_ = None self.model_tester.create_and_check_model_as_decoder( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , ) def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_for_causal_lm(*snake_case__ ) @slow def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = BertGenerationEncoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) self.assertIsNotNone(snake_case__ ) @require_torch class _lowercase (unittest.TestCase ): '''simple docstring''' @slow def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = BertGenerationEncoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) UpperCamelCase_ = torch.tensor([[101, 7592, 1010, 2026, 3899, 2003, 1_0140, 102]] ) with torch.no_grad(): UpperCamelCase_ = model(snake_case__ )[0] UpperCamelCase_ = torch.Size([1, 8, 1024] ) self.assertEqual(output.shape , snake_case__ ) UpperCamelCase_ = torch.tensor( [[[0.1_775, 0.0_083, -0.0_321], [1.6_002, 0.1_287, 0.3_912], [2.1_473, 0.5_791, 0.6_066]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case__ , atol=1e-4 ) ) @require_torch class _lowercase (unittest.TestCase ): '''simple docstring''' @slow def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = BertGenerationDecoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) UpperCamelCase_ = torch.tensor([[101, 7592, 1010, 2026, 3899, 2003, 1_0140, 102]] ) with torch.no_grad(): UpperCamelCase_ = model(snake_case__ )[0] UpperCamelCase_ = torch.Size([1, 8, 5_0358] ) self.assertEqual(output.shape , snake_case__ ) UpperCamelCase_ = torch.tensor( [[[-0.5_788, -2.5_994, -3.7_054], [0.0_438, 4.7_997, 1.8_795], [1.5_862, 6.6_409, 4.4_638]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case__ , atol=1e-4 ) )
128
0
import inspect import unittest from transformers import RegNetConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import RegNetForImageClassification, RegNetModel from transformers.models.regnet.modeling_regnet import REGNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class UpperCAmelCase : '''simple docstring''' def __init__( self : int ,A : List[Any] ,A : int=3 ,A : Optional[Any]=32 ,A : Any=3 ,A : Optional[Any]=10 ,A : Dict=[10, 20, 30, 40] ,A : Optional[int]=[1, 1, 2, 1] ,A : List[str]=True ,A : List[Any]=True ,A : Tuple="relu" ,A : Optional[int]=3 ,A : List[str]=None ,): __A = parent __A = batch_size __A = image_size __A = num_channels __A = embeddings_size __A = hidden_sizes __A = depths __A = is_training __A = use_labels __A = hidden_act __A = num_labels __A = scope __A = len(UpperCamelCase__ ) def UpperCamelCase_ ( self : Tuple ): __A = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __A = None if self.use_labels: __A = ids_tensor([self.batch_size] ,self.num_labels ) __A = self.get_config() return config, pixel_values, labels def UpperCamelCase_ ( self : Optional[Any] ): return RegNetConfig( num_channels=self.num_channels ,embeddings_size=self.embeddings_size ,hidden_sizes=self.hidden_sizes ,depths=self.depths ,hidden_act=self.hidden_act ,num_labels=self.num_labels ,) def UpperCamelCase_ ( self : List[Any] ,A : Any ,A : List[str] ,A : Dict ): __A = RegNetModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __A = model(UpperCamelCase__ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) ,) def UpperCamelCase_ ( self : Any ,A : Optional[Any] ,A : Optional[Any] ,A : Tuple ): __A = self.num_labels __A = RegNetForImageClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __A = model(UpperCamelCase__ ,labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def UpperCamelCase_ ( self : Optional[Any] ): __A = self.prepare_config_and_inputs() __A = config_and_inputs __A = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class UpperCAmelCase ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' snake_case_ = (RegNetModel, RegNetForImageClassification) if is_torch_available() else () snake_case_ = ( {"feature-extraction": RegNetModel, "image-classification": RegNetForImageClassification} if is_torch_available() else {} ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def UpperCamelCase_ ( self : Optional[int] ): __A = RegNetModelTester(self ) __A = ConfigTester(self ,config_class=UpperCamelCase__ ,has_text_modality=UpperCamelCase__ ) def UpperCamelCase_ ( self : Tuple ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase_ ( self : Dict ): return @unittest.skip(reason="RegNet does not use inputs_embeds" ) def UpperCamelCase_ ( self : Optional[Any] ): pass @unittest.skip(reason="RegNet does not support input and output embeddings" ) def UpperCamelCase_ ( self : List[Any] ): pass def UpperCamelCase_ ( self : Optional[Any] ): __A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __A = model_class(UpperCamelCase__ ) __A = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __A = [*signature.parameters.keys()] __A = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,UpperCamelCase__ ) def UpperCamelCase_ ( self : Optional[Any] ): __A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def UpperCamelCase_ ( self : Optional[int] ): __A = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __A = model_class(config=UpperCamelCase__ ) for name, module in model.named_modules(): if isinstance(UpperCamelCase__ ,(nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) ,msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' ,) self.assertTrue( torch.all(module.bias == 0 ) ,msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' ,) def UpperCamelCase_ ( self : List[str] ): def check_hidden_states_output(A : str ,A : Optional[Any] ,A : Union[str, Any] ): __A = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): __A = model(**self._prepare_for_class(UpperCamelCase__ ,UpperCamelCase__ ) ) __A = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __A = self.model_tester.num_stages self.assertEqual(len(UpperCamelCase__ ) ,expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[self.model_tester.image_size // 2, self.model_tester.image_size // 2] ,) __A = self.model_tester.prepare_config_and_inputs_for_common() __A = ['''basic''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: __A = layer_type __A = True check_hidden_states_output(UpperCamelCase__ ,UpperCamelCase__ ,UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __A = True check_hidden_states_output(UpperCamelCase__ ,UpperCamelCase__ ,UpperCamelCase__ ) def UpperCamelCase_ ( self : List[str] ): __A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase__ ) @slow def UpperCamelCase_ ( self : List[Any] ): for model_name in REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __A = RegNetModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def UpperCAmelCase ( ) -> str: """simple docstring""" __A = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase_ ( self : Union[str, Any] ): return ( AutoImageProcessor.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCamelCase_ ( self : Any ): __A = RegNetForImageClassification.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(UpperCamelCase__ ) __A = self.default_image_processor __A = prepare_img() __A = image_processor(images=UpperCamelCase__ ,return_tensors="pt" ).to(UpperCamelCase__ ) # forward pass with torch.no_grad(): __A = model(**UpperCamelCase__ ) # verify the logits __A = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape ,UpperCamelCase__ ) __A = torch.tensor([-0.41_80, -1.50_51, -3.48_36] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,UpperCamelCase__ ,atol=1E-4 ) )
367
import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class UpperCAmelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' snake_case_ = CTRLTokenizer snake_case_ = False snake_case_ = False def UpperCamelCase_ ( self : Dict ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __A = ["adapt", "re@@", "a@@", "apt", "c@@", "t", "<unk>"] __A = dict(zip(A ,range(len(A ) ) ) ) __A = ["#version: 0.2", "a p", "ap t</w>", "r e", "a d", "ad apt</w>", ""] __A = {"unk_token": "<unk>"} __A = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["vocab_file"] ) __A = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file ,"w" ,encoding="utf-8" ) as fp: fp.write(json.dumps(A ) + "\n" ) with open(self.merges_file ,"w" ,encoding="utf-8" ) as fp: fp.write("\n".join(A ) ) def UpperCamelCase_ ( self : List[str] ,**A : List[str] ): kwargs.update(self.special_tokens_map ) return CTRLTokenizer.from_pretrained(self.tmpdirname ,**A ) def UpperCamelCase_ ( self : Optional[int] ,A : Tuple ): __A = "adapt react readapt apt" __A = "adapt react readapt apt" return input_text, output_text def UpperCamelCase_ ( self : Any ): __A = CTRLTokenizer(self.vocab_file ,self.merges_file ,**self.special_tokens_map ) __A = "adapt react readapt apt" __A = "adapt re@@ a@@ c@@ t re@@ adapt apt".split() __A = tokenizer.tokenize(A ) self.assertListEqual(A ,A ) __A = tokens + [tokenizer.unk_token] __A = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) ,A )
124
0
import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): __A = get_tests_dir("fixtures/test_sentencepiece.model") if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right __A = 12_8022 __A = 12_8028 @require_sentencepiece class _SCREAMING_SNAKE_CASE ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' lowercase_ = MaMaaaTokenizer lowercase_ = False lowercase_ = False lowercase_ = True def SCREAMING_SNAKE_CASE_ (self : Dict) ->Tuple: '''simple docstring''' super().setUp() lowerCamelCase__: List[Any] =["</s>", "<unk>", "▁This", "▁is", "▁a", "▁t", "est", "\u0120", "<pad>"] lowerCamelCase__: Optional[Any] =dict(zip(UpperCAmelCase_ , range(len(UpperCAmelCase_)))) lowerCamelCase__: int =Path(self.tmpdirname) save_json(UpperCAmelCase_ , save_dir / VOCAB_FILES_NAMES["vocab_file"]) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(UpperCAmelCase_ , save_dir / VOCAB_FILES_NAMES["spm_file"]) lowerCamelCase__: List[Any] =MaMaaaTokenizer.from_pretrained(self.tmpdirname) tokenizer.save_pretrained(self.tmpdirname) def SCREAMING_SNAKE_CASE_ (self : Optional[int] , **UpperCAmelCase_ : Any) ->str: '''simple docstring''' return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : int , UpperCAmelCase_ : Any) ->Optional[Any]: '''simple docstring''' return ( "This is a test", "This is a test", ) def SCREAMING_SNAKE_CASE_ (self : str) ->Dict: '''simple docstring''' lowerCamelCase__: Dict ="</s>" lowerCamelCase__: int =0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase_) , UpperCAmelCase_) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase_) , UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : Any) ->Optional[Any]: '''simple docstring''' lowerCamelCase__: Any =self.get_tokenizer() lowerCamelCase__: Dict =list(tokenizer.get_vocab().keys()) self.assertEqual(vocab_keys[0] , "</s>") self.assertEqual(vocab_keys[1] , "<unk>") self.assertEqual(vocab_keys[-1] , "<s>") self.assertEqual(len(UpperCAmelCase_) , tokenizer.vocab_size + len(tokenizer.get_added_vocab())) @unittest.skip("Skip this test while all models are still to be uploaded.") def SCREAMING_SNAKE_CASE_ (self : List[Any]) ->List[str]: '''simple docstring''' pass def SCREAMING_SNAKE_CASE_ (self : List[str]) ->Dict: '''simple docstring''' lowerCamelCase__: Optional[Any] =self.get_tokenizer() lowerCamelCase__: Any =tokenizer.tokenize("This is a test") self.assertListEqual(UpperCAmelCase_ , ["▁This", "▁is", "▁a", "▁t", "est"]) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase_) , [2, 3, 4, 5, 6] , ) lowerCamelCase__: List[Any] =tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6]) self.assertListEqual(UpperCAmelCase_ , ["▁This", "▁is", "▁a", "▁t", "est"]) lowerCamelCase__: int =tokenizer.convert_tokens_to_string(UpperCAmelCase_) self.assertEqual(UpperCAmelCase_ , "This is a test") @slow def SCREAMING_SNAKE_CASE_ (self : int) ->Union[str, Any]: '''simple docstring''' lowerCamelCase__: Optional[Any] ={"input_ids": [[128_022, 110_108, 397, 11, 38_272, 2_247, 124_811, 285, 18_105, 1_586, 207, 7, 39_534, 4_428, 397, 1_019, 18_105, 1_586, 207, 7, 41_337, 16_786, 241, 7, 20_214, 17, 125_690, 10_398, 7, 44_378, 58_069, 68_342, 7_798, 7_343, 11, 299, 33_310, 4, 158, 37_350, 94_077, 4_569, 299, 33_310, 90, 4, 52_840, 290, 4, 31_270, 112, 299, 682, 4, 52_840, 39_953, 14_079, 193, 52_519, 90_894, 17_894, 120_697, 11, 40_445, 551, 17, 1_019, 52_519, 90_894, 17_756, 963, 11, 40_445, 480, 17, 9_792, 1_120, 5_173, 1_393, 6_240, 16_786, 241, 120_996, 28, 1_245, 1_393, 118_240, 11_123, 1_019, 93_612, 2_691, 10_618, 98_058, 120_409, 1_928, 279, 4, 40_683, 367, 178, 207, 1_019, 103, 103_121, 506, 65_296, 5, 2], [128_022, 21_217, 367, 117, 125_450, 128, 719, 7, 7_308, 40, 93_612, 12_669, 1_116, 16_704, 71, 17_785, 3_699, 15_592, 35, 144, 9_584, 241, 11_943, 713, 950, 799, 2_247, 88_427, 150, 149, 118_813, 120_706, 1_019, 106_906, 81_518, 28, 1_224, 22_799, 397, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [128_022, 1_658, 123_311, 5_155, 5_578, 4_722, 279, 14_947, 2_366, 1_120, 1_197, 14, 1_348, 9_232, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase_ , model_name="facebook/m2m100_418M" , revision="c168bae485c864188cf9aa0e4108b0b6934dc91e" , ) @require_torch @require_sentencepiece @require_tokenizers class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' lowercase_ = "facebook/m2m100_418M" lowercase_ = [ "In my opinion, there are two levels of response from the French government.", "NSA Affair Emphasizes Complete Lack of Debate on Intelligence", ] lowercase_ = [ "Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.", "L'affaire NSA souligne l'absence totale de débat sur le renseignement", ] # fmt: off lowercase_ = [EN_CODE, 593, 1949, 11_5781, 4, 7_1586, 4234, 6_0633, 12_6233, 432, 12_3808, 1_5592, 1197, 11_7132, 12_0618, 5, 2] @classmethod def SCREAMING_SNAKE_CASE_ (cls : Optional[Any]) ->Dict: '''simple docstring''' lowerCamelCase__: MaMaaaTokenizer =MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en" , tgt_lang="fr") lowerCamelCase__: List[Any] =1 return cls def SCREAMING_SNAKE_CASE_ (self : List[Any]) ->Any: '''simple docstring''' self.assertEqual(self.tokenizer.get_lang_id("ar") , 128_006) self.assertEqual(self.tokenizer.get_lang_id("en") , 128_022) self.assertEqual(self.tokenizer.get_lang_id("ro") , 128_076) self.assertEqual(self.tokenizer.get_lang_id("mr") , 128_063) def SCREAMING_SNAKE_CASE_ (self : Dict) ->List[Any]: '''simple docstring''' lowerCamelCase__: List[str] =self.tokenizer.get_vocab() self.assertEqual(len(UpperCAmelCase_) , self.tokenizer.vocab_size) self.assertEqual(vocab["<unk>"] , 3) self.assertIn(self.tokenizer.get_lang_token("en") , UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : int) ->Dict: '''simple docstring''' lowerCamelCase__: str ="en" lowerCamelCase__: int =self.tokenizer.batch_encode_plus(self.src_text).input_ids[0] self.assertListEqual(self.expected_src_tokens , UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : Dict) ->Dict: '''simple docstring''' self.assertIn(UpperCAmelCase_ , self.tokenizer.all_special_ids) # fmt: off lowerCamelCase__: Dict =[FR_CODE, 5_364, 82, 8_642, 4, 294, 47, 8, 14_028, 136, 3_286, 9_706, 6, 90_797, 6, 144_012, 162, 88_128, 30_061, 5, 2] # fmt: on lowerCamelCase__: int =self.tokenizer.decode(UpperCAmelCase_ , skip_special_tokens=UpperCAmelCase_) lowerCamelCase__: Dict =self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=UpperCAmelCase_) self.assertEqual(UpperCAmelCase_ , UpperCAmelCase_) self.assertNotIn(self.tokenizer.eos_token , UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : Optional[Any]) ->Optional[Any]: '''simple docstring''' lowerCamelCase__: int =tempfile.mkdtemp() lowerCamelCase__: Any =self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(UpperCAmelCase_) lowerCamelCase__: Tuple =MaMaaaTokenizer.from_pretrained(UpperCAmelCase_) self.assertDictEqual(new_tok.lang_token_to_id , UpperCAmelCase_) @require_torch def SCREAMING_SNAKE_CASE_ (self : List[str]) ->Union[str, Any]: '''simple docstring''' lowerCamelCase__: Union[str, Any] ="en" lowerCamelCase__: Any ="fr" lowerCamelCase__: Optional[Any] =self.tokenizer(self.src_text , text_target=self.tgt_text , padding=UpperCAmelCase_ , return_tensors="pt") lowerCamelCase__: List[str] =shift_tokens_right( batch["labels"] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id) for k in batch: lowerCamelCase__: Any =batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def SCREAMING_SNAKE_CASE_ (self : Tuple) ->Optional[int]: '''simple docstring''' lowerCamelCase__: Tuple ="mr" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr")]) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id]) lowerCamelCase__: List[Any] ="zh" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh")]) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id]) @require_torch def SCREAMING_SNAKE_CASE_ (self : Union[str, Any]) ->List[str]: '''simple docstring''' lowerCamelCase__: Any ="mr" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr")]) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id]) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang)]) lowerCamelCase__: Tuple ="zh" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh")]) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id]) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang)]) @require_torch def SCREAMING_SNAKE_CASE_ (self : Tuple) ->Union[str, Any]: '''simple docstring''' lowerCamelCase__: List[str] =self.tokenizer._build_translation_inputs("A test" , return_tensors="pt" , src_lang="en" , tgt_lang="ar") self.assertEqual( nested_simplify(UpperCAmelCase_) , { # en_XX, A, test, EOS "input_ids": [[128_022, 58, 4_183, 2]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 128_006, } , )
10
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import YolosConfig, YolosForObjectDetection, YolosImageProcessor from transformers.utils import logging logging.set_verbosity_info() __A = logging.get_logger(__name__) def lowerCAmelCase_ ( __a ) -> YolosConfig: """simple docstring""" lowerCamelCase__: str =YolosConfig() # size of the architecture if "yolos_ti" in yolos_name: lowerCamelCase__: int =192 lowerCamelCase__: Optional[int] =768 lowerCamelCase__: Any =12 lowerCamelCase__: str =3 lowerCamelCase__: Optional[int] =[800, 1333] lowerCamelCase__: Union[str, Any] =False elif yolos_name == "yolos_s_dWr": lowerCamelCase__: int =330 lowerCamelCase__: Optional[Any] =14 lowerCamelCase__: Any =6 lowerCamelCase__: List[str] =1320 elif "yolos_s" in yolos_name: lowerCamelCase__: List[str] =384 lowerCamelCase__: Union[str, Any] =1536 lowerCamelCase__: List[Any] =12 lowerCamelCase__: Any =6 elif "yolos_b" in yolos_name: lowerCamelCase__: str =[800, 1344] lowerCamelCase__: int =91 lowerCamelCase__: str ="huggingface/label-files" lowerCamelCase__: List[str] ="coco-detection-id2label.json" lowerCamelCase__: Tuple =json.load(open(hf_hub_download(__a , __a , repo_type="dataset" ) , "r" ) ) lowerCamelCase__: Dict ={int(__a ): v for k, v in idalabel.items()} lowerCamelCase__: List[str] =idalabel lowerCamelCase__: int ={v: k for k, v in idalabel.items()} return config def lowerCAmelCase_ ( __a , __a , __a = False ) -> Dict: """simple docstring""" for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCamelCase__: Optional[int] =state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) lowerCamelCase__: Dict =state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__: Union[str, Any] =in_proj_weight[: config.hidden_size, :] lowerCamelCase__: str =in_proj_bias[: config.hidden_size] lowerCamelCase__: str =in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCamelCase__: str =in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCamelCase__: Optional[int] =in_proj_weight[-config.hidden_size :, :] lowerCamelCase__: List[Any] =in_proj_bias[-config.hidden_size :] def lowerCAmelCase_ ( __a ) -> str: """simple docstring""" if "backbone" in name: lowerCamelCase__: Optional[Any] =name.replace("backbone" , "vit" ) if "cls_token" in name: lowerCamelCase__: Optional[int] =name.replace("cls_token" , "embeddings.cls_token" ) if "det_token" in name: lowerCamelCase__: str =name.replace("det_token" , "embeddings.detection_tokens" ) if "mid_pos_embed" in name: lowerCamelCase__: Tuple =name.replace("mid_pos_embed" , "encoder.mid_position_embeddings" ) if "pos_embed" in name: lowerCamelCase__: Any =name.replace("pos_embed" , "embeddings.position_embeddings" ) if "patch_embed.proj" in name: lowerCamelCase__: List[Any] =name.replace("patch_embed.proj" , "embeddings.patch_embeddings.projection" ) if "blocks" in name: lowerCamelCase__: Union[str, Any] =name.replace("blocks" , "encoder.layer" ) if "attn.proj" in name: lowerCamelCase__: Any =name.replace("attn.proj" , "attention.output.dense" ) if "attn" in name: lowerCamelCase__: Optional[int] =name.replace("attn" , "attention.self" ) if "norm1" in name: lowerCamelCase__: int =name.replace("norm1" , "layernorm_before" ) if "norm2" in name: lowerCamelCase__: int =name.replace("norm2" , "layernorm_after" ) if "mlp.fc1" in name: lowerCamelCase__: List[str] =name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: lowerCamelCase__: Any =name.replace("mlp.fc2" , "output.dense" ) if "class_embed" in name: lowerCamelCase__: Dict =name.replace("class_embed" , "class_labels_classifier" ) if "bbox_embed" in name: lowerCamelCase__: List[str] =name.replace("bbox_embed" , "bbox_predictor" ) if "vit.norm" in name: lowerCamelCase__: Any =name.replace("vit.norm" , "vit.layernorm" ) return name def lowerCAmelCase_ ( __a , __a ) -> dict: """simple docstring""" for key in orig_state_dict.copy().keys(): lowerCamelCase__: Any =orig_state_dict.pop(__a ) if "qkv" in key: lowerCamelCase__: Tuple =key.split("." ) lowerCamelCase__: List[str] =int(key_split[2] ) lowerCamelCase__: Tuple =model.vit.encoder.layer[layer_num].attention.attention.all_head_size if "weight" in key: lowerCamelCase__: int =val[:dim, :] lowerCamelCase__: str =val[ dim : dim * 2, : ] lowerCamelCase__: Any =val[-dim:, :] else: lowerCamelCase__: Tuple =val[:dim] lowerCamelCase__: Optional[Any] =val[dim : dim * 2] lowerCamelCase__: str =val[-dim:] else: lowerCamelCase__: Dict =val return orig_state_dict def lowerCAmelCase_ ( ) -> torch.Tensor: """simple docstring""" lowerCamelCase__: Any ="http://images.cocodataset.org/val2017/000000039769.jpg" lowerCamelCase__: Optional[Any] =Image.open(requests.get(__a , stream=__a ).raw ) return im @torch.no_grad() def lowerCAmelCase_ ( __a , __a , __a , __a = False ) -> List[str]: """simple docstring""" lowerCamelCase__: int =get_yolos_config(__a ) # load original state_dict lowerCamelCase__: Optional[int] =torch.load(__a , map_location="cpu" )["model"] # load 🤗 model lowerCamelCase__: int =YolosForObjectDetection(__a ) model.eval() lowerCamelCase__: Union[str, Any] =convert_state_dict(__a , __a ) model.load_state_dict(__a ) # Check outputs on an image, prepared by YolosImageProcessor lowerCamelCase__: Any =800 if yolos_name != "yolos_ti" else 512 lowerCamelCase__: Tuple =YolosImageProcessor(format="coco_detection" , size=__a ) lowerCamelCase__: str =image_processor(images=prepare_img() , return_tensors="pt" ) lowerCamelCase__: Tuple =model(**__a ) lowerCamelCase__ , lowerCamelCase__: List[str] =outputs.logits, outputs.pred_boxes lowerCamelCase__ , lowerCamelCase__: Any =None, None if yolos_name == "yolos_ti": lowerCamelCase__: Optional[Any] =torch.tensor( [[-3_9.5_0_2_2, -1_1.9_8_2_0, -1_7.6_8_8_8], [-2_9.9_5_7_4, -9.9_7_6_9, -1_7.7_6_9_1], [-4_2.3_2_8_1, -2_0.7_2_0_0, -3_0.6_2_9_4]] ) lowerCamelCase__: List[Any] =torch.tensor( [[0.4_0_2_1, 0.0_8_3_6, 0.7_9_7_9], [0.0_1_8_4, 0.2_6_0_9, 0.0_3_6_4], [0.1_7_8_1, 0.2_0_0_4, 0.2_0_9_5]] ) elif yolos_name == "yolos_s_200_pre": lowerCamelCase__: Optional[int] =torch.tensor( [[-2_4.0_2_4_8, -1_0.3_0_2_4, -1_4.8_2_9_0], [-4_2.0_3_9_2, -1_6.8_2_0_0, -2_7.4_3_3_4], [-2_7.2_7_4_3, -1_1.8_1_5_4, -1_8.7_1_4_8]] ) lowerCamelCase__: Any =torch.tensor( [[0.2_5_5_9, 0.5_4_5_5, 0.4_7_0_6], [0.2_9_8_9, 0.7_2_7_9, 0.1_8_7_5], [0.7_7_3_2, 0.4_0_1_7, 0.4_4_6_2]] ) elif yolos_name == "yolos_s_300_pre": lowerCamelCase__: str =torch.tensor( [[-3_6.2_2_2_0, -1_4.4_3_8_5, -2_3.5_4_5_7], [-3_5.6_9_7_0, -1_4.7_5_8_3, -2_1.3_9_3_5], [-3_1.5_9_3_9, -1_3.6_0_4_2, -1_6.8_0_4_9]] ) lowerCamelCase__: Optional[Any] =torch.tensor( [[0.7_6_1_4, 0.2_3_1_6, 0.4_7_2_8], [0.7_1_6_8, 0.4_4_9_5, 0.3_8_5_5], [0.4_9_9_6, 0.1_4_6_6, 0.9_9_9_6]] ) elif yolos_name == "yolos_s_dWr": lowerCamelCase__: str =torch.tensor( [[-4_2.8_6_6_8, -2_4.1_0_4_9, -4_1.1_6_9_0], [-3_4.7_4_5_6, -1_4.1_2_7_4, -2_4.9_1_9_4], [-3_3.7_8_9_8, -1_2.1_9_4_6, -2_5.6_4_9_5]] ) lowerCamelCase__: Union[str, Any] =torch.tensor( [[0.5_5_8_7, 0.2_7_7_3, 0.0_6_0_5], [0.5_0_0_4, 0.3_0_1_4, 0.9_9_9_4], [0.4_9_9_9, 0.1_5_4_8, 0.9_9_9_4]] ) elif yolos_name == "yolos_base": lowerCamelCase__: Tuple =torch.tensor( [[-4_0.6_0_6_4, -2_4.3_0_8_4, -3_2.6_4_4_7], [-5_5.1_9_9_0, -3_0.7_7_1_9, -3_5.5_8_7_7], [-5_1.4_3_1_1, -3_3.3_5_0_7, -3_5.6_4_6_2]] ) lowerCamelCase__: Optional[int] =torch.tensor( [[0.5_5_5_5, 0.2_7_9_4, 0.0_6_5_5], [0.9_0_4_9, 0.2_6_6_4, 0.1_8_9_4], [0.9_1_8_3, 0.1_9_8_4, 0.1_6_3_5]] ) else: raise ValueError(F"""Unknown yolos_name: {yolos_name}""" ) assert torch.allclose(logits[0, :3, :3] , __a , atol=1e-4 ) assert torch.allclose(pred_boxes[0, :3, :3] , __a , atol=1e-4 ) Path(__a ).mkdir(exist_ok=__a ) print(F"""Saving model {yolos_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__a ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__a ) if push_to_hub: lowerCamelCase__: Any ={ "yolos_ti": "yolos-tiny", "yolos_s_200_pre": "yolos-small", "yolos_s_300_pre": "yolos-small-300", "yolos_s_dWr": "yolos-small-dwr", "yolos_base": "yolos-base", } print("Pushing to the hub..." ) lowerCamelCase__: Optional[int] =model_mapping[yolos_name] image_processor.push_to_hub(__a , organization="hustvl" ) model.push_to_hub(__a , organization="hustvl" ) if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument( "--yolos_name", default="yolos_s_200_pre", type=str, help=( "Name of the YOLOS model you'd like to convert. Should be one of 'yolos_ti', 'yolos_s_200_pre'," " 'yolos_s_300_pre', 'yolos_s_dWr', 'yolos_base'." ), ) parser.add_argument( "--checkpoint_path", default=None, type=str, help="Path to the original state dict (.pth file)." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) __A = parser.parse_args() convert_yolos_checkpoint(args.yolos_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
10
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available SCREAMING_SNAKE_CASE_ = { """configuration_roc_bert""": ["""ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """RoCBertConfig"""], """tokenization_roc_bert""": ["""RoCBertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: pass try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ = [ """ROC_BERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """RoCBertForCausalLM""", """RoCBertForMaskedLM""", """RoCBertForMultipleChoice""", """RoCBertForPreTraining""", """RoCBertForQuestionAnswering""", """RoCBertForSequenceClassification""", """RoCBertForTokenClassification""", """RoCBertLayer""", """RoCBertModel""", """RoCBertPreTrainedModel""", """load_tf_weights_in_roc_bert""", ] if TYPE_CHECKING: from .configuration_roc_bert import ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RoCBertConfig from .tokenization_roc_bert import RoCBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: raise OptionalDependencyNotAvailable() try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roc_bert import ( ROC_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, RoCBertForCausalLM, RoCBertForMaskedLM, RoCBertForMultipleChoice, RoCBertForPreTraining, RoCBertForQuestionAnswering, RoCBertForSequenceClassification, RoCBertForTokenClassification, RoCBertLayer, RoCBertModel, RoCBertPreTrainedModel, load_tf_weights_in_roc_bert, ) else: import sys SCREAMING_SNAKE_CASE_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
193
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) SCREAMING_SNAKE_CASE_ = """\ Text data. Second line of data.""" SCREAMING_SNAKE_CASE_ = """file""" @pytest.fixture(scope="""session""" ) def __lowercase ( _SCREAMING_SNAKE_CASE ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = tmp_path_factory.mktemp("""data""" ) / (FILE_PATH + """.zstd""") SCREAMING_SNAKE_CASE = bytes(_SCREAMING_SNAKE_CASE , """utf-8""" ) with zstd.open(_SCREAMING_SNAKE_CASE , """wb""" ) as f: f.write(_SCREAMING_SNAKE_CASE ) return path @pytest.fixture def __lowercase ( _SCREAMING_SNAKE_CASE ) -> List[Any]: '''simple docstring''' with open(os.path.join(tmpfs.local_root_dir , _SCREAMING_SNAKE_CASE ) , """w""" ) as f: f.write(_SCREAMING_SNAKE_CASE ) return FILE_PATH @pytest.mark.parametrize("""compression_format""" , ["""gzip""", """xz""", """zstd"""] ) def __lowercase ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Tuple: '''simple docstring''' SCREAMING_SNAKE_CASE = {"""gzip""": gz_file, """xz""": xz_file, """zstd""": zstd_path} SCREAMING_SNAKE_CASE = input_paths[compression_format] SCREAMING_SNAKE_CASE = tmp_path / """cache""" SCREAMING_SNAKE_CASE = DownloadConfig(cache_dir=_SCREAMING_SNAKE_CASE , extract_compressed_file=_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE = cached_path(_SCREAMING_SNAKE_CASE , download_config=_SCREAMING_SNAKE_CASE ) with open(_SCREAMING_SNAKE_CASE ) as f: SCREAMING_SNAKE_CASE = f.read() with open(_SCREAMING_SNAKE_CASE ) as f: SCREAMING_SNAKE_CASE = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("""default_extracted""" , [True, False] ) @pytest.mark.parametrize("""default_cache_dir""" , [True, False] ) def __lowercase ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = """custom_cache""" SCREAMING_SNAKE_CASE = """custom_extracted_dir""" SCREAMING_SNAKE_CASE = tmp_path / """custom_extracted_path""" if default_extracted: SCREAMING_SNAKE_CASE = ("""downloads""" if default_cache_dir else custom_cache_dir, """extracted""") else: monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_DIR""" , _SCREAMING_SNAKE_CASE ) monkeypatch.setattr("""datasets.config.EXTRACTED_DATASETS_PATH""" , str(_SCREAMING_SNAKE_CASE ) ) SCREAMING_SNAKE_CASE = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) SCREAMING_SNAKE_CASE = xz_file SCREAMING_SNAKE_CASE = ( DownloadConfig(extract_compressed_file=_SCREAMING_SNAKE_CASE ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=_SCREAMING_SNAKE_CASE ) ) SCREAMING_SNAKE_CASE = cached_path(_SCREAMING_SNAKE_CASE , download_config=_SCREAMING_SNAKE_CASE ) assert Path(_SCREAMING_SNAKE_CASE ).parent.parts[-2:] == expected def __lowercase ( _SCREAMING_SNAKE_CASE ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE = str(Path(_SCREAMING_SNAKE_CASE ).resolve() ) assert cached_path(_SCREAMING_SNAKE_CASE ) == text_file # relative path SCREAMING_SNAKE_CASE = str(Path(_SCREAMING_SNAKE_CASE ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(_SCREAMING_SNAKE_CASE ) == text_file def __lowercase ( _SCREAMING_SNAKE_CASE ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = str(tmp_path.resolve() / """__missing_file__.txt""" ) with pytest.raises(_SCREAMING_SNAKE_CASE ): cached_path(_SCREAMING_SNAKE_CASE ) # relative path SCREAMING_SNAKE_CASE = """./__missing_file__.txt""" with pytest.raises(_SCREAMING_SNAKE_CASE ): cached_path(_SCREAMING_SNAKE_CASE ) def __lowercase ( _SCREAMING_SNAKE_CASE ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE = get_from_cache(F"""tmp://{tmpfs_file}""" ) with open(_SCREAMING_SNAKE_CASE ) as f: SCREAMING_SNAKE_CASE = f.read() assert output_file_content == FILE_CONTENT @patch("""datasets.config.HF_DATASETS_OFFLINE""" , _SCREAMING_SNAKE_CASE ) def __lowercase ( ) -> Dict: '''simple docstring''' with pytest.raises(_SCREAMING_SNAKE_CASE ): cached_path("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , _SCREAMING_SNAKE_CASE ) def __lowercase ( _SCREAMING_SNAKE_CASE ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(_SCREAMING_SNAKE_CASE ): http_get("""https://huggingface.co""" , temp_file=_SCREAMING_SNAKE_CASE ) with pytest.raises(_SCREAMING_SNAKE_CASE ): http_head("""https://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , _SCREAMING_SNAKE_CASE ) def __lowercase ( _SCREAMING_SNAKE_CASE ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(_SCREAMING_SNAKE_CASE ): ftp_get("""ftp://huggingface.co""" , temp_file=_SCREAMING_SNAKE_CASE ) with pytest.raises(_SCREAMING_SNAKE_CASE ): ftp_head("""ftp://huggingface.co""" ) @patch("""datasets.config.HF_DATASETS_OFFLINE""" , _SCREAMING_SNAKE_CASE ) def __lowercase ( _SCREAMING_SNAKE_CASE ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE = tmp_path_factory.mktemp("""data""" ) / """file.html""" with pytest.raises(_SCREAMING_SNAKE_CASE ): fsspec_get("""s3://huggingface.co""" , temp_file=_SCREAMING_SNAKE_CASE ) with pytest.raises(_SCREAMING_SNAKE_CASE ): fsspec_head("""s3://huggingface.co""" )
193
1
import contextlib import csv import json import os import sqlitea import tarfile import textwrap import zipfile import pyarrow as pa import pyarrow.parquet as pq import pytest import datasets import datasets.config @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( ): _A : str = 10 _A : Any = datasets.Features( { """tokens""": datasets.Sequence(datasets.Value("""string""" ) ), """labels""": datasets.Sequence(datasets.ClassLabel(names=["""negative""", """positive"""] ) ), """answers""": datasets.Sequence( { """text""": datasets.Value("""string""" ), """answer_start""": datasets.Value("""int32""" ), } ), """id""": datasets.Value("""int64""" ), } ) _A : Union[str, Any] = datasets.Dataset.from_dict( { """tokens""": [["""foo"""] * 5] * n, """labels""": [[1] * 5] * n, """answers""": [{"""answer_start""": [97], """text""": ["""1976"""]}] * 10, """id""": list(range(_a ) ), },features=_a,) return dataset @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_ ): _A : Optional[int] = str(tmp_path_factory.mktemp("""data""" ) / """file.arrow""" ) dataset.map(cache_file_name=_a ) return filename # FILE_CONTENT + files _snake_case = "\\n Text data.\n Second line of data." @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : int = tmp_path_factory.mktemp("""data""" ) / """file.txt""" _A : Dict = FILE_CONTENT with open(_a,"""w""" ) as f: f.write(_a ) return filename @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): import bza _A : Dict = tmp_path_factory.mktemp("""data""" ) / """file.txt.bz2""" _A : Tuple = bytes(_a,"""utf-8""" ) with bza.open(_a,"""wb""" ) as f: f.write(_a ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): import gzip _A : Any = str(tmp_path_factory.mktemp("""data""" ) / """file.txt.gz""" ) _A : List[str] = bytes(_a,"""utf-8""" ) with gzip.open(_a,"""wb""" ) as f: f.write(_a ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): if datasets.config.LZ4_AVAILABLE: import lza.frame _A : str = tmp_path_factory.mktemp("""data""" ) / """file.txt.lz4""" _A : List[Any] = bytes(_a,"""utf-8""" ) with lza.frame.open(_a,"""wb""" ) as f: f.write(_a ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_ ): if datasets.config.PY7ZR_AVAILABLE: import pyazr _A : int = tmp_path_factory.mktemp("""data""" ) / """file.txt.7z""" with pyazr.SevenZipFile(_a,"""w""" ) as archive: archive.write(_a,arcname=os.path.basename(_a ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_ ): import tarfile _A : List[Any] = tmp_path_factory.mktemp("""data""" ) / """file.txt.tar""" with tarfile.TarFile(_a,"""w""" ) as f: f.add(_a,arcname=os.path.basename(_a ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): import lzma _A : List[Any] = tmp_path_factory.mktemp("""data""" ) / """file.txt.xz""" _A : Dict = bytes(_a,"""utf-8""" ) with lzma.open(_a,"""wb""" ) as f: f.write(_a ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_ ): import zipfile _A : Optional[int] = tmp_path_factory.mktemp("""data""" ) / """file.txt.zip""" with zipfile.ZipFile(_a,"""w""" ) as f: f.write(_a,arcname=os.path.basename(_a ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): if datasets.config.ZSTANDARD_AVAILABLE: import zstandard as zstd _A : List[str] = tmp_path_factory.mktemp("""data""" ) / """file.txt.zst""" _A : List[str] = bytes(_a,"""utf-8""" ) with zstd.open(_a,"""wb""" ) as f: f.write(_a ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : Optional[Any] = tmp_path_factory.mktemp("""data""" ) / """file.xml""" _A : Optional[int] = textwrap.dedent( """\ <?xml version=\"1.0\" encoding=\"UTF-8\" ?> <tmx version=\"1.4\"> <header segtype=\"sentence\" srclang=\"ca\" /> <body> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 1</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 1</seg></tuv> </tu> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 2</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 2</seg></tuv> </tu> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 3</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 3</seg></tuv> </tu> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 4</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 4</seg></tuv> </tu> <tu> <tuv xml:lang=\"ca\"><seg>Contingut 5</seg></tuv> <tuv xml:lang=\"en\"><seg>Content 5</seg></tuv> </tu> </body> </tmx>""" ) with open(_a,"""w""" ) as f: f.write(_a ) return filename _snake_case = [ {"col_1": "0", "col_2": 0, "col_3": 0.0}, {"col_1": "1", "col_2": 1, "col_3": 1.0}, {"col_1": "2", "col_2": 2, "col_3": 2.0}, {"col_1": "3", "col_2": 3, "col_3": 3.0}, ] _snake_case = [ {"col_1": "4", "col_2": 4, "col_3": 4.0}, {"col_1": "5", "col_2": 5, "col_3": 5.0}, ] _snake_case = { "col_1": ["0", "1", "2", "3"], "col_2": [0, 1, 2, 3], "col_3": [0.0, 1.0, 2.0, 3.0], } _snake_case = [ {"col_3": 0.0, "col_1": "0", "col_2": 0}, {"col_3": 1.0, "col_1": "1", "col_2": 1}, ] _snake_case = [ {"col_1": "s0", "col_2": 0, "col_3": 0.0}, {"col_1": "s1", "col_2": 1, "col_3": 1.0}, {"col_1": "s2", "col_2": 2, "col_3": 2.0}, {"col_1": "s3", "col_2": 3, "col_3": 3.0}, ] @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( ): return DATA_DICT_OF_LISTS @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : Optional[Any] = datasets.Dataset.from_dict(_a ) _A : List[str] = str(tmp_path_factory.mktemp("""data""" ) / """dataset.arrow""" ) dataset.map(cache_file_name=_a ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : Optional[Any] = str(tmp_path_factory.mktemp("""data""" ) / """dataset.sqlite""" ) with contextlib.closing(sqlitea.connect(_a ) ) as con: _A : str = con.cursor() cur.execute("""CREATE TABLE dataset(col_1 text, col_2 int, col_3 real)""" ) for item in DATA: cur.execute("""INSERT INTO dataset(col_1, col_2, col_3) VALUES (?, ?, ?)""",tuple(item.values() ) ) con.commit() return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : Tuple = str(tmp_path_factory.mktemp("""data""" ) / """dataset.csv""" ) with open(_a,"""w""",newline="""""" ) as f: _A : List[str] = csv.DictWriter(_a,fieldnames=["""col_1""", """col_2""", """col_3"""] ) writer.writeheader() for item in DATA: writer.writerow(_a ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : Optional[Any] = str(tmp_path_factory.mktemp("""data""" ) / """dataset2.csv""" ) with open(_a,"""w""",newline="""""" ) as f: _A : int = csv.DictWriter(_a,fieldnames=["""col_1""", """col_2""", """col_3"""] ) writer.writeheader() for item in DATA: writer.writerow(_a ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_ ): import bza _A : int = tmp_path_factory.mktemp("""data""" ) / """dataset.csv.bz2""" with open(_a,"""rb""" ) as f: _A : int = f.read() # data = bytes(FILE_CONTENT, "utf-8") with bza.open(_a,"""wb""" ) as f: f.write(_a ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ ): _A : str = tmp_path_factory.mktemp("""data""" ) / """dataset.csv.zip""" with zipfile.ZipFile(_a,"""w""" ) as f: f.write(_a,arcname=os.path.basename(_a ) ) f.write(_a,arcname=os.path.basename(_a ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ ): _A : int = tmp_path_factory.mktemp("""data""" ) / """dataset.csv.zip""" with zipfile.ZipFile(_a,"""w""" ) as f: f.write(_a,arcname=os.path.basename(csv_path.replace(""".csv""",""".CSV""" ) ) ) f.write(_a,arcname=os.path.basename(csva_path.replace(""".csv""",""".CSV""" ) ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ ): _A : List[str] = tmp_path_factory.mktemp("""data""" ) / """dataset_with_dir.csv.zip""" with zipfile.ZipFile(_a,"""w""" ) as f: f.write(_a,arcname=os.path.join("""main_dir""",os.path.basename(_a ) ) ) f.write(_a,arcname=os.path.join("""main_dir""",os.path.basename(_a ) ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : List[Any] = str(tmp_path_factory.mktemp("""data""" ) / """dataset.parquet""" ) _A : List[Any] = pa.schema( { """col_1""": pa.string(), """col_2""": pa.intaa(), """col_3""": pa.floataa(), } ) with open(_a,"""wb""" ) as f: _A : int = pq.ParquetWriter(_a,schema=_a ) _A : Optional[Any] = pa.Table.from_pydict({k: [DATA[i][k] for i in range(len(_a ) )] for k in DATA[0]},schema=_a ) writer.write_table(_a ) writer.close() return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : List[Any] = str(tmp_path_factory.mktemp("""data""" ) / """dataset.json""" ) _A : List[str] = {"""data""": DATA} with open(_a,"""w""" ) as f: json.dump(_a,_a ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : List[Any] = str(tmp_path_factory.mktemp("""data""" ) / """dataset.json""" ) _A : List[Any] = {"""data""": DATA_DICT_OF_LISTS} with open(_a,"""w""" ) as f: json.dump(_a,_a ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : List[Any] = str(tmp_path_factory.mktemp("""data""" ) / """dataset.jsonl""" ) with open(_a,"""w""" ) as f: for item in DATA: f.write(json.dumps(_a ) + """\n""" ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : Dict = str(tmp_path_factory.mktemp("""data""" ) / """dataset2.jsonl""" ) with open(_a,"""w""" ) as f: for item in DATA: f.write(json.dumps(_a ) + """\n""" ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : Union[str, Any] = str(tmp_path_factory.mktemp("""data""" ) / """dataset_312.jsonl""" ) with open(_a,"""w""" ) as f: for item in DATA_312: f.write(json.dumps(_a ) + """\n""" ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : Union[str, Any] = str(tmp_path_factory.mktemp("""data""" ) / """dataset-str.jsonl""" ) with open(_a,"""w""" ) as f: for item in DATA_STR: f.write(json.dumps(_a ) + """\n""" ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_ ): import gzip _A : List[Any] = str(tmp_path_factory.mktemp("""data""" ) / """dataset.txt.gz""" ) with open(_a,"""rb""" ) as orig_file: with gzip.open(_a,"""wb""" ) as zipped_file: zipped_file.writelines(_a ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_ ): import gzip _A : Union[str, Any] = str(tmp_path_factory.mktemp("""data""" ) / """dataset.jsonl.gz""" ) with open(_a,"""rb""" ) as orig_file: with gzip.open(_a,"""wb""" ) as zipped_file: zipped_file.writelines(_a ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ ): _A : Optional[int] = tmp_path_factory.mktemp("""data""" ) / """dataset.jsonl.zip""" with zipfile.ZipFile(_a,"""w""" ) as f: f.write(_a,arcname=os.path.basename(_a ) ) f.write(_a,arcname=os.path.basename(_a ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_,snake_case_ ): _A : List[Any] = tmp_path_factory.mktemp("""data""" ) / """dataset_nested.jsonl.zip""" with zipfile.ZipFile(_a,"""w""" ) as f: f.write(_a,arcname=os.path.join("""nested""",os.path.basename(_a ) ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ ): _A : Any = tmp_path_factory.mktemp("""data""" ) / """dataset_with_dir.jsonl.zip""" with zipfile.ZipFile(_a,"""w""" ) as f: f.write(_a,arcname=os.path.join("""main_dir""",os.path.basename(_a ) ) ) f.write(_a,arcname=os.path.join("""main_dir""",os.path.basename(_a ) ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ ): _A : Union[str, Any] = tmp_path_factory.mktemp("""data""" ) / """dataset.jsonl.tar""" with tarfile.TarFile(_a,"""w""" ) as f: f.add(_a,arcname=os.path.basename(_a ) ) f.add(_a,arcname=os.path.basename(_a ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_,snake_case_ ): _A : List[str] = tmp_path_factory.mktemp("""data""" ) / """dataset_nested.jsonl.tar""" with tarfile.TarFile(_a,"""w""" ) as f: f.add(_a,arcname=os.path.join("""nested""",os.path.basename(_a ) ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : List[Any] = ["""0""", """1""", """2""", """3"""] _A : str = str(tmp_path_factory.mktemp("""data""" ) / """dataset.txt""" ) with open(_a,"""w""" ) as f: for item in data: f.write(item + """\n""" ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : int = ["""0""", """1""", """2""", """3"""] _A : Tuple = str(tmp_path_factory.mktemp("""data""" ) / """dataset2.txt""" ) with open(_a,"""w""" ) as f: for item in data: f.write(item + """\n""" ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : Optional[int] = ["""0""", """1""", """2""", """3"""] _A : List[str] = tmp_path_factory.mktemp("""data""" ) / """dataset.abc""" with open(_a,"""w""" ) as f: for item in data: f.write(item + """\n""" ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ ): _A : List[str] = tmp_path_factory.mktemp("""data""" ) / """dataset.text.zip""" with zipfile.ZipFile(_a,"""w""" ) as f: f.write(_a,arcname=os.path.basename(_a ) ) f.write(_a,arcname=os.path.basename(_a ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ ): _A : str = tmp_path_factory.mktemp("""data""" ) / """dataset_with_dir.text.zip""" with zipfile.ZipFile(_a,"""w""" ) as f: f.write(_a,arcname=os.path.join("""main_dir""",os.path.basename(_a ) ) ) f.write(_a,arcname=os.path.join("""main_dir""",os.path.basename(_a ) ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_,snake_case_ ): _A : Optional[int] = tmp_path_factory.mktemp("""data""" ) / """dataset.ext.zip""" with zipfile.ZipFile(_a,"""w""" ) as f: f.write(_a,arcname=os.path.basename("""unsupported.ext""" ) ) f.write(_a,arcname=os.path.basename("""unsupported_2.ext""" ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : List[Any] = """\n""".join(["""First""", """Second\u2029with Unicode new line""", """Third"""] ) _A : str = str(tmp_path_factory.mktemp("""data""" ) / """dataset_with_unicode_new_lines.txt""" ) with open(_a,"""w""",encoding="""utf-8""" ) as f: f.write(_a ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( ): return os.path.join("""tests""","""features""","""data""","""test_image_rgb.jpg""" ) @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( ): return os.path.join("""tests""","""features""","""data""","""test_audio_44100.wav""" ) @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_,snake_case_ ): _A : List[str] = tmp_path_factory.mktemp("""data""" ) / """dataset.img.zip""" with zipfile.ZipFile(_a,"""w""" ) as f: f.write(_a,arcname=os.path.basename(_a ) ) f.write(_a,arcname=os.path.basename(_a ).replace(""".jpg""","""2.jpg""" ) ) return path @pytest.fixture(scope="""session""" ) def lowerCAmelCase_ ( snake_case_ ): _A : List[str] = tmp_path_factory.mktemp("""data_dir""" ) (data_dir / "subdir").mkdir() with open(data_dir / """subdir""" / """train.txt""","""w""" ) as f: f.write("""foo\n""" * 10 ) with open(data_dir / """subdir""" / """test.txt""","""w""" ) as f: f.write("""bar\n""" * 10 ) # hidden file with open(data_dir / """subdir""" / """.test.txt""","""w""" ) as f: f.write("""bar\n""" * 10 ) # hidden directory (data_dir / ".subdir").mkdir() with open(data_dir / """.subdir""" / """train.txt""","""w""" ) as f: f.write("""foo\n""" * 10 ) with open(data_dir / """.subdir""" / """test.txt""","""w""" ) as f: f.write("""bar\n""" * 10 ) return data_dir
26
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) A =logging.getLogger(__name__) A ='Hello world! cécé herlolip' A =namedtuple( 'BertAbsConfig', [ 'temp_dir', 'large', 'use_bert_emb', 'finetune_bert', 'encoder', 'share_emb', 'max_pos', 'enc_layers', 'enc_hidden_size', 'enc_heads', 'enc_ff_size', 'enc_dropout', 'dec_layers', 'dec_hidden_size', 'dec_heads', 'dec_ff_size', 'dec_dropout', ], ) def snake_case_ (_a : List[Any] , _a : Any ): UpperCAmelCase = BertAbsConfig( temp_dir='''.''' , finetune_bert=_a , large=_a , share_emb=_a , use_bert_emb=_a , encoder='''bert''' , max_pos=5_1_2 , enc_layers=6 , enc_hidden_size=5_1_2 , enc_heads=8 , enc_ff_size=5_1_2 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=7_6_8 , dec_heads=8 , dec_ff_size=2_0_4_8 , dec_dropout=0.2 , ) UpperCAmelCase = torch.load(_a , lambda _a , _a : storage ) UpperCAmelCase = AbsSummarizer(_a , torch.device('''cpu''' ) , _a ) original.eval() UpperCAmelCase = BertAbsSummarizer(_a , torch.device('''cpu''' ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info('''convert the model''' ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info('''Make sure that the models\' outputs are identical''' ) UpperCAmelCase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) # prepare the model inputs UpperCAmelCase = tokenizer.encode('''This is sample éàalj\'-.''' ) encoder_input_ids.extend([tokenizer.pad_token_id] * (5_1_2 - len(_a )) ) UpperCAmelCase = torch.tensor(_a ).unsqueeze(0 ) UpperCAmelCase = tokenizer.encode('''This is sample 3 éàalj\'-.''' ) decoder_input_ids.extend([tokenizer.pad_token_id] * (5_1_2 - len(_a )) ) UpperCAmelCase = torch.tensor(_a ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass UpperCAmelCase = encoder_input_ids UpperCAmelCase = decoder_input_ids UpperCAmelCase = UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = UpperCAmelCase = None UpperCAmelCase = UpperCAmelCase = None UpperCAmelCase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical UpperCAmelCase = original(_a , _a , _a , _a , _a , _a , _a )[0] UpperCAmelCase = original.generator(_a ) UpperCAmelCase = new_model( _a , _a , _a , _a , _a )[0] UpperCAmelCase = new_model.generator(_a ) UpperCAmelCase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(_a ) ) UpperCAmelCase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(_a ) ) UpperCAmelCase = torch.allclose(_a , _a , atol=1E-3 ) if are_identical: logging.info('''all weights are equal up to 1e-3''' ) else: raise ValueError('''the weights are different. The new model is likely different from the original one.''' ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info('''saving the model\'s state dictionary''' ) torch.save( new_model.state_dict() , '''./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin''' ) if __name__ == "__main__": A =argparse.ArgumentParser() parser.add_argument( '--bertabs_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.', ) A =parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
34
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase : Any = logging.get_logger(__name__) def _A ( SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : List[Any]=False ): """simple docstring""" a__ : List[Any] =[] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''blocks.{i}.norm1.weight''', f'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''blocks.{i}.norm1.bias''', f'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((f'''blocks.{i}.attn.proj.weight''', f'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.attn.proj.bias''', f'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''blocks.{i}.norm2.weight''', f'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''blocks.{i}.norm2.bias''', f'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.weight''', f'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.bias''', f'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.weight''', f'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.bias''', f'''vit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "vit.embeddings.cls_token"), ("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "vit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" a__ : str =[(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) return rename_keys def _A ( SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : List[Any]=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: a__ : Any ="" else: a__ : Dict ="vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) a__ : int =state_dict.pop(f'''blocks.{i}.attn.qkv.weight''' ) a__ : Dict =state_dict.pop(f'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict a__ : Any =in_proj_weight[ : config.hidden_size, : ] a__ : Tuple =in_proj_bias[: config.hidden_size] a__ : List[str] =in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] a__ : Dict =in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] a__ : str =in_proj_weight[ -config.hidden_size :, : ] a__ : Dict =in_proj_bias[-config.hidden_size :] def _A ( SCREAMING_SNAKE_CASE : List[Any] ): """simple docstring""" a__ : List[str] =["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(__A , __A ) def _A ( SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : int ): """simple docstring""" a__ : List[Any] =dct.pop(__A ) a__ : List[Any] =val def _A ( ): """simple docstring""" a__ : List[Any] ="http://images.cocodataset.org/val2017/000000039769.jpg" a__ : List[str] =Image.open(requests.get(__A , stream=__A ).raw ) return im @torch.no_grad() def _A ( SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : Dict ): """simple docstring""" a__ : List[str] =ViTConfig() a__ : int =False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": a__ : Tuple =True a__ : List[Any] =int(vit_name[-12:-10] ) a__ : Any =int(vit_name[-9:-6] ) else: a__ : Any =1_000 a__ : int ="huggingface/label-files" a__ : List[Any] ="imagenet-1k-id2label.json" a__ : str =json.load(open(hf_hub_download(__A , __A , repo_type="dataset" ) , "r" ) ) a__ : str ={int(__A ): v for k, v in idalabel.items()} a__ : str =idalabel a__ : str ={v: k for k, v in idalabel.items()} a__ : str =int(vit_name[-6:-4] ) a__ : List[str] =int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith("tiny" ): a__ : Tuple =192 a__ : str =768 a__ : List[Any] =12 a__ : Optional[Any] =3 elif vit_name[9:].startswith("small" ): a__ : Tuple =384 a__ : str =1_536 a__ : Optional[int] =12 a__ : Union[str, Any] =6 else: pass else: if vit_name[4:].startswith("small" ): a__ : Union[str, Any] =768 a__ : List[str] =2_304 a__ : List[Any] =8 a__ : List[str] =8 elif vit_name[4:].startswith("base" ): pass elif vit_name[4:].startswith("large" ): a__ : Optional[int] =1_024 a__ : Optional[Any] =4_096 a__ : Union[str, Any] =24 a__ : Tuple =16 elif vit_name[4:].startswith("huge" ): a__ : List[Any] =1_280 a__ : Optional[int] =5_120 a__ : List[Any] =32 a__ : Tuple =16 # load original model from timm a__ : Tuple =timm.create_model(__A , pretrained=__A ) timm_model.eval() # load state_dict of original model, remove and rename some keys a__ : int =timm_model.state_dict() if base_model: remove_classification_head_(__A ) a__ : Dict =create_rename_keys(__A , __A ) for src, dest in rename_keys: rename_key(__A , __A , __A ) read_in_q_k_v(__A , __A , __A ) # load HuggingFace model if vit_name[-5:] == "in21k": a__ : Union[str, Any] =ViTModel(__A ).eval() else: a__ : Optional[Any] =ViTForImageClassification(__A ).eval() model.load_state_dict(__A ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: a__ : List[Any] =DeiTImageProcessor(size=config.image_size ) else: a__ : str =ViTImageProcessor(size=config.image_size ) a__ : Dict =image_processor(images=prepare_img() , return_tensors="pt" ) a__ : Union[str, Any] =encoding["pixel_values"] a__ : str =model(__A ) if base_model: a__ : Union[str, Any] =timm_model.forward_features(__A ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__A , outputs.pooler_output , atol=1e-3 ) else: a__ : Tuple =timm_model(__A ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__A , outputs.logits , atol=1e-3 ) Path(__A ).mkdir(exist_ok=__A ) print(f'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__A ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__A ) if __name__ == "__main__": UpperCAmelCase : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--vit_name""", default="""vit_base_patch16_224""", type=str, help="""Name of the ViT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) UpperCAmelCase : int = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
357
import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging UpperCAmelCase : Union[str, Any] = logging.get_logger(__name__) def _A ( SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : List[Any] ): """simple docstring""" a__ : Tuple =set() a__ : Optional[Any] =[] def parse_line(SCREAMING_SNAKE_CASE : Optional[int] ): for line in fp: if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): a__ : str =line.decode("UTF-8" ) if "warnings summary (final)" in line: continue # This means we are outside the body of a warning elif not line.startswith(" " ): # process a single warning and move it to `selected_warnings`. if len(SCREAMING_SNAKE_CASE ) > 0: a__ : Union[str, Any] ="\n".join(SCREAMING_SNAKE_CASE ) # Only keep the warnings specified in `targets` if any(f''': {x}: ''' in warning for x in targets ): selected_warnings.add(SCREAMING_SNAKE_CASE ) buffer.clear() continue else: a__ : Optional[Any] =line.strip() buffer.append(SCREAMING_SNAKE_CASE ) if from_gh: for filename in os.listdir(SCREAMING_SNAKE_CASE ): a__ : str =os.path.join(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if not os.path.isdir(SCREAMING_SNAKE_CASE ): # read the file if filename != "warnings.txt": continue with open(SCREAMING_SNAKE_CASE ) as fp: parse_line(SCREAMING_SNAKE_CASE ) else: try: with zipfile.ZipFile(SCREAMING_SNAKE_CASE ) as z: for filename in z.namelist(): if not os.path.isdir(SCREAMING_SNAKE_CASE ): # read the file if filename != "warnings.txt": continue with z.open(SCREAMING_SNAKE_CASE ) as fp: parse_line(SCREAMING_SNAKE_CASE ) except Exception: logger.warning( f'''{artifact_path} is either an invalid zip file or something else wrong. This file is skipped.''' ) return selected_warnings def _A ( SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : Optional[Any] ): """simple docstring""" a__ : Optional[int] =set() a__ : Any =[os.path.join(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for p in os.listdir(SCREAMING_SNAKE_CASE ) if (p.endswith(".zip" ) or from_gh)] for p in paths: selected_warnings.update(extract_warnings_from_single_artifact(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) return selected_warnings if __name__ == "__main__": def _A ( SCREAMING_SNAKE_CASE : Optional[int] ): """simple docstring""" return values.split("," ) UpperCAmelCase : Any = argparse.ArgumentParser() # Required parameters parser.add_argument("""--workflow_run_id""", type=str, required=True, help="""A GitHub Actions workflow run id.""") parser.add_argument( """--output_dir""", type=str, required=True, help="""Where to store the downloaded artifacts and other result files.""", ) parser.add_argument("""--token""", default=None, type=str, help="""A token that has actions:read permission.""") # optional parameters parser.add_argument( """--targets""", default="""DeprecationWarning,UserWarning,FutureWarning""", type=list_str, help="""Comma-separated list of target warning(s) which we want to extract.""", ) parser.add_argument( """--from_gh""", action="""store_true""", help="""If running from a GitHub action workflow and collecting warnings from its artifacts.""", ) UpperCAmelCase : List[Any] = parser.parse_args() UpperCAmelCase : str = args.from_gh if from_gh: # The artifacts have to be downloaded using `actions/download-artifact@v3` pass else: os.makedirs(args.output_dir, exist_ok=True) # get download links UpperCAmelCase : Dict = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, """artifacts.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) # download artifacts for idx, (name, url) in enumerate(artifacts.items()): print(name) print(url) print("""=""" * 80) download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) # extract warnings from artifacts UpperCAmelCase : Tuple = extract_warnings(args.output_dir, args.targets) UpperCAmelCase : Optional[Any] = sorted(selected_warnings) with open(os.path.join(args.output_dir, """selected_warnings.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(selected_warnings, fp, ensure_ascii=False, indent=4)
148
0
from math import ceil, sqrt def UpperCAmelCase ( a_ = 1_0_0_0_0_0_0 ) -> int: """simple docstring""" __A = 0 for outer_width in range(3 , (limit // 4) + 2 ): if outer_width**2 > limit: __A = max(ceil(sqrt(outer_width**2 - limit ) ) , 1 ) else: __A = 1 if (outer_width - hole_width_lower_bound) % 2: hole_width_lower_bound += 1 answer += (outer_width - hole_width_lower_bound - 2) // 2 + 1 return answer if __name__ == "__main__": print(f'''{solution() = }''')
15
import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel SCREAMING_SNAKE_CASE :Union[str, Any] = False SCREAMING_SNAKE_CASE :Any = True SCREAMING_SNAKE_CASE :Tuple = False if __name__ == "__main__": SCREAMING_SNAKE_CASE :Tuple = argparse.ArgumentParser() parser.add_argument( '--repo_path', default=None, type=str, required=True, help='The config json file corresponding to the architecture.', ) parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') SCREAMING_SNAKE_CASE :Union[str, Any] = parser.parse_args() SCREAMING_SNAKE_CASE :Dict = { 'image_size': 'sample_size', 'num_res_blocks': 'layers_per_block', 'block_channels': 'block_out_channels', 'down_blocks': 'down_block_types', 'up_blocks': 'up_block_types', 'downscale_freq_shift': 'freq_shift', 'resnet_num_groups': 'norm_num_groups', 'resnet_act_fn': 'act_fn', 'resnet_eps': 'norm_eps', 'num_head_channels': 'attention_head_dim', } SCREAMING_SNAKE_CASE :Optional[int] = { 'time_steps': 'time_proj', 'mid': 'mid_block', 'downsample_blocks': 'down_blocks', 'upsample_blocks': 'up_blocks', } SCREAMING_SNAKE_CASE :int = '' if has_file(args.repo_path, 'config.json') else 'unet' with open(os.path.join(args.repo_path, subfolder, 'config.json'), 'r', encoding='utf-8') as reader: SCREAMING_SNAKE_CASE :Dict = reader.read() SCREAMING_SNAKE_CASE :List[str] = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, 'config.json'): SCREAMING_SNAKE_CASE :Optional[int] = UNetaDModel(**config) else: SCREAMING_SNAKE_CASE :Optional[Any] = UNetaDConditionModel if 'ldm-text2im-large-256' in args.repo_path else UNetaDModel SCREAMING_SNAKE_CASE :List[str] = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) SCREAMING_SNAKE_CASE :List[str] = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: SCREAMING_SNAKE_CASE :Optional[Any] = config[key] del config[key] SCREAMING_SNAKE_CASE :Optional[Any] = [k.replace('UNetRes', '') for k in config['down_block_types']] SCREAMING_SNAKE_CASE :List[Any] = [k.replace('UNetRes', '') for k in config['up_block_types']] if do_only_weights: SCREAMING_SNAKE_CASE :Tuple = torch.load(os.path.join(args.repo_path, subfolder, 'diffusion_pytorch_model.bin')) SCREAMING_SNAKE_CASE :Any = {} for param_key, param_value in state_dict.items(): if param_key.endswith('.op.bias') or param_key.endswith('.op.weight'): continue SCREAMING_SNAKE_CASE :List[str] = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('.')[0] == key: SCREAMING_SNAKE_CASE :List[Any] = param_value SCREAMING_SNAKE_CASE :str = True if not has_changed: SCREAMING_SNAKE_CASE :List[str] = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
15
1
"""simple docstring""" import numpy as np import qiskit def __lowercase ( snake_case_ : int = 8 ,snake_case_ : int | None = None ) ->str: '''simple docstring''' __A : str = np.random.default_rng(seed=snake_case_ ) # Roughly 25% of the qubits will contribute to the key. # So we take more than we need. __A : str = 6 * key_len # Measurement basis for Alice's qubits. __A : Any = rng.integers(2 ,size=snake_case_ ) # The set of states Alice will prepare. __A : Any = rng.integers(2 ,size=snake_case_ ) # Measurement basis for Bob's qubits. __A : str = rng.integers(2 ,size=snake_case_ ) # Quantum Circuit to simulate BB84 __A : Dict = qiskit.QuantumCircuit(snake_case_ ,name='''BB84''' ) # Alice prepares her qubits according to rules above. for index, _ in enumerate(snake_case_ ): if alice_state[index] == 1: bbaa_circ.x(snake_case_ ) if alice_basis[index] == 1: bbaa_circ.h(snake_case_ ) bbaa_circ.barrier() # Bob measures the received qubits according to rules above. for index, _ in enumerate(snake_case_ ): if bob_basis[index] == 1: bbaa_circ.h(snake_case_ ) bbaa_circ.barrier() bbaa_circ.measure_all() # Simulate the quantum circuit. __A : List[str] = qiskit.Aer.get_backend('''aer_simulator''' ) # We only need to run one shot because the key is unique. # Multiple shots will produce the same key. __A : List[str] = qiskit.execute(snake_case_ ,snake_case_ ,shots=1 ,seed_simulator=snake_case_ ) # Returns the result of measurement. __A : Union[str, Any] = job.result().get_counts(snake_case_ ).most_frequent() # Extracting the generated key from the simulation results. # Only keep measurement results where Alice and Bob chose the same basis. __A : int = ''''''.join( [ result_bit for alice_basis_bit, bob_basis_bit, result_bit in zip( snake_case_ ,snake_case_ ,snake_case_ ) if alice_basis_bit == bob_basis_bit ] ) # Get final key. Pad with 0 if too short, otherwise truncate. __A : Union[str, Any] = gen_key[:key_len] if len(snake_case_ ) >= key_len else gen_key.ljust(snake_case_ ,'''0''' ) return key if __name__ == "__main__": print(f'''The generated key is : {bbaa(8, seed=0)}''') from doctest import testmod testmod()
291
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available a_ = { """configuration_bloom""": ["""BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BloomConfig""", """BloomOnnxConfig"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ["""BloomTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ """BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST""", """BloomForCausalLM""", """BloomModel""", """BloomPreTrainedModel""", """BloomForSequenceClassification""", """BloomForTokenClassification""", """BloomForQuestionAnswering""", ] if TYPE_CHECKING: from .configuration_bloom import BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP, BloomConfig, BloomOnnxConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bloom_fast import BloomTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bloom import ( BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST, BloomForCausalLM, BloomForQuestionAnswering, BloomForSequenceClassification, BloomForTokenClassification, BloomModel, BloomPreTrainedModel, ) else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
291
1
"""simple docstring""" import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class _UpperCAmelCase ( unittest.TestCase ): def A ( self : List[str] , A : Union[str, Any] ) -> int: for model_result in results.values(): for batch_size, sequence_length in zip(model_result['''bs'''] , model_result['''ss'''] ): lowercase_ : Tuple = model_result['''result'''][batch_size][sequence_length] self.assertIsNotNone(A ) def A ( self : int ) -> Optional[Any]: lowercase_ : Optional[Any] = '''sshleifer/tiny-gpt2''' lowercase_ : int = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A , inference=A , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A , ) lowercase_ : List[str] = PyTorchBenchmark(A ) lowercase_ : Optional[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def A ( self : int ) -> Optional[Any]: lowercase_ : List[Any] = '''sgugger/tiny-distilbert-classification''' lowercase_ : Optional[Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A , inference=A , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A , only_pretrain_model=A , ) lowercase_ : Tuple = PyTorchBenchmark(A ) lowercase_ : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def A ( self : List[Any] ) -> str: lowercase_ : str = '''sshleifer/tiny-gpt2''' lowercase_ : Any = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A , inference=A , torchscript=A , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A , ) lowercase_ : List[Any] = PyTorchBenchmark(A ) lowercase_ : Dict = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == '''cpu''' , '''Cant do half precision''' ) def A ( self : List[Any] ) -> str: lowercase_ : int = '''sshleifer/tiny-gpt2''' lowercase_ : Tuple = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A , inference=A , fpaa=A , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A , ) lowercase_ : Any = PyTorchBenchmark(A ) lowercase_ : Dict = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def A ( self : Union[str, Any] ) -> Optional[int]: lowercase_ : Union[str, Any] = '''sshleifer/tiny-gpt2''' lowercase_ : Optional[int] = AutoConfig.from_pretrained(A ) # set architectures equal to `None` lowercase_ : str = None lowercase_ : Tuple = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A , inference=A , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A , ) lowercase_ : Any = PyTorchBenchmark(A , configs=[config] ) lowercase_ : List[str] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def A ( self : Optional[Any] ) -> List[Any]: lowercase_ : Optional[Any] = '''sshleifer/tiny-gpt2''' lowercase_ : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A , inference=A , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A , ) lowercase_ : List[str] = PyTorchBenchmark(A ) lowercase_ : Dict = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == '''cpu''' , '''Can\'t do half precision''' ) def A ( self : Optional[Any] ) -> Dict: lowercase_ : Optional[Any] = '''sshleifer/tiny-gpt2''' lowercase_ : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A , inference=A , sequence_lengths=[8] , batch_sizes=[1] , fpaa=A , multi_process=A , ) lowercase_ : int = PyTorchBenchmark(A ) lowercase_ : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def A ( self : int ) -> Optional[Any]: lowercase_ : List[Any] = '''sshleifer/tiny-gpt2''' lowercase_ : Any = AutoConfig.from_pretrained(A ) lowercase_ : Any = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A , inference=A , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A , ) lowercase_ : Optional[Any] = PyTorchBenchmark(A , configs=[config] ) lowercase_ : List[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def A ( self : Any ) -> List[Any]: lowercase_ : Union[str, Any] = '''sshleifer/tinier_bart''' lowercase_ : Optional[Any] = AutoConfig.from_pretrained(A ) lowercase_ : int = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A , inference=A , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A , ) lowercase_ : Any = PyTorchBenchmark(A , configs=[config] ) lowercase_ : int = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def A ( self : List[str] ) -> Optional[int]: lowercase_ : str = '''sshleifer/tiny-gpt2''' lowercase_ : Union[str, Any] = AutoConfig.from_pretrained(A ) lowercase_ : List[Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A , inference=A , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A , ) lowercase_ : List[str] = PyTorchBenchmark(A , configs=[config] ) lowercase_ : str = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def A ( self : Optional[Any] ) -> int: lowercase_ : Optional[Any] = '''sshleifer/tinier_bart''' lowercase_ : Optional[Any] = AutoConfig.from_pretrained(A ) lowercase_ : Tuple = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A , inference=A , sequence_lengths=[8] , batch_sizes=[1] , multi_process=A , ) lowercase_ : str = PyTorchBenchmark(A , configs=[config] ) lowercase_ : str = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def A ( self : Union[str, Any] ) -> Optional[int]: lowercase_ : List[Any] = '''sshleifer/tiny-gpt2''' with tempfile.TemporaryDirectory() as tmp_dir: lowercase_ : Any = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A , inference=A , save_to_csv=A , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(A , '''inf_time.csv''' ) , train_memory_csv_file=os.path.join(A , '''train_mem.csv''' ) , inference_memory_csv_file=os.path.join(A , '''inf_mem.csv''' ) , train_time_csv_file=os.path.join(A , '''train_time.csv''' ) , env_info_csv_file=os.path.join(A , '''env.csv''' ) , multi_process=A , ) lowercase_ : str = PyTorchBenchmark(A ) benchmark.run() self.assertTrue(Path(os.path.join(A , '''inf_time.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(A , '''train_time.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(A , '''inf_mem.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(A , '''train_mem.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(A , '''env.csv''' ) ).exists() ) def A ( self : str ) -> Tuple: lowercase_ : str = '''sshleifer/tiny-gpt2''' def _check_summary_is_not_empty(A : List[Any] ): self.assertTrue(hasattr(A , '''sequential''' ) ) self.assertTrue(hasattr(A , '''cumulative''' ) ) self.assertTrue(hasattr(A , '''current''' ) ) self.assertTrue(hasattr(A , '''total''' ) ) with tempfile.TemporaryDirectory() as tmp_dir: lowercase_ : Any = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=A , inference=A , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(A , '''log.txt''' ) , log_print=A , trace_memory_line_by_line=A , multi_process=A , ) lowercase_ : Optional[int] = PyTorchBenchmark(A ) lowercase_ : Dict = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(A , '''log.txt''' ) ).exists() )
33
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase = 100_0000 ) -> int: lowerCamelCase__ : int = limit + 1 lowerCamelCase__ : Optional[Any] = [0] * limit for first_term in range(1 , _UpperCAmelCase ): for n in range(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): lowerCamelCase__ : Optional[Any] = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a lowerCamelCase__ : List[str] = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F"""{solution() = }""")
50
0
import json import re from typing import TYPE_CHECKING, List, Optional, Tuple, Union import numpy as np from ...utils import is_tf_available, is_torch_available, logging if TYPE_CHECKING: if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_codegen import CodeGenTokenizer lowercase__ = logging.get_logger(__name__) lowercase__ = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} lowercase__ = { """vocab_file""": { """Salesforce/codegen-350M-mono""": """https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/vocab.json""", }, """merges_file""": { """Salesforce/codegen-350M-mono""": """https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/merges.txt""", }, """tokenizer_file""": { """Salesforce/codegen-350M-mono""": ( """https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/tokenizer.json""" ), }, } lowercase__ = { """Salesforce/codegen-350M-mono""": 2048, } class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = ["""input_ids""", """attention_mask"""] lowerCamelCase__ = CodeGenTokenizer def __init__( self , lowercase=None , lowercase=None , lowercase=None , lowercase="<|endoftext|>" , lowercase="<|endoftext|>" , lowercase="<|endoftext|>" , lowercase=False , **lowercase , ): super().__init__( lowercase , lowercase , tokenizer_file=lowercase , unk_token=lowercase , bos_token=lowercase , eos_token=lowercase , add_prefix_space=lowercase , **lowercase , ) if kwargs.pop('add_bos_token' , lowercase ): _lowerCamelCase : str = kwargs.pop('name_or_path' , '' ) raise ValueError( 'Currenty GPT2\'s fast tokenizer does NOT support adding a BOS token.' 'Instead you should use GPT2\'s slow tokenizer class `CodeGenTokenizer` as follows: \n' F'''`CodeGenTokenizer.from_pretrained(\'{model_id}\')`\nor\n''' F'''`AutoTokenizer.from_pretrained(\'{model_id}\', use_fast=False)`\n''' 'This issue will be fixed soon, see: https://github.com/huggingface/tokenizers/pull/1005.' ' so that the fast tokenizer works correctly.' ) _lowerCamelCase : Optional[int] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , lowercase ) != add_prefix_space: _lowerCamelCase : Dict = getattr(lowercase , pre_tok_state.pop('type' ) ) _lowerCamelCase : Any = add_prefix_space _lowerCamelCase : List[Any] = pre_tok_class(**lowercase ) _lowerCamelCase : str = add_prefix_space def A_ ( self , *lowercase , **lowercase ): _lowerCamelCase : Any = kwargs.get('is_split_into_words' , lowercase ) assert self.add_prefix_space or not is_split_into_words, ( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*lowercase , **lowercase ) def A_ ( self , *lowercase , **lowercase ): _lowerCamelCase : List[str] = kwargs.get('is_split_into_words' , lowercase ) assert self.add_prefix_space or not is_split_into_words, ( F'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*lowercase , **lowercase ) def A_ ( self , lowercase , lowercase = None ): _lowerCamelCase : Optional[Any] = self._tokenizer.model.save(lowercase , name=lowercase ) return tuple(lowercase ) def A_ ( self , lowercase , lowercase = False , lowercase = None , lowercase = None , **lowercase , ): _lowerCamelCase : List[Any] = super().decode( token_ids=lowercase , skip_special_tokens=lowercase , clean_up_tokenization_spaces=lowercase , **lowercase , ) if truncate_before_pattern is not None and len(lowercase ) > 0: _lowerCamelCase : Dict = self.truncate(lowercase , lowercase ) return decoded_text def A_ ( self , lowercase , lowercase ): def find_re(lowercase , lowercase , lowercase ): _lowerCamelCase : List[str] = pattern.search(lowercase , lowercase ) return m.start() if m else -1 _lowerCamelCase : Dict = [re.compile(lowercase , re.MULTILINE ) for pattern in truncate_before_pattern] _lowerCamelCase : Tuple = list(re.finditer('^print' , lowercase , re.MULTILINE ) ) if len(lowercase ) > 1: _lowerCamelCase : Dict = completion[: prints[1].start()] _lowerCamelCase : Dict = list(re.finditer('^def' , lowercase , re.MULTILINE ) ) if len(lowercase ) > 1: _lowerCamelCase : Tuple = completion[: defs[1].start()] _lowerCamelCase : List[str] = 0 _lowerCamelCase : List[Any] = [ pos for pos in [find_re(lowercase , lowercase , lowercase ) for terminal in terminals] if pos != -1 ] if len(lowercase ) > 0: return completion[: min(lowercase )] else: return completion
359
"""simple docstring""" import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def A_ ( self , lowercase , lowercase , lowercase ): _lowerCamelCase : Optional[int] = hf_hub_download( repo_id='nateraw/video-demo' , filename='archery.mp4' , repo_type='dataset' ) _lowerCamelCase : Tuple = VideoClassificationPipeline(model=lowercase , image_processor=lowercase , top_k=2 ) _lowerCamelCase : List[str] = [ example_video_filepath, 'https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4', ] return video_classifier, examples def A_ ( self , lowercase , lowercase ): for example in examples: _lowerCamelCase : Tuple = video_classifier(lowercase ) self.assertEqual( lowercase , [ {'score': ANY(lowercase ), 'label': ANY(lowercase )}, {'score': ANY(lowercase ), 'label': ANY(lowercase )}, ] , ) @require_torch def A_ ( self ): _lowerCamelCase : Optional[Any] = 'hf-internal-testing/tiny-random-VideoMAEForVideoClassification' _lowerCamelCase : Tuple = VideoMAEFeatureExtractor( size={'shortest_edge': 10} , crop_size={'height': 10, 'width': 10} ) _lowerCamelCase : Dict = pipeline( 'video-classification' , model=lowercase , feature_extractor=lowercase , frame_sampling_rate=4 ) _lowerCamelCase : Any = hf_hub_download(repo_id='nateraw/video-demo' , filename='archery.mp4' , repo_type='dataset' ) _lowerCamelCase : Dict = video_classifier(lowercase , top_k=2 ) self.assertEqual( nested_simplify(lowercase , decimals=4 ) , [{'score': 0.51_99, 'label': 'LABEL_0'}, {'score': 0.48_01, 'label': 'LABEL_1'}] , ) _lowerCamelCase : str = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(lowercase , decimals=4 ) , [ [{'score': 0.51_99, 'label': 'LABEL_0'}, {'score': 0.48_01, 'label': 'LABEL_1'}], [{'score': 0.51_99, 'label': 'LABEL_0'}, {'score': 0.48_01, 'label': 'LABEL_1'}], ] , ) @require_tf def A_ ( self ): pass
12
0
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class a_ ( unittest.TestCase ): '''simple docstring''' def snake_case_( self ) -> str: _SCREAMING_SNAKE_CASE = { """task_specific_params""": { """summarization""": {"""length_penalty""": 1.0, """max_length""": 128, """min_length""": 12, """num_beams""": 4}, """summarization_cnn""": {"""length_penalty""": 2.0, """max_length""": 142, """min_length""": 56, """num_beams""": 4}, """summarization_xsum""": {"""length_penalty""": 1.0, """max_length""": 62, """min_length""": 11, """num_beams""": 6}, } } _SCREAMING_SNAKE_CASE = { """task_specific_params.summarization.length_penalty""": 1.0, """task_specific_params.summarization.max_length""": 128, """task_specific_params.summarization.min_length""": 12, """task_specific_params.summarization.num_beams""": 4, """task_specific_params.summarization_cnn.length_penalty""": 2.0, """task_specific_params.summarization_cnn.max_length""": 142, """task_specific_params.summarization_cnn.min_length""": 56, """task_specific_params.summarization_cnn.num_beams""": 4, """task_specific_params.summarization_xsum.length_penalty""": 1.0, """task_specific_params.summarization_xsum.max_length""": 62, """task_specific_params.summarization_xsum.min_length""": 11, """task_specific_params.summarization_xsum.num_beams""": 6, } self.assertEqual(flatten_dict(A ) , A ) def snake_case_( self ) -> Optional[int]: _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(A ) , x.transpose() ) ) _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(A , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def snake_case_( self ) -> Optional[int]: _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 ) _SCREAMING_SNAKE_CASE = torch.tensor(A ) self.assertTrue(np.allclose(transpose(A ) , transpose(A ).numpy() ) ) _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 , 5 ) _SCREAMING_SNAKE_CASE = torch.tensor(A ) self.assertTrue(np.allclose(transpose(A , axes=(1, 2, 0) ) , transpose(A , axes=(1, 2, 0) ).numpy() ) ) @require_tf def snake_case_( self ) -> Optional[int]: _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 ) _SCREAMING_SNAKE_CASE = tf.constant(A ) self.assertTrue(np.allclose(transpose(A ) , transpose(A ).numpy() ) ) _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 , 5 ) _SCREAMING_SNAKE_CASE = tf.constant(A ) self.assertTrue(np.allclose(transpose(A , axes=(1, 2, 0) ) , transpose(A , axes=(1, 2, 0) ).numpy() ) ) @require_flax def snake_case_( self ) -> str: _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 ) _SCREAMING_SNAKE_CASE = jnp.array(A ) self.assertTrue(np.allclose(transpose(A ) , np.asarray(transpose(A ) ) ) ) _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 , 5 ) _SCREAMING_SNAKE_CASE = jnp.array(A ) self.assertTrue(np.allclose(transpose(A , axes=(1, 2, 0) ) , np.asarray(transpose(A , axes=(1, 2, 0) ) ) ) ) def snake_case_( self ) -> List[str]: _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(A , (4, 3) ) , np.reshape(A , (4, 3) ) ) ) _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(A , (12, 5) ) , np.reshape(A , (12, 5) ) ) ) @require_torch def snake_case_( self ) -> Any: _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 ) _SCREAMING_SNAKE_CASE = torch.tensor(A ) self.assertTrue(np.allclose(reshape(A , (4, 3) ) , reshape(A , (4, 3) ).numpy() ) ) _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 , 5 ) _SCREAMING_SNAKE_CASE = torch.tensor(A ) self.assertTrue(np.allclose(reshape(A , (12, 5) ) , reshape(A , (12, 5) ).numpy() ) ) @require_tf def snake_case_( self ) -> str: _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 ) _SCREAMING_SNAKE_CASE = tf.constant(A ) self.assertTrue(np.allclose(reshape(A , (4, 3) ) , reshape(A , (4, 3) ).numpy() ) ) _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 , 5 ) _SCREAMING_SNAKE_CASE = tf.constant(A ) self.assertTrue(np.allclose(reshape(A , (12, 5) ) , reshape(A , (12, 5) ).numpy() ) ) @require_flax def snake_case_( self ) -> List[str]: _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 ) _SCREAMING_SNAKE_CASE = jnp.array(A ) self.assertTrue(np.allclose(reshape(A , (4, 3) ) , np.asarray(reshape(A , (4, 3) ) ) ) ) _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 , 5 ) _SCREAMING_SNAKE_CASE = jnp.array(A ) self.assertTrue(np.allclose(reshape(A , (12, 5) ) , np.asarray(reshape(A , (12, 5) ) ) ) ) def snake_case_( self ) -> Any: _SCREAMING_SNAKE_CASE = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(A ) , np.squeeze(A ) ) ) _SCREAMING_SNAKE_CASE = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(A , axis=2 ) , np.squeeze(A , axis=2 ) ) ) @require_torch def snake_case_( self ) -> str: _SCREAMING_SNAKE_CASE = np.random.randn(1 , 3 , 4 ) _SCREAMING_SNAKE_CASE = torch.tensor(A ) self.assertTrue(np.allclose(squeeze(A ) , squeeze(A ).numpy() ) ) _SCREAMING_SNAKE_CASE = np.random.randn(1 , 4 , 1 , 5 ) _SCREAMING_SNAKE_CASE = torch.tensor(A ) self.assertTrue(np.allclose(squeeze(A , axis=2 ) , squeeze(A , axis=2 ).numpy() ) ) @require_tf def snake_case_( self ) -> List[str]: _SCREAMING_SNAKE_CASE = np.random.randn(1 , 3 , 4 ) _SCREAMING_SNAKE_CASE = tf.constant(A ) self.assertTrue(np.allclose(squeeze(A ) , squeeze(A ).numpy() ) ) _SCREAMING_SNAKE_CASE = np.random.randn(1 , 4 , 1 , 5 ) _SCREAMING_SNAKE_CASE = tf.constant(A ) self.assertTrue(np.allclose(squeeze(A , axis=2 ) , squeeze(A , axis=2 ).numpy() ) ) @require_flax def snake_case_( self ) -> Union[str, Any]: _SCREAMING_SNAKE_CASE = np.random.randn(1 , 3 , 4 ) _SCREAMING_SNAKE_CASE = jnp.array(A ) self.assertTrue(np.allclose(squeeze(A ) , np.asarray(squeeze(A ) ) ) ) _SCREAMING_SNAKE_CASE = np.random.randn(1 , 4 , 1 , 5 ) _SCREAMING_SNAKE_CASE = jnp.array(A ) self.assertTrue(np.allclose(squeeze(A , axis=2 ) , np.asarray(squeeze(A , axis=2 ) ) ) ) def snake_case_( self ) -> str: _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(A , axis=1 ) , np.expand_dims(A , axis=1 ) ) ) @require_torch def snake_case_( self ) -> int: _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 ) _SCREAMING_SNAKE_CASE = torch.tensor(A ) self.assertTrue(np.allclose(expand_dims(A , axis=1 ) , expand_dims(A , axis=1 ).numpy() ) ) @require_tf def snake_case_( self ) -> Tuple: _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 ) _SCREAMING_SNAKE_CASE = tf.constant(A ) self.assertTrue(np.allclose(expand_dims(A , axis=1 ) , expand_dims(A , axis=1 ).numpy() ) ) @require_flax def snake_case_( self ) -> Dict: _SCREAMING_SNAKE_CASE = np.random.randn(3 , 4 ) _SCREAMING_SNAKE_CASE = jnp.array(A ) self.assertTrue(np.allclose(expand_dims(A , axis=1 ) , np.asarray(expand_dims(A , axis=1 ) ) ) )
58
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ :Any = logging.get_logger(__name__) def lowerCAmelCase__ ( a__: List[Any] , a__: Union[str, Any] , a__: Dict , a__: Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' _UpperCAmelCase = original_name.split('.' )[0] _UpperCAmelCase = key.split('.' ) _UpperCAmelCase = int(key_list[key_list.index(a__ ) - 2] ) _UpperCAmelCase = int(key_list[key_list.index(a__ ) - 1] ) _UpperCAmelCase = orig_block_num - offset _UpperCAmelCase = key.replace(F'''{orig_block_num}.{layer_num}.{original_name}''' , F'''block.{new_block_num}.{layer_num}.{new_name}''' ) return key def lowerCAmelCase__ ( a__: Tuple ) -> int: '''simple docstring''' _UpperCAmelCase = OrderedDict() _UpperCAmelCase , _UpperCAmelCase = 0, 0 for key, value in state_dict.items(): if key.startswith('network' ): _UpperCAmelCase = key.replace('network' , 'poolformer.encoder' ) if "proj" in key: # Works for the first embedding as well as the internal embedding layers if key.endswith('bias' ) and "patch_embed" not in key: patch_emb_offset += 1 _UpperCAmelCase = key[: key.find('proj' )] _UpperCAmelCase = key.replace(a__ , F'''patch_embeddings.{total_embed_found}.''' ) _UpperCAmelCase = key.replace('proj' , 'projection' ) if key.endswith('bias' ): total_embed_found += 1 if "patch_embeddings" in key: _UpperCAmelCase = 'poolformer.encoder.' + key if "mlp.fc1" in key: _UpperCAmelCase = replace_key_with_offset(a__ , a__ , 'mlp.fc1' , 'output.conv1' ) if "mlp.fc2" in key: _UpperCAmelCase = replace_key_with_offset(a__ , a__ , 'mlp.fc2' , 'output.conv2' ) if "norm1" in key: _UpperCAmelCase = replace_key_with_offset(a__ , a__ , 'norm1' , 'before_norm' ) if "norm2" in key: _UpperCAmelCase = replace_key_with_offset(a__ , a__ , 'norm2' , 'after_norm' ) if "layer_scale_1" in key: _UpperCAmelCase = replace_key_with_offset(a__ , a__ , 'layer_scale_1' , 'layer_scale_1' ) if "layer_scale_2" in key: _UpperCAmelCase = replace_key_with_offset(a__ , a__ , 'layer_scale_2' , 'layer_scale_2' ) if "head" in key: _UpperCAmelCase = key.replace('head' , 'classifier' ) _UpperCAmelCase = value return new_state_dict def lowerCAmelCase__ ( ) -> Tuple: '''simple docstring''' _UpperCAmelCase = 'http://images.cocodataset.org/val2017/000000039769.jpg' _UpperCAmelCase = Image.open(requests.get(a__ , stream=a__ ).raw ) return image @torch.no_grad() def lowerCAmelCase__ ( a__: Optional[int] , a__: Dict , a__: Any ) -> Dict: '''simple docstring''' _UpperCAmelCase = PoolFormerConfig() # set attributes based on model_name _UpperCAmelCase = 'huggingface/label-files' _UpperCAmelCase = model_name[-3:] _UpperCAmelCase = 1_0_0_0 _UpperCAmelCase = 'imagenet-1k-id2label.json' _UpperCAmelCase = (1, 1_0_0_0) # set config attributes _UpperCAmelCase = json.load(open(hf_hub_download(a__ , a__ , repo_type='dataset' ) , 'r' ) ) _UpperCAmelCase = {int(a__ ): v for k, v in idalabel.items()} _UpperCAmelCase = idalabel _UpperCAmelCase = {v: k for k, v in idalabel.items()} if size == "s12": _UpperCAmelCase = [2, 2, 6, 2] _UpperCAmelCase = [6_4, 1_2_8, 3_2_0, 5_1_2] _UpperCAmelCase = 4.0 _UpperCAmelCase = 0.9 elif size == "s24": _UpperCAmelCase = [4, 4, 1_2, 4] _UpperCAmelCase = [6_4, 1_2_8, 3_2_0, 5_1_2] _UpperCAmelCase = 4.0 _UpperCAmelCase = 0.9 elif size == "s36": _UpperCAmelCase = [6, 6, 1_8, 6] _UpperCAmelCase = [6_4, 1_2_8, 3_2_0, 5_1_2] _UpperCAmelCase = 4.0 _UpperCAmelCase = 1e-6 _UpperCAmelCase = 0.9 elif size == "m36": _UpperCAmelCase = [6, 6, 1_8, 6] _UpperCAmelCase = [9_6, 1_9_2, 3_8_4, 7_6_8] _UpperCAmelCase = 4.0 _UpperCAmelCase = 1e-6 _UpperCAmelCase = 0.95 elif size == "m48": _UpperCAmelCase = [8, 8, 2_4, 8] _UpperCAmelCase = [9_6, 1_9_2, 3_8_4, 7_6_8] _UpperCAmelCase = 4.0 _UpperCAmelCase = 1e-6 _UpperCAmelCase = 0.95 else: raise ValueError(F'''Size {size} not supported''' ) # load image processor _UpperCAmelCase = PoolFormerImageProcessor(crop_pct=a__ ) # Prepare image _UpperCAmelCase = prepare_img() _UpperCAmelCase = image_processor(images=a__ , return_tensors='pt' ).pixel_values logger.info(F'''Converting model {model_name}...''' ) # load original state dict _UpperCAmelCase = torch.load(a__ , map_location=torch.device('cpu' ) ) # rename keys _UpperCAmelCase = rename_keys(a__ ) # create HuggingFace model and load state dict _UpperCAmelCase = PoolFormerForImageClassification(a__ ) model.load_state_dict(a__ ) model.eval() # Define image processor _UpperCAmelCase = PoolFormerImageProcessor(crop_pct=a__ ) _UpperCAmelCase = image_processor(images=prepare_img() , return_tensors='pt' ).pixel_values # forward pass _UpperCAmelCase = model(a__ ) _UpperCAmelCase = outputs.logits # define expected logit slices for different models if size == "s12": _UpperCAmelCase = torch.tensor([-0.3_045, -0.6_758, -0.4_869] ) elif size == "s24": _UpperCAmelCase = torch.tensor([0.4_402, -0.1_374, -0.8_045] ) elif size == "s36": _UpperCAmelCase = torch.tensor([-0.6_080, -0.5_133, -0.5_898] ) elif size == "m36": _UpperCAmelCase = torch.tensor([0.3_952, 0.2_263, -1.2_668] ) elif size == "m48": _UpperCAmelCase = torch.tensor([0.1_167, -0.0_656, -0.3_423] ) else: raise ValueError(F'''Size {size} not supported''' ) # verify logits assert logits.shape == expected_shape assert torch.allclose(logits[0, :3] , a__ , atol=1e-2 ) # finally, save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(a__ ).mkdir(exist_ok=a__ ) model.save_pretrained(a__ ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(a__ ) if __name__ == "__main__": lowerCAmelCase__ :str = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''poolformer_s12''', type=str, help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original PyTorch checkpoint (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) lowerCAmelCase__ :Dict = parser.parse_args() convert_poolformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
329
0
"""simple docstring""" def _lowerCAmelCase ( UpperCAmelCase__ : int = 6_0_0_8_5_1_4_7_5_1_4_3 ) ->int: try: A__ : List[str] = int(lowerCAmelCase__ ) except (TypeError, ValueError): raise TypeError("""Parameter n must be int or castable to int.""" ) if n <= 0: raise ValueError("""Parameter n must be greater than or equal to one.""" ) A__ : Any = 1 A__ : str = 2 while i * i <= n: while n % i == 0: A__ : str = i n //= i i += 1 if n > 1: A__ : str = n return int(lowerCAmelCase__ ) if __name__ == "__main__": print(F'{solution() = }')
361
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def _lowerCAmelCase ( UpperCAmelCase__ : List[Any] ) ->List[str]: A__ : Union[str, Any] = DPTConfig() if "large" in checkpoint_url: A__ : int = 1_0_2_4 A__ : Union[str, Any] = 4_0_9_6 A__ : Optional[int] = 2_4 A__ : int = 1_6 A__ : Union[str, Any] = [5, 1_1, 1_7, 2_3] A__ : Tuple = [2_5_6, 5_1_2, 1_0_2_4, 1_0_2_4] A__ : Tuple = (1, 3_8_4, 3_8_4) if "ade" in checkpoint_url: A__ : Optional[int] = True A__ : int = 1_5_0 A__ : Union[str, Any] = """huggingface/label-files""" A__ : List[Any] = """ade20k-id2label.json""" A__ : Union[str, Any] = json.load(open(cached_download(hf_hub_url(UpperCAmelCase__, UpperCAmelCase__, repo_type="""dataset""" ) ), """r""" ) ) A__ : List[Any] = {int(UpperCAmelCase__ ): v for k, v in idalabel.items()} A__ : Dict = idalabel A__ : List[Any] = {v: k for k, v in idalabel.items()} A__ : Optional[Any] = [1, 1_5_0, 4_8_0, 4_8_0] return config, expected_shape def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->Any: A__ : List[Any] = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(UpperCAmelCase__, UpperCAmelCase__ ) def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any] ) ->List[str]: if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): A__ : str = name.replace("""pretrained.model""", """dpt.encoder""" ) if "pretrained.model" in name: A__ : Dict = name.replace("""pretrained.model""", """dpt.embeddings""" ) if "patch_embed" in name: A__ : List[Any] = name.replace("""patch_embed""", """patch_embeddings""" ) if "pos_embed" in name: A__ : int = name.replace("""pos_embed""", """position_embeddings""" ) if "attn.proj" in name: A__ : Tuple = name.replace("""attn.proj""", """attention.output.dense""" ) if "proj" in name and "project" not in name: A__ : List[Any] = name.replace("""proj""", """projection""" ) if "blocks" in name: A__ : Optional[Any] = name.replace("""blocks""", """layer""" ) if "mlp.fc1" in name: A__ : int = name.replace("""mlp.fc1""", """intermediate.dense""" ) if "mlp.fc2" in name: A__ : List[str] = name.replace("""mlp.fc2""", """output.dense""" ) if "norm1" in name: A__ : Any = name.replace("""norm1""", """layernorm_before""" ) if "norm2" in name: A__ : List[str] = name.replace("""norm2""", """layernorm_after""" ) if "scratch.output_conv" in name: A__ : Optional[int] = name.replace("""scratch.output_conv""", """head""" ) if "scratch" in name: A__ : List[str] = name.replace("""scratch""", """neck""" ) if "layer1_rn" in name: A__ : List[str] = name.replace("""layer1_rn""", """convs.0""" ) if "layer2_rn" in name: A__ : Optional[int] = name.replace("""layer2_rn""", """convs.1""" ) if "layer3_rn" in name: A__ : Any = name.replace("""layer3_rn""", """convs.2""" ) if "layer4_rn" in name: A__ : Any = name.replace("""layer4_rn""", """convs.3""" ) if "refinenet" in name: A__ : Union[str, Any] = int(name[len("""neck.refinenet""" ) : len("""neck.refinenet""" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 A__ : str = name.replace(f'refinenet{layer_idx}', f'fusion_stage.layers.{abs(layer_idx-4 )}' ) if "out_conv" in name: A__ : Optional[Any] = name.replace("""out_conv""", """projection""" ) if "resConfUnit1" in name: A__ : List[Any] = name.replace("""resConfUnit1""", """residual_layer1""" ) if "resConfUnit2" in name: A__ : Tuple = name.replace("""resConfUnit2""", """residual_layer2""" ) if "conv1" in name: A__ : Tuple = name.replace("""conv1""", """convolution1""" ) if "conv2" in name: A__ : List[Any] = name.replace("""conv2""", """convolution2""" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: A__ : Union[str, Any] = name.replace("""pretrained.act_postprocess1.0.project.0""", """neck.reassemble_stage.readout_projects.0.0""" ) if "pretrained.act_postprocess2.0.project.0" in name: A__ : Tuple = name.replace("""pretrained.act_postprocess2.0.project.0""", """neck.reassemble_stage.readout_projects.1.0""" ) if "pretrained.act_postprocess3.0.project.0" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess3.0.project.0""", """neck.reassemble_stage.readout_projects.2.0""" ) if "pretrained.act_postprocess4.0.project.0" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess4.0.project.0""", """neck.reassemble_stage.readout_projects.3.0""" ) # resize blocks if "pretrained.act_postprocess1.3" in name: A__ : Any = name.replace("""pretrained.act_postprocess1.3""", """neck.reassemble_stage.layers.0.projection""" ) if "pretrained.act_postprocess1.4" in name: A__ : List[Any] = name.replace("""pretrained.act_postprocess1.4""", """neck.reassemble_stage.layers.0.resize""" ) if "pretrained.act_postprocess2.3" in name: A__ : Dict = name.replace("""pretrained.act_postprocess2.3""", """neck.reassemble_stage.layers.1.projection""" ) if "pretrained.act_postprocess2.4" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess2.4""", """neck.reassemble_stage.layers.1.resize""" ) if "pretrained.act_postprocess3.3" in name: A__ : Union[str, Any] = name.replace("""pretrained.act_postprocess3.3""", """neck.reassemble_stage.layers.2.projection""" ) if "pretrained.act_postprocess4.3" in name: A__ : Optional[int] = name.replace("""pretrained.act_postprocess4.3""", """neck.reassemble_stage.layers.3.projection""" ) if "pretrained.act_postprocess4.4" in name: A__ : Dict = name.replace("""pretrained.act_postprocess4.4""", """neck.reassemble_stage.layers.3.resize""" ) if "pretrained" in name: A__ : Union[str, Any] = name.replace("""pretrained""", """dpt""" ) if "bn" in name: A__ : Union[str, Any] = name.replace("""bn""", """batch_norm""" ) if "head" in name: A__ : Dict = name.replace("""head""", """head.head""" ) if "encoder.norm" in name: A__ : Optional[int] = name.replace("""encoder.norm""", """layernorm""" ) if "auxlayer" in name: A__ : List[str] = name.replace("""auxlayer""", """auxiliary_head.head""" ) return name def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : Dict ) ->str: for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A__ : Any = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.weight' ) A__ : Tuple = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict A__ : List[str] = in_proj_weight[: config.hidden_size, :] A__ : int = in_proj_bias[: config.hidden_size] A__ : Tuple = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A__ : Any = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A__ : str = in_proj_weight[ -config.hidden_size :, : ] A__ : Optional[Any] = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( ) ->List[str]: A__ : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : int = Image.open(requests.get(UpperCAmelCase__, stream=UpperCAmelCase__ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : str, UpperCAmelCase__ : int ) ->str: A__ , A__ : Dict = get_dpt_config(UpperCAmelCase__ ) # load original state_dict from URL A__ : Any = torch.hub.load_state_dict_from_url(UpperCAmelCase__, map_location="""cpu""" ) # remove certain keys remove_ignore_keys_(UpperCAmelCase__ ) # rename keys for key in state_dict.copy().keys(): A__ : int = state_dict.pop(UpperCAmelCase__ ) A__ : str = val # read in qkv matrices read_in_q_k_v(UpperCAmelCase__, UpperCAmelCase__ ) # load HuggingFace model A__ : Optional[Any] = DPTForSemanticSegmentation(UpperCAmelCase__ ) if """ade""" in checkpoint_url else DPTForDepthEstimation(UpperCAmelCase__ ) model.load_state_dict(UpperCAmelCase__ ) model.eval() # Check outputs on an image A__ : Optional[Any] = 4_8_0 if """ade""" in checkpoint_url else 3_8_4 A__ : Dict = DPTImageProcessor(size=UpperCAmelCase__ ) A__ : Optional[int] = prepare_img() A__ : Any = image_processor(UpperCAmelCase__, return_tensors="""pt""" ) # forward pass A__ : List[str] = model(**UpperCAmelCase__ ).logits if """ade""" in checkpoint_url else model(**UpperCAmelCase__ ).predicted_depth # Assert logits A__ : Optional[Any] = torch.tensor([[6.3199, 6.3629, 6.4148], [6.3850, 6.3615, 6.4166], [6.3519, 6.3176, 6.3575]] ) if "ade" in checkpoint_url: A__ : Optional[int] = torch.tensor([[4.0480, 4.2420, 4.4360], [4.3124, 4.5693, 4.8261], [4.5768, 4.8965, 5.2163]] ) assert outputs.shape == torch.Size(UpperCAmelCase__ ) assert ( torch.allclose(outputs[0, 0, :3, :3], UpperCAmelCase__, atol=1e-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3], UpperCAmelCase__ ) ) Path(UpperCAmelCase__ ).mkdir(exist_ok=UpperCAmelCase__ ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase__ ) if push_to_hub: print("""Pushing model to hub...""" ) model.push_to_hub( repo_path_or_name=Path(UpperCAmelCase__, UpperCAmelCase__ ), organization="""nielsr""", commit_message="""Add model""", use_temp_dir=UpperCAmelCase__, ) image_processor.push_to_hub( repo_path_or_name=Path(UpperCAmelCase__, UpperCAmelCase__ ), organization="""nielsr""", commit_message="""Add image processor""", use_temp_dir=UpperCAmelCase__, ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt''', type=str, help='''URL of the original DPT checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', ) parser.add_argument( '''--model_name''', default='''dpt-large''', type=str, help='''Name of the model, in case you\'re pushing to the hub.''', ) A_ = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
296
0
import argparse import json import os import torch from transformers import LukeConfig, LukeModel, LukeTokenizer, RobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def snake_case( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Dict: '''simple docstring''' with open(__magic_name__ ) as metadata_file: lowercase : Dict = json.load(__magic_name__ ) lowercase : str = LukeConfig(use_entity_aware_attention=__magic_name__ , **metadata['''model_config'''] ) # Load in the weights from the checkpoint_path lowercase : str = torch.load(__magic_name__ , map_location='''cpu''' ) # Load the entity vocab file lowercase : Union[str, Any] = load_entity_vocab(__magic_name__ ) lowercase : Tuple = RobertaTokenizer.from_pretrained(metadata['''model_config''']['''bert_model_name'''] ) # Add special tokens to the token vocabulary for downstream tasks lowercase : int = AddedToken('''<ent>''' , lstrip=__magic_name__ , rstrip=__magic_name__ ) lowercase : Dict = AddedToken('''<ent2>''' , lstrip=__magic_name__ , rstrip=__magic_name__ ) tokenizer.add_special_tokens({'''additional_special_tokens''': [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(F"""Saving tokenizer to {pytorch_dump_folder_path}""" ) tokenizer.save_pretrained(__magic_name__ ) with open(os.path.join(__magic_name__ , LukeTokenizer.vocab_files_names['''entity_vocab_file'''] ) , '''w''' ) as f: json.dump(__magic_name__ , __magic_name__ ) lowercase : Tuple = LukeTokenizer.from_pretrained(__magic_name__ ) # Initialize the embeddings of the special tokens lowercase : Tuple = state_dict['''embeddings.word_embeddings.weight'''] lowercase : Tuple = word_emb[tokenizer.convert_tokens_to_ids(['''@'''] )[0]].unsqueeze(0 ) lowercase : str = word_emb[tokenizer.convert_tokens_to_ids(['''#'''] )[0]].unsqueeze(0 ) lowercase : List[str] = torch.cat([word_emb, ent_emb, enta_emb] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: lowercase : Any = F"""encoder.layer.{layer_index}.attention.self.""" lowercase : Any = state_dict[prefix + matrix_name] lowercase : Any = state_dict[prefix + matrix_name] lowercase : List[str] = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks lowercase : Any = state_dict['''entity_embeddings.entity_embeddings.weight'''] lowercase : List[Any] = entity_emb[entity_vocab['''[MASK]''']] lowercase : Any = LukeModel(config=__magic_name__ ).eval() lowercase , lowercase : Optional[Any] = model.load_state_dict(__magic_name__ , strict=__magic_name__ ) if not (len(__magic_name__ ) == 1 and missing_keys[0] == "embeddings.position_ids"): raise ValueError(F"""Missing keys {', '.join(__magic_name__ )}. Expected only missing embeddings.position_ids""" ) if not (all(key.startswith('''entity_predictions''' ) or key.startswith('''lm_head''' ) for key in unexpected_keys )): raise ValueError( '''Unexpected keys''' F""" {', '.join([key for key in unexpected_keys if not (key.startswith('entity_predictions' ) or key.startswith('lm_head' ))] )}""" ) # Check outputs lowercase : Any = LukeTokenizer.from_pretrained(__magic_name__ , task='''entity_classification''' ) lowercase : List[Any] = ( '''Top seed Ana Ivanovic said on Thursday she could hardly believe her luck as a fortuitous netcord helped the''' ''' new world number one avoid a humiliating second- round exit at Wimbledon .''' ) lowercase : List[Any] = (39, 42) lowercase : List[Any] = tokenizer(__magic_name__ , entity_spans=[span] , add_prefix_space=__magic_name__ , return_tensors='''pt''' ) lowercase : int = model(**__magic_name__ ) # Verify word hidden states if model_size == "large": lowercase : Optional[int] = torch.Size((1, 42, 10_24) ) lowercase : List[Any] = torch.tensor( [[0.0_1_3_3, 0.0_8_6_5, 0.0_0_9_5], [0.3_0_9_3, -0.2_5_7_6, -0.7_4_1_8], [-0.1_7_2_0, -0.2_1_1_7, -0.2_8_6_9]] ) else: # base lowercase : List[str] = torch.Size((1, 42, 7_68) ) lowercase : List[str] = torch.tensor([[0.0_0_3_7, 0.1_3_6_8, -0.0_0_9_1], [0.1_0_9_9, 0.3_3_2_9, -0.1_0_9_5], [0.0_7_6_5, 0.5_3_3_5, 0.1_1_7_9]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( F"""Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}""" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , __magic_name__ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": lowercase : List[Any] = torch.Size((1, 1, 10_24) ) lowercase : Optional[int] = torch.tensor([[0.0_4_6_6, -0.0_1_0_6, -0.0_1_7_9]] ) else: # base lowercase : Optional[Any] = torch.Size((1, 1, 7_68) ) lowercase : Any = torch.tensor([[0.1_4_5_7, 0.1_0_4_4, 0.0_1_7_4]] ) if not (outputs.entity_last_hidden_state.shape != expected_shape): raise ValueError( F"""Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is""" F""" {expected_shape}""" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , __magic_name__ , atol=1e-4 ): raise ValueError # Finally, save our PyTorch model and tokenizer print('''Saving PyTorch model to {}'''.format(__magic_name__ ) ) model.save_pretrained(__magic_name__ ) def snake_case( __magic_name__ ) -> str: '''simple docstring''' lowercase : int = {} with open(__magic_name__ , '''r''' , encoding='''utf-8''' ) as f: for index, line in enumerate(__magic_name__ ): lowercase , lowercase : Optional[Any] = line.rstrip().split('''\t''' ) lowercase : Optional[Any] = index return entity_vocab if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() # Required parameters parser.add_argument('--checkpoint_path', type=str, help='Path to a pytorch_model.bin file.') parser.add_argument( '--metadata_path', default=None, type=str, help='Path to a metadata.json file, defining the configuration.' ) parser.add_argument( '--entity_vocab_path', default=None, type=str, help='Path to an entity_vocab.tsv file, containing the entity vocabulary.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to where to dump the output PyTorch model.' ) parser.add_argument( '--model_size', default='base', type=str, choices=['base', 'large'], help='Size of the model to be converted.' ) lowerCAmelCase_ = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
308
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def snake_case( ) -> int: '''simple docstring''' lowercase : List[str] = ArgumentParser( description=( '''PyTorch TPU distributed training launch helper utility that will spawn up multiple distributed processes''' ) ) # Optional arguments for the launch helper parser.add_argument('''--num_cores''' , type=__magic_name__ , default=1 , help='''Number of TPU cores to use (1 or 8).''' ) # positional parser.add_argument( '''training_script''' , type=__magic_name__ , help=( '''The full path to the single TPU training ''' '''program/script to be launched in parallel, ''' '''followed by all the arguments for the ''' '''training script''' ) , ) # rest from the training program parser.add_argument('''training_script_args''' , nargs=__magic_name__ ) return parser.parse_args() def snake_case( ) -> Union[str, Any]: '''simple docstring''' lowercase : Optional[Any] = parse_args() # Import training_script as a module. lowercase : Optional[Any] = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) lowercase : int = script_fpath.stem lowercase : List[Any] = importlib.import_module(__magic_name__ ) # Patch sys.argv lowercase : str = [args.training_script] + args.training_script_args + ['''--tpu_num_cores''', str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
308
1
def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> int: '''simple docstring''' if not isinstance(_UpperCAmelCase, _UpperCAmelCase ): raise ValueError('multiplicative_persistence() only accepts integral values' ) if num < 0: raise ValueError('multiplicative_persistence() does not accept negative values' ) lowerCAmelCase : int = 0 lowerCAmelCase : str = str(_UpperCAmelCase ) while len(_UpperCAmelCase ) != 1: lowerCAmelCase : List[Any] = [int(_UpperCAmelCase ) for i in num_string] lowerCAmelCase : int = 1 for i in range(0, len(_UpperCAmelCase ) ): total *= numbers[i] lowerCAmelCase : Optional[Any] = str(_UpperCAmelCase ) steps += 1 return steps def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> int: '''simple docstring''' if not isinstance(_UpperCAmelCase, _UpperCAmelCase ): raise ValueError('additive_persistence() only accepts integral values' ) if num < 0: raise ValueError('additive_persistence() does not accept negative values' ) lowerCAmelCase : Optional[int] = 0 lowerCAmelCase : str = str(_UpperCAmelCase ) while len(_UpperCAmelCase ) != 1: lowerCAmelCase : Optional[Any] = [int(_UpperCAmelCase ) for i in num_string] lowerCAmelCase : Dict = 0 for i in range(0, len(_UpperCAmelCase ) ): total += numbers[i] lowerCAmelCase : List[str] = str(_UpperCAmelCase ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
323
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available from .timesteps import ( fastaa_timesteps, smartaa_timesteps, smartaa_timesteps, smartaaa_timesteps, smartaaa_timesteps, superaa_timesteps, superaa_timesteps, superaaa_timesteps, ) @dataclass class __A ( lowerCAmelCase ): lowerCAmelCase_ : Union[List[PIL.Image.Image], np.ndarray] lowerCAmelCase_ : Optional[List[bool]] lowerCAmelCase_ : Optional[List[bool]] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_if import IFPipeline from .pipeline_if_imgaimg import IFImgaImgPipeline from .pipeline_if_imgaimg_superresolution import IFImgaImgSuperResolutionPipeline from .pipeline_if_inpainting import IFInpaintingPipeline from .pipeline_if_inpainting_superresolution import IFInpaintingSuperResolutionPipeline from .pipeline_if_superresolution import IFSuperResolutionPipeline from .safety_checker import IFSafetyChecker from .watermark import IFWatermarker
323
1
from timeit import timeit def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : int ) -> int: """simple docstring""" if number < 0: raise ValueError("""the value of input must not be negative""" ) SCREAMING_SNAKE_CASE__ = 0 while number: number &= number - 1 result += 1 return result def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : int ) -> int: """simple docstring""" if number < 0: raise ValueError("""the value of input must not be negative""" ) SCREAMING_SNAKE_CASE__ = 0 while number: if number % 2 == 1: result += 1 number >>= 1 return result def __SCREAMING_SNAKE_CASE ( ) -> None: """simple docstring""" def do_benchmark(__UpperCamelCase : int ) -> None: SCREAMING_SNAKE_CASE__ = """import __main__ as z""" print(f"""Benchmark when {number = }:""" ) print(f"""{get_set_bits_count_using_modulo_operator(__UpperCamelCase ) = }""" ) SCREAMING_SNAKE_CASE__ = timeit("""z.get_set_bits_count_using_modulo_operator(25)""" , setup=__UpperCamelCase ) print(f"""timeit() runs in {timing} seconds""" ) print(f"""{get_set_bits_count_using_brian_kernighans_algorithm(__UpperCamelCase ) = }""" ) SCREAMING_SNAKE_CASE__ = timeit( """z.get_set_bits_count_using_brian_kernighans_algorithm(25)""" , setup=__UpperCamelCase , ) print(f"""timeit() runs in {timing} seconds""" ) for number in (25, 37, 58, 0): do_benchmark(__UpperCamelCase ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
219
import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseModel class __snake_case : def __init__( self : Optional[int] , _lowercase : Union[str, Any] , _lowercase : List[Any]=13 , _lowercase : List[Any]=7 , _lowercase : Optional[int]=True , _lowercase : str=True , _lowercase : str=True , _lowercase : List[str]=True , _lowercase : List[str]=99 , _lowercase : List[str]=32 , _lowercase : str=5 , _lowercase : str=4 , _lowercase : str=4 , _lowercase : Union[str, Any]="gelu" , _lowercase : str=0.0 , _lowercase : Union[str, Any]=0.1 , _lowercase : List[str]=True , _lowercase : Union[str, Any]=5_12 , _lowercase : List[str]=16 , _lowercase : Dict=2 , _lowercase : int=0.02 , _lowercase : Any=3 , _lowercase : int=4 , _lowercase : List[str]=None , ): """simple docstring""" SCREAMING_SNAKE_CASE__ = parent SCREAMING_SNAKE_CASE__ = batch_size SCREAMING_SNAKE_CASE__ = seq_length SCREAMING_SNAKE_CASE__ = is_training SCREAMING_SNAKE_CASE__ = use_input_mask SCREAMING_SNAKE_CASE__ = use_token_type_ids SCREAMING_SNAKE_CASE__ = use_labels SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = intermediate_multiple_size SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout SCREAMING_SNAKE_CASE__ = attention_dropout SCREAMING_SNAKE_CASE__ = weight_tying SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = type_vocab_size SCREAMING_SNAKE_CASE__ = type_sequence_label_size SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = num_labels SCREAMING_SNAKE_CASE__ = num_choices SCREAMING_SNAKE_CASE__ = scope def __a ( self : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE__ = None if self.use_input_mask: SCREAMING_SNAKE_CASE__ = random_attention_mask([self.batch_size, self.seq_length] ) SCREAMING_SNAKE_CASE__ = None if self.use_labels: SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) SCREAMING_SNAKE_CASE__ = self.get_config() return config, input_ids, input_mask, token_labels def __a ( self : Optional[int] ): """simple docstring""" return GPTNeoXJapaneseConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_multiple_size=self.intermediate_multiple_size , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , weight_tying=self.weight_tying , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowercase , initializer_range=self.initializer_range , ) def __a ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE__ = True return config, input_ids, input_mask, token_labels def __a ( self : List[Any] , _lowercase : Optional[Any] , _lowercase : int , _lowercase : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = GPTNeoXJapaneseModel(config=_lowercase ) model.to(_lowercase ) model.eval() SCREAMING_SNAKE_CASE__ = model(_lowercase , attention_mask=_lowercase ) SCREAMING_SNAKE_CASE__ = model(_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __a ( self : Dict , _lowercase : int , _lowercase : str , _lowercase : Any ): """simple docstring""" SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = GPTNeoXJapaneseModel(_lowercase ) model.to(_lowercase ) model.eval() SCREAMING_SNAKE_CASE__ = model(_lowercase , attention_mask=_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __a ( self : Any , _lowercase : Dict , _lowercase : List[Any] , _lowercase : Optional[Any] , _lowercase : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE__ = GPTNeoXJapaneseForCausalLM(config=_lowercase ) model.to(_lowercase ) model.eval() SCREAMING_SNAKE_CASE__ = model(_lowercase , attention_mask=_lowercase , labels=_lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __a ( self : int , _lowercase : Dict , _lowercase : Optional[Any] , _lowercase : int ): """simple docstring""" SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = GPTNeoXJapaneseForCausalLM(config=_lowercase ) model.to(_lowercase ) model.eval() # first forward pass SCREAMING_SNAKE_CASE__ = model(_lowercase , attention_mask=_lowercase , use_cache=_lowercase ) SCREAMING_SNAKE_CASE__ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids SCREAMING_SNAKE_CASE__ = ids_tensor((self.batch_size, 3) , config.vocab_size ) SCREAMING_SNAKE_CASE__ = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and SCREAMING_SNAKE_CASE__ = torch.cat([input_ids, next_tokens] , dim=-1 ) SCREAMING_SNAKE_CASE__ = torch.cat([input_mask, next_mask] , dim=-1 ) SCREAMING_SNAKE_CASE__ = model(_lowercase , attention_mask=_lowercase , output_hidden_states=_lowercase ) SCREAMING_SNAKE_CASE__ = output_from_no_past["""hidden_states"""][0] SCREAMING_SNAKE_CASE__ = model( _lowercase , attention_mask=_lowercase , past_key_values=_lowercase , output_hidden_states=_lowercase , )["""hidden_states"""][0] # select random slice SCREAMING_SNAKE_CASE__ = ids_tensor((1,) , output_from_past.shape[-1] ).item() SCREAMING_SNAKE_CASE__ = output_from_no_past[:, -3:, random_slice_idx].detach() SCREAMING_SNAKE_CASE__ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(_lowercase , _lowercase , atol=1E-3 ) ) def __a ( self : str ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = config_and_inputs SCREAMING_SNAKE_CASE__ = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __snake_case ( lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): lowerCAmelCase_ = (GPTNeoXJapaneseModel, GPTNeoXJapaneseForCausalLM) if is_torch_available() else () lowerCAmelCase_ = (GPTNeoXJapaneseForCausalLM,) if is_torch_available() else () lowerCAmelCase_ = ( {"feature-extraction": GPTNeoXJapaneseModel, "text-generation": GPTNeoXJapaneseForCausalLM} if is_torch_available() else {} ) lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False lowerCAmelCase_ = False def __a ( self : List[str] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = GPTNeoXJapaneseModelTester(self ) SCREAMING_SNAKE_CASE__ = ConfigTester(self , config_class=_lowercase , hidden_size=37 ) def __a ( self : Tuple ): """simple docstring""" self.config_tester.run_common_tests() def __a ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(_lowercase , _lowercase , _lowercase ) def __a ( self : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(_lowercase , _lowercase , _lowercase ) def __a ( self : str ): """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_decoder() SCREAMING_SNAKE_CASE__ = None self.model_tester.create_and_check_model_as_decoder(_lowercase , _lowercase , _lowercase ) def __a ( self : Union[str, Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(_lowercase , _lowercase , _lowercase ) def __a ( self : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*_lowercase ) @slow def __a ( self : Any ): """simple docstring""" SCREAMING_SNAKE_CASE__ = """abeja/gpt-neox-japanese-2.7b""" SCREAMING_SNAKE_CASE__ = ["""データサイエンティストとは、""", """100年後に必要とされる会社は、""", """フルリモートの環境で働くために必要なことは、""", """国境の長いトンネルを抜けると""", """美味しい日本食といえば、"""] SCREAMING_SNAKE_CASE__ = [ """データサイエンティストとは、データを分析し、ビジネスに役立つ知見を導き出す専門家のことです。""", """100年後に必要とされる会社は、「人」が中心の会社です。""", """フルリモートの環境で働くために必要なことは、「自分の時間をコントロールする」ことです。""", """国境の長いトンネルを抜けると、そこは雪国だった。""", """美味しい日本食といえば、やっぱりお寿司ですよね。""", ] SCREAMING_SNAKE_CASE__ = GPTNeoXJapaneseTokenizer.from_pretrained(_lowercase ) SCREAMING_SNAKE_CASE__ = GPTNeoXJapaneseForCausalLM.from_pretrained(_lowercase ) SCREAMING_SNAKE_CASE__ = [] for prompt in prompts: SCREAMING_SNAKE_CASE__ = tokenizer(_lowercase , return_tensors="""pt""" ).input_ids SCREAMING_SNAKE_CASE__ = model.generate(_lowercase , max_length=50 ) SCREAMING_SNAKE_CASE__ = tokenizer.batch_decode(_lowercase , skip_special_tokens=_lowercase ) predicted_outputs += generated_string self.assertListEqual(_lowercase , _lowercase )
219
1
from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class __a ( A__ ): _lowerCAmelCase : int = ['''vqvae'''] def __init__( self : Any , SCREAMING_SNAKE_CASE : AutoencoderKL , SCREAMING_SNAKE_CASE : UNetaDConditionModel , SCREAMING_SNAKE_CASE : Mel , SCREAMING_SNAKE_CASE : Union[DDIMScheduler, DDPMScheduler] , ): '''simple docstring''' super().__init__() self.register_modules(unet=SCREAMING_SNAKE_CASE , scheduler=SCREAMING_SNAKE_CASE , mel=SCREAMING_SNAKE_CASE , vqvae=SCREAMING_SNAKE_CASE ) def __lowercase ( self : List[Any] ): '''simple docstring''' return 50 if isinstance(self.scheduler , SCREAMING_SNAKE_CASE ) else 10_00 @torch.no_grad() def __call__( self : Optional[int] , SCREAMING_SNAKE_CASE : int = 1 , SCREAMING_SNAKE_CASE : str = None , SCREAMING_SNAKE_CASE : np.ndarray = None , SCREAMING_SNAKE_CASE : int = 0 , SCREAMING_SNAKE_CASE : int = 0 , SCREAMING_SNAKE_CASE : int = None , SCREAMING_SNAKE_CASE : torch.Generator = None , SCREAMING_SNAKE_CASE : float = 0 , SCREAMING_SNAKE_CASE : float = 0 , SCREAMING_SNAKE_CASE : torch.Generator = None , SCREAMING_SNAKE_CASE : float = 0 , SCREAMING_SNAKE_CASE : torch.Tensor = None , SCREAMING_SNAKE_CASE : torch.Tensor = None , SCREAMING_SNAKE_CASE : Optional[Any]=True , ): '''simple docstring''' UpperCamelCase__ : int = steps or self.get_default_steps() self.scheduler.set_timesteps(SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Any = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: UpperCamelCase__ : List[str] = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: UpperCamelCase__ : Optional[Any] = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=SCREAMING_SNAKE_CASE , device=self.device , ) UpperCamelCase__ : Dict = noise UpperCamelCase__ : Any = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Dict = self.mel.audio_slice_to_image(SCREAMING_SNAKE_CASE ) UpperCamelCase__ : str = np.frombuffer(input_image.tobytes() , dtype="uint8" ).reshape( (input_image.height, input_image.width) ) UpperCamelCase__ : str = (input_image / 2_55) * 2 - 1 UpperCamelCase__ : Dict = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: UpperCamelCase__ : str = self.vqvae.encode(torch.unsqueeze(SCREAMING_SNAKE_CASE , 0 ) ).latent_dist.sample( generator=SCREAMING_SNAKE_CASE )[0] UpperCamelCase__ : List[Any] = self.vqvae.config.scaling_factor * input_images if start_step > 0: UpperCamelCase__ : Optional[int] = self.scheduler.add_noise(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , self.scheduler.timesteps[start_step - 1] ) UpperCamelCase__ : List[Any] = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) UpperCamelCase__ : Optional[int] = int(mask_start_secs * pixels_per_second ) UpperCamelCase__ : List[Any] = int(mask_end_secs * pixels_per_second ) UpperCamelCase__ : int = self.scheduler.add_noise(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , SCREAMING_SNAKE_CASE ): UpperCamelCase__ : str = self.unet(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE )["sample"] else: UpperCamelCase__ : Optional[int] = self.unet(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE )["sample"] if isinstance(self.scheduler , SCREAMING_SNAKE_CASE ): UpperCamelCase__ : Optional[Any] = self.scheduler.step( model_output=SCREAMING_SNAKE_CASE , timestep=SCREAMING_SNAKE_CASE , sample=SCREAMING_SNAKE_CASE , eta=SCREAMING_SNAKE_CASE , generator=SCREAMING_SNAKE_CASE , )["prev_sample"] else: UpperCamelCase__ : List[Any] = self.scheduler.step( model_output=SCREAMING_SNAKE_CASE , timestep=SCREAMING_SNAKE_CASE , sample=SCREAMING_SNAKE_CASE , generator=SCREAMING_SNAKE_CASE , )["prev_sample"] if mask is not None: if mask_start > 0: UpperCamelCase__ : Any = mask[:, step, :, :mask_start] if mask_end > 0: UpperCamelCase__ : Optional[Any] = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance UpperCamelCase__ : int = 1 / self.vqvae.config.scaling_factor * images UpperCamelCase__ : Union[str, Any] = self.vqvae.decode(SCREAMING_SNAKE_CASE )["sample"] UpperCamelCase__ : List[Any] = (images / 2 + 0.5).clamp(0 , 1 ) UpperCamelCase__ : int = images.cpu().permute(0 , 2 , 3 , 1 ).numpy() UpperCamelCase__ : Tuple = (images * 2_55).round().astype("uint8" ) UpperCamelCase__ : int = list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(SCREAMING_SNAKE_CASE , mode="RGB" ).convert("L" ) for _ in images) ) UpperCamelCase__ : Tuple = [self.mel.image_to_audio(SCREAMING_SNAKE_CASE ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(SCREAMING_SNAKE_CASE )[:, np.newaxis, :] ) , **ImagePipelineOutput(SCREAMING_SNAKE_CASE ) ) @torch.no_grad() def __lowercase ( self : Union[str, Any] , SCREAMING_SNAKE_CASE : List[Image.Image] , SCREAMING_SNAKE_CASE : int = 50 ): '''simple docstring''' assert isinstance(self.scheduler , SCREAMING_SNAKE_CASE ) self.scheduler.set_timesteps(SCREAMING_SNAKE_CASE ) UpperCamelCase__ : List[str] = np.array( [np.frombuffer(image.tobytes() , dtype="uint8" ).reshape((1, image.height, image.width) ) for image in images] ) UpperCamelCase__ : Optional[Any] = (sample / 2_55) * 2 - 1 UpperCamelCase__ : List[Any] = torch.Tensor(SCREAMING_SNAKE_CASE ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): UpperCamelCase__ : Tuple = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps UpperCamelCase__ : Any = self.scheduler.alphas_cumprod[t] UpperCamelCase__ : str = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) UpperCamelCase__ : Tuple = 1 - alpha_prod_t UpperCamelCase__ : Tuple = self.unet(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE )["sample"] UpperCamelCase__ : List[Any] = (1 - alpha_prod_t_prev) ** 0.5 * model_output UpperCamelCase__ : int = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) UpperCamelCase__ : Optional[Any] = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def __lowercase ( SCREAMING_SNAKE_CASE : torch.Tensor , SCREAMING_SNAKE_CASE : torch.Tensor , SCREAMING_SNAKE_CASE : float ): '''simple docstring''' UpperCamelCase__ : List[str] = acos(torch.dot(torch.flatten(SCREAMING_SNAKE_CASE ) , torch.flatten(SCREAMING_SNAKE_CASE ) ) / torch.norm(SCREAMING_SNAKE_CASE ) / torch.norm(SCREAMING_SNAKE_CASE ) ) return sin((1 - alpha) * theta ) * xa / sin(SCREAMING_SNAKE_CASE ) + sin(alpha * theta ) * xa / sin(SCREAMING_SNAKE_CASE )
196
def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> int: UpperCamelCase__ : int = 1 for i in range(1 , num + 1 ): fact *= i return fact def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> int: UpperCamelCase__ : List[Any] = 0 while number > 0: UpperCamelCase__ : List[Any] = number % 10 sum_of_digits += last_digit UpperCamelCase__ : int = number // 10 # Removing the last_digit from the given number return sum_of_digits def SCREAMING_SNAKE_CASE ( __lowerCAmelCase = 100 ) -> int: UpperCamelCase__ : Optional[Any] = factorial(__lowerCAmelCase ) UpperCamelCase__ : Optional[int] = split_and_add(__lowerCAmelCase ) return result if __name__ == "__main__": print(solution(int(input('''Enter the Number: ''').strip())))
196
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer a__ : Optional[Any] = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast a__ : int = TaTokenizerFast a__ : List[Any] = {'configuration_mt5': ['MT5Config', 'MT5OnnxConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : str = [ 'MT5EncoderModel', 'MT5ForConditionalGeneration', 'MT5ForQuestionAnswering', 'MT5Model', 'MT5PreTrainedModel', 'MT5Stack', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : int = ['TFMT5EncoderModel', 'TFMT5ForConditionalGeneration', 'TFMT5Model'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Dict = ['FlaxMT5EncoderModel', 'FlaxMT5ForConditionalGeneration', 'FlaxMT5Model'] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys a__ : List[Any] = _LazyModule( __name__, globals()['__file__'], _import_structure, extra_objects={'MT5Tokenizer': MTaTokenizer, 'MT5TokenizerFast': MTaTokenizerFast}, module_spec=__spec__, )
80
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCamelCase : Optional[int] = { 'configuration_rag': ['RagConfig'], 'retrieval_rag': ['RagRetriever'], 'tokenization_rag': ['RagTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : List[Any] = [ 'RagModel', 'RagPreTrainedModel', 'RagSequenceForGeneration', 'RagTokenForGeneration', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : int = [ 'TFRagModel', 'TFRagPreTrainedModel', 'TFRagSequenceForGeneration', 'TFRagTokenForGeneration', ] if TYPE_CHECKING: from .configuration_rag import RagConfig from .retrieval_rag import RagRetriever from .tokenization_rag import RagTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rag import RagModel, RagPreTrainedModel, RagSequenceForGeneration, RagTokenForGeneration try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rag import ( TFRagModel, TFRagPreTrainedModel, TFRagSequenceForGeneration, TFRagTokenForGeneration, ) else: import sys lowerCamelCase : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
124
0
"""simple docstring""" from __future__ import annotations from decimal import Decimal from math import * # noqa: F403 from sympy import diff def _SCREAMING_SNAKE_CASE ( _lowercase : str , _lowercase : float | Decimal , _lowercase : float = 10**-10 ) ->float: '''simple docstring''' a : Tuple = a while True: a : Union[str, Any] = Decimal(_lowercase ) - ( Decimal(eval(_lowercase ) ) / Decimal(eval(str(diff(_lowercase ) ) ) ) # noqa: S307 ) # This number dictates the accuracy of the answer if abs(eval(_lowercase ) ) < precision: # noqa: S307 return float(_lowercase ) # Let's Execute if __name__ == "__main__": # Find root of trigonometric function # Find value of pi print(F'''The root of sin(x) = 0 is {newton_raphson("sin(x)", 2)}''') # Find root of polynomial print(F'''The root of x**2 - 5*x + 2 = 0 is {newton_raphson("x**2 - 5*x + 2", 0.4)}''') # Find Square Root of 5 print(F'''The root of log(x) - 1 = 0 is {newton_raphson("log(x) - 1", 2)}''') # Exponential Roots print(F'''The root of exp(x) - 1 = 0 is {newton_raphson("exp(x) - 1", 0)}''')
371
"""simple docstring""" from itertools import product def _SCREAMING_SNAKE_CASE ( _lowercase : int , _lowercase : int ) ->list[int]: '''simple docstring''' a : Dict = sides_number a : List[str] = max_face_number * dice_number a : Optional[int] = [0] * (max_total + 1) a : Dict = 1 a : Optional[Any] = range(_lowercase , max_face_number + 1 ) for dice_numbers in product(_lowercase , repeat=_lowercase ): a : Union[str, Any] = sum(_lowercase ) totals_frequencies[total] += 1 return totals_frequencies def _SCREAMING_SNAKE_CASE ( ) ->float: '''simple docstring''' a : str = total_frequency_distribution( sides_number=4 , dice_number=9 ) a : List[Any] = total_frequency_distribution( sides_number=6 , dice_number=6 ) a : Optional[Any] = 0 a : Tuple = 9 a : Union[str, Any] = 4 * 9 a : Any = 6 for peter_total in range(_lowercase , max_peter_total + 1 ): peter_wins_count += peter_totals_frequencies[peter_total] * sum( colin_totals_frequencies[min_colin_total:peter_total] ) a : List[str] = (4**9) * (6**6) a : List[Any] = peter_wins_count / total_games_number a : Any = round(_lowercase , ndigits=7 ) return rounded_peter_win_probability if __name__ == "__main__": print(F'''{solution() = }''')
79
0