code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase__ = { """configuration_x_clip""": [ """XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XCLIPConfig""", """XCLIPTextConfig""", """XCLIPVisionConfig""", ], """processing_x_clip""": ["""XCLIPProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ """XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """XCLIPModel""", """XCLIPPreTrainedModel""", """XCLIPTextModel""", """XCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
92
'''simple docstring''' def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): def update_area_of_max_square(UpperCAmelCase , UpperCAmelCase ) -> int: # BASE CASE if row >= rows or col >= cols: return 0 lowercase__ : int = update_area_of_max_square(UpperCAmelCase , col + 1 ) lowercase__ : Any = update_area_of_max_square(row + 1 , col + 1 ) lowercase__ : str = update_area_of_max_square(row + 1 , UpperCAmelCase ) if mat[row][col]: lowercase__ : List[Any] = 1 + min([right, diagonal, down] ) lowercase__ : List[Any] = max(largest_square_area[0] , UpperCAmelCase ) return sub_problem_sol else: return 0 lowercase__ : Dict = [0] update_area_of_max_square(0 , 0 ) return largest_square_area[0] def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): def update_area_of_max_square_using_dp_array( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> int: if row >= rows or col >= cols: return 0 if dp_array[row][col] != -1: return dp_array[row][col] lowercase__ : int = update_area_of_max_square_using_dp_array(UpperCAmelCase , col + 1 , UpperCAmelCase ) lowercase__ : Optional[int] = update_area_of_max_square_using_dp_array(row + 1 , col + 1 , UpperCAmelCase ) lowercase__ : Any = update_area_of_max_square_using_dp_array(row + 1 , UpperCAmelCase , UpperCAmelCase ) if mat[row][col]: lowercase__ : Optional[int] = 1 + min([right, diagonal, down] ) lowercase__ : Any = max(largest_square_area[0] , UpperCAmelCase ) lowercase__ : int = sub_problem_sol return sub_problem_sol else: return 0 lowercase__ : Any = [0] lowercase__ : List[Any] = [[-1] * cols for _ in range(UpperCAmelCase )] update_area_of_max_square_using_dp_array(0 , 0 , UpperCAmelCase ) return largest_square_area[0] def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): lowercase__ : Optional[int] = [[0] * (cols + 1) for _ in range(rows + 1 )] lowercase__ : str = 0 for row in range(rows - 1 , -1 , -1 ): for col in range(cols - 1 , -1 , -1 ): lowercase__ : str = dp_array[row][col + 1] lowercase__ : Optional[Any] = dp_array[row + 1][col + 1] lowercase__ : str = dp_array[row + 1][col] if mat[row][col] == 1: lowercase__ : Dict = 1 + min(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) lowercase__ : str = max(dp_array[row][col] , UpperCAmelCase ) else: lowercase__ : Any = 0 return largest_square_area def __UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): lowercase__ : List[str] = [0] * (cols + 1) lowercase__ : str = [0] * (cols + 1) lowercase__ : Tuple = 0 for row in range(rows - 1 , -1 , -1 ): for col in range(cols - 1 , -1 , -1 ): lowercase__ : List[Any] = current_row[col + 1] lowercase__ : Any = next_row[col + 1] lowercase__ : Optional[Any] = next_row[col] if mat[row][col] == 1: lowercase__ : List[str] = 1 + min(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) lowercase__ : List[str] = max(current_row[col] , UpperCAmelCase ) else: lowercase__ : int = 0 lowercase__ : int = current_row return largest_square_area if __name__ == "__main__": import doctest doctest.testmod() print(largest_square_area_in_matrix_bottom_up(2, 2, [[1, 1], [1, 1]]))
198
0
import random import unittest import numpy as np import torch from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionUpscalePipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class lowercase__ ( _A , unittest.TestCase): # TODO: is there an appropriate internal test set? UpperCamelCase_ = """ssube/stable-diffusion-x4-upscaler-onnx""" def __A ( self : str , UpperCamelCase__ : int=0 ): '''simple docstring''' SCREAMING_SNAKE_CASE : Dict = floats_tensor((1, 3, 128, 128) , rng=random.Random(__SCREAMING_SNAKE_CASE ) ) SCREAMING_SNAKE_CASE : Any = torch.manual_seed(__SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE : Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def __A ( self : List[str] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE : str = self.get_dummy_inputs() SCREAMING_SNAKE_CASE : Dict = pipe(**__SCREAMING_SNAKE_CASE ).images SCREAMING_SNAKE_CASE : Dict = image[0, -3:, -3:, -1].flatten() # started as 128, should now be 512 assert image.shape == (1, 512, 512, 3) SCREAMING_SNAKE_CASE : int = np.array( [0.697_4782, 0.6890_2093, 0.7013_5885, 0.758_3618, 0.780_4545, 0.785_4912, 0.7866_7426, 0.7874_3863, 0.7807_0223] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def __A ( self : int ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) SCREAMING_SNAKE_CASE : str = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE : Union[str, Any] = self.get_dummy_inputs() SCREAMING_SNAKE_CASE : int = pipe(**__SCREAMING_SNAKE_CASE ).images SCREAMING_SNAKE_CASE : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) SCREAMING_SNAKE_CASE : Tuple = np.array( [0.689_8892, 0.5924_0556, 0.5249_9527, 0.5886_6215, 0.5225_8235, 0.5257_2715, 0.6241_4473, 0.617_4387, 0.621_4964] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __A ( self : Any ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) SCREAMING_SNAKE_CASE : Tuple = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE : List[str] = self.get_dummy_inputs() SCREAMING_SNAKE_CASE : int = pipe(**__SCREAMING_SNAKE_CASE ).images SCREAMING_SNAKE_CASE : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) SCREAMING_SNAKE_CASE : List[str] = np.array( [0.765_9278, 0.7643_7664, 0.7557_9107, 0.769_1116, 0.7766_6986, 0.772_7672, 0.775_8664, 0.781_2226, 0.7694_2515] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __A ( self : Optional[int] ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) SCREAMING_SNAKE_CASE : str = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE : List[Any] = self.get_dummy_inputs() SCREAMING_SNAKE_CASE : str = pipe(**__SCREAMING_SNAKE_CASE ).images SCREAMING_SNAKE_CASE : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) SCREAMING_SNAKE_CASE : List[Any] = np.array( [0.697_4782, 0.6890_2093, 0.7013_5885, 0.758_3618, 0.780_4545, 0.785_4912, 0.7866_7426, 0.7874_3863, 0.7807_0223] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def __A ( self : Union[str, Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) SCREAMING_SNAKE_CASE : str = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE : Dict = self.get_dummy_inputs() SCREAMING_SNAKE_CASE : Dict = pipe(**__SCREAMING_SNAKE_CASE ).images SCREAMING_SNAKE_CASE : int = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) SCREAMING_SNAKE_CASE : Optional[int] = np.array( [0.7742_4496, 0.77_3601, 0.764_5288, 0.776_9598, 0.777_2739, 0.773_8688, 0.7818_7233, 0.7787_9584, 0.76_7043] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class lowercase__ ( unittest.TestCase): @property def __A ( self : List[str] ): '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def __A ( self : Dict ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[Any] = ort.SessionOptions() SCREAMING_SNAKE_CASE : Optional[Any] = False return options def __A ( self : Union[str, Any] ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) SCREAMING_SNAKE_CASE : str = init_image.resize((128, 128) ) # using the PNDM scheduler by default SCREAMING_SNAKE_CASE : Any = OnnxStableDiffusionUpscalePipeline.from_pretrained( '''ssube/stable-diffusion-x4-upscaler-onnx''' , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE : Optional[int] = '''A fantasy landscape, trending on artstation''' SCREAMING_SNAKE_CASE : Dict = torch.manual_seed(0 ) SCREAMING_SNAKE_CASE : Optional[Any] = pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , guidance_scale=7.5 , num_inference_steps=10 , generator=__SCREAMING_SNAKE_CASE , output_type='''np''' , ) SCREAMING_SNAKE_CASE : int = output.images SCREAMING_SNAKE_CASE : Tuple = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 512, 3) SCREAMING_SNAKE_CASE : List[str] = np.array([0.4883, 0.4947, 0.4980, 0.4975, 0.4982, 0.4980, 0.5000, 0.5006, 0.4972] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def __A ( self : List[str] ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) SCREAMING_SNAKE_CASE : Tuple = init_image.resize((128, 128) ) SCREAMING_SNAKE_CASE : Tuple = LMSDiscreteScheduler.from_pretrained( '''ssube/stable-diffusion-x4-upscaler-onnx''' , subfolder='''scheduler''' ) SCREAMING_SNAKE_CASE : Union[str, Any] = OnnxStableDiffusionUpscalePipeline.from_pretrained( '''ssube/stable-diffusion-x4-upscaler-onnx''' , scheduler=__SCREAMING_SNAKE_CASE , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE : Any = '''A fantasy landscape, trending on artstation''' SCREAMING_SNAKE_CASE : List[str] = torch.manual_seed(0 ) SCREAMING_SNAKE_CASE : Tuple = pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , guidance_scale=7.5 , num_inference_steps=20 , generator=__SCREAMING_SNAKE_CASE , output_type='''np''' , ) SCREAMING_SNAKE_CASE : Optional[Any] = output.images SCREAMING_SNAKE_CASE : Dict = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 512, 3) SCREAMING_SNAKE_CASE : List[Any] = np.array( [0.5017_3753, 0.5022_3356, 0.50_2039, 0.5023_3036, 0.502_3725, 0.502_2601, 0.501_8758, 0.5023_4085, 0.5024_1566] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
363
import re def A ( _lowercase ): SCREAMING_SNAKE_CASE : Any = re.compile(R'''^(\+91[\-\s]?)?[0]?(91)?[789]\d{9}$''' ) if match := re.search(_lowercase , _lowercase ): return match.string == phone return False if __name__ == "__main__": print(indian_phone_validator('+918827897895'))
258
0
'''simple docstring''' from timeit import timeit _lowerCamelCase : List[str] = { "MALAYALAM": True, "String": False, "rotor": True, "level": True, "A": True, "BB": True, "ABC": False, "amanaplanacanalpanama": True, # "a man a plan a canal panama" } # Ensure our test data is valid assert all((key == key[::-1]) is value for key, value in test_data.items()) def __lowerCamelCase ( A__ ) -> bool: """simple docstring""" UpperCamelCase = 0 UpperCamelCase = len(A__ ) - 1 while start_i < end_i: if s[start_i] == s[end_i]: start_i += 1 end_i -= 1 else: return False return True def __lowerCamelCase ( A__ ) -> bool: """simple docstring""" UpperCamelCase = len(A__ ) // 2 UpperCamelCase = len(A__ ) # We need to traverse till half of the length of string # as we can get access of the i'th last element from # i'th index. # eg: [0,1,2,3,4,5] => 4th index can be accessed # with the help of 1st index (i==n-i-1) # where n is length of string return all(s[i] == s[n - i - 1] for i in range(A__ ) ) def __lowerCamelCase ( A__ ) -> bool: """simple docstring""" if len(A__ ) <= 2: return True if s[0] == s[len(A__ ) - 1]: return is_palindrome_recursive(s[1:-1] ) else: return False def __lowerCamelCase ( A__ ) -> bool: """simple docstring""" return s == s[::-1] def __lowerCamelCase ( A__ ) -> None: """simple docstring""" UpperCamelCase = F"""all({name}(key) is value for key, value in test_data.items())""" UpperCamelCase = F"""from __main__ import test_data, {name}""" UpperCamelCase = 500_000 UpperCamelCase = timeit(stmt=A__ , setup=A__ , number=A__ ) print(F"""{name:<35} finished {number:,} runs in {result:.5f} seconds""" ) if __name__ == "__main__": for key, value in test_data.items(): assert is_palindrome(key) is is_palindrome_recursive(key) assert is_palindrome(key) is is_palindrome_slice(key) print(f'''{key:21} {value}''') print("a man a plan a canal panama") # finished 500,000 runs in 0.46793 seconds benchmark_function("is_palindrome_slice") # finished 500,000 runs in 0.85234 seconds benchmark_function("is_palindrome") # finished 500,000 runs in 1.32028 seconds benchmark_function("is_palindrome_recursive") # finished 500,000 runs in 2.08679 seconds benchmark_function("is_palindrome_traversal")
28
import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class _lowerCamelCase( unittest.TestCase ): lowercase_ : Dict = JukeboxTokenizer lowercase_ : Dict = { """artist""": """Zac Brown Band""", """genres""": """Country""", """lyrics""": """I met a traveller from an antique land, Who said \"Two vast and trunkless legs of stone Stand in the desert. . . . Near them, on the sand, Half sunk a shattered visage lies, whose frown, And wrinkled lip, and sneer of cold command, Tell that its sculptor well those passions read Which yet survive, stamped on these lifeless things, The hand that mocked them, and the heart that fed; And on the pedestal, these words appear: My name is Ozymandias, King of Kings; Look on my Works, ye Mighty, and despair! Nothing beside remains. Round the decay Of that colossal Wreck, boundless and bare The lone and level sands stretch far away """, } @require_torch def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" import torch _lowercase : str = JukeboxTokenizer.from_pretrained('openai/jukebox-1b-lyrics') _lowercase : Optional[Any] = tokenizer(**self.metas)['input_ids'] # fmt: off _lowercase : Optional[int] = [ torch.tensor([[ 0, 0, 0, 71_69, 5_07, 9, 76, 39, 31, 46, 76, 27, 76, 46, 44, 27, 48, 31, 38, 38, 31, 44, 76, 32, 44, 41, 39, 76, 27, 40, 76, 27, 40, 46, 35, 43, 47, 31, 76, 38, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 23, 34, 41, 76, 45, 27, 35, 30, 76, 71, 20, 49, 41, 76, 48, 27, 45, 46, 76, 27, 40, 30, 76, 46, 44, 47, 40, 37, 38, 31, 45, 45, 76, 38, 31, 33, 45, 76, 41, 32, 76, 45, 46, 41, 40, 31, 78, 76, 76, 76, 76, 76, 76, 76, 76, 19, 46, 27, 40, 30, 76, 35, 40, 76, 46, 34, 31, 76, 30, 31, 45, 31, 44, 46, 63, 76, 63, 76, 63, 76, 63, 76, 14, 31, 27, 44, 76, 46, 34, 31, 39, 64, 76, 41, 40, 76, 46, 34, 31, 76, 45, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 8, 27, 38, 32, 76, 45, 47, 40, 37, 76, 27, 76, 45, 34, 27, 46, 46, 31, 44, 31, 30, 76, 48, 35, 45, 27, 33, 31, 76, 38, 35, 31, 45, 64, 76, 49, 34, 41, 45, 31, 76, 32, 44, 41, 49, 40, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 1, 40, 30, 76, 49, 44, 35, 40, 37, 38, 31, 30, 76, 38, 35, 42, 64, 76, 27, 40, 30, 76, 45, 40, 31, 31, 44, 76, 41, 32, 76, 29, 41, 38, 30, 76, 29, 41, 39, 39, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 31, 38, 38, 76, 46, 34, 27, 46, 76, 35, 46, 45, 76, 45, 29, 47, 38, 42, 46, 41, 44, 76, 49, 31, 38, 38, 76, 46, 34, 41, 45, 31, 76, 42, 27, 45, 45, 35, 41, 40, 45, 76, 44, 31, 27, 30, 78, 76, 76, 76, 76, 76, 76, 76, 76, 23, 34, 35, 29, 34, 76, 51, 31, 46, 76, 45, 47, 44, 48, 35, 48, 31, 64, 76, 45, 46, 27, 39, 42, 31, 30, 76, 41, 40, 76, 46, 34, 31, 45, 31, 76, 38, 35, 32, 31, 38, 31, 45, 45, 76, 46, 34, 35, 40, 33, 45, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 34, 31, 76, 34, 27, 40, 30, 76, 46, 34, 27, 46, 76, 39, 41, 29, 37, 31, 30, 76, 46, 34, 31, 39, 64, 76, 27, 40, 30, 76, 46, 34, 31, 76, 34, 31, 27, 44, 46, 76, 46, 34, 27, 46, 76, 32, 31, 30, 66, 78, 76, 76, 76, 76, 76, 76, 76, 76, 1, 40, 30, 76, 41, 40, 76, 46, 34, 31, 76, 42, 31, 30, 31, 45, 46, 27, 38, 64, 76, 46, 34, 31, 45, 31, 76, 49, 41, 44, 30, 45, 76, 27, 42, 42, 31, 27, 44, 65, 78, 76, 76, 76, 76, 76, 76, 76, 76, 13, 51, 76, 40, 27, 39, 31, 76, 35, 45, 76, 15, 52, 51, 39, 27, 40, 30, 35, 27, 45, 64, 76, 11, 35, 40, 33, 76, 41, 32, 76, 11, 35, 40, 33, 45, 66, 78, 76, 76, 76, 76, 76, 76, 76, 76, 12, 41, 41, 37, 76, 41, 40, 76, 39, 51, 76, 23, 41, 44, 37, 45, 64, 76, 51, 31, 76, 13, 35, 33, 34, 46, 51, 64, 76, 27, 40, 30, 76, 30, 31, 45, 42, 27, 35, 44, 67, 78, 76, 76, 76, 76, 76, 76, 76, 76, 14, 41, 46, 34, 35, 40, 33, 76, 28, 31, 45, 35, 30, 31, 76, 44, 31, 39, 27, 35, 40, 45, 63, 76, 18, 41, 47, 40, 30, 76, 46, 34, 31, 76, 30, 31, 29, 27, 51, 78, 76, 76, 76, 76, 76, 76, 76, 76, 15, 32, 76, 46, 34, 27, 46, 76, 29, 41, 38, 41, 45, 45, 27, 38, 76, 23, 44, 31, 29, 37, 64, 76, 28, 41, 47, 40, 30, 38, 31, 45, 45, 76, 27, 40, 30, 76, 28, 27, 44, 31, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 34, 31, 76, 38, 41, 40, 31, 76, 27, 40, 30, 76, 38, 31, 48, 31, 38, 76, 45, 27, 40, 30, 45, 76, 45, 46, 44, 31, 46, 29, 34, 76, 32, 27, 44, 76, 27, 49, 27, 51, 78, 76, 76, 76, 76, 76, 76, 76, 76]]), torch.tensor([[0, 0, 0, 10_69, 11]]), torch.tensor([[0, 0, 0, 10_69, 11]]), ] # fmt: on self.assertTrue(torch.allclose(tokens[0], EXPECTED_OUTPUT[0])) self.assertTrue(torch.allclose(tokens[1], EXPECTED_OUTPUT[1])) self.assertTrue(torch.allclose(tokens[2], EXPECTED_OUTPUT[2])) @require_torch def UpperCamelCase ( self) -> int: """simple docstring""" import torch _lowercase : List[str] = JukeboxTokenizer.from_pretrained('openai/jukebox-5b-lyrics') _lowercase : List[str] = tokenizer(**self.metas)['input_ids'] # fmt: off _lowercase : Optional[int] = [ torch.tensor([[ 0, 0, 0, 10_69, 11, -1, -1, -1, -1, 9, 77, 39, 31, 46, 77, 27, 77, 46, 44, 27, 48, 31, 38, 38, 31, 44, 77, 32, 44, 41, 39, 77, 27, 40, 77, 27, 40, 46, 35, 43, 47, 31, 77, 38, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 23, 34, 41, 77, 45, 27, 35, 30, 77, 72, 20, 49, 41, 77, 48, 27, 45, 46, 77, 27, 40, 30, 77, 46, 44, 47, 40, 37, 38, 31, 45, 45, 77, 38, 31, 33, 45, 77, 41, 32, 77, 45, 46, 41, 40, 31, 79, 77, 77, 77, 77, 77, 77, 77, 77, 19, 46, 27, 40, 30, 77, 35, 40, 77, 46, 34, 31, 77, 30, 31, 45, 31, 44, 46, 63, 77, 63, 77, 63, 77, 63, 77, 14, 31, 27, 44, 77, 46, 34, 31, 39, 64, 77, 41, 40, 77, 46, 34, 31, 77, 45, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 8, 27, 38, 32, 77, 45, 47, 40, 37, 77, 27, 77, 45, 34, 27, 46, 46, 31, 44, 31, 30, 77, 48, 35, 45, 27, 33, 31, 77, 38, 35, 31, 45, 64, 77, 49, 34, 41, 45, 31, 77, 32, 44, 41, 49, 40, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 1, 40, 30, 77, 49, 44, 35, 40, 37, 38, 31, 30, 77, 38, 35, 42, 64, 77, 27, 40, 30, 77, 45, 40, 31, 31, 44, 77, 41, 32, 77, 29, 41, 38, 30, 77, 29, 41, 39, 39, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 31, 38, 38, 77, 46, 34, 27, 46, 77, 35, 46, 45, 77, 45, 29, 47, 38, 42, 46, 41, 44, 77, 49, 31, 38, 38, 77, 46, 34, 41, 45, 31, 77, 42, 27, 45, 45, 35, 41, 40, 45, 77, 44, 31, 27, 30, 79, 77, 77, 77, 77, 77, 77, 77, 77, 23, 34, 35, 29, 34, 77, 51, 31, 46, 77, 45, 47, 44, 48, 35, 48, 31, 64, 77, 45, 46, 27, 39, 42, 31, 30, 77, 41, 40, 77, 46, 34, 31, 45, 31, 77, 38, 35, 32, 31, 38, 31, 45, 45, 77, 46, 34, 35, 40, 33, 45, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 34, 31, 77, 34, 27, 40, 30, 77, 46, 34, 27, 46, 77, 39, 41, 29, 37, 31, 30, 77, 46, 34, 31, 39, 64, 77, 27, 40, 30, 77, 46, 34, 31, 77, 34, 31, 27, 44, 46, 77, 46, 34, 27, 46, 77, 32, 31, 30, 66, 79, 77, 77, 77, 77, 77, 77, 77, 77, 1, 40, 30, 77, 41, 40, 77, 46, 34, 31, 77, 42, 31, 30, 31, 45, 46, 27, 38, 64, 77, 46, 34, 31, 45, 31, 77, 49, 41, 44, 30, 45, 77, 27, 42, 42, 31, 27, 44, 65, 79, 77, 77, 77, 77, 77, 77, 77, 77, 13, 51, 77, 40, 27, 39, 31, 77, 35, 45, 77, 15, 52, 51, 39, 27, 40, 30, 35, 27, 45, 64, 77, 11, 35, 40, 33, 77, 41, 32, 77, 11, 35, 40, 33, 45, 66, 79, 77, 77, 77, 77, 77, 77, 77, 77, 12, 41, 41, 37, 77, 41, 40, 77, 39, 51, 77, 23, 41, 44, 37, 45, 64, 77, 51, 31, 77, 13, 35, 33, 34, 46, 51, 64, 77, 27, 40, 30, 77, 30, 31, 45, 42, 27, 35, 44, 67, 79, 77, 77, 77, 77, 77, 77, 77, 77, 14, 41, 46, 34, 35, 40, 33, 77, 28, 31, 45, 35, 30, 31, 77, 44, 31, 39, 27, 35, 40, 45, 63, 77, 18, 41, 47, 40, 30, 77, 46, 34, 31, 77, 30, 31, 29, 27, 51, 79, 77, 77, 77, 77, 77, 77, 77, 77, 15, 32, 77, 46, 34, 27, 46, 77, 29, 41, 38, 41, 45, 45, 27, 38, 77, 23, 44, 31, 29, 37, 64, 77, 28, 41, 47, 40, 30, 38, 31, 45, 45, 77, 27, 40, 30, 77, 28, 27, 44, 31, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 34, 31, 77, 38, 41, 40, 31, 77, 27, 40, 30, 77, 38, 31, 48, 31, 38, 77, 45, 27, 40, 30, 45, 77, 45, 46, 44, 31, 46, 29, 34, 77, 32, 27, 44, 77, 27, 49, 27, 51, 79, 77, 77, 77, 77, 77, 77, 77, 77]]), torch.tensor([[0, 0, 0, 10_69, 11, -1, -1, -1, -1]]), torch.tensor([[0, 0, 0, 10_69, 11, -1, -1, -1, -1]]), ] # fmt: on self.assertTrue(torch.allclose(tokens[0], EXPECTED_OUTPUT[0])) self.assertTrue(torch.allclose(tokens[1], EXPECTED_OUTPUT[1])) self.assertTrue(torch.allclose(tokens[2], EXPECTED_OUTPUT[2]))
21
0
"""simple docstring""" from typing import Dict, Optional import numpy as np import datasets __A = '\nIoU is the area of overlap between the predicted segmentation and the ground truth divided by the area of union\nbetween the predicted segmentation and the ground truth. For binary (two classes) or multi-class segmentation,\nthe mean IoU of the image is calculated by taking the IoU of each class and averaging them.\n' __A = '\nArgs:\n predictions (`List[ndarray]`):\n List of predicted segmentation maps, each of shape (height, width). Each segmentation map can be of a different size.\n references (`List[ndarray]`):\n List of ground truth segmentation maps, each of shape (height, width). Each segmentation map can be of a different size.\n num_labels (`int`):\n Number of classes (categories).\n ignore_index (`int`):\n Index that will be ignored during evaluation.\n nan_to_num (`int`, *optional*):\n If specified, NaN values will be replaced by the number defined by the user.\n label_map (`dict`, *optional*):\n If specified, dictionary mapping old label indices to new label indices.\n reduce_labels (`bool`, *optional*, defaults to `False`):\n Whether or not to reduce all label values of segmentation maps by 1. Usually used for datasets where 0 is used for background,\n and background itself is not included in all classes of a dataset (e.g. ADE20k). The background label will be replaced by 255.\n\nReturns:\n `Dict[str, float | ndarray]` comprising various elements:\n - *mean_iou* (`float`):\n Mean Intersection-over-Union (IoU averaged over all categories).\n - *mean_accuracy* (`float`):\n Mean accuracy (averaged over all categories).\n - *overall_accuracy* (`float`):\n Overall accuracy on all images.\n - *per_category_accuracy* (`ndarray` of shape `(num_labels,)`):\n Per category accuracy.\n - *per_category_iou* (`ndarray` of shape `(num_labels,)`):\n Per category IoU.\n\nExamples:\n\n >>> import numpy as np\n\n >>> mean_iou = datasets.load_metric("mean_iou")\n\n >>> # suppose one has 3 different segmentation maps predicted\n >>> predicted_1 = np.array([[1, 2], [3, 4], [5, 255]])\n >>> actual_1 = np.array([[0, 3], [5, 4], [6, 255]])\n\n >>> predicted_2 = np.array([[2, 7], [9, 2], [3, 6]])\n >>> actual_2 = np.array([[1, 7], [9, 2], [3, 6]])\n\n >>> predicted_3 = np.array([[2, 2, 3], [8, 2, 4], [3, 255, 2]])\n >>> actual_3 = np.array([[1, 2, 2], [8, 2, 1], [3, 255, 1]])\n\n >>> predicted = [predicted_1, predicted_2, predicted_3]\n >>> ground_truth = [actual_1, actual_2, actual_3]\n\n >>> results = mean_iou.compute(predictions=predicted, references=ground_truth, num_labels=10, ignore_index=255, reduce_labels=False)\n >>> print(results) # doctest: +NORMALIZE_WHITESPACE\n {\'mean_iou\': 0.47750000000000004, \'mean_accuracy\': 0.5916666666666666, \'overall_accuracy\': 0.5263157894736842, \'per_category_iou\': array([0. , 0. , 0.375, 0.4 , 0.5 , 0. , 0.5 , 1. , 1. , 1. ]), \'per_category_accuracy\': array([0. , 0. , 0.75 , 0.66666667, 1. , 0. , 0.5 , 1. , 1. , 1. ])}\n' __A = '\\n@software{MMSegmentation_Contributors_OpenMMLab_Semantic_Segmentation_2020,\nauthor = {{MMSegmentation Contributors}},\nlicense = {Apache-2.0},\nmonth = {7},\ntitle = {{OpenMMLab Semantic Segmentation Toolbox and Benchmark}},\nurl = {https://github.com/open-mmlab/mmsegmentation},\nyear = {2020}\n}' def lowerCamelCase_ ( UpperCamelCase__ : List[Any] , UpperCamelCase__ : int , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Optional[int] = None , UpperCamelCase__ : str = False , ) -> Dict: """simple docstring""" if label_map is not None: for old_id, new_id in label_map.items(): __lowerCamelCase = new_id # turn into Numpy arrays __lowerCamelCase = np.array(a__ ) __lowerCamelCase = np.array(a__ ) if reduce_labels: __lowerCamelCase = 255 __lowerCamelCase = label - 1 __lowerCamelCase = 255 __lowerCamelCase = label != ignore_index __lowerCamelCase = np.not_equal(a__ , a__ ) __lowerCamelCase = pred_label[mask] __lowerCamelCase = np.array(a__ )[mask] __lowerCamelCase = pred_label[pred_label == label] __lowerCamelCase = np.histogram(a__ , bins=a__ , range=(0, num_labels - 1) )[0] __lowerCamelCase = np.histogram(a__ , bins=a__ , range=(0, num_labels - 1) )[0] __lowerCamelCase = np.histogram(a__ , bins=a__ , range=(0, num_labels - 1) )[0] __lowerCamelCase = area_pred_label + area_label - area_intersect return area_intersect, area_union, area_pred_label, area_label def lowerCamelCase_ ( UpperCamelCase__ : Any , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Any , UpperCamelCase__ : List[str] , UpperCamelCase__ : str = None , UpperCamelCase__ : Tuple = False , ) -> List[Any]: """simple docstring""" __lowerCamelCase = np.zeros((num_labels,) , dtype=np.floataa ) __lowerCamelCase = np.zeros((num_labels,) , dtype=np.floataa ) __lowerCamelCase = np.zeros((num_labels,) , dtype=np.floataa ) __lowerCamelCase = np.zeros((num_labels,) , dtype=np.floataa ) for result, gt_seg_map in zip(a__ , a__ ): __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = intersect_and_union( a__ , a__ , a__ , a__ , a__ , a__ ) total_area_intersect += area_intersect total_area_union += area_union total_area_pred_label += area_pred_label total_area_label += area_label return total_area_intersect, total_area_union, total_area_pred_label, total_area_label def lowerCamelCase_ ( UpperCamelCase__ : Dict , UpperCamelCase__ : Any , UpperCamelCase__ : Tuple , UpperCamelCase__ : Dict , UpperCamelCase__ : Dict = None , UpperCamelCase__ : str = None , UpperCamelCase__ : Optional[int] = False , ) -> str: """simple docstring""" __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = total_intersect_and_union( a__ , a__ , a__ , a__ , a__ , a__ ) # compute metrics __lowerCamelCase = {} __lowerCamelCase = total_area_intersect.sum() / total_area_label.sum() __lowerCamelCase = total_area_intersect / total_area_union __lowerCamelCase = total_area_intersect / total_area_label __lowerCamelCase = np.nanmean(a__ ) __lowerCamelCase = np.nanmean(a__ ) __lowerCamelCase = all_acc __lowerCamelCase = iou __lowerCamelCase = acc if nan_to_num is not None: __lowerCamelCase = {metric: np.nan_to_num(a__ , nan=a__ ) for metric, metric_value in metrics.items()} return metrics @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowerCAmelCase ( datasets.Metric ): """simple docstring""" def lowercase_ ( self ) -> Optional[int]: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( # 1st Seq - height dim, 2nd - width dim { 'predictions': datasets.Sequence(datasets.Sequence(datasets.Value('uint16' ) ) ), 'references': datasets.Sequence(datasets.Sequence(datasets.Value('uint16' ) ) ), } ) , reference_urls=[ 'https://github.com/open-mmlab/mmsegmentation/blob/71c201b1813267d78764f306a297ca717827c4bf/mmseg/core/evaluation/metrics.py' ] , ) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = False , ) -> Optional[int]: '''simple docstring''' __lowerCamelCase = mean_iou( results=_snake_case , gt_seg_maps=_snake_case , num_labels=_snake_case , ignore_index=_snake_case , nan_to_num=_snake_case , label_map=_snake_case , reduce_labels=_snake_case , ) return iou_result
361
import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() __A = logging.get_logger(__name__) __A = ["model.decoder.embed_positions.weights"] def lowerCamelCase_ ( UpperCamelCase__ : List[Any] ) -> List[Any]: """simple docstring""" if "emb" in name: __lowerCamelCase = name.replace('emb' , 'model.decoder.embed_tokens' ) if "transformer" in name: __lowerCamelCase = name.replace('transformer' , 'model.decoder' ) if "cross_attention" in name: __lowerCamelCase = name.replace('cross_attention' , 'encoder_attn' ) if "linear1" in name: __lowerCamelCase = name.replace('linear1' , 'fc1' ) if "linear2" in name: __lowerCamelCase = name.replace('linear2' , 'fc2' ) if "norm1" in name: __lowerCamelCase = name.replace('norm1' , 'self_attn_layer_norm' ) if "norm_cross" in name: __lowerCamelCase = name.replace('norm_cross' , 'encoder_attn_layer_norm' ) if "norm2" in name: __lowerCamelCase = name.replace('norm2' , 'final_layer_norm' ) if "out_norm" in name: __lowerCamelCase = name.replace('out_norm' , 'model.decoder.layer_norm' ) if "linears" in name: __lowerCamelCase = name.replace('linears' , 'lm_heads' ) if "condition_provider.conditioners.description.output_proj" in name: __lowerCamelCase = name.replace('condition_provider.conditioners.description.output_proj' , 'enc_to_dec_proj' ) return name def lowerCamelCase_ ( UpperCamelCase__ : OrderedDict , UpperCamelCase__ : int ) -> Tuple[Dict, Dict]: """simple docstring""" __lowerCamelCase = list(state_dict.keys() ) __lowerCamelCase = {} for key in keys: __lowerCamelCase = state_dict.pop(UpperCamelCase__ ) __lowerCamelCase = rename_keys(UpperCamelCase__ ) if "in_proj_weight" in key: # split fused qkv proj __lowerCamelCase = val[:hidden_size, :] __lowerCamelCase = val[hidden_size : 2 * hidden_size, :] __lowerCamelCase = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: __lowerCamelCase = val else: __lowerCamelCase = val return state_dict, enc_dec_proj_state_dict def lowerCamelCase_ ( UpperCamelCase__ : str ) -> MusicgenDecoderConfig: """simple docstring""" if checkpoint == "small": # default config values __lowerCamelCase = 1024 __lowerCamelCase = 24 __lowerCamelCase = 16 elif checkpoint == "medium": __lowerCamelCase = 1536 __lowerCamelCase = 48 __lowerCamelCase = 24 elif checkpoint == "large": __lowerCamelCase = 2048 __lowerCamelCase = 48 __lowerCamelCase = 32 else: raise ValueError(F"""Checkpoint should be one of `['small', 'medium', 'large']`, got {checkpoint}.""" ) __lowerCamelCase = MusicgenDecoderConfig( hidden_size=UpperCamelCase__ , ffn_dim=hidden_size * 4 , num_hidden_layers=UpperCamelCase__ , num_attention_heads=UpperCamelCase__ , ) return config @torch.no_grad() def lowerCamelCase_ ( UpperCamelCase__ : Any , UpperCamelCase__ : Dict=None , UpperCamelCase__ : Dict=None , UpperCamelCase__ : Optional[int]="cpu" ) -> List[Any]: """simple docstring""" __lowerCamelCase = MusicGen.get_pretrained(UpperCamelCase__ , device=UpperCamelCase__ ) __lowerCamelCase = decoder_config_from_checkpoint(UpperCamelCase__ ) __lowerCamelCase = fairseq_model.lm.state_dict() __lowerCamelCase , __lowerCamelCase = rename_state_dict( UpperCamelCase__ , hidden_size=decoder_config.hidden_size ) __lowerCamelCase = TaEncoderModel.from_pretrained('t5-base' ) __lowerCamelCase = EncodecModel.from_pretrained('facebook/encodec_32khz' ) __lowerCamelCase = MusicgenForCausalLM(UpperCamelCase__ ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection __lowerCamelCase , __lowerCamelCase = decoder.load_state_dict(UpperCamelCase__ , strict=UpperCamelCase__ ) for key in missing_keys.copy(): if key.startswith(('text_encoder', 'audio_encoder') ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(UpperCamelCase__ ) if len(UpperCamelCase__ ) > 0: raise ValueError(F"""Missing key(s) in state_dict: {missing_keys}""" ) if len(UpperCamelCase__ ) > 0: raise ValueError(F"""Unexpected key(s) in state_dict: {unexpected_keys}""" ) # init the composite model __lowerCamelCase = MusicgenForConditionalGeneration(text_encoder=UpperCamelCase__ , audio_encoder=UpperCamelCase__ , decoder=UpperCamelCase__ ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(UpperCamelCase__ ) # check we can do a forward pass __lowerCamelCase = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) __lowerCamelCase = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): __lowerCamelCase = model(input_ids=UpperCamelCase__ , decoder_input_ids=UpperCamelCase__ ).logits if logits.shape != (8, 1, 2048): raise ValueError('Incorrect shape for logits' ) # now construct the processor __lowerCamelCase = AutoTokenizer.from_pretrained('t5-base' ) __lowerCamelCase = AutoFeatureExtractor.from_pretrained('facebook/encodec_32khz' , padding_side='left' ) __lowerCamelCase = MusicgenProcessor(feature_extractor=UpperCamelCase__ , tokenizer=UpperCamelCase__ ) # set the appropriate bos/pad token ids __lowerCamelCase = 2048 __lowerCamelCase = 2048 # set other default generation config params __lowerCamelCase = int(30 * audio_encoder.config.frame_rate ) __lowerCamelCase = True __lowerCamelCase = 3.0 if pytorch_dump_folder is not None: Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) logger.info(F"""Saving model {checkpoint} to {pytorch_dump_folder}""" ) model.save_pretrained(UpperCamelCase__ ) processor.save_pretrained(UpperCamelCase__ ) if repo_id: logger.info(F"""Pushing model {checkpoint} to {repo_id}""" ) model.push_to_hub(UpperCamelCase__ ) processor.push_to_hub(UpperCamelCase__ ) if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint", default="small", type=str, help="Checkpoint size of the MusicGen model you'd like to convert. Can be one of: `['small', 'medium', 'large']`.", ) parser.add_argument( "--pytorch_dump_folder", required=True, default=None, type=str, help="Path to the output PyTorch model directory.", ) parser.add_argument( "--push_to_hub", default=None, type=str, help="Where to upload the converted model on the 🤗 hub." ) parser.add_argument( "--device", default="cpu", type=str, help="Torch device to run the conversion, either cpu or cuda." ) __A = parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
348
0
import os def lowerCamelCase__ ( ) -> List[Any]: with open(os.path.dirname(_A ) + """/grid.txt""" ) as f: UpperCamelCase_ = [] # noqa: E741 for _ in range(20 ): l.append([int(_A ) for x in f.readline().split()] ) UpperCamelCase_ = 0 # right for i in range(20 ): for j in range(17 ): UpperCamelCase_ = l[i][j] * l[i][j + 1] * l[i][j + 2] * l[i][j + 3] if temp > maximum: UpperCamelCase_ = temp # down for i in range(17 ): for j in range(20 ): UpperCamelCase_ = l[i][j] * l[i + 1][j] * l[i + 2][j] * l[i + 3][j] if temp > maximum: UpperCamelCase_ = temp # diagonal 1 for i in range(17 ): for j in range(17 ): UpperCamelCase_ = l[i][j] * l[i + 1][j + 1] * l[i + 2][j + 2] * l[i + 3][j + 3] if temp > maximum: UpperCamelCase_ = temp # diagonal 2 for i in range(17 ): for j in range(3 , 20 ): UpperCamelCase_ = l[i][j] * l[i + 1][j - 1] * l[i + 2][j - 2] * l[i + 3][j - 3] if temp > maximum: UpperCamelCase_ = temp return maximum if __name__ == "__main__": print(solution())
122
from __future__ import annotations from random import choice def __UpperCamelCase ( _A : str ) ->int: """simple docstring""" return choice(_A ) def __UpperCamelCase ( _A : list[int] , _A : int ) ->int: """simple docstring""" lowerCamelCase_ =random_pivot(_A ) # partition based on pivot # linear time lowerCamelCase_ =[e for e in lst if e < pivot] lowerCamelCase_ =[e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(_A ) == k - 1: return pivot # pivot is in elements bigger than k elif len(_A ) < k - 1: return kth_number(_A , k - len(_A ) - 1 ) # pivot is in elements smaller than k else: return kth_number(_A , _A ) if __name__ == "__main__": import doctest doctest.testmod()
154
0
'''simple docstring''' from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
48
'''simple docstring''' from __future__ import annotations import math import random from typing import Any class _UpperCamelCase : '''simple docstring''' def __init__( self : List[Any]): '''simple docstring''' __lowercase =[] __lowercase =0 __lowercase =0 def __lowerCamelCase ( self : List[Any]): '''simple docstring''' return self.head == self.tail def __lowerCamelCase ( self : str , _lowerCAmelCase : Any): '''simple docstring''' self.data.append(_lowerCAmelCase) __lowercase =self.tail + 1 def __lowerCamelCase ( self : Dict): '''simple docstring''' __lowercase =self.data[self.head] __lowercase =self.head + 1 return ret def __lowerCamelCase ( self : Optional[Any]): '''simple docstring''' return self.tail - self.head def __lowerCamelCase ( self : Optional[int]): '''simple docstring''' print(self.data) print('**************') print(self.data[self.head : self.tail]) class _UpperCamelCase : '''simple docstring''' def __init__( self : Optional[int] , _lowerCAmelCase : Any): '''simple docstring''' __lowercase =data __lowercase =None __lowercase =None __lowercase =1 def __lowerCamelCase ( self : Any): '''simple docstring''' return self.data def __lowerCamelCase ( self : Optional[int]): '''simple docstring''' return self.left def __lowerCamelCase ( self : Tuple): '''simple docstring''' return self.right def __lowerCamelCase ( self : Dict): '''simple docstring''' return self.height def __lowerCamelCase ( self : int , _lowerCAmelCase : Any): '''simple docstring''' __lowercase =data def __lowerCamelCase ( self : Optional[Any] , _lowerCAmelCase : MyNode | None): '''simple docstring''' __lowercase =node def __lowerCamelCase ( self : List[Any] , _lowerCAmelCase : MyNode | None): '''simple docstring''' __lowercase =node def __lowerCamelCase ( self : Optional[int] , _lowerCAmelCase : int): '''simple docstring''' __lowercase =height def _A ( _lowerCAmelCase ): """simple docstring""" if node is None: return 0 return node.get_height() def _A ( _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" if a > b: return a return b def _A ( _lowerCAmelCase ): """simple docstring""" print('left rotation node:' , node.get_data() ) __lowercase =node.get_left() assert ret is not None node.set_left(ret.get_right() ) ret.set_right(_lowerCAmelCase ) __lowercase =my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(_lowerCAmelCase ) __lowercase =my_max(get_height(ret.get_right() ) , get_height(ret.get_left() ) ) + 1 ret.set_height(_lowerCAmelCase ) return ret def _A ( _lowerCAmelCase ): """simple docstring""" print('right rotation node:' , node.get_data() ) __lowercase =node.get_right() assert ret is not None node.set_right(ret.get_left() ) ret.set_left(_lowerCAmelCase ) __lowercase =my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(_lowerCAmelCase ) __lowercase =my_max(get_height(ret.get_right() ) , get_height(ret.get_left() ) ) + 1 ret.set_height(_lowerCAmelCase ) return ret def _A ( _lowerCAmelCase ): """simple docstring""" __lowercase =node.get_left() assert left_child is not None node.set_left(left_rotation(_lowerCAmelCase ) ) return right_rotation(_lowerCAmelCase ) def _A ( _lowerCAmelCase ): """simple docstring""" __lowercase =node.get_right() assert right_child is not None node.set_right(right_rotation(_lowerCAmelCase ) ) return left_rotation(_lowerCAmelCase ) def _A ( _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" if node is None: return MyNode(_lowerCAmelCase ) if data < node.get_data(): node.set_left(insert_node(node.get_left() , _lowerCAmelCase ) ) if ( get_height(node.get_left() ) - get_height(node.get_right() ) == 2 ): # an unbalance detected __lowercase =node.get_left() assert left_child is not None if ( data < left_child.get_data() ): # new node is the left child of the left child __lowercase =right_rotation(_lowerCAmelCase ) else: __lowercase =lr_rotation(_lowerCAmelCase ) else: node.set_right(insert_node(node.get_right() , _lowerCAmelCase ) ) if get_height(node.get_right() ) - get_height(node.get_left() ) == 2: __lowercase =node.get_right() assert right_child is not None if data < right_child.get_data(): __lowercase =rl_rotation(_lowerCAmelCase ) else: __lowercase =left_rotation(_lowerCAmelCase ) __lowercase =my_max(get_height(node.get_right() ) , get_height(node.get_left() ) ) + 1 node.set_height(_lowerCAmelCase ) return node def _A ( _lowerCAmelCase ): """simple docstring""" while True: __lowercase =root.get_right() if right_child is None: break __lowercase =right_child return root.get_data() def _A ( _lowerCAmelCase ): """simple docstring""" while True: __lowercase =root.get_left() if left_child is None: break __lowercase =left_child return root.get_data() def _A ( _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =root.get_left() __lowercase =root.get_right() if root.get_data() == data: if left_child is not None and right_child is not None: __lowercase =get_left_most(_lowerCAmelCase ) root.set_data(_lowerCAmelCase ) root.set_right(del_node(_lowerCAmelCase , _lowerCAmelCase ) ) elif left_child is not None: __lowercase =left_child elif right_child is not None: __lowercase =right_child else: return None elif root.get_data() > data: if left_child is None: print('No such data' ) return root else: root.set_left(del_node(_lowerCAmelCase , _lowerCAmelCase ) ) else: # root.get_data() < data if right_child is None: return root else: root.set_right(del_node(_lowerCAmelCase , _lowerCAmelCase ) ) if get_height(_lowerCAmelCase ) - get_height(_lowerCAmelCase ) == 2: assert right_child is not None if get_height(right_child.get_right() ) > get_height(right_child.get_left() ): __lowercase =left_rotation(_lowerCAmelCase ) else: __lowercase =rl_rotation(_lowerCAmelCase ) elif get_height(_lowerCAmelCase ) - get_height(_lowerCAmelCase ) == -2: assert left_child is not None if get_height(left_child.get_left() ) > get_height(left_child.get_right() ): __lowercase =right_rotation(_lowerCAmelCase ) else: __lowercase =lr_rotation(_lowerCAmelCase ) __lowercase =my_max(get_height(root.get_right() ) , get_height(root.get_left() ) ) + 1 root.set_height(_lowerCAmelCase ) return root class _UpperCamelCase : '''simple docstring''' def __init__( self : Tuple): '''simple docstring''' __lowercase =None def __lowerCamelCase ( self : Optional[int]): '''simple docstring''' return get_height(self.root) def __lowerCamelCase ( self : List[str] , _lowerCAmelCase : Any): '''simple docstring''' print('insert:' + str(_lowerCAmelCase)) __lowercase =insert_node(self.root , _lowerCAmelCase) def __lowerCamelCase ( self : List[Any] , _lowerCAmelCase : Any): '''simple docstring''' print('delete:' + str(_lowerCAmelCase)) if self.root is None: print('Tree is empty!') return __lowercase =del_node(self.root , _lowerCAmelCase) def __str__( self : int , ): # a level traversale, gives a more intuitive look on the tree '''simple docstring''' __lowercase ='' __lowercase =MyQueue() q.push(self.root) __lowercase =self.get_height() if layer == 0: return output __lowercase =0 while not q.is_empty(): __lowercase =q.pop() __lowercase =' ' * int(math.pow(2 , layer - 1)) output += space if node is None: output += "*" q.push(_lowerCAmelCase) q.push(_lowerCAmelCase) else: output += str(node.get_data()) q.push(node.get_left()) q.push(node.get_right()) output += space __lowercase =cnt + 1 for i in range(1_0_0): if cnt == math.pow(2 , _lowerCAmelCase) - 1: __lowercase =layer - 1 if layer == 0: output += "\n*************************************" return output output += "\n" break output += "\n*************************************" return output def _A ( ): """simple docstring""" import doctest doctest.testmod() if __name__ == "__main__": _test() lowerCamelCase = AVLtree() lowerCamelCase = list(range(10)) random.shuffle(lst) for i in lst: t.insert(i) print(str(t)) random.shuffle(lst) for i in lst: t.del_node(i) print(str(t))
48
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase_ = { "configuration_bigbird_pegasus": [ "BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP", "BigBirdPegasusConfig", "BigBirdPegasusOnnxConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = [ "BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST", "BigBirdPegasusForCausalLM", "BigBirdPegasusForConditionalGeneration", "BigBirdPegasusForQuestionAnswering", "BigBirdPegasusForSequenceClassification", "BigBirdPegasusModel", "BigBirdPegasusPreTrainedModel", ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys UpperCamelCase_ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
251
'''simple docstring''' from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. UpperCamelCase_ = 2_0_0 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. UpperCamelCase_ = 5_0 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. UpperCamelCase_ = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1_0_0_0)) def lowercase__( __UpperCamelCase: str ,__UpperCamelCase: str ): """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = len([g for position, g in enumerate(__UpperCamelCase ) if g == main_target[position]] ) return (item, float(__UpperCamelCase )) def lowercase__( __UpperCamelCase: str ,__UpperCamelCase: str ): """simple docstring""" SCREAMING_SNAKE_CASE : int = random.randint(0 ,len(__UpperCamelCase ) - 1 ) SCREAMING_SNAKE_CASE : List[str] = parent_a[:random_slice] + parent_a[random_slice:] SCREAMING_SNAKE_CASE : Tuple = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def lowercase__( __UpperCamelCase: str ,__UpperCamelCase: list[str] ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = list(__UpperCamelCase ) if random.uniform(0 ,1 ) < MUTATION_PROBABILITY: SCREAMING_SNAKE_CASE : Optional[Any] = random.choice(__UpperCamelCase ) return "".join(__UpperCamelCase ) def lowercase__( __UpperCamelCase: tuple[str, float] ,__UpperCamelCase: list[tuple[str, float]] ,__UpperCamelCase: list[str] ,): """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = [] # Generate more children proportionally to the fitness score. SCREAMING_SNAKE_CASE : Optional[Any] = int(parent_a[1] * 1_00 ) + 1 SCREAMING_SNAKE_CASE : Optional[Any] = 10 if child_n >= 10 else child_n for _ in range(__UpperCamelCase ): SCREAMING_SNAKE_CASE : List[str] = population_score[random.randint(0 ,__UpperCamelCase )][0] SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[int] = crossover(parent_a[0] ,__UpperCamelCase ) # Append new string to the population list. pop.append(mutate(__UpperCamelCase ,__UpperCamelCase ) ) pop.append(mutate(__UpperCamelCase ,__UpperCamelCase ) ) return pop def lowercase__( __UpperCamelCase: str ,__UpperCamelCase: list[str] ,__UpperCamelCase: bool = True ): """simple docstring""" if N_POPULATION < N_SELECTED: SCREAMING_SNAKE_CASE : List[str] = f"{N_POPULATION} must be bigger than {N_SELECTED}" raise ValueError(__UpperCamelCase ) # Verify that the target contains no genes besides the ones inside genes variable. SCREAMING_SNAKE_CASE : List[Any] = sorted({c for c in target if c not in genes} ) if not_in_genes_list: SCREAMING_SNAKE_CASE : List[Any] = f"{not_in_genes_list} is not in genes list, evolution cannot converge" raise ValueError(__UpperCamelCase ) # Generate random starting population. SCREAMING_SNAKE_CASE : Optional[Any] = [] for _ in range(__UpperCamelCase ): population.append(''.join([random.choice(__UpperCamelCase ) for i in range(len(__UpperCamelCase ) )] ) ) # Just some logs to know what the algorithms is doing. SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : str = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(__UpperCamelCase ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. SCREAMING_SNAKE_CASE : Optional[int] = [evaluate(__UpperCamelCase ,__UpperCamelCase ) for item in population] # Check if there is a matching evolution. SCREAMING_SNAKE_CASE : Union[str, Any] = sorted(__UpperCamelCase ,key=lambda __UpperCamelCase : x[1] ,reverse=__UpperCamelCase ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( f"\nGeneration: {generation}" f"\nTotal Population:{total_population}" f"\nBest score: {population_score[0][1]}" f"\nBest string: {population_score[0][0]}" ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. SCREAMING_SNAKE_CASE : int = population[: int(N_POPULATION / 3 )] population.clear() population.extend(__UpperCamelCase ) # Normalize population score to be between 0 and 1. SCREAMING_SNAKE_CASE : str = [ (item, score / len(__UpperCamelCase )) for item, score in population_score ] # This is selection for i in range(__UpperCamelCase ): population.extend(select(population_score[int(__UpperCamelCase )] ,__UpperCamelCase ,__UpperCamelCase ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(__UpperCamelCase ) > N_POPULATION: break if __name__ == "__main__": UpperCamelCase_ = ( "This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!" ) UpperCamelCase_ = list( " ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm" "nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\" ) UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = basic(target_str, genes_list) print( F"""\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}""" )
251
1
"""simple docstring""" import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): __lowerCAmelCase : Any = 1.5 __lowerCAmelCase : Optional[Any] = int(factor * num_class_images ) __lowerCAmelCase : List[str] = ClipClient( url='https://knn.laion.ai/knn-service' , indice_name='laion_400m' , num_images=_UpperCamelCase , aesthetic_weight=0.1 ) os.makedirs(F"{class_data_dir}/images" , exist_ok=_UpperCamelCase ) if len(list(Path(F"{class_data_dir}/images" ).iterdir() ) ) >= num_class_images: return while True: __lowerCAmelCase : List[str] = client.query(text=_UpperCamelCase ) if len(_UpperCamelCase ) >= factor * num_class_images or num_images > 1e4: break else: __lowerCAmelCase : Any = int(factor * num_images ) __lowerCAmelCase : Optional[Any] = ClipClient( url='https://knn.laion.ai/knn-service' , indice_name='laion_400m' , num_images=_UpperCamelCase , aesthetic_weight=0.1 , ) __lowerCAmelCase : Any = 0 __lowerCAmelCase : Tuple = 0 __lowerCAmelCase : Tuple = tqdm(desc='downloading real regularization images' , total=_UpperCamelCase ) with open(F"{class_data_dir}/caption.txt" , 'w' ) as fa, open(F"{class_data_dir}/urls.txt" , 'w' ) as fa, open( F"{class_data_dir}/images.txt" , 'w' ) as fa: while total < num_class_images: __lowerCAmelCase : Optional[Any] = class_images[count] count += 1 try: __lowerCAmelCase : Optional[Any] = requests.get(images['url'] ) if img.status_code == 200: __lowerCAmelCase : List[Any] = Image.open(BytesIO(img.content ) ) with open(F"{class_data_dir}/images/{total}.jpg" , 'wb' ) as f: f.write(img.content ) fa.write(images['caption'] + '\n' ) fa.write(images['url'] + '\n' ) fa.write(F"{class_data_dir}/images/{total}.jpg" + '\n' ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def __lowerCAmelCase (): __lowerCAmelCase : int = argparse.ArgumentParser('' , add_help=_UpperCamelCase ) parser.add_argument('--class_prompt' , help='text prompt to retrieve images' , required=_UpperCamelCase , type=_UpperCamelCase ) parser.add_argument('--class_data_dir' , help='path to save images' , required=_UpperCamelCase , type=_UpperCamelCase ) parser.add_argument('--num_class_images' , help='number of images to download' , default=200 , type=_UpperCamelCase ) return parser.parse_args() if __name__ == "__main__": lowerCamelCase__ = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
182
"""simple docstring""" # Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): __lowerCAmelCase : str = { 'en': 'Machine learning is great, isn\'t it?', 'ru': 'Машинное обучение - это здорово, не так ли?', 'de': 'Maschinelles Lernen ist großartig, oder?', } # BLUE scores as follows: # "pair": [fairseq, transformers] __lowerCAmelCase : Union[str, Any] = { 'ru-en': ['[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)', '39.20'], 'en-ru': ['[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)', '33.47'], 'en-de': ['[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)', '42.83'], 'de-en': ['[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)', '41.35'], } __lowerCAmelCase : List[str] = F"{src_lang}-{tgt_lang}" __lowerCAmelCase : Tuple = F"\n---\nlanguage: \n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt19\n- facebook\nlicense: apache-2.0\ndatasets:\n- wmt19\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}.\n\nFor more details, please see, [Facebook FAIR's WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616).\n\nThe abbreviation FSMT stands for FairSeqMachineTranslation\n\nAll four models are available:\n\n* [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru)\n* [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en)\n* [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de)\n* [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en)\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = \"facebook/wmt19-{src_lang}-{tgt_lang}\"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = \"{texts[src_lang]}\"\ninput_ids = tokenizer.encode(input, return_tensors=\"pt\")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n- The original (and this ported model) doesn't seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981)\n\n## Training data\n\nPretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616).\n\n## Eval results\n\npair | fairseq | transformers\n-------|---------|----------\n{pair} | {scores[pair][0]} | {scores[pair][1]}\n\nThe score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn't support:\n- model ensemble, therefore the best performing checkpoint was ported (``model4.pt``).\n- re-ranking\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=15\nmkdir -p $DATA_DIR\nsacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH=\"src:examples/seq2seq\" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\nnote: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`.\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt19/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)\n\n\n### BibTeX entry and citation info\n\n```bibtex\n@inproceedings{{...,\n year={{2020}},\n title={{Facebook FAIR's WMT19 News Translation Task Submission}},\n author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}},\n booktitle={{Proc. of WMT}},\n}}\n```\n\n\n## TODO\n\n- port model ensemble (fairseq uses 4 model checkpoints)\n\n" os.makedirs(_UpperCamelCase , exist_ok=_UpperCamelCase ) __lowerCAmelCase : Any = os.path.join(_UpperCamelCase , 'README.md' ) print(F"Generating {path}" ) with open(_UpperCamelCase , 'w' , encoding='utf-8' ) as f: f.write(_UpperCamelCase ) # make sure we are under the root of the project lowerCamelCase__ = Path(__file__).resolve().parent.parent.parent lowerCamelCase__ = repo_dir / """model_cards""" for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = model_name.split("""-""") lowerCamelCase__ = model_cards_dir / """facebook""" / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
182
1
"""simple docstring""" import requests a_ = """""" # <-- Put your OpenWeatherMap appid here! a_ = """https://api.openweathermap.org/data/2.5/""" def __lowercase ( snake_case_ : str = "Chicago" ,snake_case_ : str = APPID ) ->dict: '''simple docstring''' return requests.get(URL_BASE + '''weather''' ,params=locals() ).json() def __lowercase ( snake_case_ : str = "Kolkata, India" ,snake_case_ : str = APPID ) ->dict: '''simple docstring''' return requests.get(URL_BASE + '''forecast''' ,params=locals() ).json() def __lowercase ( snake_case_ : float = 55.68 ,snake_case_ : float = 12.57 ,snake_case_ : str = APPID ) ->dict: '''simple docstring''' return requests.get(URL_BASE + '''onecall''' ,params=locals() ).json() if __name__ == "__main__": from pprint import pprint while True: a_ = input("""Enter a location:""").strip() if location: pprint(current_weather(location)) else: break
179
"""simple docstring""" import requests a_ = """""" # <-- Put your OpenWeatherMap appid here! a_ = """https://api.openweathermap.org/data/2.5/""" def __lowercase ( snake_case_ : str = "Chicago" ,snake_case_ : str = APPID ) ->dict: '''simple docstring''' return requests.get(URL_BASE + '''weather''' ,params=locals() ).json() def __lowercase ( snake_case_ : str = "Kolkata, India" ,snake_case_ : str = APPID ) ->dict: '''simple docstring''' return requests.get(URL_BASE + '''forecast''' ,params=locals() ).json() def __lowercase ( snake_case_ : float = 55.68 ,snake_case_ : float = 12.57 ,snake_case_ : str = APPID ) ->dict: '''simple docstring''' return requests.get(URL_BASE + '''onecall''' ,params=locals() ).json() if __name__ == "__main__": from pprint import pprint while True: a_ = input("""Enter a location:""").strip() if location: pprint(current_weather(location)) else: break
179
1
'''simple docstring''' import argparse import re from flax.traverse_util import flatten_dict, unflatten_dict from tax import checkpoints from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model from transformers.utils import logging logging.set_verbosity_info() # should not include what is already done by the `from_pt` argument lowerCAmelCase__ : Optional[Any] = { "/attention/": "/0/SelfAttention/", "/self_attention/": "/0/SelfAttention/", "/encoder_decoder_attention/": "/1/EncDecAttention/", "value": "v", "query": "q", "key": "k", "out": "o", "pre_self_attention_layer_norm": "0/layer_norm", "pre_cross_attention_layer_norm": "1/layer_norm", "pre_attention_layer_norm": "0/layer_norm", # previously 1, but seems wrong "token_embedder": "shared", "encoder_norm": "final_layer_norm", "decoder_norm": "final_layer_norm", "relpos_bias/rel_embedding": "block/0/layer/0/SelfAttention/relative_attention_bias/weight", "router/router_weights/w/": "router/classifier/", "roer/roer_weights/w/": "router/classifier/", "logits_dense": "lm_head", } def __UpperCamelCase ( _UpperCAmelCase ): # 1. in HF T5, we have block.{x}.layer.{y}. which corresponds to layer.{x} in # the original model __UpperCAmelCase : List[str] = list(s_dict.keys() ) for key in keys: __UpperCAmelCase : int = R".*/layers_(\d+)" __UpperCAmelCase : List[str] = key if re.match(_UpperCAmelCase, _UpperCAmelCase ): __UpperCAmelCase : Optional[int] = re.sub(R"layers_(\d+)", R"block/\1/layer", _UpperCAmelCase ) __UpperCAmelCase : Any = R"(encoder|decoder)\/" if re.match(_UpperCAmelCase, _UpperCAmelCase ): __UpperCAmelCase : List[Any] = re.match(_UpperCAmelCase, _UpperCAmelCase ).groups() if groups[0] == "encoder": __UpperCAmelCase : Optional[Any] = re.sub(R"/mlp/", R"/1/mlp/", _UpperCAmelCase ) __UpperCAmelCase : List[Any] = re.sub(R"/pre_mlp_layer_norm/", R"/1/layer_norm/", _UpperCAmelCase ) elif groups[0] == "decoder": __UpperCAmelCase : List[Any] = re.sub(R"/mlp/", R"/2/mlp/", _UpperCAmelCase ) __UpperCAmelCase : Any = re.sub(R"/pre_mlp_layer_norm/", R"/2/layer_norm/", _UpperCAmelCase ) # 2. Convert other classic mappings for old_key, temp_key in MOE_LAYER_NAME_MAPPING.items(): if old_key in new_key: __UpperCAmelCase : List[str] = new_key.replace(_UpperCAmelCase, _UpperCAmelCase ) print(F"{key} -> {new_key}" ) __UpperCAmelCase : Any = s_dict.pop(_UpperCAmelCase ) if "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: __UpperCAmelCase : Tuple = s_dict[ "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" ].T if "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: __UpperCAmelCase : Optional[Any] = s_dict[ "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" ].T # 3. Take extra care of the EXPERTS layer for key in list(s_dict.keys() ): if "expert" in key: __UpperCAmelCase : Any = s_dict[key].shape[0] __UpperCAmelCase : str = s_dict[key] for idx in range(_UpperCAmelCase ): __UpperCAmelCase : Optional[Any] = expert_weihts[idx] print(F"{key} -> {key.replace('expert/', 'nested fstring' )}" ) s_dict.pop(_UpperCAmelCase ) return s_dict lowerCAmelCase__ : Optional[Any] = { "NUM_ENCODER_LAYERS": "num_layers", "NUM_DECODER_LAYERS": "num_decoder_layers", "NUM_HEADS": "num_heads", "HEAD_DIM": "d_kv", "EMBED_DIM": "d_model", "MLP_DIM": "d_ff", "NUM_SELECTED_EXPERTS": "num_selected_experts", "NUM_ENCODER_SPARSE_LAYERS": "num_sparse_encoder_layers", "NUM_DECODER_SPARSE_LAYERS": "num_sparse_decoder_layers", "dense.MlpBlock.activations": "feed_forward_proj", } def __UpperCamelCase ( _UpperCAmelCase, _UpperCAmelCase ): # Convert a google style config to the hugging face fromat import regex as re with open(_UpperCAmelCase, "r" ) as f: __UpperCAmelCase : List[Any] = f.read() __UpperCAmelCase : Union[str, Any] = re.findall(R"(.*) = ([0-9.]*)", _UpperCAmelCase ) __UpperCAmelCase : Dict = {} for param, value in regex_match: if param in GIN_TO_CONFIG_MAPPING and value != "": __UpperCAmelCase : Tuple = float(_UpperCAmelCase ) if "." in value else int(_UpperCAmelCase ) __UpperCAmelCase : str = re.findall(R"(.*activations) = \(\'(.*)\',\)", _UpperCAmelCase )[0] __UpperCAmelCase : int = str(activation[1] ) __UpperCAmelCase : int = num_experts __UpperCAmelCase : List[str] = SwitchTransformersConfig(**_UpperCAmelCase ) return config def __UpperCamelCase ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase=None, _UpperCAmelCase="./", _UpperCAmelCase=8 ): # Initialise PyTorch model print(F"Loading flax weights from : {flax_checkpoint_path}" ) __UpperCAmelCase : Dict = checkpoints.load_tax_checkpoint(_UpperCAmelCase ) if gin_file is not None: __UpperCAmelCase : int = convert_gin_to_config(_UpperCAmelCase, _UpperCAmelCase ) else: __UpperCAmelCase : int = SwitchTransformersConfig.from_pretrained(_UpperCAmelCase ) __UpperCAmelCase : Any = SwitchTransformersForConditionalGeneration(_UpperCAmelCase ) __UpperCAmelCase : str = flax_params["target"] __UpperCAmelCase : Any = flatten_dict(_UpperCAmelCase, sep="/" ) __UpperCAmelCase : Optional[Any] = rename_keys(_UpperCAmelCase ) __UpperCAmelCase : Any = unflatten_dict(_UpperCAmelCase, sep="/" ) # Load the flax params in the PT model load_flax_weights_in_pytorch_model(_UpperCAmelCase, _UpperCAmelCase ) print(F"Save PyTorch model to {pytorch_dump_path}" ) pt_model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": lowerCAmelCase__ : int = argparse.ArgumentParser() # Required parameters parser.add_argument( "--switch_t5x_checkpoint_path", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained SwitchTransformers model. \nThis specifies the" " model architecture. If not provided, a `gin_file` has to be provided." ), ) parser.add_argument( "--gin_file", default=None, type=str, required=False, help="Path to the gin config file. If not provided, a `config_file` has to be passed ", ) parser.add_argument( "--config_name", default=None, type=str, required=False, help="Config name of SwitchTransformers model." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output pytorch model." ) parser.add_argument("--num_experts", default=8, type=int, required=False, help="Number of experts") lowerCAmelCase__ : int = parser.parse_args() convert_flax_checkpoint_to_pytorch( args.switch_tax_checkpoint_path, args.config_name, args.gin_file, args.pytorch_dump_folder_path, args.num_experts, )
37
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class SCREAMING_SNAKE_CASE__ ( snake_case__ ): """simple docstring""" SCREAMING_SNAKE_CASE = 42 class SCREAMING_SNAKE_CASE__ ( snake_case__ ,snake_case__ ): """simple docstring""" @register_to_config def __init__( self : Union[str, Any] , UpperCAmelCase_ : int = 65_536 , UpperCAmelCase_ : Optional[int] = None , UpperCAmelCase_ : int = 2 , UpperCAmelCase_ : int = 2 , UpperCAmelCase_ : int = 0 , UpperCAmelCase_ : str = "fourier" , UpperCAmelCase_ : bool = True , UpperCAmelCase_ : bool = False , UpperCAmelCase_ : float = 0.0 , UpperCAmelCase_ : Tuple[str] = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , UpperCAmelCase_ : Tuple[str] = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , UpperCAmelCase_ : Tuple[str] = "UNetMidBlock1D" , UpperCAmelCase_ : str = None , UpperCAmelCase_ : Tuple[int] = (32, 32, 64) , UpperCAmelCase_ : str = None , UpperCAmelCase_ : int = 8 , UpperCAmelCase_ : int = 1 , UpperCAmelCase_ : bool = False , ): """simple docstring""" super().__init__() __UpperCAmelCase : str = sample_size # time if time_embedding_type == "fourier": __UpperCAmelCase : int = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=UpperCAmelCase_ , log=UpperCAmelCase_ , flip_sin_to_cos=UpperCAmelCase_ ) __UpperCAmelCase : str = 2 * block_out_channels[0] elif time_embedding_type == "positional": __UpperCAmelCase : str = Timesteps( block_out_channels[0] , flip_sin_to_cos=UpperCAmelCase_ , downscale_freq_shift=UpperCAmelCase_ ) __UpperCAmelCase : Dict = block_out_channels[0] if use_timestep_embedding: __UpperCAmelCase : Union[str, Any] = block_out_channels[0] * 4 __UpperCAmelCase : str = TimestepEmbedding( in_channels=UpperCAmelCase_ , time_embed_dim=UpperCAmelCase_ , act_fn=UpperCAmelCase_ , out_dim=block_out_channels[0] , ) __UpperCAmelCase : Tuple = nn.ModuleList([] ) __UpperCAmelCase : int = None __UpperCAmelCase : Optional[Any] = nn.ModuleList([] ) __UpperCAmelCase : Dict = None # down __UpperCAmelCase : str = in_channels for i, down_block_type in enumerate(UpperCAmelCase_ ): __UpperCAmelCase : Optional[Any] = output_channel __UpperCAmelCase : Optional[int] = block_out_channels[i] if i == 0: input_channel += extra_in_channels __UpperCAmelCase : Tuple = i == len(UpperCAmelCase_ ) - 1 __UpperCAmelCase : List[str] = get_down_block( UpperCAmelCase_ , num_layers=UpperCAmelCase_ , in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(UpperCAmelCase_ ) # mid __UpperCAmelCase : Optional[Any] = get_mid_block( UpperCAmelCase_ , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=UpperCAmelCase_ , add_downsample=UpperCAmelCase_ , ) # up __UpperCAmelCase : Tuple = list(reversed(UpperCAmelCase_ ) ) __UpperCAmelCase : Any = reversed_block_out_channels[0] if out_block_type is None: __UpperCAmelCase : Union[str, Any] = out_channels else: __UpperCAmelCase : Dict = block_out_channels[0] for i, up_block_type in enumerate(UpperCAmelCase_ ): __UpperCAmelCase : int = output_channel __UpperCAmelCase : str = ( reversed_block_out_channels[i + 1] if i < len(UpperCAmelCase_ ) - 1 else final_upsample_channels ) __UpperCAmelCase : Tuple = i == len(UpperCAmelCase_ ) - 1 __UpperCAmelCase : Dict = get_up_block( UpperCAmelCase_ , num_layers=UpperCAmelCase_ , in_channels=UpperCAmelCase_ , out_channels=UpperCAmelCase_ , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(UpperCAmelCase_ ) __UpperCAmelCase : Union[str, Any] = output_channel # out __UpperCAmelCase : Optional[int] = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) __UpperCAmelCase : List[Any] = get_out_block( out_block_type=UpperCAmelCase_ , num_groups_out=UpperCAmelCase_ , embed_dim=block_out_channels[0] , out_channels=UpperCAmelCase_ , act_fn=UpperCAmelCase_ , fc_dim=block_out_channels[-1] // 4 , ) def lowerCamelCase_ ( self : Optional[int] , UpperCAmelCase_ : torch.FloatTensor , UpperCAmelCase_ : Union[torch.Tensor, float, int] , UpperCAmelCase_ : bool = True , ): """simple docstring""" __UpperCAmelCase : Dict = timestep if not torch.is_tensor(UpperCAmelCase_ ): __UpperCAmelCase : List[str] = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(UpperCAmelCase_ ) and len(timesteps.shape ) == 0: __UpperCAmelCase : List[str] = timesteps[None].to(sample.device ) __UpperCAmelCase : List[str] = self.time_proj(UpperCAmelCase_ ) if self.config.use_timestep_embedding: __UpperCAmelCase : Any = self.time_mlp(UpperCAmelCase_ ) else: __UpperCAmelCase : Any = timestep_embed[..., None] __UpperCAmelCase : int = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) __UpperCAmelCase : Dict = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down __UpperCAmelCase : int = () for downsample_block in self.down_blocks: __UpperCAmelCase , __UpperCAmelCase : int = downsample_block(hidden_states=UpperCAmelCase_ , temb=UpperCAmelCase_ ) down_block_res_samples += res_samples # 3. mid if self.mid_block: __UpperCAmelCase : List[str] = self.mid_block(UpperCAmelCase_ , UpperCAmelCase_ ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): __UpperCAmelCase : Any = down_block_res_samples[-1:] __UpperCAmelCase : List[Any] = down_block_res_samples[:-1] __UpperCAmelCase : str = upsample_block(UpperCAmelCase_ , res_hidden_states_tuple=UpperCAmelCase_ , temb=UpperCAmelCase_ ) # 5. post-process if self.out_block: __UpperCAmelCase : Tuple = self.out_block(UpperCAmelCase_ , UpperCAmelCase_ ) if not return_dict: return (sample,) return UNetaDOutput(sample=UpperCAmelCase_ )
37
1
"""simple docstring""" from __future__ import annotations def snake_case__ ( __lowerCamelCase : list[int] , __lowerCamelCase : int ): """simple docstring""" lowerCamelCase__ : list[list[int]] =[] lowerCamelCase__ : list[int] =[] lowerCamelCase__ : Any =0 lowerCamelCase__ : Dict =sum(__lowerCamelCase ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) return result def snake_case__ ( __lowerCamelCase : list[int] , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : list[int] , __lowerCamelCase : list[list[int]] , __lowerCamelCase : int , ): """simple docstring""" if sum(__lowerCamelCase ) > max_sum or (remaining_nums_sum + sum(__lowerCamelCase )) < max_sum: return if sum(__lowerCamelCase ) == max_sum: result.append(__lowerCamelCase ) return for index in range(__lowerCamelCase , len(__lowerCamelCase ) ): create_state_space_tree( __lowerCamelCase , __lowerCamelCase , index + 1 , [*path, nums[index]] , __lowerCamelCase , remaining_nums_sum - nums[index] , ) _lowercase : Tuple = [3, 3_4, 4, 1_2, 5, 2] _lowercase : Optional[Any] = 9 _lowercase : Optional[Any] = generate_sum_of_subsets_soln(nums, max_sum) print(*result)
238
"""simple docstring""" import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def snake_case__ ( __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Tuple , __lowerCamelCase : Any , __lowerCamelCase : Tuple , __lowerCamelCase : Tuple=True , __lowerCamelCase : Dict="pt" ): """simple docstring""" lowerCamelCase__ : str ={'''add_prefix_space''': True} if isinstance(__lowerCamelCase , __lowerCamelCase ) and not line.startswith(''' ''' ) else {} lowerCamelCase__ : int =padding_side return tokenizer( [line] , max_length=__lowerCamelCase , padding='''max_length''' if pad_to_max_length else None , truncation=__lowerCamelCase , return_tensors=__lowerCamelCase , add_special_tokens=__lowerCamelCase , **__lowerCamelCase , ) def snake_case__ ( __lowerCamelCase : str , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int]=None , ): """simple docstring""" lowerCamelCase__ : Any =input_ids.ne(__lowerCamelCase ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' def __init__( self : str, lowerCamelCase : Union[str, Any], lowerCamelCase : Dict, lowerCamelCase : Union[str, Any], lowerCamelCase : Dict, lowerCamelCase : str="train", lowerCamelCase : List[Any]=None, lowerCamelCase : Tuple=None, lowerCamelCase : List[str]=None, lowerCamelCase : int="", )-> List[Any]: super().__init__() lowerCamelCase__ : Tuple =Path(lowerCamelCase ).joinpath(type_path + '''.source''' ) lowerCamelCase__ : str =Path(lowerCamelCase ).joinpath(type_path + '''.target''' ) lowerCamelCase__ : Dict =self.get_char_lens(self.src_file ) lowerCamelCase__ : Tuple =max_source_length lowerCamelCase__ : Optional[int] =max_target_length assert min(self.src_lens ) > 0, F'''found empty line in {self.src_file}''' lowerCamelCase__ : Dict =tokenizer lowerCamelCase__ : List[str] =prefix if n_obs is not None: lowerCamelCase__ : int =self.src_lens[:n_obs] lowerCamelCase__ : Dict =src_lang lowerCamelCase__ : Tuple =tgt_lang def __len__( self : Dict )-> Optional[int]: return len(self.src_lens ) def __getitem__( self : List[str], lowerCamelCase : Optional[int] )-> Dict[str, torch.Tensor]: lowerCamelCase__ : List[Any] =index + 1 # linecache starts at 1 lowerCamelCase__ : Optional[int] =self.prefix + linecache.getline(str(self.src_file ), lowerCamelCase ).rstrip('''\n''' ) lowerCamelCase__ : Optional[Any] =linecache.getline(str(self.tgt_file ), lowerCamelCase ).rstrip('''\n''' ) assert source_line, F'''empty source line for index {index}''' assert tgt_line, F'''empty tgt line for index {index}''' # Need to add eos token manually for T5 if isinstance(self.tokenizer, lowerCamelCase ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right lowerCamelCase__ : Optional[int] =( self.tokenizer.question_encoder if isinstance(self.tokenizer, lowerCamelCase ) else self.tokenizer ) lowerCamelCase__ : Tuple =self.tokenizer.generator if isinstance(self.tokenizer, lowerCamelCase ) else self.tokenizer lowerCamelCase__ : Optional[int] =encode_line(lowerCamelCase, lowerCamelCase, self.max_source_length, '''right''' ) lowerCamelCase__ : str =encode_line(lowerCamelCase, lowerCamelCase, self.max_target_length, '''right''' ) lowerCamelCase__ : str =source_inputs['''input_ids'''].squeeze() lowerCamelCase__ : str =target_inputs['''input_ids'''].squeeze() lowerCamelCase__ : Union[str, Any] =source_inputs['''attention_mask'''].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def snake_case ( lowerCamelCase : Union[str, Any] )-> Optional[int]: return [len(lowerCamelCase ) for x in Path(lowerCamelCase ).open().readlines()] def snake_case ( self : str, lowerCamelCase : str )-> Dict[str, torch.Tensor]: lowerCamelCase__ : List[Any] =torch.stack([x['''input_ids'''] for x in batch] ) lowerCamelCase__ : int =torch.stack([x['''attention_mask'''] for x in batch] ) lowerCamelCase__ : Union[str, Any] =torch.stack([x['''decoder_input_ids'''] for x in batch] ) lowerCamelCase__ : str =( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer, lowerCamelCase ) else self.tokenizer.pad_token_id ) lowerCamelCase__ : List[str] =( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer, lowerCamelCase ) else self.tokenizer.pad_token_id ) lowerCamelCase__ : Optional[int] =trim_batch(lowerCamelCase, lowerCamelCase ) lowerCamelCase__ , lowerCamelCase__ : Any =trim_batch(lowerCamelCase, lowerCamelCase, attention_mask=lowerCamelCase ) lowerCamelCase__ : List[str] ={ '''input_ids''': source_ids, '''attention_mask''': source_mask, '''decoder_input_ids''': y, } return batch _lowercase : Any = getLogger(__name__) def snake_case__ ( __lowerCamelCase : List[List] ): """simple docstring""" return list(itertools.chain.from_iterable(__lowerCamelCase ) ) def snake_case__ ( __lowerCamelCase : str ): """simple docstring""" lowerCamelCase__ : Dict =get_git_info() save_json(__lowerCamelCase , os.path.join(__lowerCamelCase , '''git_log.json''' ) ) def snake_case__ ( __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Dict=4 , **__lowerCamelCase : int ): """simple docstring""" with open(__lowerCamelCase , '''w''' ) as f: json.dump(__lowerCamelCase , __lowerCamelCase , indent=__lowerCamelCase , **__lowerCamelCase ) def snake_case__ ( __lowerCamelCase : List[Any] ): """simple docstring""" with open(__lowerCamelCase ) as f: return json.load(__lowerCamelCase ) def snake_case__ ( ): """simple docstring""" lowerCamelCase__ : List[Any] =git.Repo(search_parent_directories=__lowerCamelCase ) lowerCamelCase__ : Union[str, Any] ={ '''repo_id''': str(__lowerCamelCase ), '''repo_sha''': str(repo.head.object.hexsha ), '''repo_branch''': str(repo.active_branch ), '''hostname''': str(socket.gethostname() ), } return repo_infos def snake_case__ ( __lowerCamelCase : Callable , __lowerCamelCase : Iterable ): """simple docstring""" return list(map(__lowerCamelCase , __lowerCamelCase ) ) def snake_case__ ( __lowerCamelCase : Any , __lowerCamelCase : List[Any] ): """simple docstring""" with open(__lowerCamelCase , '''wb''' ) as f: return pickle.dump(__lowerCamelCase , __lowerCamelCase ) def snake_case__ ( __lowerCamelCase : str ): """simple docstring""" def remove_articles(__lowerCamelCase : List[Any] ): return re.sub(R'''\b(a|an|the)\b''' , ''' ''' , __lowerCamelCase ) def white_space_fix(__lowerCamelCase : Any ): return " ".join(text.split() ) def remove_punc(__lowerCamelCase : Optional[Any] ): lowerCamelCase__ : Tuple =set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(__lowerCamelCase : Any ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(__lowerCamelCase ) ) ) ) def snake_case__ ( __lowerCamelCase : List[str] , __lowerCamelCase : List[str] ): """simple docstring""" lowerCamelCase__ : List[str] =normalize_answer(__lowerCamelCase ).split() lowerCamelCase__ : List[str] =normalize_answer(__lowerCamelCase ).split() lowerCamelCase__ : Optional[int] =Counter(__lowerCamelCase ) & Counter(__lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =sum(common.values() ) if num_same == 0: return 0 lowerCamelCase__ : Dict =1.0 * num_same / len(__lowerCamelCase ) lowerCamelCase__ : List[str] =1.0 * num_same / len(__lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =(2 * precision * recall) / (precision + recall) return fa def snake_case__ ( __lowerCamelCase : Dict , __lowerCamelCase : int ): """simple docstring""" return normalize_answer(__lowerCamelCase ) == normalize_answer(__lowerCamelCase ) def snake_case__ ( __lowerCamelCase : List[str] , __lowerCamelCase : List[str] ): """simple docstring""" assert len(__lowerCamelCase ) == len(__lowerCamelCase ) lowerCamelCase__ : Any =0 for hypo, pred in zip(__lowerCamelCase , __lowerCamelCase ): em += exact_match_score(__lowerCamelCase , __lowerCamelCase ) if len(__lowerCamelCase ) > 0: em /= len(__lowerCamelCase ) return {"em": em} def snake_case__ ( __lowerCamelCase : List[str] ): """simple docstring""" return model_prefix.startswith('''rag''' ) def snake_case__ ( __lowerCamelCase : Any , __lowerCamelCase : List[str] , __lowerCamelCase : str ): """simple docstring""" lowerCamelCase__ : Any ={p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead lowerCamelCase__ : Optional[int] ='''dropout_rate''' for p in extra_params: if getattr(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if not hasattr(__lowerCamelCase , __lowerCamelCase ) and not hasattr(__lowerCamelCase , equivalent_param[p] ): logger.info('''config doesn\'t have a `{}` attribute'''.format(__lowerCamelCase ) ) delattr(__lowerCamelCase , __lowerCamelCase ) continue lowerCamelCase__ : List[Any] =p if hasattr(__lowerCamelCase , __lowerCamelCase ) else equivalent_param[p] setattr(__lowerCamelCase , __lowerCamelCase , getattr(__lowerCamelCase , __lowerCamelCase ) ) delattr(__lowerCamelCase , __lowerCamelCase ) return hparams, config
238
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { "google/realm-cc-news-pretrained-embedder": ( "https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json" ), "google/realm-cc-news-pretrained-encoder": ( "https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json" ), "google/realm-cc-news-pretrained-scorer": ( "https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json" ), "google/realm-cc-news-pretrained-openqa": ( "https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json" ), "google/realm-orqa-nq-openqa": "https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json", "google/realm-orqa-nq-reader": "https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json", "google/realm-orqa-wq-openqa": "https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json", "google/realm-orqa-wq-reader": "https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json", # See all REALM models at https://huggingface.co/models?filter=realm } class UpperCAmelCase_ ( _a ): """simple docstring""" lowercase = "realm" def __init__( self : str , snake_case_ : List[Any]=30_522 , snake_case_ : List[Any]=768 , snake_case_ : List[Any]=128 , snake_case_ : List[str]=12 , snake_case_ : Tuple=12 , snake_case_ : Any=8 , snake_case_ : str=3_072 , snake_case_ : Optional[Any]="gelu_new" , snake_case_ : Any=0.1 , snake_case_ : List[str]=0.1 , snake_case_ : Optional[int]=512 , snake_case_ : Dict=2 , snake_case_ : Tuple=0.02 , snake_case_ : Union[str, Any]=1E-1_2 , snake_case_ : int=256 , snake_case_ : List[Any]=10 , snake_case_ : int=1E-3 , snake_case_ : Tuple=5 , snake_case_ : Union[str, Any]=320 , snake_case_ : int=13_353_718 , snake_case_ : Optional[int]=5_000 , snake_case_ : Optional[int]=1 , snake_case_ : Optional[int]=0 , snake_case_ : List[str]=2 , **snake_case_ : List[str] , ): super().__init__(pad_token_id=snake_case_ , bos_token_id=snake_case_ , eos_token_id=snake_case_ , **snake_case_ ) # Common config snake_case__ : Optional[int] = vocab_size snake_case__ : Optional[Any] = max_position_embeddings snake_case__ : Any = hidden_size snake_case__ : str = retriever_proj_size snake_case__ : Optional[Any] = num_hidden_layers snake_case__ : Optional[int] = num_attention_heads snake_case__ : Optional[int] = num_candidates snake_case__ : str = intermediate_size snake_case__ : Tuple = hidden_act snake_case__ : Optional[int] = hidden_dropout_prob snake_case__ : Optional[Any] = attention_probs_dropout_prob snake_case__ : List[Any] = initializer_range snake_case__ : List[str] = type_vocab_size snake_case__ : str = layer_norm_eps # Reader config snake_case__ : int = span_hidden_size snake_case__ : int = max_span_width snake_case__ : Union[str, Any] = reader_layer_norm_eps snake_case__ : int = reader_beam_size snake_case__ : Optional[Any] = reader_seq_len # Retrieval config snake_case__ : Optional[int] = num_block_records snake_case__ : Any = searcher_beam_size
43
'''simple docstring''' import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class UpperCAmelCase_ ( _a , unittest.TestCase ): """simple docstring""" lowercase = "hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline" def lowerCamelCase ( self : Union[str, Any] , snake_case_ : Tuple=0 ): snake_case__ : Any = floats_tensor((1, 3, 128, 128) , rng=random.Random(snake_case_ ) ) snake_case__ : List[str] = np.random.RandomState(snake_case_ ) snake_case__ : Optional[int] = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """generator""": generator, """num_inference_steps""": 3, """strength""": 0.75, """guidance_scale""": 7.5, """output_type""": """numpy""", } return inputs def lowerCamelCase ( self : Optional[Any] ): snake_case__ : str = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=snake_case_ ) snake_case__ : Tuple = self.get_dummy_inputs() snake_case__ : Union[str, Any] = pipe(**snake_case_ ).images snake_case__ : List[Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 128, 128, 3) snake_case__ : int = np.array([0.69643, 0.58484, 0.50314, 0.58760, 0.55368, 0.59643, 0.51529, 0.41217, 0.49087] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def lowerCamelCase ( self : Dict ): snake_case__ : str = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) snake_case__ : int = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) snake_case__ : Dict = self.get_dummy_inputs() snake_case__ : int = pipe(**snake_case_ ).images snake_case__ : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) snake_case__ : Tuple = np.array([0.61737, 0.54642, 0.53183, 0.54465, 0.52742, 0.60525, 0.49969, 0.40655, 0.48154] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def lowerCamelCase ( self : Optional[int] ): snake_case__ : int = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) snake_case__ : Dict = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case_ ) # warmup pass to apply optimizations snake_case__ : List[Any] = pipe(**self.get_dummy_inputs() ) snake_case__ : List[str] = self.get_dummy_inputs() snake_case__ : Optional[int] = pipe(**snake_case_ ).images snake_case__ : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) snake_case__ : Any = np.array([0.52761, 0.59977, 0.49033, 0.49619, 0.54282, 0.50311, 0.47600, 0.40918, 0.45203] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def lowerCamelCase ( self : str ): snake_case__ : List[str] = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) snake_case__ : Dict = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case_ ) snake_case__ : Union[str, Any] = self.get_dummy_inputs() snake_case__ : List[Any] = pipe(**snake_case_ ).images snake_case__ : str = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) snake_case__ : Optional[Any] = np.array([0.52911, 0.60004, 0.49229, 0.49805, 0.54502, 0.50680, 0.47777, 0.41028, 0.45304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def lowerCamelCase ( self : str ): snake_case__ : List[str] = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) snake_case__ : List[Any] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case_ ) snake_case__ : Tuple = self.get_dummy_inputs() snake_case__ : Tuple = pipe(**snake_case_ ).images snake_case__ : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) snake_case__ : int = np.array([0.52911, 0.60004, 0.49229, 0.49805, 0.54502, 0.50680, 0.47777, 0.41028, 0.45304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def lowerCamelCase ( self : Dict ): snake_case__ : Optional[Any] = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) snake_case__ : Union[str, Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case_ ) snake_case__ : List[str] = self.get_dummy_inputs() snake_case__ : List[str] = pipe(**snake_case_ ).images snake_case__ : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) snake_case__ : List[str] = np.array([0.65331, 0.58277, 0.48204, 0.56059, 0.53665, 0.56235, 0.50969, 0.40009, 0.46552] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class UpperCAmelCase_ ( unittest.TestCase ): """simple docstring""" @property def lowerCamelCase ( self : Dict ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def lowerCamelCase ( self : Dict ): snake_case__ : Tuple = ort.SessionOptions() snake_case__ : Optional[Any] = False return options def lowerCamelCase ( self : List[str] ): snake_case__ : Any = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) snake_case__ : str = init_image.resize((768, 512) ) # using the PNDM scheduler by default snake_case__ : Optional[Any] = OnnxStableDiffusionImgaImgPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""onnx""" , safety_checker=snake_case_ , feature_extractor=snake_case_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=snake_case_ ) snake_case__ : Dict = """A fantasy landscape, trending on artstation""" snake_case__ : str = np.random.RandomState(0 ) snake_case__ : Union[str, Any] = pipe( prompt=snake_case_ , image=snake_case_ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=10 , generator=snake_case_ , output_type="""np""" , ) snake_case__ : str = output.images snake_case__ : Optional[Any] = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) snake_case__ : Optional[Any] = np.array([0.4909, 0.5059, 0.5372, 0.4623, 0.4876, 0.5049, 0.4820, 0.4956, 0.5019] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def lowerCamelCase ( self : int ): snake_case__ : List[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) snake_case__ : List[Any] = init_image.resize((768, 512) ) snake_case__ : Tuple = LMSDiscreteScheduler.from_pretrained( """runwayml/stable-diffusion-v1-5""" , subfolder="""scheduler""" , revision="""onnx""" ) snake_case__ : str = OnnxStableDiffusionImgaImgPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , scheduler=snake_case_ , safety_checker=snake_case_ , feature_extractor=snake_case_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=snake_case_ ) snake_case__ : Union[str, Any] = """A fantasy landscape, trending on artstation""" snake_case__ : Optional[int] = np.random.RandomState(0 ) snake_case__ : Optional[int] = pipe( prompt=snake_case_ , image=snake_case_ , strength=0.75 , guidance_scale=7.5 , num_inference_steps=20 , generator=snake_case_ , output_type="""np""" , ) snake_case__ : Any = output.images snake_case__ : Tuple = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) snake_case__ : Tuple = np.array([0.8043, 0.926, 0.9581, 0.8119, 0.8954, 0.913, 0.7209, 0.7463, 0.7431] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
43
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCAmelCase_ : Optional[int] = { '''configuration_mvp''': ['''MVP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MvpConfig''', '''MvpOnnxConfig'''], '''tokenization_mvp''': ['''MvpTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Optional[int] = ['''MvpTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Tuple = [ '''MVP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MvpForCausalLM''', '''MvpForConditionalGeneration''', '''MvpForQuestionAnswering''', '''MvpForSequenceClassification''', '''MvpModel''', '''MvpPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys UpperCAmelCase_ : Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
38
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=_a ) class _SCREAMING_SNAKE_CASE ( _a ): snake_case__ : str = field(default="""automatic-speech-recognition""" , metadata={"""include_in_asdict_even_if_is_default""": True} ) snake_case__ : ClassVar[Features] = Features({"""audio""": Audio()} ) snake_case__ : ClassVar[Features] = Features({"""transcription""": Value("""string""" )} ) snake_case__ : str = "audio" snake_case__ : str = "transcription" def _A ( self : List[str] , __lowerCamelCase : Dict ): if self.audio_column not in features: raise ValueError(F"""Column {self.audio_column} is not present in features.""" ) if not isinstance(features[self.audio_column] , __lowerCamelCase ): raise ValueError(F"""Column {self.audio_column} is not an Audio type.""" ) UpperCamelCase :int = copy.deepcopy(self ) UpperCamelCase :Any = self.input_schema.copy() UpperCamelCase :List[str] = features[self.audio_column] UpperCamelCase :List[Any] = input_schema return task_template @property def _A ( self : Optional[int] ): return {self.audio_column: "audio", self.transcription_column: "transcription"}
38
1
"""simple docstring""" from .data_collator import ( DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForSeqaSeq, DataCollatorForSOP, DataCollatorForTokenClassification, DataCollatorForWholeWordMask, DataCollatorWithPadding, DefaultDataCollator, default_data_collator, ) from .metrics import glue_compute_metrics, xnli_compute_metrics from .processors import ( DataProcessor, InputExample, InputFeatures, SingleSentenceClassificationProcessor, SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels, squad_convert_examples_to_features, xnli_output_modes, xnli_processors, xnli_tasks_num_labels, )
303
"""simple docstring""" from urllib.parse import quote import pytest from datasets.utils.hub import hf_hub_url @pytest.mark.parametrize("repo_id" , ["canonical_dataset_name", "org-name/dataset-name"] ) @pytest.mark.parametrize("path" , ["filename.csv", "filename with blanks.csv"] ) @pytest.mark.parametrize("revision" , [None, "v2"] ) def UpperCamelCase ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) ->int: """simple docstring""" a_ = hf_hub_url(repo_id=UpperCAmelCase , path=UpperCAmelCase , revision=UpperCAmelCase ) assert url == F'''https://huggingface.co/datasets/{repo_id}/resolve/{revision or "main"}/{quote(UpperCAmelCase )}'''
303
1
import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def UpperCAmelCase_ ( __UpperCAmelCase : Any , __UpperCAmelCase : Union[str, Any]=0.9_9_9 , __UpperCAmelCase : str="cosine" , ) -> List[str]: if alpha_transform_type == "cosine": def alpha_bar_fn(__UpperCAmelCase : Any ): return math.cos((t + 0.0_0_8) / 1.0_0_8 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(__UpperCAmelCase : Any ): return math.exp(t * -1_2.0 ) else: raise ValueError(f"Unsupported alpha_tranform_type: {alpha_transform_type}" ) SCREAMING_SNAKE_CASE_ = [] for i in range(__UpperCamelCase ): SCREAMING_SNAKE_CASE_ = i / num_diffusion_timesteps SCREAMING_SNAKE_CASE_ = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(__UpperCamelCase ) / alpha_bar_fn(__UpperCamelCase ) , __UpperCamelCase ) ) return torch.tensor(__UpperCamelCase , dtype=torch.floataa ) class lowerCamelCase_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' lowercase_ = [e.name for e in KarrasDiffusionSchedulers] lowercase_ = 2 @register_to_config def __init__( self : str , _lowerCAmelCase : int = 1_000 , _lowerCAmelCase : float = 0.0_0085 , _lowerCAmelCase : float = 0.012 , _lowerCAmelCase : str = "linear" , _lowerCAmelCase : Optional[Union[np.ndarray, List[float]]] = None , _lowerCAmelCase : str = "epsilon" , _lowerCAmelCase : str = "linspace" , _lowerCAmelCase : int = 0 , ): if trained_betas is not None: SCREAMING_SNAKE_CASE_ = torch.tensor(_lowerCAmelCase , dtype=torch.floataa ) elif beta_schedule == "linear": SCREAMING_SNAKE_CASE_ = torch.linspace(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. SCREAMING_SNAKE_CASE_ = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , _lowerCAmelCase , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule SCREAMING_SNAKE_CASE_ = betas_for_alpha_bar(_lowerCAmelCase ) else: raise NotImplementedError(F"{beta_schedule} does is not implemented for {self.__class__}" ) SCREAMING_SNAKE_CASE_ = 1.0 - self.betas SCREAMING_SNAKE_CASE_ = torch.cumprod(self.alphas , dim=0 ) # set all values self.set_timesteps(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def lowerCAmelCase_ ( self : str , _lowerCAmelCase : Any , _lowerCAmelCase : Dict=None ): if schedule_timesteps is None: SCREAMING_SNAKE_CASE_ = self.timesteps SCREAMING_SNAKE_CASE_ = (schedule_timesteps == timestep).nonzero() # The sigma index that is taken for the **very** first `step` # is always the second index (or the last index if there is only 1) # This way we can ensure we don't accidentally skip a sigma in # case we start in the middle of the denoising schedule (e.g. for image-to-image) if len(self._index_counter ) == 0: SCREAMING_SNAKE_CASE_ = 1 if len(_lowerCAmelCase ) > 1 else 0 else: SCREAMING_SNAKE_CASE_ = timestep.cpu().item() if torch.is_tensor(_lowerCAmelCase ) else timestep SCREAMING_SNAKE_CASE_ = self._index_counter[timestep_int] return indices[pos].item() @property def lowerCAmelCase_ ( self : Any ): # standard deviation of the initial noise distribution if self.config.timestep_spacing in ["linspace", "trailing"]: return self.sigmas.max() return (self.sigmas.max() ** 2 + 1) ** 0.5 def lowerCAmelCase_ ( self : Union[str, Any] , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : Union[float, torch.FloatTensor] , ): SCREAMING_SNAKE_CASE_ = self.index_for_timestep(_lowerCAmelCase ) if self.state_in_first_order: SCREAMING_SNAKE_CASE_ = self.sigmas[step_index] else: SCREAMING_SNAKE_CASE_ = self.sigmas_interpol[step_index] SCREAMING_SNAKE_CASE_ = sample / ((sigma**2 + 1) ** 0.5) return sample def lowerCAmelCase_ ( self : Dict , _lowerCAmelCase : int , _lowerCAmelCase : Union[str, torch.device] = None , _lowerCAmelCase : Optional[int] = None , ): SCREAMING_SNAKE_CASE_ = num_inference_steps SCREAMING_SNAKE_CASE_ = num_train_timesteps or self.config.num_train_timesteps # "linspace", "leading", "trailing" corresponds to annotation of Table 2. of https://arxiv.org/abs/2305.08891 if self.config.timestep_spacing == "linspace": SCREAMING_SNAKE_CASE_ = np.linspace(0 , num_train_timesteps - 1 , _lowerCAmelCase , dtype=_lowerCAmelCase )[::-1].copy() elif self.config.timestep_spacing == "leading": SCREAMING_SNAKE_CASE_ = num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 SCREAMING_SNAKE_CASE_ = (np.arange(0 , _lowerCAmelCase ) * step_ratio).round()[::-1].copy().astype(_lowerCAmelCase ) timesteps += self.config.steps_offset elif self.config.timestep_spacing == "trailing": SCREAMING_SNAKE_CASE_ = num_train_timesteps / self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 SCREAMING_SNAKE_CASE_ = (np.arange(_lowerCAmelCase , 0 , -step_ratio )).round().copy().astype(_lowerCAmelCase ) timesteps -= 1 else: raise ValueError( F"{self.config.timestep_spacing} is not supported. Please make sure to choose one of \'linspace\', \'leading\' or \'trailing\'." ) SCREAMING_SNAKE_CASE_ = np.array(((1 - self.alphas_cumprod) / self.alphas_cumprod) ** 0.5 ) SCREAMING_SNAKE_CASE_ = torch.from_numpy(np.log(_lowerCAmelCase ) ).to(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = np.interp(_lowerCAmelCase , np.arange(0 , len(_lowerCAmelCase ) ) , _lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = np.concatenate([sigmas, [0.0]] ).astype(np.floataa ) SCREAMING_SNAKE_CASE_ = torch.from_numpy(_lowerCAmelCase ).to(device=_lowerCAmelCase ) # interpolate sigmas SCREAMING_SNAKE_CASE_ = sigmas.log().lerp(sigmas.roll(1 ).log() , 0.5 ).exp() SCREAMING_SNAKE_CASE_ = torch.cat([sigmas[:1], sigmas[1:].repeat_interleave(2 ), sigmas[-1:]] ) SCREAMING_SNAKE_CASE_ = torch.cat( [sigmas_interpol[:1], sigmas_interpol[1:].repeat_interleave(2 ), sigmas_interpol[-1:]] ) if str(_lowerCAmelCase ).startswith('mps' ): # mps does not support float64 SCREAMING_SNAKE_CASE_ = torch.from_numpy(_lowerCAmelCase ).to(_lowerCAmelCase , dtype=torch.floataa ) else: SCREAMING_SNAKE_CASE_ = torch.from_numpy(_lowerCAmelCase ).to(_lowerCAmelCase ) # interpolate timesteps SCREAMING_SNAKE_CASE_ = self.sigma_to_t(_lowerCAmelCase ).to(_lowerCAmelCase , dtype=timesteps.dtype ) SCREAMING_SNAKE_CASE_ = torch.stack((timesteps_interpol[1:-1, None], timesteps[1:, None]) , dim=-1 ).flatten() SCREAMING_SNAKE_CASE_ = torch.cat([timesteps[:1], interleaved_timesteps] ) SCREAMING_SNAKE_CASE_ = None # for exp beta schedules, such as the one for `pipeline_shap_e.py` # we need an index counter SCREAMING_SNAKE_CASE_ = defaultdict(_lowerCAmelCase ) def lowerCAmelCase_ ( self : Optional[int] , _lowerCAmelCase : int ): # get log sigma SCREAMING_SNAKE_CASE_ = sigma.log() # get distribution SCREAMING_SNAKE_CASE_ = log_sigma - self.log_sigmas[:, None] # get sigmas range SCREAMING_SNAKE_CASE_ = dists.ge(0 ).cumsum(dim=0 ).argmax(dim=0 ).clamp(max=self.log_sigmas.shape[0] - 2 ) SCREAMING_SNAKE_CASE_ = low_idx + 1 SCREAMING_SNAKE_CASE_ = self.log_sigmas[low_idx] SCREAMING_SNAKE_CASE_ = self.log_sigmas[high_idx] # interpolate sigmas SCREAMING_SNAKE_CASE_ = (low - log_sigma) / (low - high) SCREAMING_SNAKE_CASE_ = w.clamp(0 , 1 ) # transform interpolation to time range SCREAMING_SNAKE_CASE_ = (1 - w) * low_idx + w * high_idx SCREAMING_SNAKE_CASE_ = t.view(sigma.shape ) return t @property def lowerCAmelCase_ ( self : List[Any] ): return self.sample is None def lowerCAmelCase_ ( self : Dict , _lowerCAmelCase : Union[torch.FloatTensor, np.ndarray] , _lowerCAmelCase : Union[float, torch.FloatTensor] , _lowerCAmelCase : Union[torch.FloatTensor, np.ndarray] , _lowerCAmelCase : bool = True , ): SCREAMING_SNAKE_CASE_ = self.index_for_timestep(_lowerCAmelCase ) # advance index counter by 1 SCREAMING_SNAKE_CASE_ = timestep.cpu().item() if torch.is_tensor(_lowerCAmelCase ) else timestep self._index_counter[timestep_int] += 1 if self.state_in_first_order: SCREAMING_SNAKE_CASE_ = self.sigmas[step_index] SCREAMING_SNAKE_CASE_ = self.sigmas_interpol[step_index + 1] SCREAMING_SNAKE_CASE_ = self.sigmas[step_index + 1] else: # 2nd order / KDPM2's method SCREAMING_SNAKE_CASE_ = self.sigmas[step_index - 1] SCREAMING_SNAKE_CASE_ = self.sigmas_interpol[step_index] SCREAMING_SNAKE_CASE_ = self.sigmas[step_index] # currently only gamma=0 is supported. This usually works best anyways. # We can support gamma in the future but then need to scale the timestep before # passing it to the model which requires a change in API SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = sigma * (gamma + 1) # Note: sigma_hat == sigma for now # 1. compute predicted original sample (x_0) from sigma-scaled predicted noise if self.config.prediction_type == "epsilon": SCREAMING_SNAKE_CASE_ = sigma_hat if self.state_in_first_order else sigma_interpol SCREAMING_SNAKE_CASE_ = sample - sigma_input * model_output elif self.config.prediction_type == "v_prediction": SCREAMING_SNAKE_CASE_ = sigma_hat if self.state_in_first_order else sigma_interpol SCREAMING_SNAKE_CASE_ = model_output * (-sigma_input / (sigma_input**2 + 1) ** 0.5) + ( sample / (sigma_input**2 + 1) ) elif self.config.prediction_type == "sample": raise NotImplementedError('prediction_type not implemented yet: sample' ) else: raise ValueError( F"prediction_type given as {self.config.prediction_type} must be one of `epsilon`, or `v_prediction`" ) if self.state_in_first_order: # 2. Convert to an ODE derivative for 1st order SCREAMING_SNAKE_CASE_ = (sample - pred_original_sample) / sigma_hat # 3. delta timestep SCREAMING_SNAKE_CASE_ = sigma_interpol - sigma_hat # store for 2nd order step SCREAMING_SNAKE_CASE_ = sample else: # DPM-Solver-2 # 2. Convert to an ODE derivative for 2nd order SCREAMING_SNAKE_CASE_ = (sample - pred_original_sample) / sigma_interpol # 3. delta timestep SCREAMING_SNAKE_CASE_ = sigma_next - sigma_hat SCREAMING_SNAKE_CASE_ = self.sample SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = sample + derivative * dt if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=_lowerCAmelCase ) def lowerCAmelCase_ ( self : Any , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : torch.FloatTensor , _lowerCAmelCase : torch.FloatTensor , ): # Make sure sigmas and timesteps have the same device and dtype as original_samples SCREAMING_SNAKE_CASE_ = self.sigmas.to(device=original_samples.device , dtype=original_samples.dtype ) if original_samples.device.type == "mps" and torch.is_floating_point(_lowerCAmelCase ): # mps does not support float64 SCREAMING_SNAKE_CASE_ = self.timesteps.to(original_samples.device , dtype=torch.floataa ) SCREAMING_SNAKE_CASE_ = timesteps.to(original_samples.device , dtype=torch.floataa ) else: SCREAMING_SNAKE_CASE_ = self.timesteps.to(original_samples.device ) SCREAMING_SNAKE_CASE_ = timesteps.to(original_samples.device ) SCREAMING_SNAKE_CASE_ = [self.index_for_timestep(_lowerCAmelCase , _lowerCAmelCase ) for t in timesteps] SCREAMING_SNAKE_CASE_ = sigmas[step_indices].flatten() while len(sigma.shape ) < len(original_samples.shape ): SCREAMING_SNAKE_CASE_ = sigma.unsqueeze(-1 ) SCREAMING_SNAKE_CASE_ = original_samples + noise * sigma return noisy_samples def __len__( self : int ): return self.config.num_train_timesteps
225
from functools import lru_cache @lru_cache def a__ ( __UpperCamelCase ): if num < 0: raise ValueError("Number should not be negative." ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
118
0
from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo SCREAMING_SNAKE_CASE_ = """\ @misc{wu2016googles, title={Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation}, author={Yonghui Wu and Mike Schuster and Zhifeng Chen and Quoc V. Le and Mohammad Norouzi and Wolfgang Macherey and Maxim Krikun and Yuan Cao and Qin Gao and Klaus Macherey and Jeff Klingner and Apurva Shah and Melvin Johnson and Xiaobing Liu and Łukasz Kaiser and Stephan Gouws and Yoshikiyo Kato and Taku Kudo and Hideto Kazawa and Keith Stevens and George Kurian and Nishant Patil and Wei Wang and Cliff Young and Jason Smith and Jason Riesa and Alex Rudnick and Oriol Vinyals and Greg Corrado and Macduff Hughes and Jeffrey Dean}, year={2016}, eprint={1609.08144}, archivePrefix={arXiv}, primaryClass={cs.CL} } """ SCREAMING_SNAKE_CASE_ = """\ The BLEU score has some undesirable properties when used for single sentences, as it was designed to be a corpus measure. We therefore use a slightly different score for our RL experiments which we call the 'GLEU score'. For the GLEU score, we record all sub-sequences of 1, 2, 3 or 4 tokens in output and target sequence (n-grams). We then compute a recall, which is the ratio of the number of matching n-grams to the number of total n-grams in the target (ground truth) sequence, and a precision, which is the ratio of the number of matching n-grams to the number of total n-grams in the generated output sequence. Then GLEU score is simply the minimum of recall and precision. This GLEU score's range is always between 0 (no matches) and 1 (all match) and it is symmetrical when switching output and target. According to our experiments, GLEU score correlates quite well with the BLEU metric on a corpus level but does not have its drawbacks for our per sentence reward objective. """ SCREAMING_SNAKE_CASE_ = """\ Computes corpus-level Google BLEU (GLEU) score of translated segments against one or more references. Instead of averaging the sentence level GLEU scores (i.e. macro-average precision), Wu et al. (2016) sum up the matching tokens and the max of hypothesis and reference tokens for each sentence, then compute using the aggregate values. Args: predictions (list of str): list of translations to score. Each translation should be tokenized into a list of tokens. references (list of list of str): list of lists of references for each translation. Each reference should be tokenized into a list of tokens. min_len (int): The minimum order of n-gram this function should extract. Defaults to 1. max_len (int): The maximum order of n-gram this function should extract. Defaults to 4. Returns: 'google_bleu': google_bleu score Examples: Example 1: >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always', ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat'] >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which', ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never', ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat'] >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was', ... 'interested', 'in', 'world', 'history'] >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history', ... 'because', 'he', 'read', 'the', 'book'] >>> list_of_references = [[ref1a], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric(\"google_bleu\") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references) >>> print(round(results[\"google_bleu\"], 2)) 0.44 Example 2: >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always', ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat'] >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which', ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never', ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat'] >>> ref1b = ['It', 'is', 'a', 'guide', 'to', 'action', 'that', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'will', 'never', ... 'heed', 'the', 'cat', 'commands'] >>> ref1c = ['It', 'is', 'the', 'practical', 'guide', 'for', 'the', ... 'rubber', 'duck', 'army', 'never', 'to', 'heed', 'the', 'directions', ... 'of', 'the', 'cat'] >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was', ... 'interested', 'in', 'world', 'history'] >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history', ... 'because', 'he', 'read', 'the', 'book'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric(\"google_bleu\") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references) >>> print(round(results[\"google_bleu\"], 2)) 0.61 Example 3: >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always', ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat'] >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which', ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never', ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat'] >>> ref1b = ['It', 'is', 'a', 'guide', 'to', 'action', 'that', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'will', 'never', ... 'heed', 'the', 'cat', 'commands'] >>> ref1c = ['It', 'is', 'the', 'practical', 'guide', 'for', 'the', ... 'rubber', 'duck', 'army', 'never', 'to', 'heed', 'the', 'directions', ... 'of', 'the', 'cat'] >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was', ... 'interested', 'in', 'world', 'history'] >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history', ... 'because', 'he', 'read', 'the', 'book'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric(\"google_bleu\") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references, min_len=2) >>> print(round(results[\"google_bleu\"], 2)) 0.53 Example 4: >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always', ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat'] >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which', ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never', ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat'] >>> ref1b = ['It', 'is', 'a', 'guide', 'to', 'action', 'that', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'will', 'never', ... 'heed', 'the', 'cat', 'commands'] >>> ref1c = ['It', 'is', 'the', 'practical', 'guide', 'for', 'the', ... 'rubber', 'duck', 'army', 'never', 'to', 'heed', 'the', 'directions', ... 'of', 'the', 'cat'] >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was', ... 'interested', 'in', 'world', 'history'] >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history', ... 'because', 'he', 'read', 'the', 'book'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric(\"google_bleu\") >>> results = google_bleu.compute(predictions=hypotheses,references=list_of_references, min_len=2, max_len=6) >>> print(round(results[\"google_bleu\"], 2)) 0.4 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCamelCase__ ( datasets.Metric ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self : int ) -> MetricInfo: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ,id="""token""" ) ,id="""sequence""" ), """references""": datasets.Sequence( datasets.Sequence(datasets.Value("""string""" ,id="""token""" ) ,id="""sequence""" ) ,id="""references""" ), } ) ,) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ,lowerCamelCase__ : List[List[List[str]]] ,lowerCamelCase__ : List[List[str]] ,lowerCamelCase__ : int = 1 ,lowerCamelCase__ : int = 4 ,) -> Dict[str, float]: '''simple docstring''' return { "google_bleu": gleu_score.corpus_gleu( list_of_references=lowerCamelCase__ ,hypotheses=lowerCamelCase__ ,min_len=lowerCamelCase__ ,max_len=lowerCamelCase__ ) }
193
from __future__ import annotations def __lowercase ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> list[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = len(_SCREAMING_SNAKE_CASE ) - 1 while i < j: if nums[i] + nums[j] == target: return [i, j] elif nums[i] + nums[j] < target: SCREAMING_SNAKE_CASE = i + 1 else: SCREAMING_SNAKE_CASE = j - 1 return [] if __name__ == "__main__": import doctest doctest.testmod() print(F'''{two_pointer([2, 7, 1_1, 1_5], 9) = }''')
193
1
"""simple docstring""" from string import ascii_uppercase __lowercase = {str(ord(c) - 55): c for c in ascii_uppercase} def lowercase ( A_ , A_ )-> str: '''simple docstring''' if isinstance(A_ , A_ ): raise TypeError("int() can't convert non-string with explicit base" ) if num < 0: raise ValueError("parameter must be positive int" ) if isinstance(A_ , A_ ): raise TypeError("'str' object cannot be interpreted as an integer" ) if isinstance(A_ , A_ ): raise TypeError("'float' object cannot be interpreted as an integer" ) if base in (0, 1): raise ValueError("base must be >= 2" ) if base > 36: raise ValueError("base must be <= 36" ) a : Tuple = "" a : Optional[Any] = 0 a : Optional[Any] = 0 while div != 1: a , a : Union[str, Any] = divmod(A_ , A_ ) if base >= 11 and 9 < mod < 36: a : Any = ALPHABET_VALUES[str(A_ )] else: a : Optional[Any] = str(A_ ) new_value += actual_value a : Union[str, Any] = num // base a : Union[str, Any] = div if div == 0: return str(new_value[::-1] ) elif div == 1: new_value += str(A_ ) return str(new_value[::-1] ) return new_value[::-1] if __name__ == "__main__": import doctest doctest.testmod() for base in range(2, 37): for num in range(1000): assert int(decimal_to_any(num, base), base) == num, ( num, base, decimal_to_any(num, base), int(decimal_to_any(num, base), base), )
40
"""simple docstring""" import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class snake_case ( __snake_case, unittest.TestCase ): SCREAMING_SNAKE_CASE_ : List[Any] = RoCBertTokenizer SCREAMING_SNAKE_CASE_ : Any = None SCREAMING_SNAKE_CASE_ : Any = False SCREAMING_SNAKE_CASE_ : Dict = True SCREAMING_SNAKE_CASE_ : Optional[int] = filter_non_english def lowercase_ ( self : Optional[Any])-> Any: '''simple docstring''' super().setUp() __lowerCAmelCase: Optional[Any] = ["[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "你", "好", "是", "谁", "a", "b", "c", "d"] __lowerCAmelCase: List[Any] = {} __lowerCAmelCase: Dict = {} for i, value in enumerate(UpperCamelCase__): __lowerCAmelCase: List[Any] = i __lowerCAmelCase: Union[str, Any] = i __lowerCAmelCase: List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) __lowerCAmelCase: int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["word_shape_file"]) __lowerCAmelCase: int = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["word_pronunciation_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens])) with open(self.word_shape_file , "w" , encoding="utf-8") as word_shape_writer: json.dump(UpperCamelCase__ , UpperCamelCase__ , ensure_ascii=UpperCamelCase__) with open(self.word_pronunciation_file , "w" , encoding="utf-8") as word_pronunciation_writer: json.dump(UpperCamelCase__ , UpperCamelCase__ , ensure_ascii=UpperCamelCase__) def lowercase_ ( self : Any)-> Tuple: '''simple docstring''' __lowerCAmelCase: Tuple = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file) __lowerCAmelCase: Union[str, Any] = tokenizer.tokenize("你好[SEP]你是谁") self.assertListEqual(UpperCamelCase__ , ["你", "好", "[SEP]", "你", "是", "谁"]) self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCamelCase__) , [5, 6, 2, 5, 7, 8]) self.assertListEqual(tokenizer.convert_tokens_to_shape_ids(UpperCamelCase__) , [5, 6, 2, 5, 7, 8]) self.assertListEqual(tokenizer.convert_tokens_to_pronunciation_ids(UpperCamelCase__) , [5, 6, 2, 5, 7, 8]) def lowercase_ ( self : Optional[Any])-> List[str]: '''simple docstring''' __lowerCAmelCase: int = RoCBertBasicTokenizer() self.assertListEqual(tokenizer.tokenize("ah\u535A\u63A8zz") , ["ah", "\u535A", "\u63A8", "zz"]) def lowercase_ ( self : str)-> Dict: '''simple docstring''' __lowerCAmelCase: int = RoCBertBasicTokenizer(do_lower_case=UpperCamelCase__) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? ") , ["hello", "!", "how", "are", "you", "?"]) self.assertListEqual(tokenizer.tokenize("H\u00E9llo") , ["hello"]) def lowercase_ ( self : Optional[int])-> List[Any]: '''simple docstring''' __lowerCAmelCase: Union[str, Any] = RoCBertBasicTokenizer(do_lower_case=UpperCamelCase__ , strip_accents=UpperCamelCase__) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? ") , ["hällo", "!", "how", "are", "you", "?"]) self.assertListEqual(tokenizer.tokenize("H\u00E9llo") , ["h\u00E9llo"]) def lowercase_ ( self : Optional[Any])-> Any: '''simple docstring''' __lowerCAmelCase: Tuple = RoCBertBasicTokenizer(do_lower_case=UpperCamelCase__ , strip_accents=UpperCamelCase__) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? ") , ["hallo", "!", "how", "are", "you", "?"]) self.assertListEqual(tokenizer.tokenize("H\u00E9llo") , ["hello"]) def lowercase_ ( self : str)-> List[str]: '''simple docstring''' __lowerCAmelCase: List[str] = RoCBertBasicTokenizer(do_lower_case=UpperCamelCase__) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? ") , ["hallo", "!", "how", "are", "you", "?"]) self.assertListEqual(tokenizer.tokenize("H\u00E9llo") , ["hello"]) def lowercase_ ( self : Any)-> Any: '''simple docstring''' __lowerCAmelCase: Union[str, Any] = RoCBertBasicTokenizer(do_lower_case=UpperCamelCase__) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? ") , ["HeLLo", "!", "how", "Are", "yoU", "?"]) def lowercase_ ( self : Optional[int])-> Tuple: '''simple docstring''' __lowerCAmelCase: Optional[int] = RoCBertBasicTokenizer(do_lower_case=UpperCamelCase__ , strip_accents=UpperCamelCase__) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? ") , ["HäLLo", "!", "how", "Are", "yoU", "?"]) def lowercase_ ( self : Optional[Any])-> Tuple: '''simple docstring''' __lowerCAmelCase: Optional[Any] = RoCBertBasicTokenizer(do_lower_case=UpperCamelCase__ , strip_accents=UpperCamelCase__) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? ") , ["HaLLo", "!", "how", "Are", "yoU", "?"]) def lowercase_ ( self : Tuple)-> str: '''simple docstring''' __lowerCAmelCase: Optional[Any] = RoCBertBasicTokenizer(do_lower_case=UpperCamelCase__ , never_split=["[UNK]"]) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? [UNK]") , ["HeLLo", "!", "how", "Are", "yoU", "?", "[UNK]"]) def lowercase_ ( self : List[Any])-> Any: '''simple docstring''' __lowerCAmelCase: List[str] = ["[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing"] __lowerCAmelCase: int = {} for i, token in enumerate(UpperCamelCase__): __lowerCAmelCase: Optional[Any] = i __lowerCAmelCase: str = RoCBertWordpieceTokenizer(vocab=UpperCamelCase__ , unk_token="[UNK]") self.assertListEqual(tokenizer.tokenize("") , []) self.assertListEqual(tokenizer.tokenize("unwanted running") , ["un", "##want", "##ed", "runn", "##ing"]) self.assertListEqual(tokenizer.tokenize("unwantedX running") , ["[UNK]", "runn", "##ing"]) def lowercase_ ( self : Optional[Any])-> Dict: '''simple docstring''' self.assertTrue(_is_whitespace(" ")) self.assertTrue(_is_whitespace("\t")) self.assertTrue(_is_whitespace("\r")) self.assertTrue(_is_whitespace("\n")) self.assertTrue(_is_whitespace("\u00A0")) self.assertFalse(_is_whitespace("A")) self.assertFalse(_is_whitespace("-")) def lowercase_ ( self : Dict)-> Optional[int]: '''simple docstring''' self.assertTrue(_is_control("\u0005")) self.assertFalse(_is_control("A")) self.assertFalse(_is_control(" ")) self.assertFalse(_is_control("\t")) self.assertFalse(_is_control("\r")) def lowercase_ ( self : Union[str, Any])-> str: '''simple docstring''' self.assertTrue(_is_punctuation("-")) self.assertTrue(_is_punctuation("$")) self.assertTrue(_is_punctuation("`")) self.assertTrue(_is_punctuation(".")) self.assertFalse(_is_punctuation("A")) self.assertFalse(_is_punctuation(" ")) def lowercase_ ( self : Dict)-> int: '''simple docstring''' __lowerCAmelCase: Any = self.get_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(UpperCamelCase__) for t in ["Test", "\xad", "test"]] , [["[UNK]"], [], ["[UNK]"]]) if self.test_rust_tokenizer: __lowerCAmelCase: Any = self.get_rust_tokenizer() self.assertListEqual( [rust_tokenizer.tokenize(UpperCamelCase__) for t in ["Test", "\xad", "test"]] , [["[UNK]"], [], ["[UNK]"]]) def lowercase_ ( self : Dict)-> Any: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})"): __lowerCAmelCase: str = self.rust_tokenizer_class.from_pretrained(UpperCamelCase__ , **UpperCamelCase__) __lowerCAmelCase: Optional[Any] = f"A, naïve {tokenizer_r.mask_token} AllenNLP sentence." __lowerCAmelCase: Tuple = tokenizer_r.encode_plus( UpperCamelCase__ , return_attention_mask=UpperCamelCase__ , return_token_type_ids=UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , ) __lowerCAmelCase: str = tokenizer_r.do_lower_case if hasattr(UpperCamelCase__ , "do_lower_case") else False __lowerCAmelCase: List[Any] = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "A"), ((1, 2), ","), ((3, 5), "na"), ((5, 6), "##ï"), ((6, 8), "##ve"), ((9, 1_5), tokenizer_r.mask_token), ((1_6, 2_1), "Allen"), ((2_1, 2_3), "##NL"), ((2_3, 2_4), "##P"), ((2_5, 3_3), "sentence"), ((3_3, 3_4), "."), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "a"), ((1, 2), ","), ((3, 8), "naive"), ((9, 1_5), tokenizer_r.mask_token), ((1_6, 2_1), "allen"), ((2_1, 2_3), "##nl"), ((2_3, 2_4), "##p"), ((2_5, 3_3), "sentence"), ((3_3, 3_4), "."), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["input_ids"])) self.assertEqual([e[0] for e in expected_results] , tokens["offset_mapping"]) def lowercase_ ( self : Union[str, Any])-> Optional[int]: '''simple docstring''' __lowerCAmelCase: Optional[Any] = ["的", "人", "有"] __lowerCAmelCase: int = "".join(UpperCamelCase__) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})"): __lowerCAmelCase: Tuple = True __lowerCAmelCase: str = self.tokenizer_class.from_pretrained(UpperCamelCase__ , **UpperCamelCase__) __lowerCAmelCase: Dict = self.rust_tokenizer_class.from_pretrained(UpperCamelCase__ , **UpperCamelCase__) __lowerCAmelCase: Union[str, Any] = tokenizer_p.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__) __lowerCAmelCase: List[Any] = tokenizer_r.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__) __lowerCAmelCase: Any = tokenizer_r.convert_ids_to_tokens(UpperCamelCase__) __lowerCAmelCase: List[str] = tokenizer_p.convert_ids_to_tokens(UpperCamelCase__) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(UpperCamelCase__ , UpperCamelCase__) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__) __lowerCAmelCase: int = False __lowerCAmelCase: Any = self.rust_tokenizer_class.from_pretrained(UpperCamelCase__ , **UpperCamelCase__) __lowerCAmelCase: Union[str, Any] = self.tokenizer_class.from_pretrained(UpperCamelCase__ , **UpperCamelCase__) __lowerCAmelCase: str = tokenizer_r.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__) __lowerCAmelCase: str = tokenizer_p.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__) __lowerCAmelCase: str = tokenizer_r.convert_ids_to_tokens(UpperCamelCase__) __lowerCAmelCase: Tuple = tokenizer_p.convert_ids_to_tokens(UpperCamelCase__) # it is expected that only the first Chinese character is not preceded by "##". __lowerCAmelCase: Dict = [ f"##{token}" if idx != 0 else token for idx, token in enumerate(UpperCamelCase__) ] self.assertListEqual(UpperCamelCase__ , UpperCamelCase__) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__) @slow def lowercase_ ( self : Optional[Any])-> Any: '''simple docstring''' __lowerCAmelCase: str = self.tokenizer_class(self.vocab_file , self.word_shape_file , self.word_pronunciation_file) __lowerCAmelCase: Dict = tokenizer.encode("你好" , add_special_tokens=UpperCamelCase__) __lowerCAmelCase: Union[str, Any] = tokenizer.encode("你是谁" , add_special_tokens=UpperCamelCase__) __lowerCAmelCase: Tuple = tokenizer.build_inputs_with_special_tokens(UpperCamelCase__) __lowerCAmelCase: List[Any] = tokenizer.build_inputs_with_special_tokens(UpperCamelCase__ , UpperCamelCase__) assert encoded_sentence == [1] + text + [2] assert encoded_pair == [1] + text + [2] + text_a + [2] def lowercase_ ( self : Tuple)-> Optional[Any]: '''simple docstring''' __lowerCAmelCase: int = self.get_tokenizers(do_lower_case=UpperCamelCase__) for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}"): __lowerCAmelCase: str = "你好,你是谁" __lowerCAmelCase: Dict = tokenizer.tokenize(UpperCamelCase__) __lowerCAmelCase: Union[str, Any] = tokenizer.convert_tokens_to_ids(UpperCamelCase__) __lowerCAmelCase: Optional[Any] = tokenizer.convert_tokens_to_shape_ids(UpperCamelCase__) __lowerCAmelCase: Tuple = tokenizer.convert_tokens_to_pronunciation_ids(UpperCamelCase__) __lowerCAmelCase: Dict = tokenizer.prepare_for_model( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , add_special_tokens=UpperCamelCase__) __lowerCAmelCase: Optional[Any] = tokenizer.encode_plus(UpperCamelCase__ , add_special_tokens=UpperCamelCase__) self.assertEqual(UpperCamelCase__ , UpperCamelCase__)
217
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class UpperCAmelCase_ ( lowerCAmelCase_ , unittest.TestCase ): lowerCamelCase : Optional[int] = KandinskyVaaImgaImgPipeline lowerCamelCase : List[Any] = ["""image_embeds""", """negative_image_embeds""", """image"""] lowerCamelCase : Union[str, Any] = [ """image_embeds""", """negative_image_embeds""", """image""", ] lowerCamelCase : Optional[int] = [ """generator""", """height""", """width""", """strength""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] lowerCamelCase : Optional[Any] = False @property def __UpperCAmelCase ( self : Dict ) -> Union[str, Any]: return 3_2 @property def __UpperCAmelCase ( self : Any ) -> str: return 3_2 @property def __UpperCAmelCase ( self : int ) -> Any: return self.time_input_dim @property def __UpperCAmelCase ( self : str ) -> Optional[int]: return self.time_input_dim * 4 @property def __UpperCAmelCase ( self : int ) -> Any: return 1_0_0 @property def __UpperCAmelCase ( self : Union[str, Any] ) -> List[Any]: torch.manual_seed(0 ) lowerCAmelCase = { 'in_channels': 4, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } lowerCAmelCase = UNetaDConditionModel(**__SCREAMING_SNAKE_CASE ) return model @property def __UpperCAmelCase ( self : Union[str, Any] ) -> Optional[Any]: return { "block_out_channels": [3_2, 6_4], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 1_2, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def __UpperCAmelCase ( self : Optional[int] ) -> List[str]: torch.manual_seed(0 ) lowerCAmelCase = VQModel(**self.dummy_movq_kwargs ) return model def __UpperCAmelCase ( self : Union[str, Any] ) -> Tuple: lowerCAmelCase = self.dummy_unet lowerCAmelCase = self.dummy_movq lowerCAmelCase = { 'num_train_timesteps': 1_0_0_0, 'beta_schedule': 'linear', 'beta_start': 0.00_085, 'beta_end': 0.012, 'clip_sample': False, 'set_alpha_to_one': False, 'steps_offset': 0, 'prediction_type': 'epsilon', 'thresholding': False, } lowerCAmelCase = DDIMScheduler(**__SCREAMING_SNAKE_CASE ) lowerCAmelCase = { 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def __UpperCAmelCase ( self : int , UpperCAmelCase__ : Any , UpperCAmelCase__ : Optional[Any]=0 ) -> Any: lowerCAmelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(__SCREAMING_SNAKE_CASE ) ).to(__SCREAMING_SNAKE_CASE ) lowerCAmelCase = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( __SCREAMING_SNAKE_CASE ) # create init_image lowerCAmelCase = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(__SCREAMING_SNAKE_CASE ) ).to(__SCREAMING_SNAKE_CASE ) lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowerCAmelCase = Image.fromarray(np.uinta(__SCREAMING_SNAKE_CASE ) ).convert('RGB' ).resize((2_5_6, 2_5_6) ) if str(__SCREAMING_SNAKE_CASE ).startswith('mps' ): lowerCAmelCase = torch.manual_seed(__SCREAMING_SNAKE_CASE ) else: lowerCAmelCase = torch.Generator(device=__SCREAMING_SNAKE_CASE ).manual_seed(__SCREAMING_SNAKE_CASE ) lowerCAmelCase = { 'image': init_image, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 6_4, 'width': 6_4, 'num_inference_steps': 1_0, 'guidance_scale': 7.0, 'strength': 0.2, 'output_type': 'np', } return inputs def __UpperCAmelCase ( self : Union[str, Any] ) -> Any: lowerCAmelCase = 'cpu' lowerCAmelCase = self.get_dummy_components() lowerCAmelCase = self.pipeline_class(**__SCREAMING_SNAKE_CASE ) lowerCAmelCase = pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) lowerCAmelCase = pipe(**self.get_dummy_inputs(__SCREAMING_SNAKE_CASE ) ) lowerCAmelCase = output.images lowerCAmelCase = pipe( **self.get_dummy_inputs(__SCREAMING_SNAKE_CASE ) , return_dict=__SCREAMING_SNAKE_CASE , )[0] lowerCAmelCase = image[0, -3:, -3:, -1] lowerCAmelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) lowerCAmelCase = np.array( [0.6_199_778, 0.63_984_406, 0.46_145_785, 0.62_944_984, 0.5_622_215, 0.47_306_132, 0.47_441_456, 0.4_607_606, 0.48_719_263] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), F''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), F''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class UpperCAmelCase_ ( unittest.TestCase ): def __UpperCAmelCase ( self : str ) -> Optional[Any]: super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCAmelCase ( self : Union[str, Any] ) -> int: lowerCAmelCase = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/kandinskyv22_img2img_frog.npy' ) lowerCAmelCase = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png' ) lowerCAmelCase = 'A red cartoon frog, 4k' lowerCAmelCase = KandinskyVaaPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-prior' , torch_dtype=torch.floataa ) pipe_prior.to(__SCREAMING_SNAKE_CASE ) lowerCAmelCase = KandinskyVaaImgaImgPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-decoder' , torch_dtype=torch.floataa ) lowerCAmelCase = pipeline.to(__SCREAMING_SNAKE_CASE ) pipeline.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) lowerCAmelCase = torch.Generator(device='cpu' ).manual_seed(0 ) lowerCAmelCase , lowerCAmelCase = pipe_prior( __SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE , num_inference_steps=5 , negative_prompt='' , ).to_tuple() lowerCAmelCase = pipeline( image=__SCREAMING_SNAKE_CASE , image_embeds=__SCREAMING_SNAKE_CASE , negative_image_embeds=__SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE , num_inference_steps=1_0_0 , height=7_6_8 , width=7_6_8 , strength=0.2 , output_type='np' , ) lowerCAmelCase = output.images[0] assert image.shape == (7_6_8, 7_6_8, 3) assert_mean_pixel_difference(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )
352
'''simple docstring''' from __future__ import annotations from scipy.special import comb # type: ignore class UpperCAmelCase_ : def __init__( self : Dict , UpperCAmelCase__ : list[tuple[float, float]] ) -> str: lowerCAmelCase = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. lowerCAmelCase = len(UpperCAmelCase__ ) - 1 def __UpperCAmelCase ( self : str , UpperCAmelCase__ : float ) -> list[float]: assert 0 <= t <= 1, "Time t must be between 0 and 1." lowerCAmelCase = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , UpperCAmelCase__ ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(UpperCAmelCase__ ) , 5 ) == 1 return output_values def __UpperCAmelCase ( self : Union[str, Any] , UpperCAmelCase__ : float ) -> tuple[float, float]: assert 0 <= t <= 1, "Time t must be between 0 and 1." lowerCAmelCase = self.basis_function(UpperCAmelCase__ ) lowerCAmelCase = 0.0 lowerCAmelCase = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def __UpperCAmelCase ( self : Optional[int] , UpperCAmelCase__ : float = 0.01 ) -> Optional[int]: from matplotlib import pyplot as plt # type: ignore lowerCAmelCase = [] # x coordinates of points to plot lowerCAmelCase = [] # y coordinates of points to plot lowerCAmelCase = 0.0 while t <= 1: lowerCAmelCase = self.bezier_curve_function(UpperCAmelCase__ ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size lowerCAmelCase = [i[0] for i in self.list_of_points] lowerCAmelCase = [i[1] for i in self.list_of_points] plt.plot( UpperCAmelCase__ , UpperCAmelCase__ , color='blue' , label='Curve of Degree ' + str(self.degree ) , ) plt.scatter(UpperCAmelCase__ , UpperCAmelCase__ , color='red' , label='Control Points' ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
55
0
import numpy as np def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> np.ndarray: return np.where(vector > 0 ,_SCREAMING_SNAKE_CASE ,(alpha * (np.exp(_SCREAMING_SNAKE_CASE ) - 1)) ) if __name__ == "__main__": import doctest doctest.testmod()
48
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(lowerCAmelCase__ ) , """Tatoeba directory does not exist.""" ) class UpperCamelCase__ (unittest.TestCase ): '''simple docstring''' @cached_property def _lowercase ( self ) -> int: lowerCamelCase : str = tempfile.mkdtemp() return TatoebaConverter(save_dir=UpperCamelCase__ ) @slow def _lowercase ( self ) -> List[Any]: self.resolver.convert_models(["heb-eng"] ) @slow def _lowercase ( self ) -> Tuple: lowerCamelCase , lowerCamelCase : Dict = self.resolver.write_model_card("opus-mt-he-en" , dry_run=UpperCamelCase__ ) assert mmeta["long_pair"] == "heb-eng"
48
1
def UpperCamelCase( lowercase_ ) -> Optional[int]: '''simple docstring''' if not isinstance(lowercase_ , lowercase_ ): snake_case_ = f'''Input value of [number={number}] must be an integer''' raise TypeError(lowercase_ ) if number < 1: snake_case_ = f'''Input value of [number={number}] must be > 0''' raise ValueError(lowercase_ ) snake_case_ = 1 for i in range(1 , lowercase_ ): current_number *= 4 * i - 2 current_number //= i + 1 return current_number if __name__ == "__main__": import doctest doctest.testmod()
369
import argparse import collections import json import os import re import string import sys import numpy as np lowerCamelCase_ = re.compile(R'''\b(a|an|the)\b''', re.UNICODE) lowerCamelCase_ = None def UpperCamelCase( ) -> List[Any]: '''simple docstring''' snake_case_ = argparse.ArgumentParser("""Official evaluation script for SQuAD version 2.0.""" ) parser.add_argument("""data_file""" , metavar="""data.json""" , help="""Input data JSON file.""" ) parser.add_argument("""pred_file""" , metavar="""pred.json""" , help="""Model predictions.""" ) parser.add_argument( """--out-file""" , """-o""" , metavar="""eval.json""" , help="""Write accuracy metrics to file (default is stdout).""" ) parser.add_argument( """--na-prob-file""" , """-n""" , metavar="""na_prob.json""" , help="""Model estimates of probability of no answer.""" ) parser.add_argument( """--na-prob-thresh""" , """-t""" , type=lowercase_ , default=1.0 , help="""Predict \"\" if no-answer probability exceeds this (default = 1.0).""" , ) parser.add_argument( """--out-image-dir""" , """-p""" , metavar="""out_images""" , default=lowercase_ , help="""Save precision-recall curves to directory.""" ) parser.add_argument("""--verbose""" , """-v""" , action="""store_true""" ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def UpperCamelCase( lowercase_ ) -> Union[str, Any]: '''simple docstring''' snake_case_ = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: snake_case_ = bool(qa["""answers"""]["""text"""] ) return qid_to_has_ans def UpperCamelCase( lowercase_ ) -> Tuple: '''simple docstring''' def remove_articles(lowercase_ ): return ARTICLES_REGEX.sub(""" """ , lowercase_ ) def white_space_fix(lowercase_ ): return " ".join(text.split() ) def remove_punc(lowercase_ ): snake_case_ = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(lowercase_ ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(lowercase_ ) ) ) ) def UpperCamelCase( lowercase_ ) -> Dict: '''simple docstring''' if not s: return [] return normalize_answer(lowercase_ ).split() def UpperCamelCase( lowercase_ , lowercase_ ) -> Dict: '''simple docstring''' return int(normalize_answer(lowercase_ ) == normalize_answer(lowercase_ ) ) def UpperCamelCase( lowercase_ , lowercase_ ) -> Any: '''simple docstring''' snake_case_ = get_tokens(lowercase_ ) snake_case_ = get_tokens(lowercase_ ) snake_case_ = collections.Counter(lowercase_ ) & collections.Counter(lowercase_ ) snake_case_ = sum(common.values() ) if len(lowercase_ ) == 0 or len(lowercase_ ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 snake_case_ = 1.0 * num_same / len(lowercase_ ) snake_case_ = 1.0 * num_same / len(lowercase_ ) snake_case_ = (2 * precision * recall) / (precision + recall) return fa def UpperCamelCase( lowercase_ , lowercase_ ) -> List[str]: '''simple docstring''' snake_case_ = {} snake_case_ = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: snake_case_ = qa["""id"""] snake_case_ = [t for t in qa["""answers"""]["""text"""] if normalize_answer(lowercase_ )] if not gold_answers: # For unanswerable questions, only correct answer is empty string snake_case_ = [""""""] if qid not in preds: print(f'''Missing prediction for {qid}''' ) continue snake_case_ = preds[qid] # Take max over all gold answers snake_case_ = max(compute_exact(lowercase_ , lowercase_ ) for a in gold_answers ) snake_case_ = max(compute_fa(lowercase_ , lowercase_ ) for a in gold_answers ) return exact_scores, fa_scores def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> Union[str, Any]: '''simple docstring''' snake_case_ = {} for qid, s in scores.items(): snake_case_ = na_probs[qid] > na_prob_thresh if pred_na: snake_case_ = float(not qid_to_has_ans[qid] ) else: snake_case_ = s return new_scores def UpperCamelCase( lowercase_ , lowercase_ , lowercase_=None ) -> Dict: '''simple docstring''' if not qid_list: snake_case_ = len(lowercase_ ) return collections.OrderedDict( [ ("""exact""", 1_00.0 * sum(exact_scores.values() ) / total), ("""f1""", 1_00.0 * sum(fa_scores.values() ) / total), ("""total""", total), ] ) else: snake_case_ = len(lowercase_ ) return collections.OrderedDict( [ ("""exact""", 1_00.0 * sum(exact_scores[k] for k in qid_list ) / total), ("""f1""", 1_00.0 * sum(fa_scores[k] for k in qid_list ) / total), ("""total""", total), ] ) def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ ) -> Any: '''simple docstring''' for k in new_eval: snake_case_ = new_eval[k] def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> Dict: '''simple docstring''' plt.step(lowercase_ , lowercase_ , color="""b""" , alpha=0.2 , where="""post""" ) plt.fill_between(lowercase_ , lowercase_ , step="""post""" , alpha=0.2 , color="""b""" ) plt.xlabel("""Recall""" ) plt.ylabel("""Precision""" ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(lowercase_ ) plt.savefig(lowercase_ ) plt.clf() def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_=None , lowercase_=None ) -> Dict: '''simple docstring''' snake_case_ = sorted(lowercase_ , key=lambda lowercase_ : na_probs[k] ) snake_case_ = 0.0 snake_case_ = 1.0 snake_case_ = 0.0 snake_case_ = [1.0] snake_case_ = [0.0] snake_case_ = 0.0 for i, qid in enumerate(lowercase_ ): if qid_to_has_ans[qid]: true_pos += scores[qid] snake_case_ = true_pos / float(i + 1 ) snake_case_ = true_pos / float(lowercase_ ) if i == len(lowercase_ ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(lowercase_ ) recalls.append(lowercase_ ) if out_image: plot_pr_curve(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) return {"ap": 1_00.0 * avg_prec} def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> str: '''simple docstring''' if out_image_dir and not os.path.exists(lowercase_ ): os.makedirs(lowercase_ ) snake_case_ = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return snake_case_ = make_precision_recall_eval( lowercase_ , lowercase_ , lowercase_ , lowercase_ , out_image=os.path.join(lowercase_ , """pr_exact.png""" ) , title="""Precision-Recall curve for Exact Match score""" , ) snake_case_ = make_precision_recall_eval( lowercase_ , lowercase_ , lowercase_ , lowercase_ , out_image=os.path.join(lowercase_ , """pr_f1.png""" ) , title="""Precision-Recall curve for F1 score""" , ) snake_case_ = {k: float(lowercase_ ) for k, v in qid_to_has_ans.items()} snake_case_ = make_precision_recall_eval( lowercase_ , lowercase_ , lowercase_ , lowercase_ , out_image=os.path.join(lowercase_ , """pr_oracle.png""" ) , title="""Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)""" , ) merge_eval(lowercase_ , lowercase_ , """pr_exact""" ) merge_eval(lowercase_ , lowercase_ , """pr_f1""" ) merge_eval(lowercase_ , lowercase_ , """pr_oracle""" ) def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> int: '''simple docstring''' if not qid_list: return snake_case_ = [na_probs[k] for k in qid_list] snake_case_ = np.ones_like(lowercase_ ) / float(len(lowercase_ ) ) plt.hist(lowercase_ , weights=lowercase_ , bins=20 , range=(0.0, 1.0) ) plt.xlabel("""Model probability of no-answer""" ) plt.ylabel("""Proportion of dataset""" ) plt.title(f'''Histogram of no-answer probability: {name}''' ) plt.savefig(os.path.join(lowercase_ , f'''na_prob_hist_{name}.png''' ) ) plt.clf() def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> Optional[int]: '''simple docstring''' snake_case_ = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) snake_case_ = num_no_ans snake_case_ = cur_score snake_case_ = 0.0 snake_case_ = sorted(lowercase_ , key=lambda lowercase_ : na_probs[k] ) for i, qid in enumerate(lowercase_ ): if qid not in scores: continue if qid_to_has_ans[qid]: snake_case_ = scores[qid] else: if preds[qid]: snake_case_ = -1 else: snake_case_ = 0 cur_score += diff if cur_score > best_score: snake_case_ = cur_score snake_case_ = na_probs[qid] return 1_00.0 * best_score / len(lowercase_ ), best_thresh def UpperCamelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> Any: '''simple docstring''' snake_case_ , snake_case_ = find_best_thresh(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) snake_case_ , snake_case_ = find_best_thresh(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) snake_case_ = best_exact snake_case_ = exact_thresh snake_case_ = best_fa snake_case_ = fa_thresh def UpperCamelCase( ) -> Union[str, Any]: '''simple docstring''' with open(OPTS.data_file ) as f: snake_case_ = json.load(lowercase_ ) snake_case_ = dataset_json["""data"""] with open(OPTS.pred_file ) as f: snake_case_ = json.load(lowercase_ ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: snake_case_ = json.load(lowercase_ ) else: snake_case_ = {k: 0.0 for k in preds} snake_case_ = make_qid_to_has_ans(lowercase_ ) # maps qid to True/False snake_case_ = [k for k, v in qid_to_has_ans.items() if v] snake_case_ = [k for k, v in qid_to_has_ans.items() if not v] snake_case_ , snake_case_ = get_raw_scores(lowercase_ , lowercase_ ) snake_case_ = apply_no_ans_threshold(lowercase_ , lowercase_ , lowercase_ , OPTS.na_prob_thresh ) snake_case_ = apply_no_ans_threshold(lowercase_ , lowercase_ , lowercase_ , OPTS.na_prob_thresh ) snake_case_ = make_eval_dict(lowercase_ , lowercase_ ) if has_ans_qids: snake_case_ = make_eval_dict(lowercase_ , lowercase_ , qid_list=lowercase_ ) merge_eval(lowercase_ , lowercase_ , """HasAns""" ) if no_ans_qids: snake_case_ = make_eval_dict(lowercase_ , lowercase_ , qid_list=lowercase_ ) merge_eval(lowercase_ , lowercase_ , """NoAns""" ) if OPTS.na_prob_file: find_all_best_thresh(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , OPTS.out_image_dir ) histogram_na_prob(lowercase_ , lowercase_ , OPTS.out_image_dir , """hasAns""" ) histogram_na_prob(lowercase_ , lowercase_ , OPTS.out_image_dir , """noAns""" ) if OPTS.out_file: with open(OPTS.out_file , """w""" ) as f: json.dump(lowercase_ , lowercase_ ) else: print(json.dumps(lowercase_ , indent=2 ) ) if __name__ == "__main__": lowerCamelCase_ = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use('''Agg''') import matplotlib.pyplot as plt main()
34
0
'''simple docstring''' def SCREAMING_SNAKE_CASE( __lowercase , __lowercase ) -> str: if not isinstance(__lowercase , __lowercase ): raise ValueError('''iterations must be defined as integers''' ) if not isinstance(__lowercase , __lowercase ) or not number >= 1: raise ValueError( '''starting number must be and integer and be more than 0''' ) if not iterations >= 1: raise ValueError('''Iterations must be done more than 0 times to play FizzBuzz''' ) A: Union[str, Any] = '''''' while number <= iterations: if number % 3 == 0: out += "Fizz" if number % 5 == 0: out += "Buzz" if 0 not in (number % 3, number % 5): out += str(__lowercase ) # print(out) number += 1 out += " " return out if __name__ == "__main__": import doctest doctest.testmod()
319
'''simple docstring''' from __future__ import annotations def SCREAMING_SNAKE_CASE( __lowercase ) -> bool: if len(__lowercase ) < 2: raise ValueError('''Monogons and Digons are not polygons in the Euclidean space''' ) if any(i <= 0 for i in nums ): raise ValueError('''All values must be greater than 0''' ) A: Any = nums.copy() copy_nums.sort() return copy_nums[-1] < sum(copy_nums[:-1] ) if __name__ == "__main__": import doctest doctest.testmod()
319
1
"""simple docstring""" from __future__ import annotations from collections import namedtuple from dataclasses import dataclass @dataclass class __A : '''simple docstring''' __lowercase: int __lowercase: TreeNode | None = None __lowercase: TreeNode | None = None __SCREAMING_SNAKE_CASE : Dict = namedtuple('CoinsDistribResult', 'moves excess') def _a ( _SCREAMING_SNAKE_CASE ) -> int: if root is None: return 0 # Validation def count_nodes(_SCREAMING_SNAKE_CASE ) -> int: if node is None: return 0 return count_nodes(node.left ) + count_nodes(node.right ) + 1 def count_coins(_SCREAMING_SNAKE_CASE ) -> int: if node is None: return 0 return count_coins(node.left ) + count_coins(node.right ) + node.data if count_nodes(a_ ) != count_coins(a_ ): raise ValueError("""The nodes number should be same as the number of coins""" ) # Main calculation def get_distrib(_SCREAMING_SNAKE_CASE ) -> CoinsDistribResult: if node is None: return CoinsDistribResult(0 , 1 ) snake_case_ , snake_case_ = get_distrib(node.left ) snake_case_ , snake_case_ = get_distrib(node.right ) snake_case_ = 1 - left_distrib_excess snake_case_ = 1 - right_distrib_excess snake_case_ = ( left_distrib_moves + right_distrib_moves + abs(a_ ) + abs(a_ ) ) snake_case_ = node.data - coins_to_left - coins_to_right return CoinsDistribResult(a_ , a_ ) return get_distrib(a_ )[0] if __name__ == "__main__": import doctest doctest.testmod()
364
"""simple docstring""" import secrets from random import shuffle from string import ascii_letters, ascii_lowercase, ascii_uppercase, digits, punctuation def _a ( _SCREAMING_SNAKE_CASE = 8 ) -> str: snake_case_ = ascii_letters + digits + punctuation return "".join(secrets.choice(_SCREAMING_SNAKE_CASE ) for _ in range(_SCREAMING_SNAKE_CASE ) ) def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> str: # Password Generator = full boot with random_number, random_letters, and # random_character FUNCTIONS # Put your code here... i -= len(_SCREAMING_SNAKE_CASE ) snake_case_ = i // 3 snake_case_ = i % 3 # chars = chars_incl + random_letters(ascii_letters, i / 3 + remainder) + # random_number(digits, i / 3) + random_characters(punctuation, i / 3) snake_case_ = ( chars_incl + random(_SCREAMING_SNAKE_CASE , quotient + remainder ) + random(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) + random(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) snake_case_ = list(_SCREAMING_SNAKE_CASE ) shuffle(_SCREAMING_SNAKE_CASE ) return "".join(_SCREAMING_SNAKE_CASE ) # random is a generalised function for letters, characters and numbers def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> str: return "".join(secrets.choice(_SCREAMING_SNAKE_CASE ) for _ in range(_SCREAMING_SNAKE_CASE ) ) def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Any: pass # Put your code here... def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Optional[int]: pass # Put your code here... def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[Any]: pass # Put your code here... def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 8 ) -> bool: if len(_SCREAMING_SNAKE_CASE ) < min_length: # Your Password must be at least 8 characters long return False snake_case_ = any(char in ascii_uppercase for char in password ) snake_case_ = any(char in ascii_lowercase for char in password ) snake_case_ = any(char in digits for char in password ) snake_case_ = any(char in punctuation for char in password ) return upper and lower and num and spec_char # Passwords should contain UPPERCASE, lowerase # numbers, and special characters def _a ( ) -> Union[str, Any]: snake_case_ = int(input("""Please indicate the max length of your password: """ ).strip() ) snake_case_ = input( """Please indicate the characters that must be in your password: """ ).strip() print("""Password generated:""" , password_generator(_SCREAMING_SNAKE_CASE ) ) print( """Alternative Password generated:""" , alternative_password_generator(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) , ) print("""[If you are thinking of using this passsword, You better save it.]""" ) if __name__ == "__main__": main()
233
0
'''simple docstring''' import torch from transformers import PreTrainedModel, XLMRobertaConfig, XLMRobertaModel class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' __lowercase : int = '''M-CLIP''' def __init__( self ,__UpperCAmelCase=1024 ,__UpperCAmelCase=768 ,**__UpperCAmelCase ) -> Dict: lowerCAmelCase__ : int = transformerDimSize lowerCAmelCase__ : str = imageDimSize super().__init__(**__UpperCAmelCase ) class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' __lowercase : Tuple = MCLIPConfig def __init__( self ,__UpperCAmelCase ,*__UpperCAmelCase ,**__UpperCAmelCase ) -> Tuple: super().__init__(__UpperCAmelCase ,*__UpperCAmelCase ,**__UpperCAmelCase ) lowerCAmelCase__ : Union[str, Any] = XLMRobertaModel(__UpperCAmelCase ) lowerCAmelCase__ : Tuple = torch.nn.Linear( in_features=config.transformerDimensions ,out_features=config.numDims ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase ) -> Tuple: lowerCAmelCase__ : Optional[int] = self.transformer(input_ids=__UpperCAmelCase ,attention_mask=__UpperCAmelCase )[0] lowerCAmelCase__ : Optional[int] = (embs * attention_mask.unsqueeze(2 )).sum(dim=1 ) / attention_mask.sum(dim=1 )[:, None] return self.LinearTransformation(__UpperCAmelCase ), embs
37
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = { '''microsoft/focalnet-tiny''': '''https://huggingface.co/microsoft/focalnet-tiny/resolve/main/config.json''', } class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' __lowercase : Optional[Any] = '''focalnet''' def __init__( self ,__UpperCAmelCase=224 ,__UpperCAmelCase=4 ,__UpperCAmelCase=3 ,__UpperCAmelCase=96 ,__UpperCAmelCase=False ,__UpperCAmelCase=[192, 384, 768, 768] ,__UpperCAmelCase=[2, 2, 6, 2] ,__UpperCAmelCase=[2, 2, 2, 2] ,__UpperCAmelCase=[3, 3, 3, 3] ,__UpperCAmelCase="gelu" ,__UpperCAmelCase=4.0 ,__UpperCAmelCase=0.0 ,__UpperCAmelCase=0.1 ,__UpperCAmelCase=False ,__UpperCAmelCase=1E-4 ,__UpperCAmelCase=False ,__UpperCAmelCase=False ,__UpperCAmelCase=False ,__UpperCAmelCase=0.0_2 ,__UpperCAmelCase=1E-5 ,__UpperCAmelCase=32 ,__UpperCAmelCase=None ,__UpperCAmelCase=None ,**__UpperCAmelCase ,) -> Optional[Any]: super().__init__(**__UpperCAmelCase ) lowerCAmelCase__ : Dict = image_size lowerCAmelCase__ : int = patch_size lowerCAmelCase__ : str = num_channels lowerCAmelCase__ : Dict = embed_dim lowerCAmelCase__ : List[str] = use_conv_embed lowerCAmelCase__ : List[Any] = hidden_sizes lowerCAmelCase__ : Dict = depths lowerCAmelCase__ : List[str] = focal_levels lowerCAmelCase__ : List[str] = focal_windows lowerCAmelCase__ : Dict = hidden_act lowerCAmelCase__ : Dict = mlp_ratio lowerCAmelCase__ : Tuple = hidden_dropout_prob lowerCAmelCase__ : Tuple = drop_path_rate lowerCAmelCase__ : Dict = use_layerscale lowerCAmelCase__ : Optional[Any] = layerscale_value lowerCAmelCase__ : str = use_post_layernorm lowerCAmelCase__ : Union[str, Any] = use_post_layernorm_in_modulation lowerCAmelCase__ : int = normalize_modulator lowerCAmelCase__ : Optional[Any] = initializer_range lowerCAmelCase__ : List[str] = layer_norm_eps lowerCAmelCase__ : List[Any] = encoder_stride lowerCAmelCase__ : Dict = ["""stem"""] + [F"""stage{idx}""" for idx in range(1 ,len(self.depths ) + 1 )] lowerCAmelCase__ , lowerCAmelCase__ : Any = get_aligned_output_features_output_indices( out_features=__UpperCAmelCase ,out_indices=__UpperCAmelCase ,stage_names=self.stage_names )
37
1
def __lowerCamelCase ( lowerCamelCase__ = 1_000 ): """simple docstring""" lowercase__ : List[Any] = 1, 1 lowercase__ : Union[str, Any] = [] for i in range(1 , n + 1 ): lowercase__ : int = prev_numerator + 2 * prev_denominator lowercase__ : Any = prev_numerator + prev_denominator if len(str(lowerCamelCase_ ) ) > len(str(lowerCamelCase_ ) ): result.append(lowerCamelCase_ ) lowercase__ : Tuple = numerator lowercase__ : Optional[int] = denominator return len(lowerCamelCase_ ) if __name__ == "__main__": print(f'''{solution() = }''')
358
from __future__ import annotations from math import pi # Define the Reduced Planck Constant ℏ (H bar), speed of light C, value of # Pi and the function lowerCAmelCase__ = 1.054571817e-34 # unit of ℏ : J * s lowerCAmelCase__ = 3e8 # unit of c : m * s^-1 def __lowerCamelCase ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" if (force, area, distance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if force < 0: raise ValueError("Magnitude of force can not be negative" ) if distance < 0: raise ValueError("Distance can not be negative" ) if area < 0: raise ValueError("Area can not be negative" ) if force == 0: lowercase__ : Optional[Any] = (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / ( 240 * (distance) ** 4 ) return {"force": force} elif area == 0: lowercase__ : str = (240 * force * (distance) ** 4) / ( REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 ) return {"area": area} elif distance == 0: lowercase__ : Tuple = ( (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / (240 * force) ) ** (1 / 4) return {"distance": distance} raise ValueError("One and only one argument must be 0" ) # Run doctest if __name__ == "__main__": import doctest doctest.testmod()
121
0
from scipy.stats import spearmanr import datasets __lowercase = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' __lowercase = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' __lowercase = r'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCamelCase_ ( datasets.Metric ): '''simple docstring''' def UpperCamelCase__ ( self) -> Optional[Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''float'''), '''references''': datasets.Value('''float'''), }) , reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] , ) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase=False) -> List[str]: __UpperCamelCase :Optional[Any] = spearmanr(__lowercase , __lowercase) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
43
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __lowercase = logging.get_logger(__name__) __lowercase = { '''SenseTime/deformable-detr''': '''https://huggingface.co/sensetime/deformable-detr/resolve/main/config.json''', # See all Deformable DETR models at https://huggingface.co/models?filter=deformable-detr } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : List[str] = """deformable_detr""" a__ : Union[str, Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , __lowercase=True , __lowercase=None , __lowercase=3 , __lowercase=300 , __lowercase=1_024 , __lowercase=6 , __lowercase=1_024 , __lowercase=8 , __lowercase=6 , __lowercase=1_024 , __lowercase=8 , __lowercase=0.0 , __lowercase=True , __lowercase="relu" , __lowercase=256 , __lowercase=0.1 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.02 , __lowercase=1.0 , __lowercase=True , __lowercase=False , __lowercase="sine" , __lowercase="resnet50" , __lowercase=True , __lowercase=False , __lowercase=4 , __lowercase=4 , __lowercase=4 , __lowercase=False , __lowercase=300 , __lowercase=False , __lowercase=1 , __lowercase=5 , __lowercase=2 , __lowercase=1 , __lowercase=1 , __lowercase=5 , __lowercase=2 , __lowercase=0.1 , __lowercase=0.25 , __lowercase=False , **__lowercase , ) -> int: if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''') if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''') __UpperCamelCase :str = CONFIG_MAPPING['''resnet'''](out_features=['''stage4''']) elif isinstance(__lowercase , __lowercase): __UpperCamelCase :str = backbone_config.get('''model_type''') __UpperCamelCase :Tuple = CONFIG_MAPPING[backbone_model_type] __UpperCamelCase :Any = config_class.from_dict(__lowercase) __UpperCamelCase :int = use_timm_backbone __UpperCamelCase :Dict = backbone_config __UpperCamelCase :Any = num_channels __UpperCamelCase :Optional[int] = num_queries __UpperCamelCase :Any = max_position_embeddings __UpperCamelCase :str = d_model __UpperCamelCase :Tuple = encoder_ffn_dim __UpperCamelCase :Union[str, Any] = encoder_layers __UpperCamelCase :List[Any] = encoder_attention_heads __UpperCamelCase :Any = decoder_ffn_dim __UpperCamelCase :List[str] = decoder_layers __UpperCamelCase :int = decoder_attention_heads __UpperCamelCase :str = dropout __UpperCamelCase :Any = attention_dropout __UpperCamelCase :int = activation_dropout __UpperCamelCase :List[Any] = activation_function __UpperCamelCase :List[Any] = init_std __UpperCamelCase :List[Any] = init_xavier_std __UpperCamelCase :int = encoder_layerdrop __UpperCamelCase :str = auxiliary_loss __UpperCamelCase :Optional[Any] = position_embedding_type __UpperCamelCase :Union[str, Any] = backbone __UpperCamelCase :Any = use_pretrained_backbone __UpperCamelCase :str = dilation # deformable attributes __UpperCamelCase :Optional[Any] = num_feature_levels __UpperCamelCase :str = encoder_n_points __UpperCamelCase :int = decoder_n_points __UpperCamelCase :Union[str, Any] = two_stage __UpperCamelCase :Optional[Any] = two_stage_num_proposals __UpperCamelCase :Dict = with_box_refine if two_stage is True and with_box_refine is False: raise ValueError('''If two_stage is True, with_box_refine must be True.''') # Hungarian matcher __UpperCamelCase :Optional[int] = class_cost __UpperCamelCase :List[Any] = bbox_cost __UpperCamelCase :str = giou_cost # Loss coefficients __UpperCamelCase :Tuple = mask_loss_coefficient __UpperCamelCase :Tuple = dice_loss_coefficient __UpperCamelCase :int = bbox_loss_coefficient __UpperCamelCase :Any = giou_loss_coefficient __UpperCamelCase :Dict = eos_coefficient __UpperCamelCase :Optional[Any] = focal_alpha __UpperCamelCase :Optional[Any] = disable_custom_kernels super().__init__(is_encoder_decoder=__lowercase , **__lowercase) @property def UpperCamelCase__ ( self) -> int: return self.encoder_attention_heads @property def UpperCamelCase__ ( self) -> int: return self.d_model def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Dict = copy.deepcopy(self.__dict__) if self.backbone_config is not None: __UpperCamelCase :Tuple = self.backbone_config.to_dict() __UpperCamelCase :List[Any] = self.__class__.model_type return output
43
1
"""simple docstring""" import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse("1.6"): _lowerCamelCase : List[str] = True from torch.cuda.amp import autocast _lowerCamelCase : Any = logging.getLogger(__name__) @dataclass class __UpperCAmelCase : UpperCamelCase = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) UpperCamelCase = field( default=lowerCamelCase__ , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) UpperCamelCase = field( default=lowerCamelCase__ , metadata={"""help""": """Whether to freeze the feature extractor layers of the model."""} ) UpperCamelCase = field( default=lowerCamelCase__ , metadata={"""help""": """Whether to log verbose messages or not."""} , ) UpperCamelCase = field( default=2.0 , metadata={"""help""": """Maximum temperature for gumbel softmax."""} ) UpperCamelCase = field( default=0.5 , metadata={"""help""": """Minimum temperature for gumbel softmax."""} ) UpperCamelCase = field( default=0.9_9_9_9_9_5 , metadata={"""help""": """Decay of gumbel temperature during training."""} ) def a__ ( UpperCAmelCase : ModelArguments , UpperCAmelCase : TrainingArguments ) -> Any: logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) UpperCAmelCase : Any = logging.WARNING if model_args.verbose_logging: UpperCAmelCase : Any = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank ): UpperCAmelCase : Any = logging.INFO logger.setLevel(UpperCAmelCase ) @dataclass class __UpperCAmelCase : UpperCamelCase = field( default=lowerCamelCase__ , metadata={"""help""": """The name of the dataset to use (via the datasets library)."""} ) UpperCamelCase = field( default=lowerCamelCase__ , metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) UpperCamelCase = field( default="""train""" , metadata={ """help""": """The name of the training data set split to use (via the datasets library). Defaults to 'train'""" } , ) UpperCamelCase = field( default="""validation""" , metadata={ """help""": ( """The name of the validation data set split to use (via the datasets library). Defaults to 'validation'""" ) } , ) UpperCamelCase = field( default="""file""" , metadata={"""help""": """Column in the dataset that contains speech file path. Defaults to 'file'"""} , ) UpperCamelCase = field( default=lowerCamelCase__ , metadata={"""help""": """Overwrite the cached preprocessed datasets or not."""} ) UpperCamelCase = field( default=1 , metadata={ """help""": """The percentage of the train set used as validation set in case there's no validation split""" } , ) UpperCamelCase = field( default=lowerCamelCase__ , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) UpperCamelCase = field( default=2_0.0 , metadata={"""help""": """Filter audio files that are longer than `max_duration_in_seconds` seconds"""} ) @dataclass class __UpperCAmelCase : UpperCamelCase = 4_2 UpperCamelCase = 4_2 UpperCamelCase = """longest""" UpperCamelCase = None UpperCamelCase = None def __call__( self : int, __A : List[Dict[str, Union[List[int], torch.Tensor]]] ): # reformat list to dict and set to pytorch format UpperCAmelCase : List[Any] = self.feature_extractor.pad( __A, max_length=self.max_length, padding=self.padding, pad_to_multiple_of=self.pad_to_multiple_of, return_tensors='''pt''', ) UpperCAmelCase : int = self.model._get_feat_extract_output_lengths(batch['''input_values'''].shape[-1] ) UpperCAmelCase : Tuple = batch['''input_values'''].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula UpperCAmelCase : Tuple = self.model._get_feat_extract_output_lengths(batch['''attention_mask'''].sum(-1 ) ).to( torch.long ) UpperCAmelCase : Dict = torch.zeros( (batch_size, mask_indices_seq_length), dtype=torch.long, device=batch['''input_values'''].device ) # these two operations makes sure that all values # before the output lengths indices are attended to UpperCAmelCase : Tuple = 1 UpperCAmelCase : int = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices UpperCAmelCase : Dict = _compute_mask_indices( (batch_size, mask_indices_seq_length), self.model.config.mask_time_prob, self.model.config.mask_time_length, attention_mask=__A, min_masks=2, ) return batch class __UpperCAmelCase ( lowerCamelCase__ ): def __init__( self : Union[str, Any], *__A : int, __A : Dict=1, __A : Any=0, __A : Optional[Any]=1.0, **__A : Any ): super().__init__(*__A, **__A ) UpperCAmelCase : Any = 0 UpperCAmelCase : Any = max_gumbel_temp UpperCAmelCase : Optional[Any] = min_gumbel_temp UpperCAmelCase : str = gumbel_temp_decay def __magic_name__ ( self : Dict, __A : nn.Module, __A : Dict[str, Union[torch.Tensor, Any]] ): model.train() UpperCAmelCase : List[Any] = self._prepare_inputs(__A ) if self.use_amp: with autocast(): UpperCAmelCase : Optional[Any] = self.compute_loss(__A, __A ) else: UpperCAmelCase : Optional[int] = self.compute_loss(__A, __A ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": UpperCAmelCase : Optional[Any] = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": UpperCAmelCase : str = loss.sum() / (inputs['''mask_time_indices''']).sum() else: raise ValueError(F'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: UpperCAmelCase : Any = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(__A ).backward() elif self.use_apex: with amp.scale_loss(__A, self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(__A ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step, self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step, self.min_gumbel_temp ) ) return loss.detach() def a__ ( ) -> Union[str, Any]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. UpperCAmelCase : Optional[Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) UpperCAmelCase : List[Any] = parser.parse_args_into_dataclasses() configure_logger(UpperCAmelCase , UpperCAmelCase ) # Downloading and loading a dataset from the hub. UpperCAmelCase : int = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" UpperCAmelCase : Union[str, Any] = DatasetDict() UpperCAmelCase : Any = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=f'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) UpperCAmelCase : Tuple = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=f'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" UpperCAmelCase : Optional[Any] = DatasetDict() UpperCAmelCase : List[Any] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split='''validation''' , cache_dir=model_args.cache_dir , ) UpperCAmelCase : int = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=f'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported UpperCAmelCase : Optional[Any] = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=UpperCAmelCase ) def prepare_dataset(UpperCAmelCase : Dict ): # check that all files have the correct sampling rate UpperCAmelCase : Optional[Any] = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate ) return batch # load audio files into numpy arrays UpperCAmelCase : str = datasets.map( UpperCAmelCase , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets['''train'''].column_names ) # filter audio files that are too long UpperCAmelCase : int = vectorized_datasets.filter( lambda UpperCAmelCase : len(data['''speech'''] ) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate ) ) def normalize(UpperCAmelCase : Dict ): return feature_extractor(batch['''speech'''] , sampling_rate=feature_extractor.sampling_rate ) # normalize and transform to `BatchFeatures` UpperCAmelCase : Any = vectorized_datasets.map( UpperCAmelCase , batched=UpperCAmelCase , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets['''train'''].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 UpperCAmelCase : Optional[int] = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( '''PreTraining is only supported for ``config.do_stable_layer_norm=True`` and''' ''' ``config.feat_extract_norm=\'layer\'''' ) UpperCAmelCase : Any = WavaVecaForPreTraining(UpperCAmelCase ) UpperCAmelCase : int = DataCollatorForWavaVecaPretraining(model=UpperCAmelCase , feature_extractor=UpperCAmelCase ) UpperCAmelCase : Any = WavaVecaPreTrainer( model=UpperCAmelCase , data_collator=UpperCAmelCase , args=UpperCAmelCase , train_dataset=vectorized_datasets['''train'''] , eval_dataset=vectorized_datasets['''validation'''] , tokenizer=UpperCAmelCase , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
357
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ChineseCLIPImageProcessor class __UpperCAmelCase ( unittest.TestCase ): def __init__( self : List[Any], __A : str, __A : List[str]=7, __A : List[str]=3, __A : Optional[int]=1_8, __A : List[Any]=3_0, __A : Tuple=4_0_0, __A : Tuple=True, __A : List[Any]=None, __A : str=True, __A : int=None, __A : Optional[Any]=True, __A : List[Any]=[0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3], __A : List[str]=[0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1], __A : Tuple=True, ): UpperCAmelCase : int = size if size is not None else {'''height''': 2_2_4, '''width''': 2_2_4} UpperCAmelCase : Optional[Any] = crop_size if crop_size is not None else {'''height''': 1_8, '''width''': 1_8} UpperCAmelCase : int = parent UpperCAmelCase : Union[str, Any] = batch_size UpperCAmelCase : List[Any] = num_channels UpperCAmelCase : Dict = image_size UpperCAmelCase : List[str] = min_resolution UpperCAmelCase : Optional[Any] = max_resolution UpperCAmelCase : Union[str, Any] = do_resize UpperCAmelCase : Dict = size UpperCAmelCase : Any = do_center_crop UpperCAmelCase : Union[str, Any] = crop_size UpperCAmelCase : List[str] = do_normalize UpperCAmelCase : Optional[Any] = image_mean UpperCAmelCase : Optional[Any] = image_std UpperCAmelCase : List[Any] = do_convert_rgb def __magic_name__ ( self : int ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_convert_rgb": self.do_convert_rgb, } def __magic_name__ ( self : Optional[Any], __A : Any=False, __A : str=False, __A : List[Any]=False ): assert not (numpify and torchify), "You cannot specify both numpy and PyTorch tensors at the same time" if equal_resolution: UpperCAmelCase : Dict = [] for i in range(self.batch_size ): image_inputs.append( np.random.randint( 2_5_5, size=(self.num_channels, self.max_resolution, self.max_resolution), dtype=np.uinta ) ) else: UpperCAmelCase : Tuple = [] for i in range(self.batch_size ): UpperCAmelCase , UpperCAmelCase : Tuple = np.random.choice(np.arange(self.min_resolution, self.max_resolution ), 2 ) image_inputs.append(np.random.randint(2_5_5, size=(self.num_channels, width, height), dtype=np.uinta ) ) if not numpify and not torchify: # PIL expects the channel dimension as last dimension UpperCAmelCase : Optional[int] = [Image.fromarray(np.moveaxis(__A, 0, -1 ) ) for x in image_inputs] if torchify: UpperCAmelCase : str = [torch.from_numpy(__A ) for x in image_inputs] return image_inputs @require_torch @require_vision class __UpperCAmelCase ( lowerCamelCase__ , unittest.TestCase ): UpperCamelCase = ChineseCLIPImageProcessor if is_vision_available() else None def __magic_name__ ( self : Dict ): UpperCAmelCase : Optional[Any] = ChineseCLIPImageProcessingTester(self, do_center_crop=__A ) @property def __magic_name__ ( self : Tuple ): return self.image_processor_tester.prepare_image_processor_dict() def __magic_name__ ( self : Any ): UpperCAmelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__A, '''do_resize''' ) ) self.assertTrue(hasattr(__A, '''size''' ) ) self.assertTrue(hasattr(__A, '''do_center_crop''' ) ) self.assertTrue(hasattr(__A, '''center_crop''' ) ) self.assertTrue(hasattr(__A, '''do_normalize''' ) ) self.assertTrue(hasattr(__A, '''image_mean''' ) ) self.assertTrue(hasattr(__A, '''image_std''' ) ) self.assertTrue(hasattr(__A, '''do_convert_rgb''' ) ) def __magic_name__ ( self : Optional[Any] ): UpperCAmelCase : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size, {'''height''': 2_2_4, '''width''': 2_2_4} ) self.assertEqual(image_processor.crop_size, {'''height''': 1_8, '''width''': 1_8} ) UpperCAmelCase : int = self.image_processing_class.from_dict(self.image_processor_dict, size=4_2, crop_size=8_4 ) self.assertEqual(image_processor.size, {'''shortest_edge''': 4_2} ) self.assertEqual(image_processor.crop_size, {'''height''': 8_4, '''width''': 8_4} ) def __magic_name__ ( self : Union[str, Any] ): pass def __magic_name__ ( self : Tuple ): # Initialize image_processing UpperCAmelCase : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase : Optional[int] = self.image_processor_tester.prepare_inputs(equal_resolution=__A ) for image in image_inputs: self.assertIsInstance(__A, Image.Image ) # Test not batched input UpperCAmelCase : str = image_processing(image_inputs[0], return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), ) # Test batched UpperCAmelCase : Any = image_processing(__A, return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), ) def __magic_name__ ( self : Optional[Any] ): # Initialize image_processing UpperCAmelCase : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase : List[Any] = self.image_processor_tester.prepare_inputs(equal_resolution=__A, numpify=__A ) for image in image_inputs: self.assertIsInstance(__A, np.ndarray ) # Test not batched input UpperCAmelCase : List[str] = image_processing(image_inputs[0], return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), ) # Test batched UpperCAmelCase : Dict = image_processing(__A, return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), ) def __magic_name__ ( self : Any ): # Initialize image_processing UpperCAmelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase : Optional[Any] = self.image_processor_tester.prepare_inputs(equal_resolution=__A, torchify=__A ) for image in image_inputs: self.assertIsInstance(__A, torch.Tensor ) # Test not batched input UpperCAmelCase : str = image_processing(image_inputs[0], return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), ) # Test batched UpperCAmelCase : List[str] = image_processing(__A, return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), ) @require_torch @require_vision class __UpperCAmelCase ( lowerCamelCase__ , unittest.TestCase ): UpperCamelCase = ChineseCLIPImageProcessor if is_vision_available() else None def __magic_name__ ( self : Optional[Any] ): UpperCAmelCase : str = ChineseCLIPImageProcessingTester(self, num_channels=4, do_center_crop=__A ) UpperCAmelCase : Dict = 3 @property def __magic_name__ ( self : List[Any] ): return self.image_processor_tester.prepare_image_processor_dict() def __magic_name__ ( self : Dict ): UpperCAmelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__A, '''do_resize''' ) ) self.assertTrue(hasattr(__A, '''size''' ) ) self.assertTrue(hasattr(__A, '''do_center_crop''' ) ) self.assertTrue(hasattr(__A, '''center_crop''' ) ) self.assertTrue(hasattr(__A, '''do_normalize''' ) ) self.assertTrue(hasattr(__A, '''image_mean''' ) ) self.assertTrue(hasattr(__A, '''image_std''' ) ) self.assertTrue(hasattr(__A, '''do_convert_rgb''' ) ) def __magic_name__ ( self : List[Any] ): pass def __magic_name__ ( self : Tuple ): # Initialize image_processing UpperCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase : Dict = self.image_processor_tester.prepare_inputs(equal_resolution=__A ) for image in image_inputs: self.assertIsInstance(__A, Image.Image ) # Test not batched input UpperCAmelCase : Tuple = image_processing(image_inputs[0], return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), ) # Test batched UpperCAmelCase : Tuple = image_processing(__A, return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), )
99
0
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase : int = {"configuration_focalnet": ["FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "FocalNetConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Dict = [ "FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST", "FocalNetForImageClassification", "FocalNetForMaskedImageModeling", "FocalNetBackbone", "FocalNetModel", "FocalNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_focalnet import FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FocalNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_focalnet import ( FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST, FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, FocalNetPreTrainedModel, ) else: import sys UpperCAmelCase : int = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
136
"""simple docstring""" import torch from diffusers import CMStochasticIterativeScheduler from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ): lowercase__ = (CMStochasticIterativeScheduler,) lowercase__ = 10 def _UpperCAmelCase ( self : Optional[int] , **lowerCAmelCase_ : str): """simple docstring""" lowercase_ = { """num_train_timesteps""": 2_0_1, """sigma_min""": 0.002, """sigma_max""": 80.0, } config.update(**lowerCAmelCase_) return config def _UpperCAmelCase ( self : Optional[int]): """simple docstring""" lowercase_ = 1_0 lowercase_ = self.get_scheduler_config() lowercase_ = self.scheduler_classes[0](**lowerCAmelCase_) scheduler.set_timesteps(lowerCAmelCase_) lowercase_ = scheduler.timesteps[0] lowercase_ = scheduler.timesteps[1] lowercase_ = self.dummy_sample lowercase_ = 0.1 * sample lowercase_ = scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_).prev_sample lowercase_ = scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_).prev_sample self.assertEqual(output_a.shape , sample.shape) self.assertEqual(output_a.shape , output_a.shape) def _UpperCAmelCase ( self : Optional[Any]): """simple docstring""" for timesteps in [1_0, 5_0, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=lowerCAmelCase_) def _UpperCAmelCase ( self : Optional[Any]): """simple docstring""" for clip_denoised in [True, False]: self.check_over_configs(clip_denoised=lowerCAmelCase_) def _UpperCAmelCase ( self : List[Any]): """simple docstring""" lowercase_ = self.scheduler_classes[0] lowercase_ = self.get_scheduler_config() lowercase_ = scheduler_class(**lowerCAmelCase_) lowercase_ = 1 scheduler.set_timesteps(lowerCAmelCase_) lowercase_ = scheduler.timesteps lowercase_ = torch.manual_seed(0) lowercase_ = self.dummy_model() lowercase_ = self.dummy_sample_deter * scheduler.init_noise_sigma for i, t in enumerate(lowerCAmelCase_): # 1. scale model input lowercase_ = scheduler.scale_model_input(lowerCAmelCase_ , lowerCAmelCase_) # 2. predict noise residual lowercase_ = model(lowerCAmelCase_ , lowerCAmelCase_) # 3. predict previous sample x_t-1 lowercase_ = scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , generator=lowerCAmelCase_).prev_sample lowercase_ = pred_prev_sample lowercase_ = torch.sum(torch.abs(lowerCAmelCase_)) lowercase_ = torch.mean(torch.abs(lowerCAmelCase_)) assert abs(result_sum.item() - 192.7_614) < 1E-2 assert abs(result_mean.item() - 0.2_510) < 1E-3 def _UpperCAmelCase ( self : List[Any]): """simple docstring""" lowercase_ = self.scheduler_classes[0] lowercase_ = self.get_scheduler_config() lowercase_ = scheduler_class(**lowerCAmelCase_) lowercase_ = [1_0_6, 0] scheduler.set_timesteps(timesteps=lowerCAmelCase_) lowercase_ = scheduler.timesteps lowercase_ = torch.manual_seed(0) lowercase_ = self.dummy_model() lowercase_ = self.dummy_sample_deter * scheduler.init_noise_sigma for t in timesteps: # 1. scale model input lowercase_ = scheduler.scale_model_input(lowerCAmelCase_ , lowerCAmelCase_) # 2. predict noise residual lowercase_ = model(lowerCAmelCase_ , lowerCAmelCase_) # 3. predict previous sample x_t-1 lowercase_ = scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , generator=lowerCAmelCase_).prev_sample lowercase_ = pred_prev_sample lowercase_ = torch.sum(torch.abs(lowerCAmelCase_)) lowercase_ = torch.mean(torch.abs(lowerCAmelCase_)) assert abs(result_sum.item() - 347.6_357) < 1E-2 assert abs(result_mean.item() - 0.4_527) < 1E-3 def _UpperCAmelCase ( self : Union[str, Any]): """simple docstring""" lowercase_ = self.scheduler_classes[0] lowercase_ = self.get_scheduler_config() lowercase_ = scheduler_class(**lowerCAmelCase_) lowercase_ = [3_9, 3_0, 1_2, 1_5, 0] with self.assertRaises(lowerCAmelCase_ , msg="""`timesteps` must be in descending order."""): scheduler.set_timesteps(timesteps=lowerCAmelCase_) def _UpperCAmelCase ( self : Optional[Any]): """simple docstring""" lowercase_ = self.scheduler_classes[0] lowercase_ = self.get_scheduler_config() lowercase_ = scheduler_class(**lowerCAmelCase_) lowercase_ = [3_9, 3_0, 1_2, 1, 0] lowercase_ = len(lowerCAmelCase_) with self.assertRaises(lowerCAmelCase_ , msg="""Can only pass one of `num_inference_steps` or `timesteps`."""): scheduler.set_timesteps(num_inference_steps=lowerCAmelCase_ , timesteps=lowerCAmelCase_) def _UpperCAmelCase ( self : int): """simple docstring""" lowercase_ = self.scheduler_classes[0] lowercase_ = self.get_scheduler_config() lowercase_ = scheduler_class(**lowerCAmelCase_) lowercase_ = [scheduler.config.num_train_timesteps] with self.assertRaises( lowerCAmelCase_ , msg="""`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}""" , ): scheduler.set_timesteps(timesteps=lowerCAmelCase_)
136
1
"""simple docstring""" import glob import os import random from string import ascii_lowercase, digits import cva UpperCamelCase : Optional[Any] = "" UpperCamelCase : int = "" UpperCamelCase : Optional[Any] = "" UpperCamelCase : Any = 1 # (0 is vertical, 1 is horizontal) def A ( ) -> None: __UpperCamelCase , __UpperCamelCase = get_dataset(snake_case , snake_case ) print('Processing...' ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = update_image_and_anno(snake_case , snake_case , snake_case ) for index, image in enumerate(snake_case ): # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __UpperCamelCase = random_chars(3_2 ) __UpperCamelCase = paths[index].split(os.sep )[-1].rsplit('.' , 1 )[0] __UpperCamelCase = f'{OUTPUT_DIR}/{file_name}_FLIP_{letter_code}' cva.imwrite(f'/{file_root}.jpg' , snake_case , [cva.IMWRITE_JPEG_QUALITY, 8_5] ) print(f'Success {index+1}/{len(snake_case )} with {file_name}' ) __UpperCamelCase = [] for anno in new_annos[index]: __UpperCamelCase = f'{anno[0]} {anno[1]} {anno[2]} {anno[3]} {anno[4]}' annos_list.append(snake_case ) with open(f'/{file_root}.txt' , 'w' ) as outfile: outfile.write('\n'.join(line for line in annos_list ) ) def A ( snake_case :str , snake_case :str ) -> tuple[list, list]: __UpperCamelCase = [] __UpperCamelCase = [] for label_file in glob.glob(os.path.join(snake_case , '*.txt' ) ): __UpperCamelCase = label_file.split(os.sep )[-1].rsplit('.' , 1 )[0] with open(snake_case ) as in_file: __UpperCamelCase = in_file.readlines() __UpperCamelCase = os.path.join(snake_case , f'{label_name}.jpg' ) __UpperCamelCase = [] for obj_list in obj_lists: __UpperCamelCase = obj_list.rstrip('\n' ).split(' ' ) boxes.append( [ int(obj[0] ), float(obj[1] ), float(obj[2] ), float(obj[3] ), float(obj[4] ), ] ) if not boxes: continue img_paths.append(snake_case ) labels.append(snake_case ) return img_paths, labels def A ( snake_case :list , snake_case :list , snake_case :int = 1 ) -> tuple[list, list, list]: __UpperCamelCase = [] __UpperCamelCase = [] __UpperCamelCase = [] for idx in range(len(snake_case ) ): __UpperCamelCase = [] __UpperCamelCase = img_list[idx] path_list.append(snake_case ) __UpperCamelCase = anno_list[idx] __UpperCamelCase = cva.imread(snake_case ) if flip_type == 1: __UpperCamelCase = cva.flip(snake_case , snake_case ) for bbox in img_annos: __UpperCamelCase = 1 - bbox[1] new_annos.append([bbox[0], x_center_new, bbox[2], bbox[3], bbox[4]] ) elif flip_type == 0: __UpperCamelCase = cva.flip(snake_case , snake_case ) for bbox in img_annos: __UpperCamelCase = 1 - bbox[2] new_annos.append([bbox[0], bbox[1], y_center_new, bbox[3], bbox[4]] ) new_annos_lists.append(snake_case ) new_imgs_list.append(snake_case ) return new_imgs_list, new_annos_lists, path_list def A ( snake_case :int = 3_2 ) -> str: assert number_char > 1, "The number of character should greater than 1" __UpperCamelCase = ascii_lowercase + digits return "".join(random.choice(snake_case ) for _ in range(snake_case ) ) if __name__ == "__main__": main() print("DONE ✅")
263
"""simple docstring""" import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = None , __UpperCAmelCase = None , **__UpperCAmelCase , ): '''simple docstring''' super().__init__( __UpperCAmelCase , split=__UpperCAmelCase , features=__UpperCAmelCase , cache_dir=__UpperCAmelCase , keep_in_memory=__UpperCAmelCase , streaming=__UpperCAmelCase , num_proc=__UpperCAmelCase , **__UpperCAmelCase , ) __UpperCamelCase = field __UpperCamelCase = path_or_paths if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else {self.split: path_or_paths} __UpperCamelCase = Json( cache_dir=__UpperCAmelCase , data_files=__UpperCAmelCase , features=__UpperCAmelCase , field=__UpperCAmelCase , **__UpperCAmelCase , ) def UpperCAmelCase ( self ): '''simple docstring''' if self.streaming: __UpperCamelCase = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None __UpperCamelCase = None self.builder.download_and_prepare( download_config=__UpperCAmelCase , download_mode=__UpperCAmelCase , verification_mode=__UpperCAmelCase , base_path=__UpperCAmelCase , num_proc=self.num_proc , ) __UpperCamelCase = self.builder.as_dataset( split=self.split , verification_mode=__UpperCAmelCase , in_memory=self.keep_in_memory ) return dataset class __lowerCAmelCase : def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , **__UpperCAmelCase , ): '''simple docstring''' if num_proc is not None and num_proc <= 0: raise ValueError(F'num_proc {num_proc} must be an integer > 0.' ) __UpperCamelCase = dataset __UpperCamelCase = path_or_buf __UpperCamelCase = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE __UpperCamelCase = num_proc __UpperCamelCase = 'utf-8' __UpperCamelCase = to_json_kwargs def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = self.to_json_kwargs.pop('path_or_buf' , __UpperCAmelCase ) __UpperCamelCase = self.to_json_kwargs.pop('orient' , 'records' ) __UpperCamelCase = self.to_json_kwargs.pop('lines' , True if orient == 'records' else False ) __UpperCamelCase = self.to_json_kwargs.pop('index' , False if orient in ['split', 'table'] else True ) __UpperCamelCase = self.to_json_kwargs.pop('compression' , __UpperCAmelCase ) if compression not in [None, "infer", "gzip", "bz2", "xz"]: raise NotImplementedError(F'`datasets` currently does not support {compression} compression' ) if isinstance(self.path_or_buf , (str, bytes, os.PathLike) ): with fsspec.open(self.path_or_buf , 'wb' , compression=__UpperCAmelCase ) as buffer: __UpperCamelCase = self._write(file_obj=__UpperCAmelCase , orient=__UpperCAmelCase , lines=__UpperCAmelCase , index=__UpperCAmelCase , **self.to_json_kwargs ) else: if compression: raise NotImplementedError( F'The compression parameter is not supported when writing to a buffer, but compression={compression}' ' was passed. Please provide a local path instead.' ) __UpperCamelCase = self._write( file_obj=self.path_or_buf , orient=__UpperCAmelCase , lines=__UpperCAmelCase , index=__UpperCAmelCase , **self.to_json_kwargs ) return written def UpperCAmelCase ( self , __UpperCAmelCase ): '''simple docstring''' __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = args __UpperCamelCase = query_table( table=self.dataset.data , key=slice(__UpperCAmelCase , offset + self.batch_size ) , indices=self.dataset._indices , ) __UpperCamelCase = batch.to_pandas().to_json( path_or_buf=__UpperCAmelCase , orient=__UpperCAmelCase , lines=__UpperCAmelCase , index=__UpperCAmelCase , **__UpperCAmelCase ) if not json_str.endswith('\n' ): json_str += "\n" return json_str.encode(self.encoding ) def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ): '''simple docstring''' __UpperCamelCase = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit='ba' , disable=not logging.is_progress_bar_enabled() , desc='Creating json from Arrow format' , ): __UpperCamelCase = self._batch_json((offset, orient, lines, index, to_json_kwargs) ) written += file_obj.write(__UpperCAmelCase ) else: __UpperCamelCase , __UpperCamelCase = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for json_str in logging.tqdm( pool.imap( self._batch_json , [(offset, orient, lines, index, to_json_kwargs) for offset in range(0 , __UpperCAmelCase , __UpperCAmelCase )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit='ba' , disable=not logging.is_progress_bar_enabled() , desc='Creating json from Arrow format' , ): written += file_obj.write(__UpperCAmelCase ) return written
263
1
from collections import defaultdict from pathlib import Path import pandas as pd from rouge_cli import calculate_rouge_path from utils import calculate_rouge a__: str = [ 'Prosecutor: "No videos were used in the crash investigation" German papers say they saw a cell phone video of the' ' final seconds on board Flight 9525. The Germanwings co-pilot says he had a "previous episode of severe' ' depression\" German airline confirms it knew of Andreas Lubitz\'s depression years before he took control.', 'The Palestinian Authority officially becomes the 123rd member of the International Criminal Court. The formal' ' accession was marked with a ceremony at The Hague, in the Netherlands. The Palestinians signed the ICC\'s' ' founding Rome Statute in January. Israel and the United States opposed the Palestinians\' efforts to join the' ' body.', 'Amnesty International releases its annual report on the death penalty. The report catalogs the use of' ' state-sanctioned killing as a punitive measure across the globe. At least 607 people were executed around the' ' world in 2014, compared to 778 in 2013. The U.S. remains one of the worst offenders for imposing capital' ' punishment.', ] a__: Tuple = [ 'Marseille prosecutor says "so far no videos were used in the crash investigation" despite media reports .' ' Journalists at Bild and Paris Match are "very confident" the video clip is real, an editor says . Andreas Lubitz' ' had informed his Lufthansa training school of an episode of severe depression, airline says .', 'Membership gives the ICC jurisdiction over alleged crimes committed in Palestinian territories since last June .' ' Israel and the United States opposed the move, which could open the door to war crimes investigations against' ' Israelis .', 'Amnesty\'s annual death penalty report catalogs encouraging signs, but setbacks in numbers of those sentenced to' ' death . Organization claims that governments around the world are using the threat of terrorism to advance' ' executions . The number of executions worldwide has gone down by almost 22% compared with 2013, but death' ' sentences up by 28% .', ] def UpperCamelCase__( )->Optional[int]: A__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , bootstrap_aggregation=UpperCamelCase__ , rouge_keys=['''rouge2''', '''rougeL'''] ) assert isinstance(UpperCamelCase__ , UpperCamelCase__ ) A__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , bootstrap_aggregation=UpperCamelCase__ , rouge_keys=['''rouge2'''] ) assert ( pd.DataFrame(no_aggregation['''rouge2'''] ).fmeasure.mean() == pd.DataFrame(no_aggregation_just_ra['''rouge2'''] ).fmeasure.mean() ) def UpperCamelCase__( )->Dict: A__ = '''rougeLsum''' A__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , newline_sep=UpperCamelCase__ , rouge_keys=[k] )[k] A__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , newline_sep=UpperCamelCase__ , rouge_keys=[k] )[k] assert score > score_no_sep def UpperCamelCase__( )->Any: A__ = ['''rouge1''', '''rouge2''', '''rougeL'''] A__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , newline_sep=UpperCamelCase__ , rouge_keys=UpperCamelCase__ ) A__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , newline_sep=UpperCamelCase__ , rouge_keys=UpperCamelCase__ ) assert score_sep == score_no_sep def UpperCamelCase__( )->int: A__ = [ '''Her older sister, Margot Frank, died in 1945, a month earlier than previously thought.''', '''Marseille prosecutor says "so far no videos were used in the crash investigation" despite media reports .''', ] A__ = [ '''Margot Frank, died in 1945, a month earlier than previously thought.''', '''Prosecutor: "No videos were used in the crash investigation" German papers say they saw a cell phone video of''' ''' the final seconds on board Flight 9525.''', ] assert calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , newline_sep=UpperCamelCase__ ) == calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , newline_sep=UpperCamelCase__ ) def UpperCamelCase__( )->Dict: A__ = [ '''" "a person who has such a video needs to immediately give it to the investigators," prosecutor says .<n> "it is a very disturbing scene," editor-in-chief of bild online tells "erin burnett: outfront" ''' ] A__ = [ ''' Marseille prosecutor says "so far no videos were used in the crash investigation" despite media reports . Journalists at Bild and Paris Match are "very confident" the video clip is real, an editor says . Andreas Lubitz had informed his Lufthansa training school of an episode of severe depression, airline says .''' ] A__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , rouge_keys=['''rougeLsum'''] , newline_sep=UpperCamelCase__ )['''rougeLsum'''] A__ = calculate_rouge(UpperCamelCase__ , UpperCamelCase__ , rouge_keys=['''rougeLsum'''] )['''rougeLsum'''] assert new_score > prev_score def UpperCamelCase__( )->Any: A__ = Path('''examples/seq2seq/test_data/wmt_en_ro''' ) A__ = calculate_rouge_path(data_dir.joinpath('''test.source''' ) , data_dir.joinpath('''test.target''' ) ) assert isinstance(UpperCamelCase__ , UpperCamelCase__ ) A__ = calculate_rouge_path( data_dir.joinpath('''test.source''' ) , data_dir.joinpath('''test.target''' ) , bootstrap_aggregation=UpperCamelCase__ ) assert isinstance(UpperCamelCase__ , UpperCamelCase__ )
193
import torch from diffusers import DDIMParallelScheduler from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE__ ( UpperCamelCase__ ): __SCREAMING_SNAKE_CASE = (DDIMParallelScheduler,) __SCREAMING_SNAKE_CASE = (('''eta''', 0.0), ('''num_inference_steps''', 50)) def UpperCamelCase ( self,**__lowerCamelCase ): A__ = { '''num_train_timesteps''': 1000, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''clip_sample''': True, } config.update(**__lowerCamelCase ) return config def UpperCamelCase ( self,**__lowerCamelCase ): A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config(**__lowerCamelCase ) A__ = scheduler_class(**__lowerCamelCase ) A__ , A__ = 10, 0.0 A__ = self.dummy_model() A__ = self.dummy_sample_deter scheduler.set_timesteps(__lowerCamelCase ) for t in scheduler.timesteps: A__ = model(__lowerCamelCase,__lowerCamelCase ) A__ = scheduler.step(__lowerCamelCase,__lowerCamelCase,__lowerCamelCase,__lowerCamelCase ).prev_sample return sample def UpperCamelCase ( self ): for timesteps in [100, 500, 1000]: self.check_over_configs(num_train_timesteps=__lowerCamelCase ) def UpperCamelCase ( self ): for steps_offset in [0, 1]: self.check_over_configs(steps_offset=__lowerCamelCase ) A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config(steps_offset=1 ) A__ = scheduler_class(**__lowerCamelCase ) scheduler.set_timesteps(5 ) assert torch.equal(scheduler.timesteps,torch.LongTensor([801, 601, 401, 201, 1] ) ) def UpperCamelCase ( self ): for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1],[0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=__lowerCamelCase,beta_end=__lowerCamelCase ) def UpperCamelCase ( self ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=__lowerCamelCase ) def UpperCamelCase ( self ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=__lowerCamelCase ) def UpperCamelCase ( self ): for clip_sample in [True, False]: self.check_over_configs(clip_sample=__lowerCamelCase ) def UpperCamelCase ( self ): for timestep_spacing in ["trailing", "leading"]: self.check_over_configs(timestep_spacing=__lowerCamelCase ) def UpperCamelCase ( self ): for rescale_betas_zero_snr in [True, False]: self.check_over_configs(rescale_betas_zero_snr=__lowerCamelCase ) def UpperCamelCase ( self ): self.check_over_configs(thresholding=__lowerCamelCase ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs( thresholding=__lowerCamelCase,prediction_type=__lowerCamelCase,sample_max_value=__lowerCamelCase,) def UpperCamelCase ( self ): for t in [1, 10, 49]: self.check_over_forward(time_step=__lowerCamelCase ) def UpperCamelCase ( self ): for t, num_inference_steps in zip([1, 10, 50],[10, 50, 500] ): self.check_over_forward(time_step=__lowerCamelCase,num_inference_steps=__lowerCamelCase ) def UpperCamelCase ( self ): for t, eta in zip([1, 10, 49],[0.0, 0.5, 1.0] ): self.check_over_forward(time_step=__lowerCamelCase,eta=__lowerCamelCase ) def UpperCamelCase ( self ): A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config() A__ = scheduler_class(**__lowerCamelCase ) assert torch.sum(torch.abs(scheduler._get_variance(0,0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(420,400 ) - 0.14771 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(980,960 ) - 0.32460 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(0,0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487,486 ) - 0.00979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999,998 ) - 0.02 ) ) < 1E-5 def UpperCamelCase ( self ): A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config() A__ = scheduler_class(**__lowerCamelCase ) A__ , A__ = 10, 0.0 scheduler.set_timesteps(__lowerCamelCase ) A__ = self.dummy_model() A__ = self.dummy_sample_deter A__ = self.dummy_sample_deter + 0.1 A__ = self.dummy_sample_deter - 0.1 A__ = samplea.shape[0] A__ = torch.stack([samplea, samplea, samplea],dim=0 ) A__ = torch.arange(__lowerCamelCase )[0:3, None].repeat(1,__lowerCamelCase ) A__ = model(samples.flatten(0,1 ),timesteps.flatten(0,1 ) ) A__ = scheduler.batch_step_no_noise(__lowerCamelCase,timesteps.flatten(0,1 ),samples.flatten(0,1 ),__lowerCamelCase ) A__ = torch.sum(torch.abs(__lowerCamelCase ) ) A__ = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 1147.7904 ) < 1E-2 assert abs(result_mean.item() - 0.4982 ) < 1E-3 def UpperCamelCase ( self ): A__ = self.full_loop() A__ = torch.sum(torch.abs(__lowerCamelCase ) ) A__ = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 172.0067 ) < 1E-2 assert abs(result_mean.item() - 0.223967 ) < 1E-3 def UpperCamelCase ( self ): A__ = self.full_loop(prediction_type='''v_prediction''' ) A__ = torch.sum(torch.abs(__lowerCamelCase ) ) A__ = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 52.5302 ) < 1E-2 assert abs(result_mean.item() - 0.0684 ) < 1E-3 def UpperCamelCase ( self ): # We specify different beta, so that the first alpha is 0.99 A__ = self.full_loop(set_alpha_to_one=__lowerCamelCase,beta_start=0.01 ) A__ = torch.sum(torch.abs(__lowerCamelCase ) ) A__ = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 149.8295 ) < 1E-2 assert abs(result_mean.item() - 0.1951 ) < 1E-3 def UpperCamelCase ( self ): # We specify different beta, so that the first alpha is 0.99 A__ = self.full_loop(set_alpha_to_one=__lowerCamelCase,beta_start=0.01 ) A__ = torch.sum(torch.abs(__lowerCamelCase ) ) A__ = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 149.0784 ) < 1E-2 assert abs(result_mean.item() - 0.1941 ) < 1E-3
193
1
from math import pow def __lowerCamelCase (UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int , ): if current_sum == needed_sum: # If the sum of the powers is equal to needed_sum, then we have a solution. solutions_count += 1 return current_sum, solutions_count SCREAMING_SNAKE_CASE = int(pow(UpperCAmelCase__ , UpperCAmelCase__ ) ) if current_sum + i_to_n <= needed_sum: # If the sum of the powers is less than needed_sum, then continue adding powers. current_sum += i_to_n SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = backtrack( UpperCAmelCase__ , UpperCAmelCase__ , current_number + 1 , UpperCAmelCase__ , UpperCAmelCase__ ) current_sum -= i_to_n if i_to_n < needed_sum: # If the power of i is less than needed_sum, then try with the next power. SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = backtrack( UpperCAmelCase__ , UpperCAmelCase__ , current_number + 1 , UpperCAmelCase__ , UpperCAmelCase__ ) return current_sum, solutions_count def __lowerCamelCase (UpperCAmelCase__ : int , UpperCAmelCase__ : int ): if not (1 <= needed_sum <= 1_0_0_0 and 2 <= power <= 1_0): raise ValueError( "Invalid input\n" "needed_sum must be between 1 and 1000, power between 2 and 10." ) return backtrack(UpperCAmelCase__ , UpperCAmelCase__ , 1 , 0 , 0 )[1] # Return the solutions_count if __name__ == "__main__": import doctest doctest.testmod()
206
import contextlib from multiprocessing import Pool, RLock from tqdm.auto import tqdm from ..utils import experimental, logging _lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) class lowercase : lowercase__ : str = None @experimental def __lowerCamelCase (UpperCAmelCase__ : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Any ): if ParallelBackendConfig.backend_name is None: return _map_with_multiprocessing_pool( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) return _map_with_joblib(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) def __lowerCamelCase (UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[Any] ): SCREAMING_SNAKE_CASE = num_proc if num_proc <= len(UpperCAmelCase__ ) else len(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = [] # We organize the splits ourselve (contiguous splits) for index in range(UpperCAmelCase__ ): SCREAMING_SNAKE_CASE = len(UpperCAmelCase__ ) // num_proc SCREAMING_SNAKE_CASE = len(UpperCAmelCase__ ) % num_proc SCREAMING_SNAKE_CASE = div * index + min(UpperCAmelCase__ , UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = start + div + (1 if index < mod else 0) split_kwds.append((function, iterable[start:end], types, index, disable_tqdm, desc) ) if len(UpperCAmelCase__ ) != sum(len(i[1] ) for i in split_kwds ): raise ValueError( F"Error dividing inputs iterable among processes. " F"Total number of objects {len(UpperCAmelCase__ )}, " F"length: {sum(len(i[1] ) for i in split_kwds )}" ) logger.info( F"Spawning {num_proc} processes for {len(UpperCAmelCase__ )} objects in slices of {[len(i[1] ) for i in split_kwds]}" ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None, None if not disable_tqdm: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = (RLock(),), tqdm.set_lock with Pool(UpperCAmelCase__ , initargs=UpperCAmelCase__ , initializer=UpperCAmelCase__ ) as pool: SCREAMING_SNAKE_CASE = pool.map(UpperCAmelCase__ , UpperCAmelCase__ ) logger.info(F"Finished {num_proc} processes" ) SCREAMING_SNAKE_CASE = [obj for proc_res in mapped for obj in proc_res] logger.info(F"Unpacked {len(UpperCAmelCase__ )} objects" ) return mapped def __lowerCamelCase (UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Any , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Any ): # progress bar is not yet supported for _map_with_joblib, because tqdm couldn't accurately be applied to joblib, # and it requires monkey-patching joblib internal classes which is subject to change import joblib with joblib.parallel_backend(ParallelBackendConfig.backend_name , n_jobs=UpperCAmelCase__ ): return joblib.Parallel()( joblib.delayed(UpperCAmelCase__ )((function, obj, types, None, True, None) ) for obj in iterable ) @experimental @contextlib.contextmanager def __lowerCamelCase (UpperCAmelCase__ : str ): SCREAMING_SNAKE_CASE = backend_name if backend_name == "spark": from joblibspark import register_spark register_spark() # TODO: call create_cache_and_write_probe if "download" in steps # TODO: raise NotImplementedError when Dataset.map etc is called try: yield finally: SCREAMING_SNAKE_CASE = None
206
1
"""simple docstring""" import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class __A ( unittest.TestCase ): def __A ( self ): _lowerCAmelCase : int = 10 def __A ( self ): _lowerCAmelCase : str = [1, 2, 3, 4] _lowerCAmelCase : int = [1, 2, 3, 4, 0, 0, 0, 0, 0, 0] self.assertEqual(truncate_or_pad(a__ , self.block_size , 0 ) , a__ ) def __A ( self ): _lowerCAmelCase : int = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] _lowerCAmelCase : Any = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(a__ , self.block_size , 0 ) , a__ ) def __A ( self ): _lowerCAmelCase : str = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13] _lowerCAmelCase : Optional[Any] = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(a__ , self.block_size , 0 ) , a__ ) def __A ( self ): _lowerCAmelCase : Union[str, Any] = """It was the year of Our Lord one thousand seven hundred and seventy-five.\n\nSpiritual revelations were conceded to England at that favoured period, as at this.""" _lowerCAmelCase , _lowerCAmelCase : Optional[int] = process_story(a__ ) self.assertEqual(a__ , [] ) def __A ( self ): _lowerCAmelCase : List[Any] = """""" _lowerCAmelCase , _lowerCAmelCase : Union[str, Any] = process_story(a__ ) self.assertEqual(a__ , [] ) self.assertEqual(a__ , [] ) def __A ( self ): _lowerCAmelCase : str = ( """It was the year of Our Lord one thousand seven hundred and """ """seventy-five\n\nSpiritual revelations were conceded to England """ """at that favoured period, as at this.\n@highlight\n\nIt was the best of times""" ) _lowerCAmelCase , _lowerCAmelCase : List[str] = process_story(a__ ) _lowerCAmelCase : Union[str, Any] = [ """It was the year of Our Lord one thousand seven hundred and seventy-five.""", """Spiritual revelations were conceded to England at that favoured period, as at this.""", ] self.assertEqual(a__ , a__ ) _lowerCAmelCase : List[str] = ["""It was the best of times."""] self.assertEqual(a__ , a__ ) def __A ( self ): _lowerCAmelCase : Any = torch.tensor([1, 2, 3, 4] ) _lowerCAmelCase : Union[str, Any] = torch.tensor([1, 1, 1, 1] ) np.testing.assert_array_equal(build_mask(a__ , 0 ).numpy() , expected.numpy() ) def __A ( self ): _lowerCAmelCase : Tuple = torch.tensor([1, 2, 3, 4, 23, 23, 23] ) _lowerCAmelCase : Any = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(a__ , 23 ).numpy() , expected.numpy() ) def __A ( self ): _lowerCAmelCase : Tuple = torch.tensor([8, 2, 3, 4, 1, 1, 1] ) _lowerCAmelCase : Dict = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(a__ , 1 ).numpy() , expected.numpy() ) def __A ( self ): _lowerCAmelCase : int = 101 _lowerCAmelCase : Tuple = torch.tensor([[1, 2, 3, 4, 5, 6], [1, 2, 3, 101, 5, 6], [1, 101, 3, 4, 101, 6]] ) _lowerCAmelCase : Dict = torch.tensor([[1, 1, 1, 1, 1, 1], [1, 1, 1, 0, 0, 0], [1, 0, 0, 0, 1, 1]] ) _lowerCAmelCase : int = compute_token_type_ids(a__ , a__ ) np.testing.assert_array_equal(a__ , a__ )
44
'''simple docstring''' from __future__ import annotations def __snake_case ( UpperCAmelCase_ : int ): lowerCamelCase_ = 2 lowerCamelCase_ = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(UpperCAmelCase_ ) if n > 1: factors.append(UpperCAmelCase_ ) return factors if __name__ == "__main__": import doctest doctest.testmod()
55
0
import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class lowerCamelCase : def __init__(self : Tuple , _A : List[Any] , _A : Tuple=1_3 , _A : Any=3_0 , _A : Any=2 , _A : List[str]=3 , _A : Any=True , _A : Union[str, Any]=True , _A : Any=3_2 , _A : Dict=5 , _A : Optional[int]=4 , _A : List[Any]=3_7 , _A : Any="gelu" , _A : Tuple=0.1 , _A : str=0.1 , _A : int=1_0 , _A : str=0.02 , _A : Tuple=3 , _A : Tuple=None , _A : str=2 , ) -> str: snake_case = parent snake_case = batch_size snake_case = image_size snake_case = patch_size snake_case = num_channels snake_case = is_training snake_case = use_labels snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = intermediate_size snake_case = hidden_act snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = type_sequence_label_size snake_case = initializer_range snake_case = scope snake_case = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) snake_case = (image_size // patch_size) ** 2 snake_case = num_patches + 2 def UpperCAmelCase(self : List[str] ) -> List[Any]: snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case = None if self.use_labels: snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case = self.get_config() return config, pixel_values, labels def UpperCAmelCase(self : Optional[Any] ) -> Optional[int]: return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=a__ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def UpperCAmelCase(self : List[str] , _A : str , _A : str , _A : int ) -> Optional[int]: snake_case = DeiTModel(config=a__ ) model.to(a__ ) model.eval() snake_case = model(a__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase(self : str , _A : Union[str, Any] , _A : Optional[Any] , _A : Union[str, Any] ) -> List[str]: snake_case = DeiTForMaskedImageModeling(config=a__ ) model.to(a__ ) model.eval() snake_case = model(a__ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images snake_case = 1 snake_case = DeiTForMaskedImageModeling(a__ ) model.to(a__ ) model.eval() snake_case = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case = model(a__ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def UpperCAmelCase(self : Dict , _A : List[str] , _A : str , _A : int ) -> List[str]: snake_case = self.type_sequence_label_size snake_case = DeiTForImageClassification(a__ ) model.to(a__ ) model.eval() snake_case = model(a__ , labels=a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images snake_case = 1 snake_case = DeiTForImageClassification(a__ ) model.to(a__ ) model.eval() snake_case = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case = model(a__ , labels=a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCAmelCase(self : List[Any] ) -> List[str]: snake_case = self.prepare_config_and_inputs() ( ( snake_case ) , ( snake_case ) , ( snake_case ) , ) = config_and_inputs snake_case = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowerCamelCase ( __a , __a , unittest.TestCase ): UpperCAmelCase__ : Optional[Any] = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) UpperCAmelCase__ : Optional[Any] = ( { """feature-extraction""": DeiTModel, """image-classification""": (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) UpperCAmelCase__ : Tuple = False UpperCAmelCase__ : str = False UpperCAmelCase__ : Union[str, Any] = False def UpperCAmelCase(self : Optional[Any] ) -> Union[str, Any]: snake_case = DeiTModelTester(self ) snake_case = ConfigTester(self , config_class=a__ , has_text_modality=a__ , hidden_size=3_7 ) def UpperCAmelCase(self : Optional[int] ) -> Union[str, Any]: self.config_tester.run_common_tests() @unittest.skip(reason="DeiT does not use inputs_embeds" ) def UpperCAmelCase(self : List[Any] ) -> Optional[Any]: pass def UpperCAmelCase(self : Union[str, Any] ) -> Union[str, Any]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(a__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a__ , nn.Linear ) ) def UpperCAmelCase(self : Dict ) -> Dict: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(a__ ) snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case = [*signature.parameters.keys()] snake_case = ["pixel_values"] self.assertListEqual(arg_names[:1] , a__ ) def UpperCAmelCase(self : Dict ) -> List[str]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a__ ) def UpperCAmelCase(self : Dict ) -> Optional[int]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*a__ ) def UpperCAmelCase(self : List[Any] ) -> Tuple: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a__ ) def UpperCAmelCase(self : int , _A : Tuple , _A : Tuple , _A : Dict=False ) -> Union[str, Any]: snake_case = super()._prepare_for_class(a__ , a__ , return_labels=a__ ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def UpperCAmelCase(self : Dict ) -> Any: if not self.model_tester.is_training: return snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() snake_case = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(a__ ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue snake_case = model_class(a__ ) model.to(a__ ) model.train() snake_case = self._prepare_for_class(a__ , a__ , return_labels=a__ ) snake_case = model(**a__ ).loss loss.backward() def UpperCAmelCase(self : Optional[Any] ) -> Optional[int]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return snake_case = False snake_case = True for model_class in self.all_model_classes: if model_class in get_values(a__ ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue snake_case = model_class(a__ ) model.gradient_checkpointing_enable() model.to(a__ ) model.train() snake_case = self._prepare_for_class(a__ , a__ , return_labels=a__ ) snake_case = model(**a__ ).loss loss.backward() def UpperCAmelCase(self : Any ) -> List[Any]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() snake_case = [ {"title": "multi_label_classification", "num_labels": 2, "dtype": torch.float}, {"title": "single_label_classification", "num_labels": 1, "dtype": torch.long}, {"title": "regression", "num_labels": 1, "dtype": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(a__ ), *get_values(a__ ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=f'Testing {model_class} with {problem_type["title"]}' ): snake_case = problem_type["title"] snake_case = problem_type["num_labels"] snake_case = model_class(a__ ) model.to(a__ ) model.train() snake_case = self._prepare_for_class(a__ , a__ , return_labels=a__ ) if problem_type["num_labels"] > 1: snake_case = inputs["labels"].unsqueeze(1 ).repeat(1 , problem_type["num_labels"] ) snake_case = inputs["labels"].to(problem_type["dtype"] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=a__ ) as warning_list: snake_case = model(**a__ ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( f'Something is going wrong in the regression problem: intercepted {w.message}' ) loss.backward() @slow def UpperCAmelCase(self : Optional[Any] ) -> Tuple: for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case = DeiTModel.from_pretrained(a__ ) self.assertIsNotNone(a__ ) def lowercase_ ( ) -> Optional[int]: """simple docstring""" snake_case = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class lowerCamelCase ( unittest.TestCase ): @cached_property def UpperCAmelCase(self : List[Any] ) -> List[Any]: return ( DeiTImageProcessor.from_pretrained("facebook/deit-base-distilled-patch16-224" ) if is_vision_available() else None ) @slow def UpperCAmelCase(self : Optional[Any] ) -> Dict: snake_case = DeiTForImageClassificationWithTeacher.from_pretrained("facebook/deit-base-distilled-patch16-224" ).to( a__ ) snake_case = self.default_image_processor snake_case = prepare_img() snake_case = image_processor(images=a__ , return_tensors="pt" ).to(a__ ) # forward pass with torch.no_grad(): snake_case = model(**a__ ) # verify the logits snake_case = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , a__ ) snake_case = torch.tensor([-1.02_66, 0.19_12, -1.28_61] ).to(a__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a__ , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def UpperCAmelCase(self : Union[str, Any] ) -> Optional[Any]: snake_case = DeiTModel.from_pretrained( "facebook/deit-base-distilled-patch16-224" , torch_dtype=torch.floataa , device_map="auto" ) snake_case = self.default_image_processor snake_case = prepare_img() snake_case = image_processor(images=a__ , return_tensors="pt" ) snake_case = inputs.pixel_values.to(a__ ) # forward pass to make sure inference works in fp16 with torch.no_grad(): snake_case = model(a__ )
360
def lowercase_ ( A__ = 1000 ) -> int: """simple docstring""" return sum(2 * a * ((a - 1) // 2) for a in range(3 , n + 1 ) ) if __name__ == "__main__": print(solution())
137
0
'''simple docstring''' def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> Optional[int]: stooge(UpperCamelCase , 0 , len(UpperCamelCase ) - 1 ) return arr def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Any: if i >= h: return # If first element is smaller than the last then swap them if arr[i] > arr[h]: lowerCamelCase__ , lowerCamelCase__ : List[Any] = arr[h], arr[i] # If there are more than 2 elements in the array if h - i + 1 > 2: lowerCamelCase__ : Optional[int] = (int)((h - i + 1) / 3 ) # Recursively sort first 2/3 elements stooge(UpperCamelCase , UpperCamelCase , (h - t) ) # Recursively sort last 2/3 elements stooge(UpperCamelCase , i + t , (UpperCamelCase) ) # Recursively sort first 2/3 elements stooge(UpperCamelCase , UpperCamelCase , (h - t) ) if __name__ == "__main__": _A : Union[str, Any] =input('''Enter numbers separated by a comma:\n''').strip() _A : int =[int(item) for item in user_input.split(''',''')] print(stooge_sort(unsorted))
41
'''simple docstring''' from queue import PriorityQueue from typing import Any import numpy as np def snake_case_ (_a : dict , _a : str , _a : set , _a : set , _a : dict , _a : dict , _a : PriorityQueue , _a : dict , _a : float | int , ): for nxt, d in graph[v]: if nxt in visited_forward: continue UpperCAmelCase = cst_fwd.get(_a , np.inf ) UpperCAmelCase = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) UpperCAmelCase = new_cost_f UpperCAmelCase = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: UpperCAmelCase = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def snake_case_ (_a : str , _a : str , _a : dict , _a : dict ): UpperCAmelCase = -1 UpperCAmelCase = set() UpperCAmelCase = set() UpperCAmelCase = {source: 0} UpperCAmelCase = {destination: 0} UpperCAmelCase = {source: None} UpperCAmelCase = {destination: None} UpperCAmelCase = PriorityQueue() UpperCAmelCase = PriorityQueue() UpperCAmelCase = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): UpperCAmelCase , UpperCAmelCase = queue_forward.get() visited_forward.add(_a ) UpperCAmelCase , UpperCAmelCase = queue_backward.get() visited_backward.add(_a ) UpperCAmelCase = pass_and_relaxation( _a , _a , _a , _a , _a , _a , _a , _a , _a , ) UpperCAmelCase = pass_and_relaxation( _a , _a , _a , _a , _a , _a , _a , _a , _a , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: UpperCAmelCase = shortest_distance return shortest_path_distance A ={ 'B': [['C', 1]], 'C': [['D', 1]], 'D': [['F', 1]], 'E': [['B', 1], ['G', 2]], 'F': [], 'G': [['F', 1]], } A ={ 'B': [['E', 1]], 'C': [['B', 1]], 'D': [['C', 1]], 'F': [['D', 1], ['G', 1]], 'E': [[None, np.inf]], 'G': [['E', 2]], } if __name__ == "__main__": import doctest doctest.testmod()
34
0
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor lowerCamelCase = random.Random() def lowerCamelCase_ ( _a , _a=1.0 , _a=None , _a=None ): """simple docstring""" if rng is None: lowerCAmelCase__ : List[Any] = global_rng lowerCAmelCase__ : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class _a ( unittest.TestCase): def __init__( self : Optional[Any] , _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : str=7 , _SCREAMING_SNAKE_CASE : Optional[Any]=400 , _SCREAMING_SNAKE_CASE : int=2000 , _SCREAMING_SNAKE_CASE : Optional[Any]=24 , _SCREAMING_SNAKE_CASE : Optional[int]=24 , _SCREAMING_SNAKE_CASE : Union[str, Any]=0.0 , _SCREAMING_SNAKE_CASE : str=1_6000 , _SCREAMING_SNAKE_CASE : Tuple=True , _SCREAMING_SNAKE_CASE : Tuple=True , )-> Optional[Any]: lowerCAmelCase__ : Tuple = parent lowerCAmelCase__ : int = batch_size lowerCAmelCase__ : int = min_seq_length lowerCAmelCase__ : Optional[int] = max_seq_length lowerCAmelCase__ : Any = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowerCAmelCase__ : List[Any] = feature_size lowerCAmelCase__ : int = num_mel_bins lowerCAmelCase__ : Union[str, Any] = padding_value lowerCAmelCase__ : str = sampling_rate lowerCAmelCase__ : Any = return_attention_mask lowerCAmelCase__ : Dict = do_normalize def UpperCAmelCase__( self : Dict )-> int: return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCAmelCase__( self : Optional[int] , _SCREAMING_SNAKE_CASE : Union[str, Any]=False , _SCREAMING_SNAKE_CASE : Dict=False )-> Any: def _flatten(_SCREAMING_SNAKE_CASE : List[Any] ): return list(itertools.chain(*_SCREAMING_SNAKE_CASE ) ) if equal_length: lowerCAmelCase__ : List[Any] = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size lowerCAmelCase__ : Any = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCAmelCase__ : List[str] = [np.asarray(_SCREAMING_SNAKE_CASE ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class _a ( _lowercase , unittest.TestCase): _a : Tuple = SpeechaTextFeatureExtractor if is_speech_available() else None def UpperCAmelCase__( self : Union[str, Any] )-> Any: lowerCAmelCase__ : str = SpeechaTextFeatureExtractionTester(self ) def UpperCAmelCase__( self : List[Any] , _SCREAMING_SNAKE_CASE : str )-> Optional[int]: self.assertTrue(np.all(np.mean(_SCREAMING_SNAKE_CASE , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(_SCREAMING_SNAKE_CASE , axis=0 ) - 1 ) < 1E-3 ) ) def UpperCAmelCase__( self : Optional[int] )-> int: # Tests that all call wrap to encode_plus and batch_encode_plus lowerCAmelCase__ : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCAmelCase__ : Optional[int] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase__ : Optional[int] = [np.asarray(_SCREAMING_SNAKE_CASE ) for speech_input in speech_inputs] # Test feature size lowerCAmelCase__ : Dict = feature_extractor(_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input lowerCAmelCase__ : str = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features lowerCAmelCase__ : Optional[Any] = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1E-3 ) ) # Test batched lowerCAmelCase__ : int = feature_extractor(_SCREAMING_SNAKE_CASE , return_tensors='''np''' ).input_features lowerCAmelCase__ : Union[str, Any] = feature_extractor(_SCREAMING_SNAKE_CASE , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): self.assertTrue(np.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. lowerCAmelCase__ : Optional[int] = [floats_list((1, x) )[0] for x in (800, 800, 800)] lowerCAmelCase__ : str = np.asarray(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Tuple = feature_extractor(_SCREAMING_SNAKE_CASE , return_tensors='''np''' ).input_features lowerCAmelCase__ : List[Any] = feature_extractor(_SCREAMING_SNAKE_CASE , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): self.assertTrue(np.allclose(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , atol=1E-3 ) ) def UpperCAmelCase__( self : Any )-> Union[str, Any]: lowerCAmelCase__ : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase__ : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase__ : Dict = ['''longest''', '''max_length''', '''do_not_pad'''] lowerCAmelCase__ : Any = [None, 16, None] for max_length, padding in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): lowerCAmelCase__ : Union[str, Any] = feature_extractor( _SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE , return_attention_mask=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : int = inputs.input_features lowerCAmelCase__ : Optional[int] = inputs.attention_mask lowerCAmelCase__ : str = [np.sum(_SCREAMING_SNAKE_CASE ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def UpperCAmelCase__( self : List[Any] )-> int: lowerCAmelCase__ : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase__ : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase__ : Optional[Any] = ['''longest''', '''max_length''', '''do_not_pad'''] lowerCAmelCase__ : Optional[int] = [None, 16, None] for max_length, padding in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): lowerCAmelCase__ : str = feature_extractor( _SCREAMING_SNAKE_CASE , max_length=_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE , return_tensors='''np''' , return_attention_mask=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Union[str, Any] = inputs.input_features lowerCAmelCase__ : Optional[int] = inputs.attention_mask lowerCAmelCase__ : str = [np.sum(_SCREAMING_SNAKE_CASE ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def UpperCAmelCase__( self : str )-> Optional[int]: lowerCAmelCase__ : Optional[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase__ : List[str] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase__ : Optional[Any] = feature_extractor( _SCREAMING_SNAKE_CASE , padding='''max_length''' , max_length=4 , truncation=_SCREAMING_SNAKE_CASE , return_tensors='''np''' , return_attention_mask=_SCREAMING_SNAKE_CASE , ) lowerCAmelCase__ : Union[str, Any] = inputs.input_features lowerCAmelCase__ : Any = inputs.attention_mask lowerCAmelCase__ : str = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def UpperCAmelCase__( self : Dict )-> Optional[int]: lowerCAmelCase__ : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase__ : List[str] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase__ : Tuple = feature_extractor( _SCREAMING_SNAKE_CASE , padding='''longest''' , max_length=4 , truncation=_SCREAMING_SNAKE_CASE , return_tensors='''np''' , return_attention_mask=_SCREAMING_SNAKE_CASE , ) lowerCAmelCase__ : Optional[int] = inputs.input_features lowerCAmelCase__ : Optional[Any] = inputs.attention_mask lowerCAmelCase__ : Tuple = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) lowerCAmelCase__ : List[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase__ : Optional[int] = feature_extractor( _SCREAMING_SNAKE_CASE , padding='''longest''' , max_length=16 , truncation=_SCREAMING_SNAKE_CASE , return_tensors='''np''' , return_attention_mask=_SCREAMING_SNAKE_CASE , ) lowerCAmelCase__ : Any = inputs.input_features lowerCAmelCase__ : int = inputs.attention_mask lowerCAmelCase__ : List[Any] = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def UpperCAmelCase__( self : str )-> Union[str, Any]: import torch lowerCAmelCase__ : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase__ : Optional[Any] = np.random.rand(100 , 32 ).astype(np.floataa ) lowerCAmelCase__ : Any = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowerCAmelCase__ : List[Any] = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) lowerCAmelCase__ : int = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def UpperCAmelCase__( self : List[str] , _SCREAMING_SNAKE_CASE : Tuple )-> List[str]: from datasets import load_dataset lowerCAmelCase__ : str = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech lowerCAmelCase__ : List[str] = ds.sort('''id''' ).select(range(_SCREAMING_SNAKE_CASE ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def UpperCAmelCase__( self : Optional[int] )-> int: # fmt: off lowerCAmelCase__ : Optional[Any] = np.array([ -1.5745, -1.7713, -1.7020, -1.6069, -1.2250, -1.1105, -0.9072, -0.8241, -1.2310, -0.8098, -0.3320, -0.4101, -0.7985, -0.4996, -0.8213, -0.9128, -1.0420, -1.1286, -1.0440, -0.7999, -0.8405, -1.2275, -1.5443, -1.4625, ] ) # fmt: on lowerCAmelCase__ : str = self._load_datasamples(1 ) lowerCAmelCase__ : int = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase__ : int = feature_extractor(_SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).input_features self.assertEquals(input_features.shape , (1, 584, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , _SCREAMING_SNAKE_CASE , atol=1E-4 ) )
211
import math class _a : def __init__( self : List[Any] , _SCREAMING_SNAKE_CASE : Any=0 )-> Optional[Any]: # a graph with Node 0,1,...,N-1 lowerCAmelCase__ : Optional[int] = n lowerCAmelCase__ : List[Any] = [ [math.inf for j in range(0 , _SCREAMING_SNAKE_CASE )] for i in range(0 , _SCREAMING_SNAKE_CASE ) ] # adjacency matrix for weight lowerCAmelCase__ : str = [ [math.inf for j in range(0 , _SCREAMING_SNAKE_CASE )] for i in range(0 , _SCREAMING_SNAKE_CASE ) ] # dp[i][j] stores minimum distance from i to j def UpperCAmelCase__( self : Optional[Any] , _SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : List[str] , _SCREAMING_SNAKE_CASE : str )-> List[str]: lowerCAmelCase__ : Optional[int] = w def UpperCAmelCase__( self : List[Any] )-> Optional[int]: for k in range(0 , self.n ): for i in range(0 , self.n ): for j in range(0 , self.n ): lowerCAmelCase__ : Dict = min(self.dp[i][j] , self.dp[i][k] + self.dp[k][j] ) def UpperCAmelCase__( self : str , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : str )-> str: return self.dp[u][v] if __name__ == "__main__": lowerCamelCase = Graph(5) graph.add_edge(0, 2, 9) graph.add_edge(0, 4, 10) graph.add_edge(1, 3, 5) graph.add_edge(2, 3, 7) graph.add_edge(3, 0, 10) graph.add_edge(3, 1, 2) graph.add_edge(3, 2, 1) graph.add_edge(3, 4, 6) graph.add_edge(4, 1, 3) graph.add_edge(4, 2, 4) graph.add_edge(4, 3, 9) graph.floyd_warshall() graph.show_min(1, 4) graph.show_min(0, 3)
211
1
def __UpperCamelCase ( _A : str ) ->int: """simple docstring""" return " ".join( """""".join(word[::-1] ) if len(lowercase__ ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words('Hey wollef sroirraw'))
154
def _lowercase ( lowercase__ , lowercase__ ): return (pointa[0] - pointa[0]) ** 2 + (pointa[1] - pointa[1]) ** 2 def _lowercase ( lowercase__ , lowercase__=0 ): return sorted(lowercase__ , key=lambda lowercase__ : x[column] ) def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(points_counts - 1 ): for j in range(i + 1 , lowercase__ ): __lowerCAmelCase : List[str] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : Tuple = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__=float('''inf''' ) ): for i in range(min(6 , points_counts - 1 ) , lowercase__ ): for j in range(max(0 , i - 6 ) , lowercase__ ): __lowerCAmelCase : Union[str, Any] = euclidean_distance_sqr(points[i] , points[j] ) if current_dis < min_dis: __lowerCAmelCase : int = current_dis return min_dis def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): # base case if points_counts <= 3: return dis_between_closest_pair(lowercase__ , lowercase__ ) # recursion __lowerCAmelCase : Optional[Any] = points_counts // 2 __lowerCAmelCase : Optional[Any] = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[:mid] , lowercase__ ) __lowerCAmelCase : str = closest_pair_of_points_sqr( lowercase__ , points_sorted_on_y[mid:] , points_counts - mid ) __lowerCAmelCase : Optional[int] = min(lowercase__ , lowercase__ ) __lowerCAmelCase : Tuple = [] for point in points_sorted_on_x: if abs(point[0] - points_sorted_on_x[mid][0] ) < closest_pair_dis: cross_strip.append(lowercase__ ) __lowerCAmelCase : List[Any] = dis_between_closest_in_strip( lowercase__ , len(lowercase__ ) , lowercase__ ) return min(lowercase__ , lowercase__ ) def _lowercase ( lowercase__ , lowercase__ ): __lowerCAmelCase : Union[str, Any] = column_based_sort(lowercase__ , column=0 ) __lowerCAmelCase : Any = column_based_sort(lowercase__ , column=1 ) return ( closest_pair_of_points_sqr( lowercase__ , lowercase__ , lowercase__ ) ) ** 0.5 if __name__ == "__main__": _UpperCamelCase = [(2, 3), (12, 30), (40, 50), (5, 1), (12, 10), (3, 4)] print("Distance:", closest_pair_of_points(points, len(points)))
275
0
from pathlib import PurePosixPath from typing import Optional import fsspec from fsspec import AbstractFileSystem from huggingface_hub.hf_api import DatasetInfo from ..utils.file_utils import get_authentication_headers_for_url from ..utils.hub import hf_hub_url class lowercase_ ( snake_case_ ): UpperCamelCase_ : int = "" UpperCamelCase_ : int = "hf-legacy" # "hf://"" is reserved for hffs def __init__( self : str , A__ : Optional[DatasetInfo] = None , A__ : Optional[str] = None , **A__ : Optional[Any] , ) -> Tuple: super().__init__(self , **_A ) _snake_case = repo_info _snake_case = token _snake_case = None def UpperCamelCase_ ( self : List[str] ) -> List[str]: if self.dir_cache is None: _snake_case = {} for hf_file in self.repo_info.siblings: # TODO(QL): add sizes _snake_case = { 'name': hf_file.rfilename, 'size': None, 'type': 'file', } self.dir_cache.update( { str(_A ): {'''name''': str(_A ), '''size''': None, '''type''': '''directory'''} for d in list(PurePosixPath(hf_file.rfilename ).parents )[:-1] } ) def UpperCamelCase_ ( self : Union[str, Any] , A__ : str , A__ : str = "rb" , **A__ : Tuple , ) -> Dict: if not isinstance(self.repo_info , _A ): raise NotImplementedError(f"""Open is only implemented for dataset repositories, but got {self.repo_info}""" ) _snake_case = hf_hub_url(self.repo_info.id , _A , revision=self.repo_info.sha ) return fsspec.open( _A , mode=_A , headers=get_authentication_headers_for_url(_A , use_auth_token=self.token ) , client_kwargs={'''trust_env''': True} , ).open() def UpperCamelCase_ ( self : int , A__ : Tuple , **A__ : str ) -> List[str]: self._get_dirs() _snake_case = self._strip_protocol(_A ) if path in self.dir_cache: return self.dir_cache[path] else: raise FileNotFoundError(_A ) def UpperCamelCase_ ( self : Dict , A__ : List[str] , A__ : Union[str, Any]=False , **A__ : Any ) -> Tuple: self._get_dirs() _snake_case = PurePosixPath(path.strip('''/''' ) ) _snake_case = {} for p, f in self.dir_cache.items(): _snake_case = PurePosixPath(p.strip('''/''' ) ) _snake_case = p.parent if root == path: _snake_case = f _snake_case = list(paths.values() ) if detail: return out else: return sorted(f['''name'''] for f in out )
350
__A = '''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def snake_case_(_UpperCamelCase ) -> bytes: """simple docstring""" if not isinstance(_UpperCamelCase , _UpperCamelCase ): _snake_case = F"""a bytes-like object is required, not '{data.__class__.__name__}'""" raise TypeError(_UpperCamelCase ) _snake_case = ''''''.join(bin(_UpperCamelCase )[2:].zfill(8 ) for byte in data ) _snake_case = len(_UpperCamelCase ) % 6 != 0 if padding_needed: # The padding that will be added later _snake_case = b'''=''' * ((6 - len(_UpperCamelCase ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(_UpperCamelCase ) % 6) else: _snake_case = b'''''' # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(_UpperCamelCase ) , 6 ) ).encode() + padding ) def snake_case_(_UpperCamelCase ) -> bytes: """simple docstring""" if not isinstance(_UpperCamelCase , _UpperCamelCase ) and not isinstance(_UpperCamelCase , _UpperCamelCase ): _snake_case = ( '''argument should be a bytes-like object or ASCII string, ''' F"""not '{encoded_data.__class__.__name__}'""" ) raise TypeError(_UpperCamelCase ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(_UpperCamelCase , _UpperCamelCase ): try: _snake_case = encoded_data.decode('''utf-8''' ) except UnicodeDecodeError: raise ValueError('''base64 encoded data should only contain ASCII characters''' ) _snake_case = encoded_data.count('''=''' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(_UpperCamelCase ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one _snake_case = encoded_data[:-padding] _snake_case = ''''''.join( bin(B64_CHARSET.index(_UpperCamelCase ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: _snake_case = ''''''.join( bin(B64_CHARSET.index(_UpperCamelCase ) )[2:].zfill(6 ) for char in encoded_data ) _snake_case = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(_UpperCamelCase ) , 8 ) ] return bytes(_UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
278
0
"""simple docstring""" def _lowerCAmelCase ( UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = len(UpperCamelCase_ ) for i in range(length - 1 ): __SCREAMING_SNAKE_CASE = i for k in range(i + 1 , UpperCamelCase_ ): if collection[k] < collection[least]: __SCREAMING_SNAKE_CASE = k if least != i: __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE = (collection[i], collection[least]) return collection if __name__ == "__main__": __magic_name__ = input("Enter numbers separated by a comma:\n").strip() __magic_name__ = [int(item) for item in user_input.split(",")] print(selection_sort(unsorted))
100
"""simple docstring""" import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, ) from transformers.models.albert.modeling_albert import ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE_ : """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=1_3 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=9_9 , lowerCAmelCase__=1_6 , lowerCAmelCase__=3_6 , lowerCAmelCase__=6 , lowerCAmelCase__=6 , lowerCAmelCase__=6 , lowerCAmelCase__=3_7 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=5_1_2 , lowerCAmelCase__=1_6 , lowerCAmelCase__=2 , lowerCAmelCase__=0.02 , lowerCAmelCase__=3 , lowerCAmelCase__=4 , lowerCAmelCase__=None , ): __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = seq_length __SCREAMING_SNAKE_CASE = is_training __SCREAMING_SNAKE_CASE = use_input_mask __SCREAMING_SNAKE_CASE = use_token_type_ids __SCREAMING_SNAKE_CASE = use_labels __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = embedding_size __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_hidden_groups __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = max_position_embeddings __SCREAMING_SNAKE_CASE = type_vocab_size __SCREAMING_SNAKE_CASE = type_sequence_label_size __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = num_labels __SCREAMING_SNAKE_CASE = num_choices __SCREAMING_SNAKE_CASE = scope def snake_case_ ( self): __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __SCREAMING_SNAKE_CASE = None if self.use_input_mask: __SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length]) __SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None if self.use_labels: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.type_sequence_label_size) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_choices) __SCREAMING_SNAKE_CASE = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case_ ( self): return AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , num_hidden_groups=self.num_hidden_groups , ) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = AlbertModel(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , token_type_ids=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = AlbertForPreTraining(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__ , sentence_order_label=lowerCAmelCase__ , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) self.parent.assertEqual(result.sop_logits.shape , (self.batch_size, config.num_labels)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = AlbertForMaskedLM(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = AlbertForQuestionAnswering(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , start_positions=lowerCAmelCase__ , end_positions=lowerCAmelCase__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self.num_labels __SCREAMING_SNAKE_CASE = AlbertForSequenceClassification(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self.num_labels __SCREAMING_SNAKE_CASE = AlbertForTokenClassification(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self.num_choices __SCREAMING_SNAKE_CASE = AlbertForMultipleChoice(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() __SCREAMING_SNAKE_CASE = input_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __SCREAMING_SNAKE_CASE = token_type_ids.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __SCREAMING_SNAKE_CASE = input_mask.unsqueeze(1).expand(-1 , self.num_choices , -1).contiguous() __SCREAMING_SNAKE_CASE = model( lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , token_type_ids=lowerCAmelCase__ , labels=lowerCAmelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices)) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() ( ( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) ,( __SCREAMING_SNAKE_CASE ) , ) = config_and_inputs __SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE_ ( __a , __a , unittest.TestCase ): """simple docstring""" __lowercase : Union[str, Any] = ( ( AlbertModel, AlbertForPreTraining, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertForQuestionAnswering, ) if is_torch_available() else () ) __lowercase : Optional[int] = ( { '''feature-extraction''': AlbertModel, '''fill-mask''': AlbertForMaskedLM, '''question-answering''': AlbertForQuestionAnswering, '''text-classification''': AlbertForSequenceClassification, '''token-classification''': AlbertForTokenClassification, '''zero-shot''': AlbertForSequenceClassification, } if is_torch_available() else {} ) __lowercase : List[Any] = True def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__=False): __SCREAMING_SNAKE_CASE = super()._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__) if return_labels: if model_class in get_values(lowerCAmelCase__): __SCREAMING_SNAKE_CASE = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCAmelCase__) return inputs_dict def snake_case_ ( self): __SCREAMING_SNAKE_CASE = AlbertModelTester(self) __SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=lowerCAmelCase__ , hidden_size=3_7) def snake_case_ ( self): self.config_tester.run_common_tests() def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __SCREAMING_SNAKE_CASE = type self.model_tester.create_and_check_model(*lowerCAmelCase__) @slow def snake_case_ ( self): for model_name in ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __SCREAMING_SNAKE_CASE = AlbertModel.from_pretrained(lowerCAmelCase__) self.assertIsNotNone(lowerCAmelCase__) @require_torch class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): """simple docstring""" @slow def snake_case_ ( self): __SCREAMING_SNAKE_CASE = AlbertModel.from_pretrained("""albert-base-v2""") __SCREAMING_SNAKE_CASE = torch.tensor([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]]) __SCREAMING_SNAKE_CASE = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]) with torch.no_grad(): __SCREAMING_SNAKE_CASE = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__)[0] __SCREAMING_SNAKE_CASE = torch.Size((1, 1_1, 7_6_8)) self.assertEqual(output.shape , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = torch.tensor( [[[-0.65_13, 1.50_35, -0.27_66], [-0.65_15, 1.50_46, -0.27_80], [-0.65_12, 1.50_49, -0.27_84]]]) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , lowerCAmelCase__ , atol=1E-4))
100
1
'''simple docstring''' def a__ ( _SCREAMING_SNAKE_CASE : str ) -> int: """simple docstring""" assert column_title.isupper() UpperCAmelCase_ : Tuple = 0 UpperCAmelCase_ : Optional[Any] = len(lowerCAmelCase__ ) - 1 UpperCAmelCase_ : str = 0 while index >= 0: UpperCAmelCase_ : Optional[Any] = (ord(column_title[index] ) - 64) * pow(26 , lowerCAmelCase__ ) answer += value power += 1 index -= 1 return answer if __name__ == "__main__": from doctest import testmod testmod()
365
'''simple docstring''' from ..utils import DummyObject, requires_backends class _snake_case (metaclass=__SCREAMING_SNAKE_CASE): __A : Union[str, Any] =["torch", "torchsde"] def __init__( self ,*_snake_case ,**_snake_case ): requires_backends(self ,["torch", "torchsde"] ) @classmethod def UpperCamelCase__ ( cls ,*_snake_case ,**_snake_case ): requires_backends(cls ,["torch", "torchsde"] ) @classmethod def UpperCamelCase__ ( cls ,*_snake_case ,**_snake_case ): requires_backends(cls ,["torch", "torchsde"] )
67
0
import argparse import collections import json import os import re import string import sys import numpy as np snake_case : int = re.compile(R'''\b(a|an|the)\b''', re.UNICODE) snake_case : Optional[Any] = None def __lowerCamelCase ( ): """simple docstring""" a :Optional[int] = argparse.ArgumentParser('''Official evaluation script for SQuAD version 2.0.''' ) parser.add_argument('''data_file''' , metavar='''data.json''' , help='''Input data JSON file.''' ) parser.add_argument('''pred_file''' , metavar='''pred.json''' , help='''Model predictions.''' ) parser.add_argument( '''--out-file''' , '''-o''' , metavar='''eval.json''' , help='''Write accuracy metrics to file (default is stdout).''' ) parser.add_argument( '''--na-prob-file''' , '''-n''' , metavar='''na_prob.json''' , help='''Model estimates of probability of no answer.''' ) parser.add_argument( '''--na-prob-thresh''' , '''-t''' , type=UpperCAmelCase_ , default=1.0 , help='''Predict "" if no-answer probability exceeds this (default = 1.0).''' , ) parser.add_argument( '''--out-image-dir''' , '''-p''' , metavar='''out_images''' , default=UpperCAmelCase_ , help='''Save precision-recall curves to directory.''' ) parser.add_argument('''--verbose''' , '''-v''' , action='''store_true''' ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def __lowerCamelCase ( UpperCAmelCase_ : Any ): """simple docstring""" a :Tuple = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: a :List[str] = bool(qa['''answers''']['''text'''] ) return qid_to_has_ans def __lowerCamelCase ( UpperCAmelCase_ : Tuple ): """simple docstring""" def remove_articles(UpperCAmelCase_ : Optional[int] ): return ARTICLES_REGEX.sub(''' ''' , UpperCAmelCase_ ) def white_space_fix(UpperCAmelCase_ : Tuple ): return " ".join(text.split() ) def remove_punc(UpperCAmelCase_ : List[str] ): a :List[Any] = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(UpperCAmelCase_ : int ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(UpperCAmelCase_ ) ) ) ) def __lowerCamelCase ( UpperCAmelCase_ : Any ): """simple docstring""" if not s: return [] return normalize_answer(UpperCAmelCase_ ).split() def __lowerCamelCase ( UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Optional[Any] ): """simple docstring""" return int(normalize_answer(UpperCAmelCase_ ) == normalize_answer(UpperCAmelCase_ ) ) def __lowerCamelCase ( UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : List[str] ): """simple docstring""" a :Dict = get_tokens(UpperCAmelCase_ ) a :Dict = get_tokens(UpperCAmelCase_ ) a :Dict = collections.Counter(UpperCAmelCase_ ) & collections.Counter(UpperCAmelCase_ ) a :List[str] = sum(common.values() ) if len(UpperCAmelCase_ ) == 0 or len(UpperCAmelCase_ ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 a :Dict = 1.0 * num_same / len(UpperCAmelCase_ ) a :int = 1.0 * num_same / len(UpperCAmelCase_ ) a :Tuple = (2 * precision * recall) / (precision + recall) return fa def __lowerCamelCase ( UpperCAmelCase_ : Dict , UpperCAmelCase_ : Tuple ): """simple docstring""" a :Union[str, Any] = {} a :int = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: a :str = qa['''id'''] a :Any = [t for t in qa['''answers''']['''text'''] if normalize_answer(UpperCAmelCase_ )] if not gold_answers: # For unanswerable questions, only correct answer is empty string a :Optional[Any] = [''''''] if qid not in preds: print(F'''Missing prediction for {qid}''' ) continue a :List[Any] = preds[qid] # Take max over all gold answers a :Any = max(compute_exact(UpperCAmelCase_ , UpperCAmelCase_ ) for a in gold_answers ) a :Optional[int] = max(compute_fa(UpperCAmelCase_ , UpperCAmelCase_ ) for a in gold_answers ) return exact_scores, fa_scores def __lowerCamelCase ( UpperCAmelCase_ : Any , UpperCAmelCase_ : str , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Optional[Any] ): """simple docstring""" a :Dict = {} for qid, s in scores.items(): a :Any = na_probs[qid] > na_prob_thresh if pred_na: a :int = float(not qid_to_has_ans[qid] ) else: a :Union[str, Any] = s return new_scores def __lowerCamelCase ( UpperCAmelCase_ : int , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Union[str, Any]=None ): """simple docstring""" if not qid_list: a :Optional[Any] = len(UpperCAmelCase_ ) return collections.OrderedDict( [ ('''exact''', 100.0 * sum(exact_scores.values() ) / total), ('''f1''', 100.0 * sum(fa_scores.values() ) / total), ('''total''', total), ] ) else: a :Optional[int] = len(UpperCAmelCase_ ) return collections.OrderedDict( [ ('''exact''', 100.0 * sum(exact_scores[k] for k in qid_list ) / total), ('''f1''', 100.0 * sum(fa_scores[k] for k in qid_list ) / total), ('''total''', total), ] ) def __lowerCamelCase ( UpperCAmelCase_ : Any , UpperCAmelCase_ : Any , UpperCAmelCase_ : Tuple ): """simple docstring""" for k in new_eval: a :List[Any] = new_eval[k] def __lowerCamelCase ( UpperCAmelCase_ : Any , UpperCAmelCase_ : int , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : List[Any] ): """simple docstring""" plt.step(UpperCAmelCase_ , UpperCAmelCase_ , color='''b''' , alpha=0.2 , where='''post''' ) plt.fill_between(UpperCAmelCase_ , UpperCAmelCase_ , step='''post''' , alpha=0.2 , color='''b''' ) plt.xlabel('''Recall''' ) plt.ylabel('''Precision''' ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(UpperCAmelCase_ ) plt.savefig(UpperCAmelCase_ ) plt.clf() def __lowerCamelCase ( UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Tuple=None , UpperCAmelCase_ : str=None ): """simple docstring""" a :Optional[Any] = sorted(UpperCAmelCase_ , key=lambda UpperCAmelCase_ : na_probs[k] ) a :List[str] = 0.0 a :str = 1.0 a :Any = 0.0 a :Optional[Any] = [1.0] a :Any = [0.0] a :Tuple = 0.0 for i, qid in enumerate(UpperCAmelCase_ ): if qid_to_has_ans[qid]: true_pos += scores[qid] a :List[str] = true_pos / float(i + 1 ) a :str = true_pos / float(UpperCAmelCase_ ) if i == len(UpperCAmelCase_ ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(UpperCAmelCase_ ) recalls.append(UpperCAmelCase_ ) if out_image: plot_pr_curve(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) return {"ap": 100.0 * avg_prec} def __lowerCamelCase ( UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : str , UpperCAmelCase_ : str , UpperCAmelCase_ : Any , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Any ): """simple docstring""" if out_image_dir and not os.path.exists(UpperCAmelCase_ ): os.makedirs(UpperCAmelCase_ ) a :Optional[int] = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return a :Union[str, Any] = make_precision_recall_eval( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , out_image=os.path.join(UpperCAmelCase_ , '''pr_exact.png''' ) , title='''Precision-Recall curve for Exact Match score''' , ) a :Union[str, Any] = make_precision_recall_eval( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , out_image=os.path.join(UpperCAmelCase_ , '''pr_f1.png''' ) , title='''Precision-Recall curve for F1 score''' , ) a :Union[str, Any] = {k: float(UpperCAmelCase_ ) for k, v in qid_to_has_ans.items()} a :int = make_precision_recall_eval( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , out_image=os.path.join(UpperCAmelCase_ , '''pr_oracle.png''' ) , title='''Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)''' , ) merge_eval(UpperCAmelCase_ , UpperCAmelCase_ , '''pr_exact''' ) merge_eval(UpperCAmelCase_ , UpperCAmelCase_ , '''pr_f1''' ) merge_eval(UpperCAmelCase_ , UpperCAmelCase_ , '''pr_oracle''' ) def __lowerCamelCase ( UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Any , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : List[str] ): """simple docstring""" if not qid_list: return a :List[Any] = [na_probs[k] for k in qid_list] a :List[str] = np.ones_like(UpperCAmelCase_ ) / float(len(UpperCAmelCase_ ) ) plt.hist(UpperCAmelCase_ , weights=UpperCAmelCase_ , bins=20 , range=(0.0, 1.0) ) plt.xlabel('''Model probability of no-answer''' ) plt.ylabel('''Proportion of dataset''' ) plt.title(F'''Histogram of no-answer probability: {name}''' ) plt.savefig(os.path.join(UpperCAmelCase_ , F'''na_prob_hist_{name}.png''' ) ) plt.clf() def __lowerCamelCase ( UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Any , UpperCAmelCase_ : str , UpperCAmelCase_ : Dict ): """simple docstring""" a :str = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) a :List[Any] = num_no_ans a :str = cur_score a :List[Any] = 0.0 a :Tuple = sorted(UpperCAmelCase_ , key=lambda UpperCAmelCase_ : na_probs[k] ) for i, qid in enumerate(UpperCAmelCase_ ): if qid not in scores: continue if qid_to_has_ans[qid]: a :Dict = scores[qid] else: if preds[qid]: a :Optional[int] = -1 else: a :Optional[Any] = 0 cur_score += diff if cur_score > best_score: a :Tuple = cur_score a :Dict = na_probs[qid] return 100.0 * best_score / len(UpperCAmelCase_ ), best_thresh def __lowerCamelCase ( UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Any , UpperCAmelCase_ : str ): """simple docstring""" a , a :List[Any] = find_best_thresh(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) a , a :Optional[int] = find_best_thresh(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) a :Optional[Any] = best_exact a :int = exact_thresh a :Any = best_fa a :int = fa_thresh def __lowerCamelCase ( ): """simple docstring""" with open(OPTS.data_file ) as f: a :Dict = json.load(UpperCAmelCase_ ) a :List[str] = dataset_json['''data'''] with open(OPTS.pred_file ) as f: a :Dict = json.load(UpperCAmelCase_ ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: a :Dict = json.load(UpperCAmelCase_ ) else: a :Any = {k: 0.0 for k in preds} a :int = make_qid_to_has_ans(UpperCAmelCase_ ) # maps qid to True/False a :Optional[Any] = [k for k, v in qid_to_has_ans.items() if v] a :Dict = [k for k, v in qid_to_has_ans.items() if not v] a , a :List[Any] = get_raw_scores(UpperCAmelCase_ , UpperCAmelCase_ ) a :Any = apply_no_ans_threshold(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , OPTS.na_prob_thresh ) a :Any = apply_no_ans_threshold(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , OPTS.na_prob_thresh ) a :List[Any] = make_eval_dict(UpperCAmelCase_ , UpperCAmelCase_ ) if has_ans_qids: a :Tuple = make_eval_dict(UpperCAmelCase_ , UpperCAmelCase_ , qid_list=UpperCAmelCase_ ) merge_eval(UpperCAmelCase_ , UpperCAmelCase_ , '''HasAns''' ) if no_ans_qids: a :Optional[int] = make_eval_dict(UpperCAmelCase_ , UpperCAmelCase_ , qid_list=UpperCAmelCase_ ) merge_eval(UpperCAmelCase_ , UpperCAmelCase_ , '''NoAns''' ) if OPTS.na_prob_file: find_all_best_thresh(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , OPTS.out_image_dir ) histogram_na_prob(UpperCAmelCase_ , UpperCAmelCase_ , OPTS.out_image_dir , '''hasAns''' ) histogram_na_prob(UpperCAmelCase_ , UpperCAmelCase_ , OPTS.out_image_dir , '''noAns''' ) if OPTS.out_file: with open(OPTS.out_file , '''w''' ) as f: json.dump(UpperCAmelCase_ , UpperCAmelCase_ ) else: print(json.dumps(UpperCAmelCase_ , indent=2 ) ) if __name__ == "__main__": snake_case : Optional[Any] = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use('''Agg''') import matplotlib.pyplot as plt main()
94
def __lowerCamelCase ( UpperCAmelCase_ : int = 1000 ): """simple docstring""" a , a :int = 1, 1 a :Any = 2 while True: a :Optional[int] = 0 a :str = fa + fa a , a :List[Any] = fa, f index += 1 for _ in str(UpperCAmelCase_ ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
94
1
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __snake_case ( __lowerCamelCase ): '''simple docstring''' lowerCAmelCase__ = ["""image_processor""", """tokenizer"""] lowerCAmelCase__ = """ChineseCLIPImageProcessor""" lowerCAmelCase__ = ("""BertTokenizer""", """BertTokenizerFast""") def __init__( self : Optional[int] , A : List[Any]=None , A : Tuple=None , **A : str ): __snake_case: List[str] = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , A , ) __snake_case: Dict = kwargs.pop("""feature_extractor""" ) __snake_case: str = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(A , A ) __snake_case: int = self.image_processor def __call__( self : List[Any] , A : int=None , A : str=None , A : int=None , **A : Tuple ): if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: __snake_case: Any = self.tokenizer(A , return_tensors=A , **A ) if images is not None: __snake_case: Tuple = self.image_processor(A , return_tensors=A , **A ) if text is not None and images is not None: __snake_case: Union[str, Any] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**A ) , tensor_type=A ) def UpperCAmelCase__ ( self : List[Any] , *A : Optional[Any] , **A : Optional[Any] ): return self.tokenizer.batch_decode(*A , **A ) def UpperCAmelCase__ ( self : Optional[int] , *A : Optional[Any] , **A : Optional[int] ): return self.tokenizer.decode(*A , **A ) @property def UpperCAmelCase__ ( self : List[Any] ): __snake_case: Dict = self.tokenizer.model_input_names __snake_case: str = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def UpperCAmelCase__ ( self : Union[str, Any] ): warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" , A , ) return self.image_processor_class
293
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __UpperCAmelCase : List[str] = { "configuration_roberta": ["ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP", "RobertaConfig", "RobertaOnnxConfig"], "tokenization_roberta": ["RobertaTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase : Optional[Any] = ["RobertaTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase : Tuple = [ "ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST", "RobertaForCausalLM", "RobertaForMaskedLM", "RobertaForMultipleChoice", "RobertaForQuestionAnswering", "RobertaForSequenceClassification", "RobertaForTokenClassification", "RobertaModel", "RobertaPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase : Optional[int] = [ "TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST", "TFRobertaForCausalLM", "TFRobertaForMaskedLM", "TFRobertaForMultipleChoice", "TFRobertaForQuestionAnswering", "TFRobertaForSequenceClassification", "TFRobertaForTokenClassification", "TFRobertaMainLayer", "TFRobertaModel", "TFRobertaPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase : List[Any] = [ "FlaxRobertaForCausalLM", "FlaxRobertaForMaskedLM", "FlaxRobertaForMultipleChoice", "FlaxRobertaForQuestionAnswering", "FlaxRobertaForSequenceClassification", "FlaxRobertaForTokenClassification", "FlaxRobertaModel", "FlaxRobertaPreTrainedModel", ] if TYPE_CHECKING: from .configuration_roberta import ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaConfig, RobertaOnnxConfig from .tokenization_roberta import RobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roberta_fast import RobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta import ( ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaForCausalLM, RobertaForMaskedLM, RobertaForMultipleChoice, RobertaForQuestionAnswering, RobertaForSequenceClassification, RobertaForTokenClassification, RobertaModel, RobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta import ( TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForMultipleChoice, TFRobertaForQuestionAnswering, TFRobertaForSequenceClassification, TFRobertaForTokenClassification, TFRobertaMainLayer, TFRobertaModel, TFRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, FlaxRobertaPreTrainedModel, ) else: import sys __UpperCAmelCase : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
293
1
"""simple docstring""" import random import timeit from functools import wraps from typing import Callable, Optional from ..configuration_utils import PretrainedConfig from ..models.auto.modeling_tf_auto import TF_MODEL_MAPPING, TF_MODEL_WITH_LM_HEAD_MAPPING from ..utils import is_pyanvml_available, is_tf_available, logging from .benchmark_utils import ( Benchmark, Memory, MemorySummary, measure_peak_memory_cpu, start_memory_tracing, stop_memory_tracing, ) if is_tf_available(): import tensorflow as tf from tensorflow.python.framework.errors_impl import ResourceExhaustedError from .benchmark_args_tf import TensorFlowBenchmarkArguments if is_pyanvml_available(): import pyanvml.pyanvml as nvml _lowerCAmelCase :Any = logging.get_logger(__name__) def lowerCamelCase_ (UpperCamelCase__ : bool , UpperCamelCase__ : bool ): def run_func(UpperCamelCase__ : List[Any] ): @wraps(UpperCamelCase__ ) def run_in_eager_mode(*UpperCamelCase__ : List[str] , **UpperCamelCase__ : Any ): return func(*UpperCamelCase__ , **UpperCamelCase__ ) @wraps(UpperCamelCase__ ) @tf.function(experimental_compile=UpperCamelCase__ ) def run_in_graph_mode(*UpperCamelCase__ : Any , **UpperCamelCase__ : Dict ): return func(*UpperCamelCase__ , **UpperCamelCase__ ) if do_eager_mode is True: if use_xla is not False: raise ValueError( '''Cannot run model in XLA, if `args.eager_mode` is set to `True`. Please set `args.eager_mode=False`.''' ) return run_in_eager_mode else: return run_in_graph_mode return run_func def lowerCamelCase_ (UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : int ): _UpperCAmelCase : Tuple = random.Random() _UpperCAmelCase : int = [rng.randint(0 , vocab_size - 1 ) for i in range(batch_size * sequence_length )] return tf.constant(UpperCamelCase__ , shape=(batch_size, sequence_length) , dtype=tf.intaa ) class _UpperCAmelCase ( a ): '''simple docstring''' a__ =42 a__ =42 a__ ="TensorFlow" @property def __lowerCAmelCase ( self ) -> List[str]: return tf.__version__ def __lowerCAmelCase ( self , A , A , A ) -> float: # initialize GPU on separate process _UpperCAmelCase : str = self.args.strategy if strategy is None: raise ValueError('''A device strategy has to be initialized before using TensorFlow.''' ) _UpperCAmelCase : str = self._prepare_inference_func(A , A , A ) return self._measure_speed(_inference ) def __lowerCAmelCase ( self , A , A , A ) -> float: _UpperCAmelCase : Optional[Any] = self.args.strategy if strategy is None: raise ValueError('''A device strategy has to be initialized before using TensorFlow.''' ) _UpperCAmelCase : str = self._prepare_train_func(A , A , A ) return self._measure_speed(_train ) def __lowerCAmelCase ( self , A , A , A ) -> [Memory, Optional[MemorySummary]]: # initialize GPU on separate process if self.args.is_gpu: tf.config.experimental.set_memory_growth(self.args.gpu_list[self.args.device_idx] , A ) _UpperCAmelCase : List[str] = self.args.strategy if strategy is None: raise ValueError('''A device strategy has to be initialized before using TensorFlow.''' ) _UpperCAmelCase : Tuple = self._prepare_inference_func(A , A , A ) return self._measure_memory(_inference ) def __lowerCAmelCase ( self , A , A , A ) -> [Memory, Optional[MemorySummary]]: if self.args.is_gpu: tf.config.experimental.set_memory_growth(self.args.gpu_list[self.args.device_idx] , A ) _UpperCAmelCase : Dict = self.args.strategy if strategy is None: raise ValueError('''A device strategy has to be initialized before using TensorFlow.''' ) _UpperCAmelCase : Optional[int] = self._prepare_train_func(A , A , A ) return self._measure_memory(_train ) def __lowerCAmelCase ( self , A , A , A ) -> Callable[[], None]: _UpperCAmelCase : Dict = self.config_dict[model_name] if self.args.fpaa: raise NotImplementedError('''Mixed precision is currently not supported.''' ) _UpperCAmelCase : Dict = ( hasattr(A , '''architectures''' ) and isinstance(config.architectures , A ) and len(config.architectures ) > 0 ) if not self.args.only_pretrain_model and has_model_class_in_config: try: _UpperCAmelCase : Tuple = '''TF''' + config.architectures[0] # prepend 'TF' for tensorflow model _UpperCAmelCase : Optional[Any] = __import__('''transformers''' , fromlist=[model_class] ) _UpperCAmelCase : Any = getattr(A , A ) _UpperCAmelCase : Any = model_cls(A ) except ImportError: raise ImportError( f'{model_class} does not exist. If you just want to test the pretrained model, you might want to' ''' set `--only_pretrain_model` or `args.only_pretrain_model=True`.''' ) else: _UpperCAmelCase : str = TF_MODEL_MAPPING[config.__class__](A ) # encoder-decoder has vocab size saved differently _UpperCAmelCase : Optional[Any] = config.vocab_size if hasattr(A , '''vocab_size''' ) else config.encoder.vocab_size _UpperCAmelCase : List[Any] = random_input_ids(A , A , A ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_decoder_forward(): return model(A , decoder_input_ids=A , training=A ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_forward(): return model(A , training=A ) _UpperCAmelCase : Dict = encoder_decoder_forward if config.is_encoder_decoder else encoder_forward return _inference def __lowerCAmelCase ( self , A , A , A ) -> Callable[[], None]: _UpperCAmelCase : Union[str, Any] = self.config_dict[model_name] if self.args.eager_mode is not False: raise ValueError('''Training cannot be done in eager mode. Please make sure that `args.eager_mode = False`.''' ) if self.args.fpaa: raise NotImplementedError('''Mixed precision is currently not supported.''' ) _UpperCAmelCase : List[Any] = ( hasattr(A , '''architectures''' ) and isinstance(config.architectures , A ) and len(config.architectures ) > 0 ) if not self.args.only_pretrain_model and has_model_class_in_config: try: _UpperCAmelCase : Optional[int] = '''TF''' + config.architectures[0] # prepend 'TF' for tensorflow model _UpperCAmelCase : Any = __import__('''transformers''' , fromlist=[model_class] ) _UpperCAmelCase : List[str] = getattr(A , A ) _UpperCAmelCase : Union[str, Any] = model_cls(A ) except ImportError: raise ImportError( f'{model_class} does not exist. If you just want to test the pretrained model, you might want to' ''' set `--only_pretrain_model` or `args.only_pretrain_model=True`.''' ) else: _UpperCAmelCase : Union[str, Any] = TF_MODEL_WITH_LM_HEAD_MAPPING[config.__class__](A ) # encoder-decoder has vocab size saved differently _UpperCAmelCase : str = config.vocab_size if hasattr(A , '''vocab_size''' ) else config.encoder.vocab_size _UpperCAmelCase : int = random_input_ids(A , A , A ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_decoder_train(): _UpperCAmelCase : Tuple = model(A , decoder_input_ids=A , labels=A , training=A )[0] _UpperCAmelCase : Union[str, Any] = tf.gradients(A , model.trainable_variables ) return gradients @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_train(): _UpperCAmelCase : Dict = model(A , labels=A , training=A )[0] _UpperCAmelCase : List[Any] = tf.gradients(A , model.trainable_variables ) return gradients _UpperCAmelCase : Any = encoder_decoder_train if config.is_encoder_decoder else encoder_train return _train def __lowerCAmelCase ( self , A ) -> float: with self.args.strategy.scope(): try: if self.args.is_tpu or self.args.use_xla: # run additional 10 times to stabilize compilation for tpu logger.info('''Do inference on TPU. Running model 5 times to stabilize compilation''' ) timeit.repeat(A , repeat=1 , number=5 ) # as written in https://docs.python.org/2/library/timeit.html#timeit.Timer.repeat, min should be taken rather than the average _UpperCAmelCase : List[str] = timeit.repeat( A , repeat=self.args.repeat , number=1_0 , ) return min(A ) / 10.0 except ResourceExhaustedError as e: self.print_fn(f'Doesn\'t fit on GPU. {e}' ) def __lowerCAmelCase ( self , A ) -> [Memory, MemorySummary]: logger.info( '''Note that TensorFlow allocates more memory than ''' '''it might need to speed up computation. ''' '''The memory reported here corresponds to the memory ''' '''reported by `nvidia-smi`, which can vary depending ''' '''on total available memory on the GPU that is used.''' ) with self.args.strategy.scope(): try: if self.args.trace_memory_line_by_line: if not self.args.eager_mode: raise ValueError( '''`args.eager_mode` is set to `False`. Make sure to run model in eager mode to measure memory''' ''' consumption line by line.''' ) _UpperCAmelCase : Optional[Any] = start_memory_tracing('''transformers''' ) if self.args.is_tpu: # tpu raise NotImplementedError( '''Memory Benchmarking is currently not implemented for TPU. Please disable memory benchmarking''' ''' with `args.memory=False`''' ) elif self.args.is_gpu: # gpu if not is_pyanvml_available(): logger.warning( '''py3nvml not installed, we won\'t log GPU memory usage. ''' '''Install py3nvml (pip install py3nvml) to log information about GPU.''' ) _UpperCAmelCase : Optional[Any] = '''N/A''' else: logger.info( '''Measuring total GPU usage on GPU device. Make sure to not have additional processes''' ''' running on the same GPU.''' ) # init nvml nvml.nvmlInit() func() _UpperCAmelCase : int = nvml.nvmlDeviceGetHandleByIndex(self.args.device_idx ) _UpperCAmelCase : int = nvml.nvmlDeviceGetMemoryInfo(A ) _UpperCAmelCase : List[Any] = meminfo.used _UpperCAmelCase : Any = Memory(A ) # shutdown nvml nvml.nvmlShutdown() else: # cpu if self.args.trace_memory_line_by_line: logger.info( '''When enabling line by line tracing, the max peak memory for CPU is inaccurate in''' ''' TensorFlow.''' ) _UpperCAmelCase : Union[str, Any] = None else: _UpperCAmelCase : Optional[int] = measure_peak_memory_cpu(A ) _UpperCAmelCase : int = Memory(A ) if isinstance(A , A ) else memory_bytes if self.args.trace_memory_line_by_line: _UpperCAmelCase : List[Any] = stop_memory_tracing(A ) if memory is None: _UpperCAmelCase : Tuple = summary.total else: _UpperCAmelCase : int = None return memory, summary except ResourceExhaustedError as e: self.print_fn(f'Doesn\'t fit on GPU. {e}' ) return "N/A", None
263
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowerCAmelCase :int = {'configuration_vit_msn': ['VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ViTMSNConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase :Any = [ 'VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST', 'ViTMSNModel', 'ViTMSNForImageClassification', 'ViTMSNPreTrainedModel', ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys _lowerCAmelCase :int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
263
1
import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class lowercase_ ( unittest.TestCase ): @slow def UpperCamelCase_ ( self : str ) -> List[str]: _snake_case = AutoImageProcessor.from_pretrained('''microsoft/dit-base-finetuned-rvlcdip''' ) _snake_case = AutoModelForImageClassification.from_pretrained('''microsoft/dit-base-finetuned-rvlcdip''' ) model.to(A__ ) from datasets import load_dataset _snake_case = load_dataset('''nielsr/rvlcdip-demo''' ) _snake_case = dataset['''train'''][0]['''image'''].convert('''RGB''' ) _snake_case = image_processor(A__ , return_tensors='''pt''' ).to(A__ ) # forward pass with torch.no_grad(): _snake_case = model(**A__ ) _snake_case = outputs.logits _snake_case = torch.Size((1, 16) ) self.assertEqual(logits.shape , A__ ) _snake_case = torch.tensor( [-0.4158, -0.4092, -0.4347] , device=A__ , dtype=torch.float , ) self.assertTrue(torch.allclose(logits[0, :3] , A__ , atol=1e-4 ) )
359
import cmath import math def snake_case_(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> complex: """simple docstring""" _snake_case = math.radians(_UpperCamelCase ) _snake_case = math.radians(_UpperCamelCase ) # Convert voltage and current to rectangular form _snake_case = cmath.rect(_UpperCamelCase , _UpperCamelCase ) _snake_case = cmath.rect(_UpperCamelCase , _UpperCamelCase ) # Calculate apparent power return voltage_rect * current_rect if __name__ == "__main__": import doctest doctest.testmod()
278
0
"""simple docstring""" import numpy as np import datasets __snake_case : str = '\nCompute the Mahalanobis Distance\n\nMahalonobis distance is the distance between a point and a distribution.\nAnd not between two distinct points. It is effectively a multivariate equivalent of the Euclidean distance.\nIt was introduced by Prof. P. C. Mahalanobis in 1936\nand has been used in various statistical applications ever since\n[source: https://www.machinelearningplus.com/statistics/mahalanobis-distance/]\n' __snake_case : Optional[Any] = '\\n@article{de2000mahalanobis,\n title={The mahalanobis distance},\n author={De Maesschalck, Roy and Jouan-Rimbaud, Delphine and Massart, D{\'e}sir{\'e} L},\n journal={Chemometrics and intelligent laboratory systems},\n volume={50},\n number={1},\n pages={1--18},\n year={2000},\n publisher={Elsevier}\n}\n' __snake_case : List[Any] = '\nArgs:\n X: List of datapoints to be compared with the `reference_distribution`.\n reference_distribution: List of datapoints from the reference distribution we want to compare to.\nReturns:\n mahalanobis: The Mahalonobis distance for each datapoint in `X`.\nExamples:\n\n >>> mahalanobis_metric = datasets.load_metric("mahalanobis")\n >>> results = mahalanobis_metric.compute(reference_distribution=[[0, 1], [1, 0]], X=[[0, 1]])\n >>> print(results)\n {\'mahalanobis\': array([0.5])}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: Any) -> Any: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "X": datasets.Sequence(datasets.Value("float" , id="sequence") , id="X"), }) , ) def _SCREAMING_SNAKE_CASE ( self: Optional[int] , _SCREAMING_SNAKE_CASE: Union[str, Any] , _SCREAMING_SNAKE_CASE: Any) -> Dict: """simple docstring""" __lowerCAmelCase : Any = np.array(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : Any = np.array(_SCREAMING_SNAKE_CASE) # Assert that arrays are 2D if len(X.shape) != 2: raise ValueError("Expected `X` to be a 2D vector") if len(reference_distribution.shape) != 2: raise ValueError("Expected `reference_distribution` to be a 2D vector") if reference_distribution.shape[0] < 2: raise ValueError( "Expected `reference_distribution` to be a 2D vector with more than one element in the first dimension") # Get mahalanobis distance for each prediction __lowerCAmelCase : Any = X - np.mean(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = np.cov(reference_distribution.T) try: __lowerCAmelCase : Union[str, Any] = np.linalg.inv(_SCREAMING_SNAKE_CASE) except np.linalg.LinAlgError: __lowerCAmelCase : Optional[int] = np.linalg.pinv(_SCREAMING_SNAKE_CASE) __lowerCAmelCase : List[Any] = np.dot(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) __lowerCAmelCase : Dict = np.dot(_SCREAMING_SNAKE_CASE , X_minus_mu.T).diagonal() return {"mahalanobis": mahal_dist}
269
"""simple docstring""" import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets __snake_case : Optional[int] = '\\n @inproceedings{kakwani2020indicnlpsuite,\n title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}},\n author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar},\n year={2020},\n booktitle={Findings of EMNLP},\n}\n' __snake_case : str = '\\n IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide\n variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te.\n' __snake_case : str = '\nCompute IndicGLUE evaluation metric associated to each IndicGLUE dataset.\nArgs:\n predictions: list of predictions to score (as int64),\n except for \'cvit-mkb-clsr\' where each prediction is a vector (of float32).\n references: list of ground truth labels corresponding to the predictions (as int64),\n except for \'cvit-mkb-clsr\' where each reference is a vector (of float32).\nReturns: depending on the IndicGLUE subset, one or several of:\n "accuracy": Accuracy\n "f1": F1 score\n "precision": Precision@10\nExamples:\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wnli\') # \'wnli\' or any of ["copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wiki-ner\')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0, \'f1\': 1.0}\n\n >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'cvit-mkb-clsr\')\n >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]]\n >>> results = indic_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'precision@10\': 1.0}\n\n' def _lowercase ( __snake_case ,__snake_case ) -> Union[str, Any]: return float((preds == labels).mean() ) def _lowercase ( __snake_case ,__snake_case ) -> str: __lowerCAmelCase : str = simple_accuracy(__snake_case ,__snake_case ) __lowerCAmelCase : Any = float(fa_score(y_true=__snake_case ,y_pred=__snake_case ) ) return { "accuracy": acc, "f1": fa, } def _lowercase ( __snake_case ,__snake_case ) -> int: __lowerCAmelCase : Union[str, Any] = np.array(__snake_case ) __lowerCAmelCase : Tuple = np.array(__snake_case ) __lowerCAmelCase : List[Any] = en_sentvecs.shape[0] # mean centering __lowerCAmelCase : Union[str, Any] = en_sentvecs - np.mean(__snake_case ,axis=0 ) __lowerCAmelCase : int = in_sentvecs - np.mean(__snake_case ,axis=0 ) __lowerCAmelCase : Optional[Any] = cdist(__snake_case ,__snake_case ,"cosine" ) __lowerCAmelCase : int = np.array(range(__snake_case ) ) __lowerCAmelCase : int = sim.argsort(axis=1 )[:, :10] __lowerCAmelCase : Optional[Any] = np.any(preds == actual[:, None] ,axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A__ ( datasets.Metric ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self: int) -> str: """simple docstring""" if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]") return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), "references": datasets.Value("int64") if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32")), }) , codebase_urls=[] , reference_urls=[] , format="numpy" if self.config_name != "cvit-mkb-clsr" else None , ) def _SCREAMING_SNAKE_CASE ( self: List[str] , _SCREAMING_SNAKE_CASE: int , _SCREAMING_SNAKE_CASE: Optional[Any]) -> int: """simple docstring""" if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)} elif self.config_name in ["wiki-ner"]: return acc_and_fa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE)} else: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]")
269
1
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) __A ='''\ Text data. Second line of data.''' __A ='''file''' @pytest.fixture(scope="session" ) def lowerCamelCase_ ( lowerCamelCase__ ): lowerCamelCase_ = tmp_path_factory.mktemp("data" ) / (FILE_PATH + ".zstd") lowerCamelCase_ = bytes(lowerCamelCase__ , "utf-8" ) with zstd.open(lowerCamelCase__ , "wb" ) as f: f.write(lowerCamelCase__ ) return path @pytest.fixture def lowerCamelCase_ ( lowerCamelCase__ ): with open(os.path.join(tmpfs.local_root_dir , lowerCamelCase__ ) , "w" ) as f: f.write(lowerCamelCase__ ) return FILE_PATH @pytest.mark.parametrize("compression_format" , ["gzip", "xz", "zstd"] ) def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} lowerCamelCase_ = input_paths[compression_format] lowerCamelCase_ = tmp_path / "cache" lowerCamelCase_ = DownloadConfig(cache_dir=lowerCamelCase__ , extract_compressed_file=lowerCamelCase__ ) lowerCamelCase_ = cached_path(lowerCamelCase__ , download_config=lowerCamelCase__ ) with open(lowerCamelCase__ ) as f: lowerCamelCase_ = f.read() with open(lowerCamelCase__ ) as f: lowerCamelCase_ = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("default_extracted" , [True, False] ) @pytest.mark.parametrize("default_cache_dir" , [True, False] ) def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = "custom_cache" lowerCamelCase_ = "custom_extracted_dir" lowerCamelCase_ = tmp_path / "custom_extracted_path" if default_extracted: lowerCamelCase_ = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_DIR" , lowerCamelCase__ ) monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(lowerCamelCase__ ) ) lowerCamelCase_ = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) lowerCamelCase_ = xz_file lowerCamelCase_ = ( DownloadConfig(extract_compressed_file=lowerCamelCase__ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=lowerCamelCase__ ) ) lowerCamelCase_ = cached_path(lowerCamelCase__ , download_config=lowerCamelCase__ ) assert Path(lowerCamelCase__ ).parent.parts[-2:] == expected def lowerCamelCase_ ( lowerCamelCase__ ): # absolute path lowerCamelCase_ = str(Path(lowerCamelCase__ ).resolve() ) assert cached_path(lowerCamelCase__ ) == text_file # relative path lowerCamelCase_ = str(Path(lowerCamelCase__ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(lowerCamelCase__ ) == text_file def lowerCamelCase_ ( lowerCamelCase__ ): # absolute path lowerCamelCase_ = str(tmp_path.resolve() / "__missing_file__.txt" ) with pytest.raises(lowerCamelCase__ ): cached_path(lowerCamelCase__ ) # relative path lowerCamelCase_ = "./__missing_file__.txt" with pytest.raises(lowerCamelCase__ ): cached_path(lowerCamelCase__ ) def lowerCamelCase_ ( lowerCamelCase__ ): lowerCamelCase_ = get_from_cache(F'tmp://{tmpfs_file}' ) with open(lowerCamelCase__ ) as f: lowerCamelCase_ = f.read() assert output_file_content == FILE_CONTENT @patch("datasets.config.HF_DATASETS_OFFLINE" , lowerCamelCase__ ) def lowerCamelCase_ ( ): with pytest.raises(lowerCamelCase__ ): cached_path("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , lowerCamelCase__ ) def lowerCamelCase_ ( lowerCamelCase__ ): lowerCamelCase_ = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(lowerCamelCase__ ): http_get("https://huggingface.co" , temp_file=lowerCamelCase__ ) with pytest.raises(lowerCamelCase__ ): http_head("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , lowerCamelCase__ ) def lowerCamelCase_ ( lowerCamelCase__ ): lowerCamelCase_ = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(lowerCamelCase__ ): ftp_get("ftp://huggingface.co" , temp_file=lowerCamelCase__ ) with pytest.raises(lowerCamelCase__ ): ftp_head("ftp://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , lowerCamelCase__ ) def lowerCamelCase_ ( lowerCamelCase__ ): lowerCamelCase_ = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(lowerCamelCase__ ): fsspec_get("s3://huggingface.co" , temp_file=lowerCamelCase__ ) with pytest.raises(lowerCamelCase__ ): fsspec_head("s3://huggingface.co" )
47
from __future__ import annotations import math def lowerCamelCase_ ( lowerCamelCase__ ): if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(lowerCamelCase__ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def lowerCamelCase_ ( lowerCamelCase__ ): lowerCamelCase_ = str(lowerCamelCase__ ) lowerCamelCase_ = [n] for i in range(1 , len(lowerCamelCase__ ) ): list_nums.append(int(str_num[i:] ) ) list_nums.append(int(str_num[:-i] ) ) return list_nums def lowerCamelCase_ ( lowerCamelCase__ ): if len(str(lowerCamelCase__ ) ) > 3: if not is_prime(int(str(lowerCamelCase__ )[-3:] ) ) or not is_prime(int(str(lowerCamelCase__ )[:3] ) ): return False return True def lowerCamelCase_ ( lowerCamelCase__ = 1_1 ): lowerCamelCase_ = [] lowerCamelCase_ = 1_3 while len(lowerCamelCase__ ) != count: if validate(lowerCamelCase__ ): lowerCamelCase_ = list_truncated_nums(lowerCamelCase__ ) if all(is_prime(lowerCamelCase__ ) for i in list_nums ): list_truncated_primes.append(lowerCamelCase__ ) num += 2 return list_truncated_primes def lowerCamelCase_ ( ): return sum(compute_truncated_primes(1_1 ) ) if __name__ == "__main__": print(F"""{sum(compute_truncated_primes(1_1)) = }""")
47
1
import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class __lowerCAmelCase ( A__ , unittest.TestCase): _a = '''hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline''' def SCREAMING_SNAKE_CASE ( self: Tuple , _lowerCAmelCase: List[Any]=0 ): lowercase :Any = floats_tensor((1, 3, 1_28, 1_28) , rng=random.Random(_lowerCAmelCase ) ) lowercase :Dict = np.random.RandomState(_lowerCAmelCase ) lowercase :List[Any] = { "prompt": "A painting of a squirrel eating a burger", "image": image, "generator": generator, "num_inference_steps": 3, "strength": 0.75, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def SCREAMING_SNAKE_CASE ( self: Dict ): lowercase :Any = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) lowercase :Any = self.get_dummy_inputs() lowercase :Optional[Any] = pipe(**_lowerCAmelCase ).images lowercase :Union[str, Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 1_28, 1_28, 3) lowercase :Optional[int] = np.array([0.6_96_43, 0.5_84_84, 0.5_03_14, 0.5_87_60, 0.5_53_68, 0.5_96_43, 0.5_15_29, 0.4_12_17, 0.4_90_87] ) assert np.abs(image_slice - expected_slice ).max() < 1e-1 def SCREAMING_SNAKE_CASE ( self: Optional[int] ): lowercase :Tuple = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) lowercase :List[str] = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) lowercase :int = self.get_dummy_inputs() lowercase :Tuple = pipe(**_lowerCAmelCase ).images lowercase :List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowercase :Optional[int] = np.array([0.6_17_37, 0.5_46_42, 0.5_31_83, 0.5_44_65, 0.5_27_42, 0.6_05_25, 0.4_99_69, 0.4_06_55, 0.4_81_54] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def SCREAMING_SNAKE_CASE ( self: Optional[int] ): lowercase :Any = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) lowercase :List[Any] = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) # warmup pass to apply optimizations lowercase :int = pipe(**self.get_dummy_inputs() ) lowercase :Optional[int] = self.get_dummy_inputs() lowercase :int = pipe(**_lowerCAmelCase ).images lowercase :Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowercase :Union[str, Any] = np.array([0.5_27_61, 0.5_99_77, 0.4_90_33, 0.4_96_19, 0.5_42_82, 0.5_03_11, 0.4_76_00, 0.4_09_18, 0.4_52_03] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def SCREAMING_SNAKE_CASE ( self: List[Any] ): lowercase :Any = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) lowercase :int = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) lowercase :Optional[int] = self.get_dummy_inputs() lowercase :str = pipe(**_lowerCAmelCase ).images lowercase :Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowercase :Tuple = np.array([0.5_29_11, 0.6_00_04, 0.4_92_29, 0.4_98_05, 0.5_45_02, 0.5_06_80, 0.4_77_77, 0.4_10_28, 0.4_53_04] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def SCREAMING_SNAKE_CASE ( self: int ): lowercase :Optional[int] = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) lowercase :List[str] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) lowercase :Tuple = self.get_dummy_inputs() lowercase :Optional[Any] = pipe(**_lowerCAmelCase ).images lowercase :Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowercase :Any = np.array([0.5_29_11, 0.6_00_04, 0.4_92_29, 0.4_98_05, 0.5_45_02, 0.5_06_80, 0.4_77_77, 0.4_10_28, 0.4_53_04] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def SCREAMING_SNAKE_CASE ( self: Tuple ): lowercase :Any = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) lowercase :List[Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) lowercase :Any = self.get_dummy_inputs() lowercase :List[Any] = pipe(**_lowerCAmelCase ).images lowercase :Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 1_28, 1_28, 3) lowercase :Dict = np.array([0.6_53_31, 0.5_82_77, 0.4_82_04, 0.5_60_59, 0.5_36_65, 0.5_62_35, 0.5_09_69, 0.4_00_09, 0.4_65_52] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 @nightly @require_onnxruntime @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase): @property def SCREAMING_SNAKE_CASE ( self: Tuple ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def SCREAMING_SNAKE_CASE ( self: List[str] ): lowercase :Union[str, Any] = ort.SessionOptions() lowercase :Optional[Any] = False return options def SCREAMING_SNAKE_CASE ( self: Optional[int] ): lowercase :Dict = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) lowercase :List[Any] = init_image.resize((7_68, 5_12) ) # using the PNDM scheduler by default lowercase :int = OnnxStableDiffusionImgaImgPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="onnx" , safety_checker=_lowerCAmelCase , feature_extractor=_lowerCAmelCase , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) lowercase :Optional[int] = "A fantasy landscape, trending on artstation" lowercase :int = np.random.RandomState(0 ) lowercase :Any = pipe( prompt=_lowerCAmelCase , image=_lowerCAmelCase , strength=0.75 , guidance_scale=7.5 , num_inference_steps=10 , generator=_lowerCAmelCase , output_type="np" , ) lowercase :Any = output.images lowercase :int = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) lowercase :Optional[int] = np.array([0.49_09, 0.50_59, 0.53_72, 0.46_23, 0.48_76, 0.50_49, 0.48_20, 0.49_56, 0.50_19] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def SCREAMING_SNAKE_CASE ( self: List[Any] ): lowercase :Optional[int] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) lowercase :Tuple = init_image.resize((7_68, 5_12) ) lowercase :List[str] = LMSDiscreteScheduler.from_pretrained( "runwayml/stable-diffusion-v1-5" , subfolder="scheduler" , revision="onnx" ) lowercase :int = OnnxStableDiffusionImgaImgPipeline.from_pretrained( "runwayml/stable-diffusion-v1-5" , revision="onnx" , scheduler=_lowerCAmelCase , safety_checker=_lowerCAmelCase , feature_extractor=_lowerCAmelCase , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) lowercase :Union[str, Any] = "A fantasy landscape, trending on artstation" lowercase :Optional[int] = np.random.RandomState(0 ) lowercase :Optional[int] = pipe( prompt=_lowerCAmelCase , image=_lowerCAmelCase , strength=0.75 , guidance_scale=7.5 , num_inference_steps=20 , generator=_lowerCAmelCase , output_type="np" , ) lowercase :Optional[int] = output.images lowercase :Dict = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 7_68, 3) lowercase :Tuple = np.array([0.80_43, 0.9_26, 0.95_81, 0.81_19, 0.89_54, 0.9_13, 0.72_09, 0.74_63, 0.74_31] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2
236
import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) a_ : Tuple = logging.getLogger() def lowerCamelCase__ (): SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument('-f') SCREAMING_SNAKE_CASE = parser.parse_args() return args.f class _snake_case ( A__ ): def SCREAMING_SNAKE_CASE__ ( self) -> None: SCREAMING_SNAKE_CASE = logging.StreamHandler(sys.stdout) logger.addHandler(a) def SCREAMING_SNAKE_CASE__ ( self , a) -> str: SCREAMING_SNAKE_CASE = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , 'run_glue_deebert.py') with patch.object(a , 'argv' , a): SCREAMING_SNAKE_CASE = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(a , 0.6_66) @slow @require_torch_non_multi_gpu def SCREAMING_SNAKE_CASE__ ( self) -> Optional[Any]: SCREAMING_SNAKE_CASE = '\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n '.split() self.run_and_check(a) SCREAMING_SNAKE_CASE = '\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n '.split() self.run_and_check(a) SCREAMING_SNAKE_CASE = '\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n '.split() self.run_and_check(a)
137
0
"""simple docstring""" def lowerCAmelCase__ ( _UpperCamelCase : str , _UpperCamelCase : Any , _UpperCamelCase : List[str] , _UpperCamelCase : List[str] ) -> Dict: """simple docstring""" snake_case = [False] * len(__lowerCamelCase ) snake_case = [] queue.append(__lowerCamelCase ) snake_case = True while queue: snake_case = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(__lowerCamelCase ) snake_case = True snake_case = u return visited[t] def lowerCAmelCase__ ( _UpperCamelCase : str , _UpperCamelCase : int , _UpperCamelCase : Union[str, Any] ) -> Dict: """simple docstring""" snake_case = [-1] * (len(__lowerCamelCase )) snake_case = 0 while bfs(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): snake_case = float('Inf' ) snake_case = sink while s != source: # Find the minimum value in select path snake_case = min(__lowerCamelCase , graph[parent[s]][s] ) snake_case = parent[s] max_flow += path_flow snake_case = sink while v != source: snake_case = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow snake_case = parent[v] return max_flow SCREAMING_SNAKE_CASE__ = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = 0, 5 print(ford_fulkerson(graph, source, sink))
351
"""simple docstring""" import os import sys import unittest SCREAMING_SNAKE_CASE__ = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, create_dummy_object, find_backend, read_init # noqa: E402 # Align TRANSFORMERS_PATH in check_dummies with the current path SCREAMING_SNAKE_CASE__ = os.path.join(git_repo_path, "src", "transformers") SCREAMING_SNAKE_CASE__ = "\n{0} = None\n" SCREAMING_SNAKE_CASE__ = "\nclass {0}(metaclass=DummyObject):\n _backends = {1}\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, {1})\n" SCREAMING_SNAKE_CASE__ = "\ndef {0}(*args, **kwargs):\n requires_backends({0}, {1})\n" class lowerCAmelCase_ ( unittest.TestCase ): """simple docstring""" def snake_case ( self ): """simple docstring""" snake_case = find_backend(' _import_structure["models.albert"].append("AlbertTokenizerFast")' ) self.assertIsNone(lowerCAmelCase ) snake_case = find_backend(' if not is_tokenizers_available():' ) self.assertEqual(lowerCAmelCase , 'tokenizers' ) snake_case = find_backend(' if not is_tensorflow_text_available():' ) self.assertEqual(lowerCAmelCase , 'tensorflow_text' ) snake_case = find_backend(' if not (is_sentencepiece_available() and is_tokenizers_available()):' ) self.assertEqual(lowerCAmelCase , 'sentencepiece_and_tokenizers' ) snake_case = find_backend( ' if not (is_sentencepiece_available() and is_tensorflow_text_available()):' ) self.assertEqual(lowerCAmelCase , 'sentencepiece_and_tensorflow_text' ) snake_case = find_backend( ' if not (is_sentencepiece_available() and is_tokenizers_available() and is_vision_available()):' ) self.assertEqual(lowerCAmelCase , 'sentencepiece_and_tokenizers_and_vision' ) def snake_case ( self ): """simple docstring""" snake_case = read_init() # We don't assert on the exact list of keys to allow for smooth grow of backend-specific objects self.assertIn('torch' , lowerCAmelCase ) self.assertIn('tensorflow_text' , lowerCAmelCase ) self.assertIn('sentencepiece_and_tokenizers' , lowerCAmelCase ) # Likewise, we can't assert on the exact content of a key self.assertIn('BertModel' , objects['torch'] ) self.assertIn('TFBertModel' , objects['tf'] ) self.assertIn('FlaxBertModel' , objects['flax'] ) self.assertIn('BertModel' , objects['torch'] ) self.assertIn('TFBertTokenizer' , objects['tensorflow_text'] ) self.assertIn('convert_slow_tokenizer' , objects['sentencepiece_and_tokenizers'] ) def snake_case ( self ): """simple docstring""" snake_case = create_dummy_object('CONSTANT' , '\'torch\'' ) self.assertEqual(lowerCAmelCase , '\nCONSTANT = None\n' ) snake_case = create_dummy_object('function' , '\'torch\'' ) self.assertEqual( lowerCAmelCase , '\ndef function(*args, **kwargs):\n requires_backends(function, \'torch\')\n' ) snake_case = '\nclass FakeClass(metaclass=DummyObject):\n _backends = \'torch\'\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, \'torch\')\n' snake_case = create_dummy_object('FakeClass' , '\'torch\'' ) self.assertEqual(lowerCAmelCase , lowerCAmelCase ) def snake_case ( self ): """simple docstring""" snake_case = '# This file is autogenerated by the command `make fix-copies`, do not edit.\nfrom ..utils import DummyObject, requires_backends\n\n\nCONSTANT = None\n\n\ndef function(*args, **kwargs):\n requires_backends(function, ["torch"])\n\n\nclass FakeClass(metaclass=DummyObject):\n _backends = ["torch"]\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, ["torch"])\n' snake_case = create_dummy_files({'torch': ['CONSTANT', 'function', 'FakeClass']} ) self.assertEqual(dummy_files['torch'] , lowerCAmelCase )
149
0
def _A ( _lowercase ) -> list: """simple docstring""" def merge(_lowercase , _lowercase ) -> list: def _merge(): while left and right: yield (left if left[0] <= right[0] else right).pop(0 ) yield from left yield from right return list(_merge() ) if len(_lowercase ) <= 1: return collection __UpperCamelCase = len(_lowercase ) // 2 return merge(merge_sort(collection[:mid] ) , merge_sort(collection[mid:] ) ) if __name__ == "__main__": import doctest doctest.testmod() __snake_case = input('''Enter numbers separated by a comma:\n''').strip() __snake_case = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
310
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __snake_case = logging.get_logger(__name__) __snake_case = { '''xlm-roberta-base''': '''https://huggingface.co/xlm-roberta-base/resolve/main/config.json''', '''xlm-roberta-large''': '''https://huggingface.co/xlm-roberta-large/resolve/main/config.json''', '''xlm-roberta-large-finetuned-conll02-dutch''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll02-spanish''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll03-english''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll03-german''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/config.json''' ), } class __lowerCamelCase (_a ): _lowercase = """xlm-roberta""" def __init__( self: Union[str, Any],A_: Union[str, Any]=3_0522,A_: Dict=768,A_: Union[str, Any]=12,A_: Any=12,A_: str=3072,A_: Union[str, Any]="gelu",A_: str=0.1,A_: Optional[int]=0.1,A_: List[Any]=512,A_: Optional[Any]=2,A_: Dict=0.0_2,A_: List[Any]=1E-12,A_: Optional[int]=1,A_: str=0,A_: str=2,A_: Optional[Any]="absolute",A_: Union[str, Any]=True,A_: int=None,**A_: Optional[Any],): '''simple docstring''' super().__init__(pad_token_id=A_,bos_token_id=A_,eos_token_id=A_,**A_ ) __UpperCamelCase = vocab_size __UpperCamelCase = hidden_size __UpperCamelCase = num_hidden_layers __UpperCamelCase = num_attention_heads __UpperCamelCase = hidden_act __UpperCamelCase = intermediate_size __UpperCamelCase = hidden_dropout_prob __UpperCamelCase = attention_probs_dropout_prob __UpperCamelCase = max_position_embeddings __UpperCamelCase = type_vocab_size __UpperCamelCase = initializer_range __UpperCamelCase = layer_norm_eps __UpperCamelCase = position_embedding_type __UpperCamelCase = use_cache __UpperCamelCase = classifier_dropout class __lowerCamelCase (_a ): @property def snake_case_ ( self: Optional[Any] ): '''simple docstring''' if self.task == "multiple-choice": __UpperCamelCase = {0: 'batch', 1: 'choice', 2: 'sequence'} else: __UpperCamelCase = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ] )
310
1
"""simple docstring""" import os from typing import Optional import fsspec from fsspec.archive import AbstractArchiveFileSystem from fsspec.utils import DEFAULT_BLOCK_SIZE class _snake_case ( a__ ): snake_case__ = "" snake_case__ = ( None # protocol passed in prefix to the url. ex: "gzip", for gzip://file.txt::http://foo.bar/file.txt.gz ) snake_case__ = None # compression type in fsspec. ex: "gzip" snake_case__ = None # extension of the filename to strip. ex: "".gz" to get file.txt from file.txt.gz def __init__( self : Tuple , UpperCAmelCase : str = "" , UpperCAmelCase : Optional[str] = None , UpperCAmelCase : Optional[dict] = None , **UpperCAmelCase : Dict ): super().__init__(self , **UpperCAmelCase ) # always open as "rb" since fsspec can then use the TextIOWrapper to make it work for "r" mode __lowerCamelCase : int = fsspec.open( UpperCAmelCase , mode="rb" , protocol=UpperCAmelCase , compression=self.compression , client_kwargs={ "requote_redirect_url": False, # see https://github.com/huggingface/datasets/pull/5459 "trust_env": True, # Enable reading proxy env variables. **(target_options or {}).pop("client_kwargs" , {} ), # To avoid issues if it was already passed. } , **(target_options or {}) , ) __lowerCamelCase : Optional[int] = os.path.basename(self.file.path.split("::" )[0] ) __lowerCamelCase : Union[str, Any] = ( self.compressed_name[: self.compressed_name.rindex("." )] if "." in self.compressed_name else self.compressed_name ) __lowerCamelCase : Union[str, Any] = None @classmethod def lowerCamelCase__ ( cls : Dict , UpperCAmelCase : List[str] ): # compressed file paths are always relative to the archive root return super()._strip_protocol(UpperCAmelCase ).lstrip("/" ) def lowerCamelCase__ ( self : Union[str, Any] ): if self.dir_cache is None: __lowerCamelCase : int = {**self.file.fs.info(self.file.path ), "name": self.uncompressed_name} __lowerCamelCase : str = {f["name"]: f} def lowerCamelCase__ ( self : List[str] , UpperCAmelCase : str ): return self.file.open().read() def lowerCamelCase__ ( self : Union[str, Any] , UpperCAmelCase : str , UpperCAmelCase : str = "rb" , UpperCAmelCase : Tuple=None , UpperCAmelCase : int=True , UpperCAmelCase : List[str]=None , **UpperCAmelCase : str , ): __lowerCamelCase : Dict = self._strip_protocol(UpperCAmelCase ) if mode != "rb": raise ValueError(F"""Tried to read with mode {mode} on file {self.file.path} opened with mode 'rb'""" ) return self.file.open() class _snake_case ( a__ ): snake_case__ = "bz2" snake_case__ = "bz2" snake_case__ = ".bz2" class _snake_case ( a__ ): snake_case__ = "gzip" snake_case__ = "gzip" snake_case__ = ".gz" class _snake_case ( a__ ): snake_case__ = "lz4" snake_case__ = "lz4" snake_case__ = ".lz4" class _snake_case ( a__ ): snake_case__ = "xz" snake_case__ = "xz" snake_case__ = ".xz" class _snake_case ( a__ ): snake_case__ = "zstd" snake_case__ = "zstd" snake_case__ = ".zst" def __init__( self : List[Any] , UpperCAmelCase : str , UpperCAmelCase : str = "rb" , UpperCAmelCase : Optional[str] = None , UpperCAmelCase : Optional[dict] = None , UpperCAmelCase : int = DEFAULT_BLOCK_SIZE , **UpperCAmelCase : List[str] , ): super().__init__( fo=UpperCAmelCase , mode=UpperCAmelCase , target_protocol=UpperCAmelCase , target_options=UpperCAmelCase , block_size=UpperCAmelCase , **UpperCAmelCase , ) # We need to wrap the zstd decompressor to avoid this error in fsspec==2021.7.0 and zstandard==0.15.2: # # File "/Users/user/.virtualenvs/hf-datasets/lib/python3.7/site-packages/fsspec/core.py", line 145, in open # out.close = close # AttributeError: 'zstd.ZstdDecompressionReader' object attribute 'close' is read-only # # see https://github.com/intake/filesystem_spec/issues/725 __lowerCamelCase : Union[str, Any] = self.file.__enter__ class _snake_case : def __init__( self : Tuple , UpperCAmelCase : str ): __lowerCamelCase : List[str] = file_ def __enter__( self : Optional[int] ): self._file.__enter__() return self def __exit__( self : Tuple , *UpperCAmelCase : Optional[int] , **UpperCAmelCase : List[Any] ): self._file.__exit__(*UpperCAmelCase , **UpperCAmelCase ) def __iter__( self : Optional[Any] ): return iter(self._file ) def lowerCamelCase__ ( self : str ): return next(self._file ) def __getattr__( self : Union[str, Any] , UpperCAmelCase : Tuple ): return getattr(self._file , UpperCAmelCase ) def fixed_enter(*UpperCAmelCase : Union[str, Any] , **UpperCAmelCase : str ): return WrappedFile(_enter(*UpperCAmelCase , **UpperCAmelCase ) ) __lowerCamelCase : List[Any] = fixed_enter
64
"""simple docstring""" import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all image processors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...image_processing_utils import ImageProcessingMixin from ...utils import CONFIG_NAME, IMAGE_PROCESSOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) __A = logging.get_logger(__name__) __A = OrderedDict( [ ('''align''', '''EfficientNetImageProcessor'''), ('''beit''', '''BeitImageProcessor'''), ('''bit''', '''BitImageProcessor'''), ('''blip''', '''BlipImageProcessor'''), ('''blip-2''', '''BlipImageProcessor'''), ('''bridgetower''', '''BridgeTowerImageProcessor'''), ('''chinese_clip''', '''ChineseCLIPImageProcessor'''), ('''clip''', '''CLIPImageProcessor'''), ('''clipseg''', '''ViTImageProcessor'''), ('''conditional_detr''', '''ConditionalDetrImageProcessor'''), ('''convnext''', '''ConvNextImageProcessor'''), ('''convnextv2''', '''ConvNextImageProcessor'''), ('''cvt''', '''ConvNextImageProcessor'''), ('''data2vec-vision''', '''BeitImageProcessor'''), ('''deformable_detr''', '''DeformableDetrImageProcessor'''), ('''deit''', '''DeiTImageProcessor'''), ('''deta''', '''DetaImageProcessor'''), ('''detr''', '''DetrImageProcessor'''), ('''dinat''', '''ViTImageProcessor'''), ('''donut-swin''', '''DonutImageProcessor'''), ('''dpt''', '''DPTImageProcessor'''), ('''efficientformer''', '''EfficientFormerImageProcessor'''), ('''efficientnet''', '''EfficientNetImageProcessor'''), ('''flava''', '''FlavaImageProcessor'''), ('''focalnet''', '''BitImageProcessor'''), ('''git''', '''CLIPImageProcessor'''), ('''glpn''', '''GLPNImageProcessor'''), ('''groupvit''', '''CLIPImageProcessor'''), ('''imagegpt''', '''ImageGPTImageProcessor'''), ('''instructblip''', '''BlipImageProcessor'''), ('''layoutlmv2''', '''LayoutLMv2ImageProcessor'''), ('''layoutlmv3''', '''LayoutLMv3ImageProcessor'''), ('''levit''', '''LevitImageProcessor'''), ('''mask2former''', '''Mask2FormerImageProcessor'''), ('''maskformer''', '''MaskFormerImageProcessor'''), ('''mgp-str''', '''ViTImageProcessor'''), ('''mobilenet_v1''', '''MobileNetV1ImageProcessor'''), ('''mobilenet_v2''', '''MobileNetV2ImageProcessor'''), ('''mobilevit''', '''MobileViTImageProcessor'''), ('''mobilevit''', '''MobileViTImageProcessor'''), ('''mobilevitv2''', '''MobileViTImageProcessor'''), ('''nat''', '''ViTImageProcessor'''), ('''oneformer''', '''OneFormerImageProcessor'''), ('''owlvit''', '''OwlViTImageProcessor'''), ('''perceiver''', '''PerceiverImageProcessor'''), ('''pix2struct''', '''Pix2StructImageProcessor'''), ('''poolformer''', '''PoolFormerImageProcessor'''), ('''regnet''', '''ConvNextImageProcessor'''), ('''resnet''', '''ConvNextImageProcessor'''), ('''sam''', '''SamImageProcessor'''), ('''segformer''', '''SegformerImageProcessor'''), ('''swiftformer''', '''ViTImageProcessor'''), ('''swin''', '''ViTImageProcessor'''), ('''swin2sr''', '''Swin2SRImageProcessor'''), ('''swinv2''', '''ViTImageProcessor'''), ('''table-transformer''', '''DetrImageProcessor'''), ('''timesformer''', '''VideoMAEImageProcessor'''), ('''tvlt''', '''TvltImageProcessor'''), ('''upernet''', '''SegformerImageProcessor'''), ('''van''', '''ConvNextImageProcessor'''), ('''videomae''', '''VideoMAEImageProcessor'''), ('''vilt''', '''ViltImageProcessor'''), ('''vit''', '''ViTImageProcessor'''), ('''vit_hybrid''', '''ViTHybridImageProcessor'''), ('''vit_mae''', '''ViTImageProcessor'''), ('''vit_msn''', '''ViTImageProcessor'''), ('''xclip''', '''CLIPImageProcessor'''), ('''yolos''', '''YolosImageProcessor'''), ] ) __A = _LazyAutoMapping(CONFIG_MAPPING_NAMES, IMAGE_PROCESSOR_MAPPING_NAMES) def lowercase_ ( _lowerCamelCase: str ) -> int: '''simple docstring''' for module_name, extractors in IMAGE_PROCESSOR_MAPPING_NAMES.items(): if class_name in extractors: __lowerCamelCase : int = model_type_to_module_name(_lowerCamelCase ) __lowerCamelCase : Union[str, Any] = importlib.import_module(F""".{module_name}""" , "transformers.models" ) try: return getattr(_lowerCamelCase , _lowerCamelCase ) except AttributeError: continue for _, extractor in IMAGE_PROCESSOR_MAPPING._extra_content.items(): if getattr(_lowerCamelCase , "__name__" , _lowerCamelCase ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. __lowerCamelCase : int = importlib.import_module("transformers" ) if hasattr(_lowerCamelCase , _lowerCamelCase ): return getattr(_lowerCamelCase , _lowerCamelCase ) return None def lowercase_ ( _lowerCamelCase: Union[str, os.PathLike] , _lowerCamelCase: Optional[Union[str, os.PathLike]] = None , _lowerCamelCase: bool = False , _lowerCamelCase: bool = False , _lowerCamelCase: Optional[Dict[str, str]] = None , _lowerCamelCase: Optional[Union[bool, str]] = None , _lowerCamelCase: Optional[str] = None , _lowerCamelCase: bool = False , **_lowerCamelCase: Tuple , ) -> List[str]: '''simple docstring''' __lowerCamelCase : List[str] = get_file_from_repo( _lowerCamelCase , _lowerCamelCase , cache_dir=_lowerCamelCase , force_download=_lowerCamelCase , resume_download=_lowerCamelCase , proxies=_lowerCamelCase , use_auth_token=_lowerCamelCase , revision=_lowerCamelCase , local_files_only=_lowerCamelCase , ) if resolved_config_file is None: logger.info( "Could not locate the image processor configuration file, will try to use the model config instead." ) return {} with open(_lowerCamelCase , encoding="utf-8" ) as reader: return json.load(_lowerCamelCase ) class _snake_case : def __init__( self : Tuple ): raise EnvironmentError( "AutoImageProcessor is designed to be instantiated " "using the `AutoImageProcessor.from_pretrained(pretrained_model_name_or_path)` method." ) @classmethod @replace_list_option_in_docstrings(UpperCAmelCase ) def lowerCamelCase__ ( cls : Dict , UpperCAmelCase : Optional[int] , **UpperCAmelCase : Any ): __lowerCamelCase : int = kwargs.pop("config" , UpperCAmelCase ) __lowerCamelCase : Dict = kwargs.pop("trust_remote_code" , UpperCAmelCase ) __lowerCamelCase : Any = True __lowerCamelCase , __lowerCamelCase : str = ImageProcessingMixin.get_image_processor_dict(UpperCAmelCase , **UpperCAmelCase ) __lowerCamelCase : Optional[int] = config_dict.get("image_processor_type" , UpperCAmelCase ) __lowerCamelCase : List[Any] = None if "AutoImageProcessor" in config_dict.get("auto_map" , {} ): __lowerCamelCase : List[str] = config_dict["auto_map"]["AutoImageProcessor"] # If we still don't have the image processor class, check if we're loading from a previous feature extractor config # and if so, infer the image processor class from there. if image_processor_class is None and image_processor_auto_map is None: __lowerCamelCase : Dict = config_dict.pop("feature_extractor_type" , UpperCAmelCase ) if feature_extractor_class is not None: logger.warning( "Could not find image processor class in the image processor config or the model config. Loading" " based on pattern matching with the model's feature extractor configuration." ) __lowerCamelCase : Tuple = feature_extractor_class.replace("FeatureExtractor" , "ImageProcessor" ) if "AutoFeatureExtractor" in config_dict.get("auto_map" , {} ): __lowerCamelCase : Any = config_dict["auto_map"]["AutoFeatureExtractor"] __lowerCamelCase : Optional[int] = feature_extractor_auto_map.replace("FeatureExtractor" , "ImageProcessor" ) logger.warning( "Could not find image processor auto map in the image processor config or the model config." " Loading based on pattern matching with the model's feature extractor configuration." ) # If we don't find the image processor class in the image processor config, let's try the model config. if image_processor_class is None and image_processor_auto_map is None: if not isinstance(UpperCAmelCase , UpperCAmelCase ): __lowerCamelCase : int = AutoConfig.from_pretrained(UpperCAmelCase , **UpperCAmelCase ) # It could be in `config.image_processor_type`` __lowerCamelCase : int = getattr(UpperCAmelCase , "image_processor_type" , UpperCAmelCase ) if hasattr(UpperCAmelCase , "auto_map" ) and "AutoImageProcessor" in config.auto_map: __lowerCamelCase : Optional[int] = config.auto_map["AutoImageProcessor"] if image_processor_class is not None: __lowerCamelCase : Any = image_processor_class_from_name(UpperCAmelCase ) __lowerCamelCase : str = image_processor_auto_map is not None __lowerCamelCase : Optional[Any] = image_processor_class is not None or type(UpperCAmelCase ) in IMAGE_PROCESSOR_MAPPING __lowerCamelCase : Dict = resolve_trust_remote_code( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) if has_remote_code and trust_remote_code: __lowerCamelCase : Optional[Any] = get_class_from_dynamic_module( UpperCAmelCase , UpperCAmelCase , **UpperCAmelCase ) __lowerCamelCase : List[Any] = kwargs.pop("code_revision" , UpperCAmelCase ) if os.path.isdir(UpperCAmelCase ): image_processor_class.register_for_auto_class() return image_processor_class.from_dict(UpperCAmelCase , **UpperCAmelCase ) elif image_processor_class is not None: return image_processor_class.from_dict(UpperCAmelCase , **UpperCAmelCase ) # Last try: we use the IMAGE_PROCESSOR_MAPPING. elif type(UpperCAmelCase ) in IMAGE_PROCESSOR_MAPPING: __lowerCamelCase : Tuple = IMAGE_PROCESSOR_MAPPING[type(UpperCAmelCase )] return image_processor_class.from_dict(UpperCAmelCase , **UpperCAmelCase ) raise ValueError( F"""Unrecognized image processor in {pretrained_model_name_or_path}. Should have a """ F"""`image_processor_type` key in its {IMAGE_PROCESSOR_NAME} of {CONFIG_NAME}, or one of the following """ F"""`model_type` keys in its {CONFIG_NAME}: {", ".join(c for c in IMAGE_PROCESSOR_MAPPING_NAMES.keys() )}""" ) @staticmethod def lowerCamelCase__ ( UpperCAmelCase : Any , UpperCAmelCase : Union[str, Any] ): IMAGE_PROCESSOR_MAPPING.register(UpperCAmelCase , UpperCAmelCase )
64
1
"""simple docstring""" from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging __magic_name__ = logging.get_logger(__name__) __magic_name__ = { "huggingface/time-series-transformer-tourism-monthly": ( "https://huggingface.co/huggingface/time-series-transformer-tourism-monthly/resolve/main/config.json" ), # See all TimeSeriesTransformer models at https://huggingface.co/models?filter=time_series_transformer } class SCREAMING_SNAKE_CASE_ ( __a ): """simple docstring""" __lowercase : Tuple = '''time_series_transformer''' __lowercase : List[str] = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', '''num_hidden_layers''': '''encoder_layers''', } def __init__( self , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = "student_t" , lowerCAmelCase__ = "nll" , lowerCAmelCase__ = 1 , lowerCAmelCase__ = [1, 2, 3, 4, 5, 6, 7] , lowerCAmelCase__ = "mean" , lowerCAmelCase__ = 0 , lowerCAmelCase__ = 0 , lowerCAmelCase__ = 0 , lowerCAmelCase__ = 0 , lowerCAmelCase__ = None , lowerCAmelCase__ = None , lowerCAmelCase__ = 3_2 , lowerCAmelCase__ = 3_2 , lowerCAmelCase__ = 2 , lowerCAmelCase__ = 2 , lowerCAmelCase__ = 2 , lowerCAmelCase__ = 2 , lowerCAmelCase__ = True , lowerCAmelCase__ = "gelu" , lowerCAmelCase__ = 6_4 , lowerCAmelCase__ = 0.1 , lowerCAmelCase__ = 0.1 , lowerCAmelCase__ = 0.1 , lowerCAmelCase__ = 0.1 , lowerCAmelCase__ = 0.1 , lowerCAmelCase__ = 1_0_0 , lowerCAmelCase__ = 0.02 , lowerCAmelCase__=True , **lowerCAmelCase__ , ): # time series specific configuration __SCREAMING_SNAKE_CASE = prediction_length __SCREAMING_SNAKE_CASE = context_length or prediction_length __SCREAMING_SNAKE_CASE = distribution_output __SCREAMING_SNAKE_CASE = loss __SCREAMING_SNAKE_CASE = input_size __SCREAMING_SNAKE_CASE = num_time_features __SCREAMING_SNAKE_CASE = lags_sequence __SCREAMING_SNAKE_CASE = scaling __SCREAMING_SNAKE_CASE = num_dynamic_real_features __SCREAMING_SNAKE_CASE = num_static_real_features __SCREAMING_SNAKE_CASE = num_static_categorical_features if cardinality and num_static_categorical_features > 0: if len(lowerCAmelCase__) != num_static_categorical_features: raise ValueError( """The cardinality should be a list of the same length as `num_static_categorical_features`""") __SCREAMING_SNAKE_CASE = cardinality else: __SCREAMING_SNAKE_CASE = [0] if embedding_dimension and num_static_categorical_features > 0: if len(lowerCAmelCase__) != num_static_categorical_features: raise ValueError( """The embedding dimension should be a list of the same length as `num_static_categorical_features`""") __SCREAMING_SNAKE_CASE = embedding_dimension else: __SCREAMING_SNAKE_CASE = [min(5_0 , (cat + 1) // 2) for cat in self.cardinality] __SCREAMING_SNAKE_CASE = num_parallel_samples # Transformer architecture configuration __SCREAMING_SNAKE_CASE = input_size * len(lowerCAmelCase__) + self._number_of_features __SCREAMING_SNAKE_CASE = d_model __SCREAMING_SNAKE_CASE = encoder_attention_heads __SCREAMING_SNAKE_CASE = decoder_attention_heads __SCREAMING_SNAKE_CASE = encoder_ffn_dim __SCREAMING_SNAKE_CASE = decoder_ffn_dim __SCREAMING_SNAKE_CASE = encoder_layers __SCREAMING_SNAKE_CASE = decoder_layers __SCREAMING_SNAKE_CASE = dropout __SCREAMING_SNAKE_CASE = attention_dropout __SCREAMING_SNAKE_CASE = activation_dropout __SCREAMING_SNAKE_CASE = encoder_layerdrop __SCREAMING_SNAKE_CASE = decoder_layerdrop __SCREAMING_SNAKE_CASE = activation_function __SCREAMING_SNAKE_CASE = init_std __SCREAMING_SNAKE_CASE = use_cache super().__init__(is_encoder_decoder=lowerCAmelCase__ , **lowerCAmelCase__) @property def snake_case_ ( self): return ( sum(self.embedding_dimension) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
100
import argparse import os import re import torch from flax.traverse_util import flatten_dict from tax import checkpoints from transformers import ( AutoTokenizer, PixaStructConfig, PixaStructForConditionalGeneration, PixaStructImageProcessor, PixaStructProcessor, PixaStructTextConfig, PixaStructVisionConfig, ) def __UpperCamelCase ( _A ): lowerCAmelCase_ = checkpoints.load_tax_checkpoint(_A ) lowerCAmelCase_ = flatten_dict(_A ) return flax_params def __UpperCamelCase ( _A ): lowerCAmelCase_ = {} lowerCAmelCase_ = { '''token_embedder''': '''embeddings''', '''encoder_norm''': '''layernorm''', '''kernel''': '''weight''', '''.out''': '''.output''', '''scale''': '''weight''', '''embedders_0.pos_embedding''': '''row_embedder.weight''', '''embedders_1.pos_embedding''': '''column_embedder.weight''', } lowerCAmelCase_ = { '''query''': '''attention.query''', '''key''': '''attention.key''', '''value''': '''attention.value''', '''output.dense''': '''output''', '''encoder_decoder_attention.o''': '''encoder_decoder_attention.attention.o''', '''pre_self_attention_layer_norm''': '''self_attention.layer_norm''', '''pre_cross_attention_layer_norm''': '''encoder_decoder_attention.layer_norm''', '''mlp.''': '''mlp.DenseReluDense.''', '''pre_mlp_layer_norm''': '''mlp.layer_norm''', '''self_attention.o''': '''self_attention.attention.o''', '''decoder.embeddings.embedding''': '''decoder.embed_tokens.weight''', '''decoder.relpos_bias.rel_embedding''': '''decoder.layer.0.self_attention.attention.relative_attention_bias.weight''', '''decoder.decoder_norm.weight''': '''decoder.final_layer_norm.weight''', '''decoder.logits_dense.weight''': '''decoder.lm_head.weight''', } for key in flax_dict.keys(): if "target" in key: # remove the first prefix from the key lowerCAmelCase_ = '''.'''.join(key[1:] ) # rename the key for old, new in CONVERSION_MAPPING.items(): lowerCAmelCase_ = new_key.replace(_A , _A ) if "decoder" in new_key: for old, new in DECODER_CONVERSION_MAPPING.items(): lowerCAmelCase_ = new_key.replace(_A , _A ) if "layers" in new_key and "decoder" not in new_key: # use regex to replace the layer number lowerCAmelCase_ = re.sub(r'''layers_(\d+)''' , r'''layer.\1''' , _A ) lowerCAmelCase_ = new_key.replace('''encoder''' , '''encoder.encoder''' ) elif "layers" in new_key and "decoder" in new_key: # use regex to replace the layer number lowerCAmelCase_ = re.sub(r'''layers_(\d+)''' , r'''layer.\1''' , _A ) lowerCAmelCase_ = flax_dict[key] lowerCAmelCase_ = {} # convert converted_dict into torch format for key in converted_dict.keys(): if ("embed_tokens" not in key) and ("embedder" not in key): lowerCAmelCase_ = torch.from_numpy(converted_dict[key].T ) else: lowerCAmelCase_ = torch.from_numpy(converted_dict[key] ) return converted_torch_dict def __UpperCamelCase ( _A , _A , _A=False , _A=False ): lowerCAmelCase_ = get_flax_param(_A ) if not use_large: lowerCAmelCase_ = PixaStructVisionConfig() lowerCAmelCase_ = PixaStructTextConfig() else: lowerCAmelCase_ = PixaStructVisionConfig( hidden_size=1536 , d_ff=3968 , num_attention_heads=24 , num_hidden_layers=18 ) lowerCAmelCase_ = PixaStructTextConfig(hidden_size=1536 , d_ff=3968 , num_heads=24 , num_layers=18 ) lowerCAmelCase_ = PixaStructConfig( vision_config=encoder_config.to_dict() , text_config=decoder_config.to_dict() , is_vqa=_A ) lowerCAmelCase_ = PixaStructForConditionalGeneration(_A ) lowerCAmelCase_ = rename_and_convert_flax_params(_A ) model.load_state_dict(_A ) lowerCAmelCase_ = AutoTokenizer.from_pretrained('''ybelkada/test-pix2struct-tokenizer''' ) lowerCAmelCase_ = PixaStructImageProcessor() lowerCAmelCase_ = PixaStructProcessor(image_processor=_A , tokenizer=_A ) if use_large: lowerCAmelCase_ = 4096 lowerCAmelCase_ = True # mkdir if needed os.makedirs(_A , exist_ok=_A ) model.save_pretrained(_A ) processor.save_pretrained(_A ) print('''Model saved in {}'''.format(_A ) ) if __name__ == "__main__": _A = argparse.ArgumentParser() parser.add_argument('''--t5x_checkpoint_path''', default=None, type=str, help='''Path to the original T5x checkpoint.''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--use_large''', action='''store_true''', help='''Use large model.''') parser.add_argument('''--is_vqa''', action='''store_true''', help='''Use large model.''') _A = parser.parse_args() convert_pixastruct_original_pytorch_checkpoint_to_hf( args.tax_checkpoint_path, args.pytorch_dump_folder_path, args.use_large )
278
0
'''simple docstring''' import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed lowerCAmelCase_ : str = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"{bindir}/../../examples/pytorch/translation"): from run_translation import main # noqa set_seed(42) lowerCAmelCase_ : Tuple = '''sshleifer/student_marian_en_ro_6_1''' lowerCAmelCase_ : List[str] = '''sshleifer/tiny-mbart''' @require_torch class __lowerCAmelCase ( __a ): def snake_case_ (self , lowerCAmelCase__=False , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=True , ): _UpperCAmelCase : List[Any] = self.run_trainer( eval_steps=1 , max_len=1_2 , model_name=lowerCAmelCase__ , num_train_epochs=1 , distributed=lowerCAmelCase__ , extra_args_str=lowerCAmelCase__ , predict_with_generate=lowerCAmelCase__ , do_train=lowerCAmelCase__ , do_eval=lowerCAmelCase__ , do_predict=lowerCAmelCase__ , ) _UpperCAmelCase : Any = TrainerState.load_from_json(os.path.join(lowerCAmelCase__ , """trainer_state.json""" ) ).log_history if not do_eval: return _UpperCAmelCase : Optional[int] = [log for log in logs if """eval_loss""" in log.keys()] _UpperCAmelCase : Dict = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats _UpperCAmelCase : str = eval_metrics[-1] assert isinstance(last_step_stats["""eval_bleu"""] , lowerCAmelCase__ ) assert not math.isnan(float(last_step_stats["""eval_loss"""] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def snake_case_ (self ): self.run_seqaseq_quick() @require_torch_multi_gpu def snake_case_ (self ): self.run_seqaseq_quick(distributed=lowerCAmelCase__ ) @require_torch_multi_gpu def snake_case_ (self ): self.run_seqaseq_quick(distributed=lowerCAmelCase__ ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def snake_case_ (self ): self.run_seqaseq_quick(distributed=lowerCAmelCase__ , extra_args_str="""--sharded_ddp simple""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def snake_case_ (self ): self.run_seqaseq_quick(distributed=lowerCAmelCase__ , extra_args_str="""--sharded_ddp simple --fp16""" ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def snake_case_ (self ): self.run_seqaseq_quick(distributed=lowerCAmelCase__ , extra_args_str="""--sharded_ddp zero_dp_2""" , predict_with_generate=lowerCAmelCase__ ) @unittest.skip("""Requires an update of the env running those tests""" ) @require_torch_multi_gpu @require_fairscale def snake_case_ (self ): self.run_seqaseq_quick( distributed=lowerCAmelCase__ , extra_args_str="""--sharded_ddp zero_dp_2 --fp16""" , predict_with_generate=lowerCAmelCase__ ) @require_apex @require_torch_gpu def snake_case_ (self ): # XXX: apex breaks the trainer if it's run twice e.g. run_seq2seq.main() from the same # program and it breaks other tests that run from the same pytest worker, therefore until this is # sorted out it must be run only in an external program, that is distributed=True in this # test and only under one or more gpus - if we want cpu will need to make a special test # # specifically to the problem traced it to self.optimizer.step() - if it's run 2nd time via # 2nd main() call it botches the future eval. # self.run_seqaseq_quick(distributed=lowerCAmelCase__ , extra_args_str="""--fp16 --fp16_backend=apex""" ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=lowerCAmelCase__ , extra_args_str="""--fp16 --fp16_backend=apex""" ) @parameterized.expand(["""base""", """low""", """high""", """mixed"""] ) @require_torch_multi_gpu def snake_case_ (self , lowerCAmelCase__ ): # as each sub-test is slow-ish split into multiple sub-tests to avoid CI timeout _UpperCAmelCase : Optional[int] = { # test with the default log_level - should be info and thus log info once """base""": {"""extra_args_str""": """""", """n_matches""": 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes """low""": {"""extra_args_str""": """--log_level debug --log_level_replica debug""", """n_matches""": 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica """high""": {"""extra_args_str""": """--log_level error --log_level_replica debug""", """n_matches""": 1}, # test with high log_level and log_level_replica - should be quiet on all processes """mixed""": {"""extra_args_str""": """--log_level error --log_level_replica error""", """n_matches""": 0}, } _UpperCAmelCase : Dict = experiments[experiment_id] _UpperCAmelCase : Tuple = {"""distributed""": True, """predict_with_generate""": False, """do_eval""": False, """do_predict""": False} _UpperCAmelCase : Union[str, Any] = """Running training""" with CaptureStderr() as cl: self.run_seqaseq_quick(**lowerCAmelCase__ , extra_args_str=data["""extra_args_str"""] ) _UpperCAmelCase : Union[str, Any] = len(re.findall(lowerCAmelCase__ , cl.err ) ) self.assertEqual(lowerCAmelCase__ , data["""n_matches"""] ) @slow def snake_case_ (self ): _UpperCAmelCase : Union[str, Any] = self.run_trainer( eval_steps=2 , max_len=1_2_8 , model_name=lowerCAmelCase__ , learning_rate=3e-4 , num_train_epochs=1_0 , distributed=lowerCAmelCase__ , ) # Check metrics _UpperCAmelCase : Union[str, Any] = TrainerState.load_from_json(os.path.join(lowerCAmelCase__ , """trainer_state.json""" ) ).log_history _UpperCAmelCase : Union[str, Any] = [log for log in logs if """eval_loss""" in log.keys()] _UpperCAmelCase : Tuple = eval_metrics[0] _UpperCAmelCase : Dict = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats["""eval_bleu"""] , lowerCAmelCase__ ) # test if do_predict saves generations and metrics _UpperCAmelCase : List[str] = os.listdir(lowerCAmelCase__ ) _UpperCAmelCase : Dict = {os.path.basename(lowerCAmelCase__ ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def snake_case_ (self ): from transformers.training_args import OptimizerNames def train_and_return_metrics(lowerCAmelCase__ ) -> Tuple[int, float]: _UpperCAmelCase : Tuple = """--skip_memory_metrics 0""" _UpperCAmelCase : str = self.run_trainer( max_len=1_2_8 , model_name=lowerCAmelCase__ , learning_rate=3e-4 , num_train_epochs=1 , optim=lowerCAmelCase__ , distributed=lowerCAmelCase__ , extra_args_str=lowerCAmelCase__ , do_eval=lowerCAmelCase__ , do_predict=lowerCAmelCase__ , n_gpus_to_use=1 , ) # Check metrics _UpperCAmelCase : str = TrainerState.load_from_json(Path(lowerCAmelCase__ , """trainer_state.json""" ) ).log_history _UpperCAmelCase : int = int(logs[0]["""train_mem_gpu_peaked_delta"""] / 2**2_0 ) _UpperCAmelCase : Optional[int] = int(logs[0]["""train_mem_gpu_alloc_delta"""] / 2**2_0 ) _UpperCAmelCase : str = logs[0]["""train_loss"""] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : str = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : List[Any] = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) _UpperCAmelCase : int = gpu_alloc_mem_orig - gpu_alloc_mem_bnb _UpperCAmelCase : str = gpu_peak_mem_orig + gpu_alloc_mem_orig _UpperCAmelCase : Dict = gpu_peak_mem_bnb + gpu_alloc_mem_bnb _UpperCAmelCase : Any = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings _UpperCAmelCase : Dict = 1_2_0 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( lowerCAmelCase__ , lowerCAmelCase__ , """should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got""" F" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and" F" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB" , ) self.assertGreater( lowerCAmelCase__ , lowerCAmelCase__ , """should use ~150MB less total gpu memory with BNB, compared to without it for this model but got""" F" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and" F" gpu_total_mem_bnb={gpu_total_mem_bnb}MB" , ) self.assertEqual( lowerCAmelCase__ , lowerCAmelCase__ , F"loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}" ) def snake_case_ (self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 3e-3 , lowerCAmelCase__ = "adafactor" , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = 0 , lowerCAmelCase__ = True , lowerCAmelCase__ = True , lowerCAmelCase__ = True , lowerCAmelCase__ = True , lowerCAmelCase__ = None , ): _UpperCAmelCase : Optional[Any] = self.test_file_dir / """../fixtures/tests_samples/wmt_en_ro""" _UpperCAmelCase : Any = self.get_auto_remove_tmp_dir() _UpperCAmelCase : str = F"\n --model_name_or_path {model_name}\n --train_file {data_dir}/train.json\n --validation_file {data_dir}/val.json\n --test_file {data_dir}/test.json\n --output_dir {output_dir}\n --overwrite_output_dir\n --max_train_samples 8\n --max_source_length {max_len}\n --max_target_length {max_len}\n --do_train\n --num_train_epochs {str(lowerCAmelCase__ )}\n --per_device_train_batch_size 4\n --learning_rate {learning_rate}\n --warmup_steps 8\n --logging_steps 0\n --logging_strategy no\n --save_steps {str(lowerCAmelCase__ )}\n --group_by_length\n --label_smoothing_factor 0.1\n --target_lang ro_RO\n --source_lang en_XX\n ".split() _UpperCAmelCase : Any = F"\n --do_eval\n --per_device_eval_batch_size 4\n --max_eval_samples 8\n --val_max_target_length {max_len}\n --evaluation_strategy steps\n --eval_steps {str(lowerCAmelCase__ )}\n ".split() _UpperCAmelCase : Optional[int] = """ --do_predict """.split() _UpperCAmelCase : Optional[int] = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F"--optim {optim}".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: _UpperCAmelCase : Optional[int] = get_gpu_count() _UpperCAmelCase : Dict = get_torch_dist_unique_port() _UpperCAmelCase : Union[str, Any] = F"\n -m torch.distributed.run\n --nproc_per_node={n_gpus_to_use}\n --master_port={master_port}\n {self.examples_dir_str}/pytorch/translation/run_translation.py\n ".split() _UpperCAmelCase : str = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(lowerCAmelCase__ , env=self.get_env() ) else: _UpperCAmelCase : Union[str, Any] = ["""run_translation.py"""] + args with patch.object(lowerCAmelCase__ , """argv""" , lowerCAmelCase__ ): main() return output_dir
170
'''simple docstring''' import gc import random import unittest import numpy as np import torch from diffusers import ( DDIMScheduler, KandinskyVaaControlnetPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __lowerCAmelCase ( __a , unittest.TestCase ): snake_case : Union[str, Any] = KandinskyVaaControlnetPipeline snake_case : Dict = ["""image_embeds""", """negative_image_embeds""", """hint"""] snake_case : str = ["""image_embeds""", """negative_image_embeds""", """hint"""] snake_case : Optional[int] = [ """generator""", """height""", """width""", """latents""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] snake_case : str = False @property def snake_case_ (self ): return 3_2 @property def snake_case_ (self ): return 3_2 @property def snake_case_ (self ): return self.time_input_dim @property def snake_case_ (self ): return self.time_input_dim * 4 @property def snake_case_ (self ): return 1_0_0 @property def snake_case_ (self ): torch.manual_seed(0 ) _UpperCAmelCase : str = { """in_channels""": 8, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """image_hint""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } _UpperCAmelCase : Union[str, Any] = UNetaDConditionModel(**lowerCAmelCase__ ) return model @property def snake_case_ (self ): return { "block_out_channels": [3_2, 3_2, 6_4, 6_4], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 1_2, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def snake_case_ (self ): torch.manual_seed(0 ) _UpperCAmelCase : Optional[int] = VQModel(**self.dummy_movq_kwargs ) return model def snake_case_ (self ): _UpperCAmelCase : List[Any] = self.dummy_unet _UpperCAmelCase : str = self.dummy_movq _UpperCAmelCase : Any = DDIMScheduler( num_train_timesteps=1_0_0_0 , beta_schedule="""linear""" , beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , clip_sample=lowerCAmelCase__ , set_alpha_to_one=lowerCAmelCase__ , steps_offset=1 , prediction_type="""epsilon""" , thresholding=lowerCAmelCase__ , ) _UpperCAmelCase : List[Any] = { """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def snake_case_ (self , lowerCAmelCase__ , lowerCAmelCase__=0 ): _UpperCAmelCase : Union[str, Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) _UpperCAmelCase : Optional[Any] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( lowerCAmelCase__ ) # create hint _UpperCAmelCase : List[Any] = floats_tensor((1, 3, 6_4, 6_4) , rng=random.Random(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) if str(lowerCAmelCase__ ).startswith("""mps""" ): _UpperCAmelCase : int = torch.manual_seed(lowerCAmelCase__ ) else: _UpperCAmelCase : Tuple = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) _UpperCAmelCase : Union[str, Any] = { """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """hint""": hint, """generator""": generator, """height""": 6_4, """width""": 6_4, """guidance_scale""": 4.0, """num_inference_steps""": 2, """output_type""": """np""", } return inputs def snake_case_ (self ): _UpperCAmelCase : Union[str, Any] = """cpu""" _UpperCAmelCase : List[str] = self.get_dummy_components() _UpperCAmelCase : str = self.pipeline_class(**lowerCAmelCase__ ) _UpperCAmelCase : int = pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) _UpperCAmelCase : Any = pipe(**self.get_dummy_inputs(lowerCAmelCase__ ) ) _UpperCAmelCase : str = output.images _UpperCAmelCase : Optional[int] = pipe( **self.get_dummy_inputs(lowerCAmelCase__ ) , return_dict=lowerCAmelCase__ , )[0] _UpperCAmelCase : int = image[0, -3:, -3:, -1] _UpperCAmelCase : Union[str, Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) _UpperCAmelCase : Union[str, Any] = np.array( [0.6_9_5_9_8_2_6, 0.8_6_8_2_7_9, 0.7_5_5_8_0_9_2, 0.6_8_7_6_9_4_6_7, 0.8_5_8_0_5_8_0_4, 0.6_5_9_7_7_4_9_6, 0.4_4_8_8_5_3_0_2, 0.5_9_5_9_1_1_1, 0.4_2_5_1_5_9_5] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), F" expected_slice {expected_slice}, but got {image_slice.flatten()}" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), F" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}" @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def snake_case_ (self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def snake_case_ (self ): _UpperCAmelCase : Tuple = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/kandinskyv22_controlnet_robotcat_fp16.npy""" ) _UpperCAmelCase : Optional[Any] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinskyv22/hint_image_cat.png""" ) _UpperCAmelCase : Union[str, Any] = torch.from_numpy(np.array(lowerCAmelCase__ ) ).float() / 2_5_5.0 _UpperCAmelCase : Union[str, Any] = hint.permute(2 , 0 , 1 ).unsqueeze(0 ) _UpperCAmelCase : Union[str, Any] = KandinskyVaaPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(lowerCAmelCase__ ) _UpperCAmelCase : str = KandinskyVaaControlnetPipeline.from_pretrained( """kandinsky-community/kandinsky-2-2-controlnet-depth""" , torch_dtype=torch.floataa ) _UpperCAmelCase : int = pipeline.to(lowerCAmelCase__ ) pipeline.set_progress_bar_config(disable=lowerCAmelCase__ ) _UpperCAmelCase : str = """A robot, 4k photo""" _UpperCAmelCase : Dict = torch.Generator(device="""cuda""" ).manual_seed(0 ) _UpperCAmelCase , _UpperCAmelCase : Tuple = pipe_prior( lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() _UpperCAmelCase : Dict = torch.Generator(device="""cuda""" ).manual_seed(0 ) _UpperCAmelCase : Tuple = pipeline( image_embeds=lowerCAmelCase__ , negative_image_embeds=lowerCAmelCase__ , hint=lowerCAmelCase__ , generator=lowerCAmelCase__ , num_inference_steps=1_0_0 , output_type="""np""" , ) _UpperCAmelCase : str = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ )
170
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available SCREAMING_SNAKE_CASE : int = {"""configuration_glpn""": ["""GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GLPNConfig"""]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Tuple = ["""GLPNFeatureExtractor"""] SCREAMING_SNAKE_CASE : Dict = ["""GLPNImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : str = [ """GLPN_PRETRAINED_MODEL_ARCHIVE_LIST""", """GLPNForDepthEstimation""", """GLPNLayer""", """GLPNModel""", """GLPNPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
102
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCAmelCase ={"configuration_vit_msn": ["VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP", "ViTMSNConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase =[ "VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST", "ViTMSNModel", "ViTMSNForImageClassification", "ViTMSNPreTrainedModel", ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys __UpperCAmelCase =_LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
67
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( BertTokenizer, ViltConfig, ViltForImageAndTextRetrieval, ViltForImagesAndTextClassification, ViltForMaskedLM, ViltForQuestionAnswering, ViltImageProcessor, ViltProcessor, ) from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase = logging.get_logger(__name__) def lowerCamelCase (a_ :Dict , a_ :List[str]=False , a_ :Tuple=False , a_ :Any=False) -> Union[str, Any]: lowercase :Any = [] for i in range(config.num_hidden_layers): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""transformer.blocks.{i}.norm1.weight""", F"""vilt.encoder.layer.{i}.layernorm_before.weight""")) rename_keys.append((F"""transformer.blocks.{i}.norm1.bias""", F"""vilt.encoder.layer.{i}.layernorm_before.bias""")) rename_keys.append( (F"""transformer.blocks.{i}.attn.proj.weight""", F"""vilt.encoder.layer.{i}.attention.output.dense.weight""")) rename_keys.append( (F"""transformer.blocks.{i}.attn.proj.bias""", F"""vilt.encoder.layer.{i}.attention.output.dense.bias""")) rename_keys.append((F"""transformer.blocks.{i}.norm2.weight""", F"""vilt.encoder.layer.{i}.layernorm_after.weight""")) rename_keys.append((F"""transformer.blocks.{i}.norm2.bias""", F"""vilt.encoder.layer.{i}.layernorm_after.bias""")) rename_keys.append( (F"""transformer.blocks.{i}.mlp.fc1.weight""", F"""vilt.encoder.layer.{i}.intermediate.dense.weight""")) rename_keys.append((F"""transformer.blocks.{i}.mlp.fc1.bias""", F"""vilt.encoder.layer.{i}.intermediate.dense.bias""")) rename_keys.append((F"""transformer.blocks.{i}.mlp.fc2.weight""", F"""vilt.encoder.layer.{i}.output.dense.weight""")) rename_keys.append((F"""transformer.blocks.{i}.mlp.fc2.bias""", F"""vilt.encoder.layer.{i}.output.dense.bias""")) # embeddings rename_keys.extend( [ # text embeddings ('''text_embeddings.word_embeddings.weight''', '''vilt.embeddings.text_embeddings.word_embeddings.weight'''), ( '''text_embeddings.position_embeddings.weight''', '''vilt.embeddings.text_embeddings.position_embeddings.weight''', ), ('''text_embeddings.position_ids''', '''vilt.embeddings.text_embeddings.position_ids'''), ( '''text_embeddings.token_type_embeddings.weight''', '''vilt.embeddings.text_embeddings.token_type_embeddings.weight''', ), ('''text_embeddings.LayerNorm.weight''', '''vilt.embeddings.text_embeddings.LayerNorm.weight'''), ('''text_embeddings.LayerNorm.bias''', '''vilt.embeddings.text_embeddings.LayerNorm.bias'''), # patch embeddings ('''transformer.cls_token''', '''vilt.embeddings.cls_token'''), ('''transformer.patch_embed.proj.weight''', '''vilt.embeddings.patch_embeddings.projection.weight'''), ('''transformer.patch_embed.proj.bias''', '''vilt.embeddings.patch_embeddings.projection.bias'''), ('''transformer.pos_embed''', '''vilt.embeddings.position_embeddings'''), # token type embeddings ('''token_type_embeddings.weight''', '''vilt.embeddings.token_type_embeddings.weight'''), ]) # final layernorm + pooler rename_keys.extend( [ ('''transformer.norm.weight''', '''vilt.layernorm.weight'''), ('''transformer.norm.bias''', '''vilt.layernorm.bias'''), ('''pooler.dense.weight''', '''vilt.pooler.dense.weight'''), ('''pooler.dense.bias''', '''vilt.pooler.dense.bias'''), ]) # classifier head(s) if vqa_model: # classification head rename_keys.extend( [ ('''vqa_classifier.0.weight''', '''classifier.0.weight'''), ('''vqa_classifier.0.bias''', '''classifier.0.bias'''), ('''vqa_classifier.1.weight''', '''classifier.1.weight'''), ('''vqa_classifier.1.bias''', '''classifier.1.bias'''), ('''vqa_classifier.3.weight''', '''classifier.3.weight'''), ('''vqa_classifier.3.bias''', '''classifier.3.bias'''), ]) elif nlvr_model: # classification head rename_keys.extend( [ ('''nlvr2_classifier.0.weight''', '''classifier.0.weight'''), ('''nlvr2_classifier.0.bias''', '''classifier.0.bias'''), ('''nlvr2_classifier.1.weight''', '''classifier.1.weight'''), ('''nlvr2_classifier.1.bias''', '''classifier.1.bias'''), ('''nlvr2_classifier.3.weight''', '''classifier.3.weight'''), ('''nlvr2_classifier.3.bias''', '''classifier.3.bias'''), ]) else: pass return rename_keys def lowerCamelCase (a_ :List[str] , a_ :Any) -> Union[str, Any]: for i in range(config.num_hidden_layers): lowercase :Dict = '''vilt.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowercase :Dict = state_dict.pop(F"""transformer.blocks.{i}.attn.qkv.weight""") lowercase :Optional[int] = state_dict.pop(F"""transformer.blocks.{i}.attn.qkv.bias""") # next, add query, keys and values (in that order) to the state dict lowercase :Optional[int] = in_proj_weight[ : config.hidden_size, : ] lowercase :str = in_proj_bias[: config.hidden_size] lowercase :Optional[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowercase :Optional[int] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowercase :Union[str, Any] = in_proj_weight[ -config.hidden_size :, : ] lowercase :Dict = in_proj_bias[-config.hidden_size :] def lowerCamelCase (a_ :int) -> Optional[Any]: lowercase :Optional[Any] = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(a_ , a_) def lowerCamelCase (a_ :Union[str, Any] , a_ :List[str] , a_ :Optional[int]) -> Optional[Any]: lowercase :int = dct.pop(a_) lowercase :Tuple = val @torch.no_grad() def lowerCamelCase (a_ :List[str] , a_ :Tuple) -> List[str]: lowercase :Dict = ViltConfig(image_size=384 , patch_size=32 , tie_word_embeddings=a_) lowercase :Union[str, Any] = False lowercase :Union[str, Any] = False lowercase :List[Any] = False lowercase :Optional[Any] = False if "vqa" in checkpoint_url: lowercase :int = True lowercase :Optional[Any] = 3129 lowercase :Union[str, Any] = '''huggingface/label-files''' lowercase :Union[str, Any] = '''vqa2-id2label.json''' lowercase :str = json.load(open(hf_hub_download(a_ , a_ , repo_type='''dataset''') , '''r''')) lowercase :Dict = {int(a_): v for k, v in idalabel.items()} lowercase :List[str] = idalabel lowercase :List[str] = {v: k for k, v in idalabel.items()} lowercase :int = ViltForQuestionAnswering(a_) elif "nlvr" in checkpoint_url: lowercase :Optional[Any] = True lowercase :Union[str, Any] = 2 lowercase :Union[str, Any] = {0: '''False''', 1: '''True'''} lowercase :Union[str, Any] = {v: k for k, v in config.idalabel.items()} lowercase :Tuple = 3 lowercase :Union[str, Any] = ViltForImagesAndTextClassification(a_) elif "irtr" in checkpoint_url: lowercase :Tuple = True lowercase :str = ViltForImageAndTextRetrieval(a_) elif "mlm_itm" in checkpoint_url: lowercase :int = True lowercase :Tuple = ViltForMaskedLM(a_) else: raise ValueError('''Unknown model type''') # load state_dict of original model, remove and rename some keys lowercase :Optional[Any] = torch.hub.load_state_dict_from_url(a_ , map_location='''cpu''')['''state_dict'''] lowercase :int = create_rename_keys(a_ , a_ , a_ , a_) for src, dest in rename_keys: rename_key(a_ , a_ , a_) read_in_q_k_v(a_ , a_) if mlm_model or irtr_model: lowercase :List[str] = ['''itm_score.fc.weight''', '''itm_score.fc.bias'''] for k in ignore_keys: state_dict.pop(a_ , a_) # load state dict into HuggingFace model model.eval() if mlm_model: lowercase , lowercase :Dict = model.load_state_dict(a_ , strict=a_) assert missing_keys == ["mlm_score.decoder.bias"] else: model.load_state_dict(a_) # Define processor lowercase :List[str] = ViltImageProcessor(size=384) lowercase :Optional[int] = BertTokenizer.from_pretrained('''bert-base-uncased''') lowercase :Any = ViltProcessor(a_ , a_) # Forward pass on example inputs (image + text) if nlvr_model: lowercase :Tuple = Image.open(requests.get('''https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg''' , stream=a_).raw) lowercase :int = Image.open(requests.get('''https://lil.nlp.cornell.edu/nlvr/exs/ex0_0.jpg''' , stream=a_).raw) lowercase :Tuple = ( '''The left image contains twice the number of dogs as the right image, and at least two dogs in total are''' ''' standing.''' ) lowercase :List[Any] = processor(a_ , a_ , return_tensors='''pt''') lowercase :List[str] = processor(a_ , a_ , return_tensors='''pt''') lowercase :Optional[int] = model( input_ids=encoding_a.input_ids , pixel_values=encoding_a.pixel_values , pixel_values_a=encoding_a.pixel_values , ) else: lowercase :List[str] = Image.open(requests.get('''http://images.cocodataset.org/val2017/000000039769.jpg''' , stream=a_).raw) if mlm_model: lowercase :List[Any] = '''a bunch of [MASK] laying on a [MASK].''' else: lowercase :Optional[int] = '''How many cats are there?''' lowercase :Any = processor(a_ , a_ , return_tensors='''pt''') lowercase :Union[str, Any] = model(**a_) # Verify outputs if mlm_model: lowercase :Optional[Any] = torch.Size([1, 11, 3_0522]) lowercase :int = torch.tensor([-12.50_61, -12.51_23, -12.51_74]) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , a_ , atol=1E-4) # verify masked token prediction equals "cats" lowercase :int = outputs.logits[0, 4, :].argmax(-1).item() assert tokenizer.decode([predicted_id]) == "cats" elif vqa_model: lowercase :str = torch.Size([1, 3129]) lowercase :Dict = torch.tensor([-15.94_95, -18.14_72, -10.30_41]) assert torch.allclose(outputs.logits[0, :3] , a_ , atol=1E-4) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, 0, :3] , a_ , atol=1E-4) # verify vqa prediction equals "2" lowercase :Any = outputs.logits.argmax(-1).item() assert model.config.idalabel[predicted_idx] == "2" elif nlvr_model: lowercase :Optional[int] = torch.Size([1, 2]) lowercase :Union[str, Any] = torch.tensor([-2.87_21, 2.12_91]) assert torch.allclose(outputs.logits[0, :3] , a_ , atol=1E-4) assert outputs.logits.shape == expected_shape Path(a_).mkdir(exist_ok=a_) print(F"""Saving model and processor to {pytorch_dump_folder_path}""") model.save_pretrained(a_) processor.save_pretrained(a_) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/dandelin/ViLT/releases/download/200k/vilt_200k_mlm_itm.ckpt''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) UpperCAmelCase = parser.parse_args() convert_vilt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
172
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import YolosConfig, YolosForObjectDetection, YolosImageProcessor from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase = logging.get_logger(__name__) def lowerCamelCase (a_ :str) -> YolosConfig: lowercase :Union[str, Any] = YolosConfig() # size of the architecture if "yolos_ti" in yolos_name: lowercase :List[str] = 192 lowercase :List[str] = 768 lowercase :int = 12 lowercase :str = 3 lowercase :List[Any] = [800, 1333] lowercase :Any = False elif yolos_name == "yolos_s_dWr": lowercase :List[str] = 330 lowercase :List[Any] = 14 lowercase :int = 6 lowercase :List[Any] = 1320 elif "yolos_s" in yolos_name: lowercase :int = 384 lowercase :Union[str, Any] = 1536 lowercase :int = 12 lowercase :str = 6 elif "yolos_b" in yolos_name: lowercase :Dict = [800, 1344] lowercase :List[str] = 91 lowercase :List[Any] = '''huggingface/label-files''' lowercase :Union[str, Any] = '''coco-detection-id2label.json''' lowercase :int = json.load(open(hf_hub_download(a_ , a_ , repo_type='''dataset''') , '''r''')) lowercase :List[Any] = {int(a_): v for k, v in idalabel.items()} lowercase :Dict = idalabel lowercase :Tuple = {v: k for k, v in idalabel.items()} return config def lowerCamelCase (a_ :dict , a_ :YolosConfig , a_ :bool = False) -> Optional[int]: for i in range(config.num_hidden_layers): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowercase :Dict = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""") lowercase :List[Any] = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""") # next, add query, keys and values (in that order) to the state dict lowercase :int = in_proj_weight[: config.hidden_size, :] lowercase :List[str] = in_proj_bias[: config.hidden_size] lowercase :Optional[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowercase :int = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowercase :Any = in_proj_weight[-config.hidden_size :, :] lowercase :Union[str, Any] = in_proj_bias[-config.hidden_size :] def lowerCamelCase (a_ :str) -> str: if "backbone" in name: lowercase :Optional[int] = name.replace('''backbone''' , '''vit''') if "cls_token" in name: lowercase :List[Any] = name.replace('''cls_token''' , '''embeddings.cls_token''') if "det_token" in name: lowercase :int = name.replace('''det_token''' , '''embeddings.detection_tokens''') if "mid_pos_embed" in name: lowercase :List[Any] = name.replace('''mid_pos_embed''' , '''encoder.mid_position_embeddings''') if "pos_embed" in name: lowercase :List[str] = name.replace('''pos_embed''' , '''embeddings.position_embeddings''') if "patch_embed.proj" in name: lowercase :Any = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''') if "blocks" in name: lowercase :Any = name.replace('''blocks''' , '''encoder.layer''') if "attn.proj" in name: lowercase :Dict = name.replace('''attn.proj''' , '''attention.output.dense''') if "attn" in name: lowercase :Tuple = name.replace('''attn''' , '''attention.self''') if "norm1" in name: lowercase :List[Any] = name.replace('''norm1''' , '''layernorm_before''') if "norm2" in name: lowercase :List[Any] = name.replace('''norm2''' , '''layernorm_after''') if "mlp.fc1" in name: lowercase :Union[str, Any] = name.replace('''mlp.fc1''' , '''intermediate.dense''') if "mlp.fc2" in name: lowercase :Dict = name.replace('''mlp.fc2''' , '''output.dense''') if "class_embed" in name: lowercase :Dict = name.replace('''class_embed''' , '''class_labels_classifier''') if "bbox_embed" in name: lowercase :Dict = name.replace('''bbox_embed''' , '''bbox_predictor''') if "vit.norm" in name: lowercase :Dict = name.replace('''vit.norm''' , '''vit.layernorm''') return name def lowerCamelCase (a_ :dict , a_ :YolosForObjectDetection) -> dict: for key in orig_state_dict.copy().keys(): lowercase :List[Any] = orig_state_dict.pop(a_) if "qkv" in key: lowercase :str = key.split('''.''') lowercase :List[str] = int(key_split[2]) lowercase :List[str] = model.vit.encoder.layer[layer_num].attention.attention.all_head_size if "weight" in key: lowercase :List[Any] = val[:dim, :] lowercase :Optional[int] = val[ dim : dim * 2, : ] lowercase :Any = val[-dim:, :] else: lowercase :List[str] = val[:dim] lowercase :Union[str, Any] = val[dim : dim * 2] lowercase :List[Any] = val[-dim:] else: lowercase :List[str] = val return orig_state_dict def lowerCamelCase () -> torch.Tensor: lowercase :Tuple = '''http://images.cocodataset.org/val2017/000000039769.jpg''' lowercase :Dict = Image.open(requests.get(a_ , stream=a_).raw) return im @torch.no_grad() def lowerCamelCase (a_ :str , a_ :str , a_ :str , a_ :bool = False) -> List[Any]: lowercase :Union[str, Any] = get_yolos_config(a_) # load original state_dict lowercase :List[str] = torch.load(a_ , map_location='''cpu''')['''model'''] # load 🤗 model lowercase :Tuple = YolosForObjectDetection(a_) model.eval() lowercase :Dict = convert_state_dict(a_ , a_) model.load_state_dict(a_) # Check outputs on an image, prepared by YolosImageProcessor lowercase :Tuple = 800 if yolos_name != '''yolos_ti''' else 512 lowercase :Dict = YolosImageProcessor(format='''coco_detection''' , size=a_) lowercase :Optional[int] = image_processor(images=prepare_img() , return_tensors='''pt''') lowercase :List[Any] = model(**a_) lowercase , lowercase :Dict = outputs.logits, outputs.pred_boxes lowercase , lowercase :int = None, None if yolos_name == "yolos_ti": lowercase :Dict = torch.tensor( [[-39.50_22, -11.98_20, -17.68_88], [-29.95_74, -9.97_69, -17.76_91], [-42.32_81, -20.72_00, -30.62_94]]) lowercase :Dict = torch.tensor( [[0.40_21, 0.08_36, 0.79_79], [0.01_84, 0.26_09, 0.03_64], [0.17_81, 0.20_04, 0.20_95]]) elif yolos_name == "yolos_s_200_pre": lowercase :Union[str, Any] = torch.tensor( [[-24.02_48, -10.30_24, -14.82_90], [-42.03_92, -16.82_00, -27.43_34], [-27.27_43, -11.81_54, -18.71_48]]) lowercase :List[str] = torch.tensor( [[0.25_59, 0.54_55, 0.47_06], [0.29_89, 0.72_79, 0.18_75], [0.77_32, 0.40_17, 0.44_62]]) elif yolos_name == "yolos_s_300_pre": lowercase :int = torch.tensor( [[-36.22_20, -14.43_85, -23.54_57], [-35.69_70, -14.75_83, -21.39_35], [-31.59_39, -13.60_42, -16.80_49]]) lowercase :Optional[Any] = torch.tensor( [[0.76_14, 0.23_16, 0.47_28], [0.71_68, 0.44_95, 0.38_55], [0.49_96, 0.14_66, 0.99_96]]) elif yolos_name == "yolos_s_dWr": lowercase :int = torch.tensor( [[-42.86_68, -24.10_49, -41.16_90], [-34.74_56, -14.12_74, -24.91_94], [-33.78_98, -12.19_46, -25.64_95]]) lowercase :Dict = torch.tensor( [[0.55_87, 0.27_73, 0.06_05], [0.50_04, 0.30_14, 0.99_94], [0.49_99, 0.15_48, 0.99_94]]) elif yolos_name == "yolos_base": lowercase :Dict = torch.tensor( [[-40.60_64, -24.30_84, -32.64_47], [-55.19_90, -30.77_19, -35.58_77], [-51.43_11, -33.35_07, -35.64_62]]) lowercase :Tuple = torch.tensor( [[0.55_55, 0.27_94, 0.06_55], [0.90_49, 0.26_64, 0.18_94], [0.91_83, 0.19_84, 0.16_35]]) else: raise ValueError(F"""Unknown yolos_name: {yolos_name}""") assert torch.allclose(logits[0, :3, :3] , a_ , atol=1E-4) assert torch.allclose(pred_boxes[0, :3, :3] , a_ , atol=1E-4) Path(a_).mkdir(exist_ok=a_) print(F"""Saving model {yolos_name} to {pytorch_dump_folder_path}""") model.save_pretrained(a_) print(F"""Saving image processor to {pytorch_dump_folder_path}""") image_processor.save_pretrained(a_) if push_to_hub: lowercase :Optional[int] = { '''yolos_ti''': '''yolos-tiny''', '''yolos_s_200_pre''': '''yolos-small''', '''yolos_s_300_pre''': '''yolos-small-300''', '''yolos_s_dWr''': '''yolos-small-dwr''', '''yolos_base''': '''yolos-base''', } print('''Pushing to the hub...''') lowercase :Optional[Any] = model_mapping[yolos_name] image_processor.push_to_hub(a_ , organization='''hustvl''') model.push_to_hub(a_ , organization='''hustvl''') if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--yolos_name''', default='''yolos_s_200_pre''', type=str, help=( '''Name of the YOLOS model you\'d like to convert. Should be one of \'yolos_ti\', \'yolos_s_200_pre\',''' ''' \'yolos_s_300_pre\', \'yolos_s_dWr\', \'yolos_base\'.''' ), ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original state dict (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) UpperCAmelCase = parser.parse_args() convert_yolos_checkpoint(args.yolos_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
172
1
import argparse import re from typing import Dict import torch from datasets import Audio, Dataset, load_dataset, load_metric from transformers import AutoFeatureExtractor, pipeline def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase ) -> Optional[int]: """simple docstring""" snake_case__ : int = args.log_outputs snake_case__ : Tuple = '''_'''.join(args.dataset.split('''/''' ) + [args.config, args.split] ) # load metric snake_case__ : Optional[Any] = load_metric('''wer''' ) snake_case__ : List[str] = load_metric('''cer''' ) # compute metrics snake_case__ : List[Any] = wer.compute(references=result['''target'''] , predictions=result['''prediction'''] ) snake_case__ : str = cer.compute(references=result['''target'''] , predictions=result['''prediction'''] ) # print & log results snake_case__ : str = f"""WER: {wer_result}\nCER: {cer_result}""" print(_UpperCamelCase ) with open(f"""{dataset_id}_eval_results.txt""" , '''w''' ) as f: f.write(_UpperCamelCase ) # log all results in text file. Possibly interesting for analysis if log_outputs is not None: snake_case__ : str = f"""log_{dataset_id}_predictions.txt""" snake_case__ : str = f"""log_{dataset_id}_targets.txt""" with open(_UpperCamelCase , '''w''' ) as p, open(_UpperCamelCase , '''w''' ) as t: # mapping function to write output def write_to_file(__lowerCAmelCase , __lowerCAmelCase ): p.write(f"""{i}""" + '''\n''' ) p.write(batch['''prediction'''] + '''\n''' ) t.write(f"""{i}""" + '''\n''' ) t.write(batch['''target'''] + '''\n''' ) result.map(_UpperCamelCase , with_indices=_UpperCamelCase ) def _lowerCAmelCase ( __lowerCAmelCase ) -> Tuple: """simple docstring""" snake_case__ : str = '''[,?.!\-\;\:"“%‘”�—’…–]''' # noqa: W605 IMPORTANT: this should correspond to the chars that were ignored during training snake_case__ : Tuple = re.sub(_UpperCamelCase , '''''' , text.lower() ) # In addition, we can normalize the target text, e.g. removing new lines characters etc... # note that order is important here! snake_case__ : Optional[Any] = ['''\n\n''', '''\n''', ''' ''', ''' '''] for t in token_sequences_to_ignore: snake_case__ : List[Any] = ''' '''.join(text.split(_UpperCamelCase ) ) return text def _lowerCAmelCase ( __lowerCAmelCase ) -> Optional[int]: """simple docstring""" snake_case__ : Union[str, Any] = load_dataset(args.dataset , args.config , split=args.split , use_auth_token=_UpperCamelCase ) # for testing: only process the first two examples as a test # dataset = dataset.select(range(10)) # load processor snake_case__ : Any = AutoFeatureExtractor.from_pretrained(args.model_id ) snake_case__ : List[str] = feature_extractor.sampling_rate # resample audio snake_case__ : List[str] = dataset.cast_column('''audio''' , Audio(sampling_rate=_UpperCamelCase ) ) # load eval pipeline if args.device is None: snake_case__ : Dict = 0 if torch.cuda.is_available() else -1 snake_case__ : Any = pipeline('''automatic-speech-recognition''' , model=args.model_id , device=args.device ) # map function to decode audio def map_to_pred(__lowerCAmelCase ): snake_case__ : List[str] = asr( batch['''audio''']['''array'''] , chunk_length_s=args.chunk_length_s , stride_length_s=args.stride_length_s ) snake_case__ : Optional[int] = prediction['''text'''] snake_case__ : int = normalize_text(batch['''sentence'''] ) return batch # run inference on all examples snake_case__ : Union[str, Any] = dataset.map(_UpperCamelCase , remove_columns=dataset.column_names ) # compute and log_results # do not change function below log_results(_UpperCamelCase , _UpperCamelCase ) if __name__ == "__main__": A__ = argparse.ArgumentParser() parser.add_argument( '''--model_id''', type=str, required=True, help='''Model identifier. Should be loadable with 🤗 Transformers''' ) parser.add_argument( '''--dataset''', type=str, required=True, help='''Dataset name to evaluate the `model_id`. Should be loadable with 🤗 Datasets''', ) parser.add_argument( '''--config''', type=str, required=True, help='''Config of the dataset. *E.g.* `\'en\'` for Common Voice''' ) parser.add_argument('''--split''', type=str, required=True, help='''Split of the dataset. *E.g.* `\'test\'`''') parser.add_argument( '''--chunk_length_s''', type=float, default=None, help='''Chunk length in seconds. Defaults to 5 seconds.''' ) parser.add_argument( '''--stride_length_s''', type=float, default=None, help='''Stride of the audio chunks. Defaults to 1 second.''' ) parser.add_argument( '''--log_outputs''', action='''store_true''', help='''If defined, write outputs to log file for analysis.''' ) parser.add_argument( '''--device''', type=int, default=None, help='''The device to run the pipeline on. -1 for CPU (default), 0 for the first GPU and so on.''', ) A__ = parser.parse_args() main(args)
230
import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class __lowerCAmelCase : def __init__( self , __UpperCAmelCase , __UpperCAmelCase=2 , __UpperCAmelCase=8 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=99 , __UpperCAmelCase=16 , __UpperCAmelCase=5 , __UpperCAmelCase=2 , __UpperCAmelCase=36 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.0 , __UpperCAmelCase=512 , __UpperCAmelCase=16 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=3 , __UpperCAmelCase=4 , __UpperCAmelCase=None , ): '''simple docstring''' __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_input_mask __lowerCamelCase = use_token_type_ids __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = type_sequence_label_size __lowerCamelCase = initializer_range __lowerCamelCase = num_labels __lowerCamelCase = num_choices __lowerCamelCase = scope def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = None if self.use_input_mask: __lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCamelCase = None if self.use_token_type_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCamelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase ( self ): '''simple docstring''' return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__UpperCAmelCase , initializer_range=self.initializer_range , ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.get_config() __lowerCamelCase = 300 return config def lowerCamelCase ( self ): '''simple docstring''' ( ( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) , ) = self.prepare_config_and_inputs() __lowerCamelCase = True __lowerCamelCase = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = MraModel(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase , token_type_ids=__UpperCAmelCase ) __lowerCamelCase = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ): '''simple docstring''' __lowerCamelCase = True __lowerCamelCase = MraModel(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() __lowerCamelCase = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=__UpperCAmelCase , ) __lowerCamelCase = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , ) __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = MraForMaskedLM(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = MraForQuestionAnswering(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() __lowerCamelCase = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , start_positions=__UpperCAmelCase , end_positions=__UpperCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.num_labels __lowerCamelCase = MraForSequenceClassification(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.num_labels __lowerCamelCase = MraForTokenClassification(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() __lowerCamelCase = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' __lowerCamelCase = self.num_choices __lowerCamelCase = MraForMultipleChoice(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() __lowerCamelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCamelCase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCamelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCamelCase = model( __UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.prepare_config_and_inputs() ( ( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) ,( __lowerCamelCase ) , ) = config_and_inputs __lowerCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __lowerCAmelCase ( lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = False lowerCAmelCase__ = () def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = MraModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=__UpperCAmelCase , hidden_size=37 ) def lowerCamelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __lowerCamelCase = type self.model_tester.create_and_check_model(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCAmelCase ) def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCAmelCase ) @slow def lowerCamelCase ( self ): '''simple docstring''' for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = MraModel.from_pretrained(__UpperCAmelCase ) self.assertIsNotNone(__UpperCAmelCase ) @unittest.skip(reason='''MRA does not output attentions''' ) def lowerCamelCase ( self ): '''simple docstring''' return @require_torch class __lowerCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = MraModel.from_pretrained('''uw-madison/mra-base-512-4''' ) __lowerCamelCase = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): __lowerCamelCase = model(__UpperCAmelCase )[0] __lowerCamelCase = torch.Size((1, 256, 768) ) self.assertEqual(output.shape , __UpperCAmelCase ) __lowerCamelCase = torch.tensor( [[[-0.0_140, 0.0_830, -0.0_381], [0.1_546, 0.1_402, 0.0_220], [0.1_162, 0.0_851, 0.0_165]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=1E-4 ) ) @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = MraForMaskedLM.from_pretrained('''uw-madison/mra-base-512-4''' ) __lowerCamelCase = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): __lowerCamelCase = model(__UpperCAmelCase )[0] __lowerCamelCase = 50265 __lowerCamelCase = torch.Size((1, 256, vocab_size) ) self.assertEqual(output.shape , __UpperCAmelCase ) __lowerCamelCase = torch.tensor( [[[9.2_595, -3.6_038, 11.8_819], [9.3_869, -3.2_693, 11.0_956], [11.8_524, -3.4_938, 13.1_210]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=1E-4 ) ) @slow def lowerCamelCase ( self ): '''simple docstring''' __lowerCamelCase = MraForMaskedLM.from_pretrained('''uw-madison/mra-base-4096-8-d3''' ) __lowerCamelCase = torch.arange(4096 ).unsqueeze(0 ) with torch.no_grad(): __lowerCamelCase = model(__UpperCAmelCase )[0] __lowerCamelCase = 50265 __lowerCamelCase = torch.Size((1, 4096, vocab_size) ) self.assertEqual(output.shape , __UpperCAmelCase ) __lowerCamelCase = torch.tensor( [[[5.4_789, -2.3_564, 7.5_064], [7.9_067, -1.3_369, 9.9_668], [9.0_712, -1.8_106, 7.0_380]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCAmelCase , atol=1E-4 ) )
330
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __a = { "configuration_data2vec_audio": ["DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP", "Data2VecAudioConfig"], "configuration_data2vec_text": [ "DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "Data2VecTextConfig", "Data2VecTextOnnxConfig", ], "configuration_data2vec_vision": [ "DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP", "Data2VecVisionConfig", "Data2VecVisionOnnxConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST", "Data2VecAudioForAudioFrameClassification", "Data2VecAudioForCTC", "Data2VecAudioForSequenceClassification", "Data2VecAudioForXVector", "Data2VecAudioModel", "Data2VecAudioPreTrainedModel", ] __a = [ "DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "Data2VecTextForCausalLM", "Data2VecTextForMaskedLM", "Data2VecTextForMultipleChoice", "Data2VecTextForQuestionAnswering", "Data2VecTextForSequenceClassification", "Data2VecTextForTokenClassification", "Data2VecTextModel", "Data2VecTextPreTrainedModel", ] __a = [ "DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST", "Data2VecVisionForImageClassification", "Data2VecVisionForMaskedImageModeling", "Data2VecVisionForSemanticSegmentation", "Data2VecVisionModel", "Data2VecVisionPreTrainedModel", ] if is_tf_available(): __a = [ "TFData2VecVisionForImageClassification", "TFData2VecVisionForSemanticSegmentation", "TFData2VecVisionModel", "TFData2VecVisionPreTrainedModel", ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
43
'''simple docstring''' from queue import Queue from typing import TYPE_CHECKING, Optional if TYPE_CHECKING: from ..models.auto import AutoTokenizer class UpperCAmelCase_ : """simple docstring""" def lowerCamelCase ( self : Optional[Any] , snake_case_ : Optional[int] ): raise NotImplementedError() def lowerCamelCase ( self : Optional[int] ): raise NotImplementedError() class UpperCAmelCase_ ( _a ): """simple docstring""" def __init__( self : Tuple , snake_case_ : "AutoTokenizer" , snake_case_ : bool = False , **snake_case_ : Tuple ): snake_case__ : Tuple = tokenizer snake_case__ : List[str] = skip_prompt snake_case__ : Optional[int] = decode_kwargs # variables used in the streaming process snake_case__ : Optional[int] = [] snake_case__ : Optional[int] = 0 snake_case__ : List[Any] = True def lowerCamelCase ( self : List[str] , snake_case_ : int ): if len(value.shape ) > 1 and value.shape[0] > 1: raise ValueError("""TextStreamer only supports batch size 1""" ) elif len(value.shape ) > 1: snake_case__ : Optional[Any] = value[0] if self.skip_prompt and self.next_tokens_are_prompt: snake_case__ : List[Any] = False return # Add the new token to the cache and decodes the entire thing. self.token_cache.extend(value.tolist() ) snake_case__ : Tuple = self.tokenizer.decode(self.token_cache , **self.decode_kwargs ) # After the symbol for a new line, we flush the cache. if text.endswith("""\n""" ): snake_case__ : int = text[self.print_len :] snake_case__ : Optional[int] = [] snake_case__ : int = 0 # If the last token is a CJK character, we print the characters. elif len(snake_case_ ) > 0 and self._is_chinese_char(ord(text[-1] ) ): snake_case__ : str = text[self.print_len :] self.print_len += len(snake_case_ ) # Otherwise, prints until the last space char (simple heuristic to avoid printing incomplete words, # which may change with the subsequent token -- there are probably smarter ways to do this!) else: snake_case__ : Dict = text[self.print_len : text.rfind(""" """ ) + 1] self.print_len += len(snake_case_ ) self.on_finalized_text(snake_case_ ) def lowerCamelCase ( self : int ): # Flush the cache, if it exists if len(self.token_cache ) > 0: snake_case__ : Union[str, Any] = self.tokenizer.decode(self.token_cache , **self.decode_kwargs ) snake_case__ : Optional[Any] = text[self.print_len :] snake_case__ : Tuple = [] snake_case__ : int = 0 else: snake_case__ : int = """""" snake_case__ : Union[str, Any] = True self.on_finalized_text(snake_case_ , stream_end=snake_case_ ) def lowerCamelCase ( self : Optional[int] , snake_case_ : str , snake_case_ : bool = False ): print(snake_case_ , flush=snake_case_ , end="""""" if not stream_end else None ) def lowerCamelCase ( self : int , snake_case_ : Optional[int] ): # This defines a "chinese character" as anything in the CJK Unicode block: # https://en.wikipedia.org/wiki/CJK_Unified_Ideographs_(Unicode_block) # # Note that the CJK Unicode block is NOT all Japanese and Korean characters, # despite its name. The modern Korean Hangul alphabet is a different block, # as is Japanese Hiragana and Katakana. Those alphabets are used to write # space-separated words, so they are not treated specially and handled # like the all of the other languages. if ( (cp >= 0x4E00 and cp <= 0x9FFF) or (cp >= 0x3400 and cp <= 0x4DBF) # or (cp >= 0x20000 and cp <= 0x2A6DF) # or (cp >= 0x2A700 and cp <= 0x2B73F) # or (cp >= 0x2B740 and cp <= 0x2B81F) # or (cp >= 0x2B820 and cp <= 0x2CEAF) # or (cp >= 0xF900 and cp <= 0xFAFF) or (cp >= 0x2F800 and cp <= 0x2FA1F) # ): # return True return False class UpperCAmelCase_ ( _a ): """simple docstring""" def __init__( self : Optional[int] , snake_case_ : "AutoTokenizer" , snake_case_ : bool = False , snake_case_ : Optional[float] = None , **snake_case_ : List[Any] ): super().__init__(snake_case_ , snake_case_ , **snake_case_ ) snake_case__ : Dict = Queue() snake_case__ : List[Any] = None snake_case__ : int = timeout def lowerCamelCase ( self : Dict , snake_case_ : str , snake_case_ : bool = False ): self.text_queue.put(snake_case_ , timeout=self.timeout ) if stream_end: self.text_queue.put(self.stop_signal , timeout=self.timeout ) def __iter__( self : List[str] ): return self def lowerCamelCase ( self : str ): snake_case__ : List[Any] = self.text_queue.get(timeout=self.timeout ) if value == self.stop_signal: raise StopIteration() else: return value
43
1
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import ( DiffusionPipeline, UnCLIPImageVariationPipeline, UnCLIPScheduler, UNetaDConditionModel, UNetaDModel, ) from diffusers.pipelines.unclip.text_proj import UnCLIPTextProjModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, load_image, require_torch_gpu, skip_mps from ..pipeline_params import IMAGE_VARIATION_BATCH_PARAMS, IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class a ( UpperCAmelCase__ , unittest.TestCase ): UpperCamelCase : Union[str, Any] = UnCLIPImageVariationPipeline UpperCamelCase : Optional[Any] = IMAGE_VARIATION_PARAMS - {'height', 'width', 'guidance_scale'} UpperCamelCase : Any = IMAGE_VARIATION_BATCH_PARAMS UpperCamelCase : Any = [ 'generator', 'return_dict', 'decoder_num_inference_steps', 'super_res_num_inference_steps', ] UpperCamelCase : Optional[int] = False @property def lowerCamelCase__ ( self : Tuple ) -> List[Any]: '''simple docstring''' return 32 @property def lowerCamelCase__ ( self : List[str] ) -> Tuple: '''simple docstring''' return 32 @property def lowerCamelCase__ ( self : str ) -> Optional[int]: '''simple docstring''' return self.time_input_dim @property def lowerCamelCase__ ( self : Tuple ) -> List[str]: '''simple docstring''' return self.time_input_dim * 4 @property def lowerCamelCase__ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' return 100 @property def lowerCamelCase__ ( self : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE_: Union[str, Any] =CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) return tokenizer @property def lowerCamelCase__ ( self : List[str] ) -> List[str]: '''simple docstring''' torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_: Dict =CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(_lowerCAmelCase ) @property def lowerCamelCase__ ( self : int ) -> Union[str, Any]: '''simple docstring''' torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_: List[Any] =CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , ) return CLIPVisionModelWithProjection(_lowerCAmelCase ) @property def lowerCamelCase__ ( self : List[str] ) -> str: '''simple docstring''' torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_: List[Any] ={ """clip_embeddings_dim""": self.text_embedder_hidden_size, """time_embed_dim""": self.time_embed_dim, """cross_attention_dim""": self.cross_attention_dim, } SCREAMING_SNAKE_CASE_: Optional[Any] =UnCLIPTextProjModel(**_lowerCAmelCase ) return model @property def lowerCamelCase__ ( self : Tuple ) -> Any: '''simple docstring''' torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_: str ={ """sample_size""": 32, # RGB in channels """in_channels""": 3, # Out channels is double in channels because predicts mean and variance """out_channels""": 6, """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": """identity""", } SCREAMING_SNAKE_CASE_: Optional[int] =UNetaDConditionModel(**_lowerCAmelCase ) return model @property def lowerCamelCase__ ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' return { "sample_size": 64, "layers_per_block": 1, "down_block_types": ("ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D"), "up_block_types": ("ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D"), "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "in_channels": 6, "out_channels": 3, } @property def lowerCamelCase__ ( self : int ) -> List[str]: '''simple docstring''' torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_: Optional[int] =UNetaDModel(**self.dummy_super_res_kwargs ) return model @property def lowerCamelCase__ ( self : int ) -> Union[str, Any]: '''simple docstring''' torch.manual_seed(1 ) SCREAMING_SNAKE_CASE_: int =UNetaDModel(**self.dummy_super_res_kwargs ) return model def lowerCamelCase__ ( self : int ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE_: List[Any] =self.dummy_decoder SCREAMING_SNAKE_CASE_: Dict =self.dummy_text_proj SCREAMING_SNAKE_CASE_: Optional[int] =self.dummy_text_encoder SCREAMING_SNAKE_CASE_: Dict =self.dummy_tokenizer SCREAMING_SNAKE_CASE_: Tuple =self.dummy_super_res_first SCREAMING_SNAKE_CASE_: int =self.dummy_super_res_last SCREAMING_SNAKE_CASE_: Any =UnCLIPScheduler( variance_type="""learned_range""" , prediction_type="""epsilon""" , num_train_timesteps=1000 , ) SCREAMING_SNAKE_CASE_: str =UnCLIPScheduler( variance_type="""fixed_small_log""" , prediction_type="""epsilon""" , num_train_timesteps=1000 , ) SCREAMING_SNAKE_CASE_: Union[str, Any] =CLIPImageProcessor(crop_size=32 , size=32 ) SCREAMING_SNAKE_CASE_: Dict =self.dummy_image_encoder return { "decoder": decoder, "text_encoder": text_encoder, "tokenizer": tokenizer, "text_proj": text_proj, "feature_extractor": feature_extractor, "image_encoder": image_encoder, "super_res_first": super_res_first, "super_res_last": super_res_last, "decoder_scheduler": decoder_scheduler, "super_res_scheduler": super_res_scheduler, } def lowerCamelCase__ ( self : Any , lowerCAmelCase : Optional[int] , lowerCAmelCase : Union[str, Any]=0 , lowerCAmelCase : Optional[Any]=True ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE_: str =floats_tensor((1, 3, 32, 32) , rng=random.Random(_lowerCAmelCase ) ).to(_lowerCAmelCase ) if str(_lowerCAmelCase ).startswith("""mps""" ): SCREAMING_SNAKE_CASE_: int =torch.manual_seed(_lowerCAmelCase ) else: SCREAMING_SNAKE_CASE_: Tuple =torch.Generator(device=_lowerCAmelCase ).manual_seed(_lowerCAmelCase ) if pil_image: SCREAMING_SNAKE_CASE_: Any =input_image * 0.5 + 0.5 SCREAMING_SNAKE_CASE_: List[Any] =input_image.clamp(0 , 1 ) SCREAMING_SNAKE_CASE_: int =input_image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() SCREAMING_SNAKE_CASE_: str =DiffusionPipeline.numpy_to_pil(_lowerCAmelCase )[0] return { "image": input_image, "generator": generator, "decoder_num_inference_steps": 2, "super_res_num_inference_steps": 2, "output_type": "np", } def lowerCamelCase__ ( self : int ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE_: Tuple ="""cpu""" SCREAMING_SNAKE_CASE_: Any =self.get_dummy_components() SCREAMING_SNAKE_CASE_: Any =self.pipeline_class(**_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: List[str] =pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: List[Any] =self.get_dummy_inputs(_lowerCAmelCase , pil_image=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Dict =pipe(**_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] =output.images SCREAMING_SNAKE_CASE_: Dict =self.get_dummy_inputs(_lowerCAmelCase , pil_image=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: List[str] =pipe( **_lowerCAmelCase , return_dict=_lowerCAmelCase , )[0] SCREAMING_SNAKE_CASE_: str =image[0, -3:, -3:, -1] SCREAMING_SNAKE_CASE_: str =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) SCREAMING_SNAKE_CASE_: List[Any] =np.array( [ 0.9_9_9_7, 0.0_0_0_2, 0.9_9_9_7, 0.9_9_9_7, 0.9_9_6_9, 0.0_0_2_3, 0.9_9_9_7, 0.9_9_6_9, 0.9_9_7_0, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase__ ( self : Tuple ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE_: int ="""cpu""" SCREAMING_SNAKE_CASE_: List[str] =self.get_dummy_components() SCREAMING_SNAKE_CASE_: int =self.pipeline_class(**_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Any =pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Dict =self.get_dummy_inputs(_lowerCAmelCase , pil_image=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Union[str, Any] =pipe(**_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Tuple =output.images SCREAMING_SNAKE_CASE_: Any =self.get_dummy_inputs(_lowerCAmelCase , pil_image=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Tuple =pipe( **_lowerCAmelCase , return_dict=_lowerCAmelCase , )[0] SCREAMING_SNAKE_CASE_: List[str] =image[0, -3:, -3:, -1] SCREAMING_SNAKE_CASE_: List[Any] =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) SCREAMING_SNAKE_CASE_: Optional[int] =np.array([0.9_9_9_7, 0.0_0_0_3, 0.9_9_9_7, 0.9_9_9_7, 0.9_9_7_0, 0.0_0_2_4, 0.9_9_9_7, 0.9_9_7_1, 0.9_9_7_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase__ ( self : List[str] ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE_: Dict ="""cpu""" SCREAMING_SNAKE_CASE_: int =self.get_dummy_components() SCREAMING_SNAKE_CASE_: List[Any] =self.pipeline_class(**_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: int =pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: List[str] =self.get_dummy_inputs(_lowerCAmelCase , pil_image=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: str =[ pipeline_inputs["""image"""], pipeline_inputs["""image"""], ] SCREAMING_SNAKE_CASE_: Any =pipe(**_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: List[Any] =output.images SCREAMING_SNAKE_CASE_: Optional[Any] =self.get_dummy_inputs(_lowerCAmelCase , pil_image=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Dict =[ tuple_pipeline_inputs["""image"""], tuple_pipeline_inputs["""image"""], ] SCREAMING_SNAKE_CASE_: List[str] =pipe( **_lowerCAmelCase , return_dict=_lowerCAmelCase , )[0] SCREAMING_SNAKE_CASE_: int =image[0, -3:, -3:, -1] SCREAMING_SNAKE_CASE_: Optional[Any] =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (2, 64, 64, 3) SCREAMING_SNAKE_CASE_: List[str] =np.array( [ 0.9_9_9_7, 0.9_9_8_9, 0.0_0_0_8, 0.0_0_2_1, 0.9_9_6_0, 0.0_0_1_8, 0.0_0_1_4, 0.0_0_0_2, 0.9_9_3_3, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase__ ( self : Any ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE_: Union[str, Any] =torch.device("""cpu""" ) class a : UpperCamelCase : Optional[Any] = 1 SCREAMING_SNAKE_CASE_: str =self.get_dummy_components() SCREAMING_SNAKE_CASE_: Optional[int] =self.pipeline_class(**_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Tuple =pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] =torch.Generator(device=_lowerCAmelCase ).manual_seed(0 ) SCREAMING_SNAKE_CASE_: int =pipe.decoder.dtype SCREAMING_SNAKE_CASE_: Tuple =1 SCREAMING_SNAKE_CASE_: Optional[int] =( batch_size, pipe.decoder.config.in_channels, pipe.decoder.config.sample_size, pipe.decoder.config.sample_size, ) SCREAMING_SNAKE_CASE_: Optional[int] =pipe.prepare_latents( _lowerCAmelCase , dtype=_lowerCAmelCase , device=_lowerCAmelCase , generator=_lowerCAmelCase , latents=_lowerCAmelCase , scheduler=DummyScheduler() ) SCREAMING_SNAKE_CASE_: Optional[int] =( batch_size, pipe.super_res_first.config.in_channels // 2, pipe.super_res_first.config.sample_size, pipe.super_res_first.config.sample_size, ) SCREAMING_SNAKE_CASE_: Optional[int] =pipe.prepare_latents( _lowerCAmelCase , dtype=_lowerCAmelCase , device=_lowerCAmelCase , generator=_lowerCAmelCase , latents=_lowerCAmelCase , scheduler=DummyScheduler() ) SCREAMING_SNAKE_CASE_: Optional[int] =self.get_dummy_inputs(_lowerCAmelCase , pil_image=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] =pipe( **_lowerCAmelCase , decoder_latents=_lowerCAmelCase , super_res_latents=_lowerCAmelCase ).images SCREAMING_SNAKE_CASE_: Any =self.get_dummy_inputs(_lowerCAmelCase , pil_image=_lowerCAmelCase ) # Don't pass image, instead pass embedding SCREAMING_SNAKE_CASE_: Union[str, Any] =pipeline_inputs.pop("""image""" ) SCREAMING_SNAKE_CASE_: Tuple =pipe.image_encoder(_lowerCAmelCase ).image_embeds SCREAMING_SNAKE_CASE_: List[str] =pipe( **_lowerCAmelCase , decoder_latents=_lowerCAmelCase , super_res_latents=_lowerCAmelCase , image_embeddings=_lowerCAmelCase , ).images # make sure passing text embeddings manually is identical assert np.abs(img_out_a - img_out_a ).max() < 1E-4 @skip_mps def lowerCamelCase__ ( self : Optional[Any] ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE_: int =torch_device == """cpu""" # Check is relaxed because there is not a torch 2.0 sliced attention added kv processor SCREAMING_SNAKE_CASE_: List[Any] =1E-2 self._test_attention_slicing_forward_pass( test_max_difference=_lowerCAmelCase , expected_max_diff=_lowerCAmelCase ) @skip_mps def lowerCamelCase__ ( self : Optional[int] ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE_: Optional[int] =torch_device == """cpu""" SCREAMING_SNAKE_CASE_: int =True SCREAMING_SNAKE_CASE_: Dict =[ """decoder_num_inference_steps""", """super_res_num_inference_steps""", ] self._test_inference_batch_single_identical( test_max_difference=_lowerCAmelCase , relax_max_difference=_lowerCAmelCase , additional_params_copy_to_batched_inputs=_lowerCAmelCase , ) def lowerCamelCase__ ( self : Optional[int] ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE_: Union[str, Any] =[ """decoder_num_inference_steps""", """super_res_num_inference_steps""", ] if torch_device == "mps": # TODO: MPS errors with larger batch sizes SCREAMING_SNAKE_CASE_: Optional[int] =[2, 3] self._test_inference_batch_consistent( batch_sizes=_lowerCAmelCase , additional_params_copy_to_batched_inputs=_lowerCAmelCase , ) else: self._test_inference_batch_consistent( additional_params_copy_to_batched_inputs=_lowerCAmelCase ) @skip_mps def lowerCamelCase__ ( self : Tuple ) -> List[Any]: '''simple docstring''' return super().test_dict_tuple_outputs_equivalent() @skip_mps def lowerCamelCase__ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' return super().test_save_load_local() @skip_mps def lowerCamelCase__ ( self : Tuple ) -> Any: '''simple docstring''' return super().test_save_load_optional_components() @slow @require_torch_gpu class a ( unittest.TestCase ): def lowerCamelCase__ ( self : Optional[Any] ) -> int: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase__ ( self : List[Any] ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE_: int =load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/unclip/cat.png""" ) SCREAMING_SNAKE_CASE_: List[str] =load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/unclip/karlo_v1_alpha_cat_variation_fp16.npy""" ) SCREAMING_SNAKE_CASE_: Optional[int] =UnCLIPImageVariationPipeline.from_pretrained( """kakaobrain/karlo-v1-alpha-image-variations""" , torch_dtype=torch.floataa ) SCREAMING_SNAKE_CASE_: Optional[int] =pipeline.to(_lowerCAmelCase ) pipeline.set_progress_bar_config(disable=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] =torch.Generator(device="""cpu""" ).manual_seed(0 ) SCREAMING_SNAKE_CASE_: Dict =pipeline( _lowerCAmelCase , generator=_lowerCAmelCase , output_type="""np""" , ) SCREAMING_SNAKE_CASE_: int =output.images[0] assert image.shape == (256, 256, 3) assert_mean_pixel_difference(_lowerCAmelCase , _lowerCAmelCase , 15 )
173
'''simple docstring''' # HF Trainer benchmarking tool # # This tool can be used to run and compare multiple dimensions of the HF Trainers args. # # It then prints a report once in github format with all the information that needs to be shared # with others and second time in a console-friendly format, so it's easier to use for tuning things up. # # The main idea is: # # ./trainer-benchmark.py --base-cmd '<cmd args that don't change>' \ # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' \ # --target-metric-key train_samples_per_second # # The variations can be any command line argument that you want to compare and not just dtype as in # the example. # # --variations allows you to compare variations in multiple dimensions. # # as the first dimention has 2 options and the second 3 in our example, this will run the trainer 6 # times adding one of: # # 1. --tf32 0 --fp16 0 # 2. --tf32 0 --fp16 1 # 3. --tf32 0 --bf16 1 # 4. --tf32 1 --fp16 0 # 5. --tf32 1 --fp16 1 # 6. --tf32 1 --bf16 1 # # and print the results. This is just a cartesian product - and more than 2 dimensions can be used. # # If you want to rely on defaults, this: # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' # is identical to this: # --variations '--tf32 0|--tf32 1' '|--fp16|--bf16' # # the leading empty variation in the 2nd dimension is a valid variation. # # So here we get the following 6 variations: # # 1. --tf32 0 # 2. --tf32 0 --fp16 # 3. --tf32 0 --bf16 # 4. --tf32 1 # 5. --tf32 1 --fp16 # 6. --tf32 1 --bf16 # # In this particular case we don't know what the default tf32 setting is as it's normally # pytorch-version dependent). That's why it's best to do an explicit setting of each variation: # `--tf32 0|--tf32 1` # # Here is a full example of a train: # # CUDA_VISIBLE_DEVICES=0 python ./scripts/benchmark/trainer-benchmark.py \ # --base-cmd \ # ' examples/pytorch/translation/run_translation.py --model_name_or_path t5-small \ # --output_dir output_dir --do_train --label_smoothing 0.1 --logging_strategy no \ # --save_strategy no --per_device_train_batch_size 32 --max_source_length 512 \ # --max_target_length 512 --num_train_epochs 1 --overwrite_output_dir \ # --source_lang en --target_lang ro --dataset_name wmt16 --dataset_config "ro-en" \ # --source_prefix "translate English to Romanian: " --warmup_steps 50 \ # --max_train_samples 20000 --dataloader_num_workers 2 ' \ # --target-metric-key train_samples_per_second --repeat-times 1 --variations \ # '|--fp16|--bf16' '--tf32 0|--tf32 1' --report-metric-keys train_loss \ # --repeat-times 1 --base-variation '--tf32 0' # # and here is a possible output: # # # | Variation | Train | Diff | Train | # | | samples | % | loss | # | | per | | | # | | second | | | # |:----------------|----------:|-------:|--------:| # | --tf32 0 | 285.11 | 0 | 2.51 | # | --tf32 1 | 342.09 | 20 | 2.51 | # | --fp16 --tf32 0 | 423.49 | 49 | 2.51 | # | --fp16 --tf32 1 | 423.13 | 48 | 2.51 | # | --bf16 --tf32 0 | 416.80 | 46 | 2.52 | # | --bf16 --tf32 1 | 415.87 | 46 | 2.52 | # # # So you can quickly compare the different outcomes. # # Typically running each experiment once is enough, but if the environment is unstable you can # re-run each multiple times, e.g., 3 using --repeat-times 3 and it will report the averaged results. # # By default it'll use the lowest result as the base line to use as 100% and then compare the rest to # it as can be seen from the table above, but you can also specify which combination is the one to use as # the baseline, e.g., to change to another entry use: --base-variation '--tf32 1 --fp16 0' # # --target-metric-key is there to tell the program which metrics to compare - the different metric keys are # inside output_dir/all_results.json. e.g., to measure eval performance instead of train use: # --target-metric-key eval_samples_per_second # but of course you will need to adjust the --base-cmd value in the example to perform evaluation as # well (as currently it doesn't) # import argparse import datetime import io import itertools import json import math import os import platform import re import shlex import subprocess import sys from pathlib import Path from statistics import fmean import pandas as pd import torch from tqdm import tqdm import transformers lowerCamelCase = float("""nan""") class _UpperCamelCase : '''simple docstring''' def __init__( self : Tuple , _lowerCAmelCase : Optional[int]): '''simple docstring''' __lowercase =sys.stdout __lowercase =open(_lowerCAmelCase , 'a') def __getattr__( self : Any , _lowerCAmelCase : Union[str, Any]): '''simple docstring''' return getattr(self.stdout , _lowerCAmelCase) def __lowerCamelCase ( self : str , _lowerCAmelCase : int): '''simple docstring''' self.stdout.write(_lowerCAmelCase) # strip tqdm codes self.file.write(re.sub(R'^.*\r' , '' , _lowerCAmelCase , 0 , re.M)) def _A ( _lowerCAmelCase=80 , _lowerCAmelCase=False ): """simple docstring""" __lowercase =[] # deal with critical env vars __lowercase =['CUDA_VISIBLE_DEVICES'] for key in env_keys: __lowercase =os.environ.get(_lowerCAmelCase , _lowerCAmelCase ) if val is not None: cmd.append(f"""{key}={val}""" ) # python executable (not always needed if the script is executable) __lowercase =sys.executable if full_python_path else sys.executable.split('/' )[-1] cmd.append(_lowerCAmelCase ) # now the normal args cmd += list(map(shlex.quote , sys.argv ) ) # split up into up to MAX_WIDTH lines with shell multi-line escapes __lowercase =[] __lowercase ='' while len(_lowerCAmelCase ) > 0: current_line += f"""{cmd.pop(0 )} """ if len(_lowerCAmelCase ) == 0 or len(_lowerCAmelCase ) + len(cmd[0] ) + 1 > max_width - 1: lines.append(_lowerCAmelCase ) __lowercase ='' return "\\\n".join(_lowerCAmelCase ) def _A ( _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =re.sub(r'[\\\n]+' , ' ' , args.base_cmd ) # remove --output_dir if any and set our own __lowercase =re.sub('--output_dir\s+[^\s]+' , '' , args.base_cmd ) args.base_cmd += f""" --output_dir {output_dir}""" # ensure we have --overwrite_output_dir __lowercase =re.sub('--overwrite_output_dir\s+' , '' , args.base_cmd ) args.base_cmd += " --overwrite_output_dir" return [sys.executable] + shlex.split(args.base_cmd ) def _A ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" if 0: import random from time import sleep sleep(0 ) return dict( {k: random.uniform(0 , 100 ) for k in metric_keys} , **{target_metric_key: random.choice([nan, 10.31, 1_00.2, 55.66_66, 2_22.22_22_22_22] )} , ) __lowercase =subprocess.run(_lowerCAmelCase , capture_output=_lowerCAmelCase , text=_lowerCAmelCase ) if verbose: print('STDOUT' , result.stdout ) print('STDERR' , result.stderr ) # save the streams __lowercase =variation.replace(' ' , '-' ) with open(Path(_lowerCAmelCase ) / f"""log.{prefix}.stdout.txt""" , 'w' ) as f: f.write(result.stdout ) with open(Path(_lowerCAmelCase ) / f"""log.{prefix}.stderr.txt""" , 'w' ) as f: f.write(result.stderr ) if result.returncode != 0: if verbose: print('failed' ) return {target_metric_key: nan} with io.open(f"""{output_dir}/all_results.json""" , 'r' , encoding='utf-8' ) as f: __lowercase =json.load(_lowerCAmelCase ) # filter out just the keys we want return {k: v for k, v in metrics.items() if k in metric_keys} def _A ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ): """simple docstring""" __lowercase =[] __lowercase =[] __lowercase =f"""{id}: {variation:<{longest_variation_len}}""" __lowercase =f"""{preamble}: """ __lowercase =set(report_metric_keys + [target_metric_key] ) for i in tqdm(range(_lowerCAmelCase ) , desc=_lowerCAmelCase , leave=_lowerCAmelCase ): __lowercase =process_run_single( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) __lowercase =single_run_metrics[target_metric_key] if not math.isnan(_lowerCAmelCase ): metrics.append(_lowerCAmelCase ) results.append(_lowerCAmelCase ) outcome += "✓" else: outcome += "✘" __lowercase =f"""\33[2K\r{outcome}""" if len(_lowerCAmelCase ) > 0: __lowercase ={k: fmean([x[k] for x in metrics] ) for k in metrics[0].keys()} __lowercase =round(mean_metrics[target_metric_key] , 2 ) __lowercase =f"""{outcome} {mean_target}""" if len(_lowerCAmelCase ) > 1: results_str += f""" {tuple(round(_lowerCAmelCase , 2 ) for x in results )}""" print(_lowerCAmelCase ) __lowercase =variation return mean_metrics else: print(_lowerCAmelCase ) return {variation_key: variation, target_metric_key: nan} def _A ( ): """simple docstring""" __lowercase =torch.cuda.get_device_properties(torch.device('cuda' ) ) return f""" Datetime : {datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S' )} Software: transformers: {transformers.__version__} torch : {torch.__version__} cuda : {torch.version.cuda} python : {platform.python_version()} Hardware: {torch.cuda.device_count()} GPUs : {properties.name}, {properties.total_memory/2**30:0.2f}GB """ def _A ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =pd.DataFrame(_lowerCAmelCase ) __lowercase ='variation' __lowercase ='diff_%' __lowercase =nan if base_variation is not None and len(df[df[variation_key] == base_variation] ): # this may still return nan __lowercase =df.loc[df[variation_key] == base_variation][target_metric_key].item() if math.isnan(_lowerCAmelCase ): # as a fallback, use the minimal value as the sentinel __lowercase =df.loc[df[target_metric_key] != nan][target_metric_key].min() # create diff column if possible if not math.isnan(_lowerCAmelCase ): __lowercase =df.apply( lambda _lowerCAmelCase : round(100 * (r[target_metric_key] - sentinel_value) / sentinel_value ) if not math.isnan(r[target_metric_key] ) else 0 , axis='columns' , ) # re-order columns __lowercase =[variation_key, target_metric_key, diff_key, *report_metric_keys] __lowercase =df.reindex(_lowerCAmelCase , axis='columns' ) # reorder cols # capitalize __lowercase =df.rename(str.capitalize , axis='columns' ) # make the cols as narrow as possible __lowercase =df.rename(lambda _lowerCAmelCase : c.replace('_' , '<br>' ) , axis='columns' ) __lowercase =df.rename(lambda _lowerCAmelCase : c.replace('_' , '\n' ) , axis='columns' ) __lowercase =['', 'Copy between the cut-here-lines and paste as is to github or a forum'] report += ["----------8<-----------------8<--------"] report += ["*** Results:", df_github.to_markdown(index=_lowerCAmelCase , floatfmt='.2f' )] report += ["```"] report += ["*** Setup:", get_versions()] report += ["*** The benchmark command line was:", get_original_command()] report += ["```"] report += ["----------8<-----------------8<--------"] report += ["*** Results (console):", df_console.to_markdown(index=_lowerCAmelCase , floatfmt='.2f' )] print('\n\n'.join(_lowerCAmelCase ) ) def _A ( ): """simple docstring""" __lowercase =argparse.ArgumentParser() parser.add_argument( '--base-cmd' , default=_lowerCAmelCase , type=_lowerCAmelCase , required=_lowerCAmelCase , help='Base cmd' , ) parser.add_argument( '--variations' , default=_lowerCAmelCase , type=_lowerCAmelCase , nargs='+' , required=_lowerCAmelCase , help='Multi-dimensional variations, example: \'|--fp16|--bf16\' \'|--tf32\'' , ) parser.add_argument( '--base-variation' , default=_lowerCAmelCase , type=_lowerCAmelCase , help='Baseline variation to compare to. if None the minimal target value will be used to compare against' , ) parser.add_argument( '--target-metric-key' , default=_lowerCAmelCase , type=_lowerCAmelCase , required=_lowerCAmelCase , help='Target metric key in output_dir/all_results.json, e.g., train_samples_per_second' , ) parser.add_argument( '--report-metric-keys' , default='' , type=_lowerCAmelCase , help='Report metric keys - other metric keys from output_dir/all_results.json to report, e.g., train_loss. Use a single argument e.g., \'train_loss train_samples' , ) parser.add_argument( '--repeat-times' , default=1 , type=_lowerCAmelCase , help='How many times to re-run each variation - an average will be reported' , ) parser.add_argument( '--output_dir' , default='output_benchmark' , type=_lowerCAmelCase , help='The output directory where all the benchmark reports will go to and additionally this directory will be used to override --output_dir in the script that is being benchmarked' , ) parser.add_argument( '--verbose' , default=_lowerCAmelCase , action='store_true' , help='Whether to show the outputs of each run or just the benchmark progress' , ) __lowercase =parser.parse_args() __lowercase =args.output_dir Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) __lowercase =get_base_command(_lowerCAmelCase , _lowerCAmelCase ) # split each dimension into its --foo variations __lowercase =[list(map(str.strip , re.split(r'\|' , _lowerCAmelCase ) ) ) for x in args.variations] # build a cartesian product of dimensions and convert those back into cmd-line arg strings, # while stripping white space for inputs that were empty __lowercase =list(map(str.strip , map(' '.join , itertools.product(*_lowerCAmelCase ) ) ) ) __lowercase =max(len(_lowerCAmelCase ) for x in variations ) # split wanted keys __lowercase =args.report_metric_keys.split() # capture prints into a log file for convenience __lowercase =f"""benchmark-report-{datetime.datetime.now().strftime('%Y-%m-%d-%H-%M-%S' )}.txt""" print(f"""\nNote: each run's output is also logged under {output_dir}/log.*.std*.txt""" ) print(f"""and this script's output is also piped into {report_fn}""" ) __lowercase =Tee(_lowerCAmelCase ) print(f"""\n*** Running {len(_lowerCAmelCase )} benchmarks:""" ) print(f"""Base command: {' '.join(_lowerCAmelCase )}""" ) __lowercase ='variation' __lowercase =[] for id, variation in enumerate(tqdm(_lowerCAmelCase , desc='Total completion: ' , leave=_lowerCAmelCase ) ): __lowercase =base_cmd + variation.split() results.append( process_run( id + 1 , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , args.target_metric_key , _lowerCAmelCase , args.repeat_times , _lowerCAmelCase , args.verbose , ) ) process_results(_lowerCAmelCase , args.target_metric_key , _lowerCAmelCase , args.base_variation , _lowerCAmelCase ) if __name__ == "__main__": main()
166
0
def __lowerCamelCase (UpperCAmelCase__ : int , UpperCAmelCase__ : int ): while b: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = b, a % b return a def __lowerCamelCase (UpperCAmelCase__ : int , UpperCAmelCase__ : int ): return a if b == 0 else euclidean_gcd_recursive(UpperCAmelCase__ , a % b ) def __lowerCamelCase (): print(F"euclidean_gcd(3, 5) = {euclidean_gcd(3 , 5 )}" ) print(F"euclidean_gcd(5, 3) = {euclidean_gcd(5 , 3 )}" ) print(F"euclidean_gcd(1, 3) = {euclidean_gcd(1 , 3 )}" ) print(F"euclidean_gcd(3, 6) = {euclidean_gcd(3 , 6 )}" ) print(F"euclidean_gcd(6, 3) = {euclidean_gcd(6 , 3 )}" ) print(F"euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3 , 5 )}" ) print(F"euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5 , 3 )}" ) print(F"euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1 , 3 )}" ) print(F"euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3 , 6 )}" ) print(F"euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6 , 3 )}" ) if __name__ == "__main__": main()
206
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class lowercase ( a ): def __snake_case( self : Optional[int] ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_UpperCamelCase , "tf_padding" ) ) self.parent.assertTrue(hasattr(_UpperCamelCase , "depth_multiplier" ) ) class lowercase : def __init__( self : Tuple , _UpperCamelCase : Optional[Any] , _UpperCamelCase : Any=13 , _UpperCamelCase : Any=3 , _UpperCamelCase : Union[str, Any]=32 , _UpperCamelCase : Optional[Any]=0.2_5 , _UpperCamelCase : int=8 , _UpperCamelCase : str=True , _UpperCamelCase : Any=1_024 , _UpperCamelCase : Tuple=32 , _UpperCamelCase : List[str]="relu6" , _UpperCamelCase : Tuple=0.1 , _UpperCamelCase : List[str]=0.0_2 , _UpperCamelCase : int=True , _UpperCamelCase : int=True , _UpperCamelCase : Optional[Any]=10 , _UpperCamelCase : List[str]=None , ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = parent SCREAMING_SNAKE_CASE = batch_size SCREAMING_SNAKE_CASE = num_channels SCREAMING_SNAKE_CASE = image_size SCREAMING_SNAKE_CASE = depth_multiplier SCREAMING_SNAKE_CASE = min_depth SCREAMING_SNAKE_CASE = tf_padding SCREAMING_SNAKE_CASE = int(last_hidden_size * depth_multiplier ) SCREAMING_SNAKE_CASE = output_stride SCREAMING_SNAKE_CASE = hidden_act SCREAMING_SNAKE_CASE = classifier_dropout_prob SCREAMING_SNAKE_CASE = use_labels SCREAMING_SNAKE_CASE = is_training SCREAMING_SNAKE_CASE = num_labels SCREAMING_SNAKE_CASE = initializer_range SCREAMING_SNAKE_CASE = scope def __snake_case( self : Dict ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = None if self.use_labels: SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size] , self.num_labels ) SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) SCREAMING_SNAKE_CASE = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case( self : Optional[Any] ) -> str: '''simple docstring''' return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , min_depth=self.min_depth , tf_padding=self.tf_padding , hidden_act=self.hidden_act , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case( self : int , _UpperCamelCase : Optional[int] , _UpperCamelCase : Any , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : int ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = MobileNetVaModel(config=_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() SCREAMING_SNAKE_CASE = model(_UpperCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case( self : Dict , _UpperCamelCase : Tuple , _UpperCamelCase : Any , _UpperCamelCase : str , _UpperCamelCase : List[Any] ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.num_labels SCREAMING_SNAKE_CASE = MobileNetVaForImageClassification(_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() SCREAMING_SNAKE_CASE = model(_UpperCamelCase , labels=_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case( self : Optional[Any] ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = config_and_inputs SCREAMING_SNAKE_CASE = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowercase ( a , a , unittest.TestCase ): lowercase__ : Dict = (MobileNetVaModel, MobileNetVaForImageClassification) if is_torch_available() else () lowercase__ : Tuple = ( {"""feature-extraction""": MobileNetVaModel, """image-classification""": MobileNetVaForImageClassification} if is_torch_available() else {} ) lowercase__ : Union[str, Any] = False lowercase__ : Union[str, Any] = False lowercase__ : Tuple = False lowercase__ : List[str] = False def __snake_case( self : List[str] ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = MobileNetVaModelTester(self ) SCREAMING_SNAKE_CASE = MobileNetVaConfigTester(self , config_class=_UpperCamelCase , has_text_modality=_UpperCamelCase ) def __snake_case( self : Union[str, Any] ) -> List[Any]: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="MobileNetV1 does not use inputs_embeds" ) def __snake_case( self : Tuple ) -> Tuple: '''simple docstring''' pass @unittest.skip(reason="MobileNetV1 does not support input and output embeddings" ) def __snake_case( self : Optional[int] ) -> Dict: '''simple docstring''' pass @unittest.skip(reason="MobileNetV1 does not output attentions" ) def __snake_case( self : Any ) -> List[str]: '''simple docstring''' pass def __snake_case( self : List[Any] ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE = model_class(_UpperCamelCase ) SCREAMING_SNAKE_CASE = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE = ["pixel_values"] self.assertListEqual(arg_names[:1] , _UpperCamelCase ) def __snake_case( self : List[Any] ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_UpperCamelCase ) def __snake_case( self : List[str] ) -> Optional[int]: '''simple docstring''' def check_hidden_states_output(_UpperCamelCase : Any , _UpperCamelCase : Dict , _UpperCamelCase : Tuple ): SCREAMING_SNAKE_CASE = model_class(_UpperCamelCase ) model.to(_UpperCamelCase ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE = model(**self._prepare_for_class(_UpperCamelCase , _UpperCamelCase ) ) SCREAMING_SNAKE_CASE = outputs.hidden_states SCREAMING_SNAKE_CASE = 26 self.assertEqual(len(_UpperCamelCase ) , _UpperCamelCase ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE = True check_hidden_states_output(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] SCREAMING_SNAKE_CASE = True check_hidden_states_output(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) def __snake_case( self : Any ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_UpperCamelCase ) @slow def __snake_case( self : int ) -> str: '''simple docstring''' for model_name in MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE = MobileNetVaModel.from_pretrained(_UpperCamelCase ) self.assertIsNotNone(_UpperCamelCase ) def __lowerCamelCase (): SCREAMING_SNAKE_CASE = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class lowercase ( unittest.TestCase ): @cached_property def __snake_case( self : List[Any] ) -> Optional[int]: '''simple docstring''' return ( MobileNetVaImageProcessor.from_pretrained("google/mobilenet_v1_1.0_224" ) if is_vision_available() else None ) @slow def __snake_case( self : str ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = MobileNetVaForImageClassification.from_pretrained("google/mobilenet_v1_1.0_224" ).to(_UpperCamelCase ) SCREAMING_SNAKE_CASE = self.default_image_processor SCREAMING_SNAKE_CASE = prepare_img() SCREAMING_SNAKE_CASE = image_processor(images=_UpperCamelCase , return_tensors="pt" ).to(_UpperCamelCase ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE = model(**_UpperCamelCase ) # verify the logits SCREAMING_SNAKE_CASE = torch.Size((1, 1_001) ) self.assertEqual(outputs.logits.shape , _UpperCamelCase ) SCREAMING_SNAKE_CASE = torch.tensor([-4.1_7_3_9, -1.1_2_3_3, 3.1_2_0_5] ).to(_UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _UpperCamelCase , atol=1e-4 ) )
206
1
'''simple docstring''' from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES lowerCamelCase : Any = logging.get_logger(__name__) lowerCamelCase : Dict = OrderedDict( [ # Base model mapping ("albert", "FlaxAlbertModel"), ("bart", "FlaxBartModel"), ("beit", "FlaxBeitModel"), ("bert", "FlaxBertModel"), ("big_bird", "FlaxBigBirdModel"), ("blenderbot", "FlaxBlenderbotModel"), ("blenderbot-small", "FlaxBlenderbotSmallModel"), ("clip", "FlaxCLIPModel"), ("distilbert", "FlaxDistilBertModel"), ("electra", "FlaxElectraModel"), ("gpt-sw3", "FlaxGPT2Model"), ("gpt2", "FlaxGPT2Model"), ("gpt_neo", "FlaxGPTNeoModel"), ("gptj", "FlaxGPTJModel"), ("longt5", "FlaxLongT5Model"), ("marian", "FlaxMarianModel"), ("mbart", "FlaxMBartModel"), ("mt5", "FlaxMT5Model"), ("opt", "FlaxOPTModel"), ("pegasus", "FlaxPegasusModel"), ("regnet", "FlaxRegNetModel"), ("resnet", "FlaxResNetModel"), ("roberta", "FlaxRobertaModel"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormModel"), ("roformer", "FlaxRoFormerModel"), ("t5", "FlaxT5Model"), ("vision-text-dual-encoder", "FlaxVisionTextDualEncoderModel"), ("vit", "FlaxViTModel"), ("wav2vec2", "FlaxWav2Vec2Model"), ("whisper", "FlaxWhisperModel"), ("xglm", "FlaxXGLMModel"), ("xlm-roberta", "FlaxXLMRobertaModel"), ] ) lowerCamelCase : Union[str, Any] = OrderedDict( [ # Model for pre-training mapping ("albert", "FlaxAlbertForPreTraining"), ("bart", "FlaxBartForConditionalGeneration"), ("bert", "FlaxBertForPreTraining"), ("big_bird", "FlaxBigBirdForPreTraining"), ("electra", "FlaxElectraForPreTraining"), ("longt5", "FlaxLongT5ForConditionalGeneration"), ("mbart", "FlaxMBartForConditionalGeneration"), ("mt5", "FlaxMT5ForConditionalGeneration"), ("roberta", "FlaxRobertaForMaskedLM"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormForMaskedLM"), ("roformer", "FlaxRoFormerForMaskedLM"), ("t5", "FlaxT5ForConditionalGeneration"), ("wav2vec2", "FlaxWav2Vec2ForPreTraining"), ("whisper", "FlaxWhisperForConditionalGeneration"), ("xlm-roberta", "FlaxXLMRobertaForMaskedLM"), ] ) lowerCamelCase : Dict = OrderedDict( [ # Model for Masked LM mapping ("albert", "FlaxAlbertForMaskedLM"), ("bart", "FlaxBartForConditionalGeneration"), ("bert", "FlaxBertForMaskedLM"), ("big_bird", "FlaxBigBirdForMaskedLM"), ("distilbert", "FlaxDistilBertForMaskedLM"), ("electra", "FlaxElectraForMaskedLM"), ("mbart", "FlaxMBartForConditionalGeneration"), ("roberta", "FlaxRobertaForMaskedLM"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormForMaskedLM"), ("roformer", "FlaxRoFormerForMaskedLM"), ("xlm-roberta", "FlaxXLMRobertaForMaskedLM"), ] ) lowerCamelCase : Any = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ("bart", "FlaxBartForConditionalGeneration"), ("blenderbot", "FlaxBlenderbotForConditionalGeneration"), ("blenderbot-small", "FlaxBlenderbotSmallForConditionalGeneration"), ("encoder-decoder", "FlaxEncoderDecoderModel"), ("longt5", "FlaxLongT5ForConditionalGeneration"), ("marian", "FlaxMarianMTModel"), ("mbart", "FlaxMBartForConditionalGeneration"), ("mt5", "FlaxMT5ForConditionalGeneration"), ("pegasus", "FlaxPegasusForConditionalGeneration"), ("t5", "FlaxT5ForConditionalGeneration"), ] ) lowerCamelCase : Union[str, Any] = OrderedDict( [ # Model for Image-classsification ("beit", "FlaxBeitForImageClassification"), ("regnet", "FlaxRegNetForImageClassification"), ("resnet", "FlaxResNetForImageClassification"), ("vit", "FlaxViTForImageClassification"), ] ) lowerCamelCase : Any = OrderedDict( [ ("vision-encoder-decoder", "FlaxVisionEncoderDecoderModel"), ] ) lowerCamelCase : List[Any] = OrderedDict( [ # Model for Causal LM mapping ("bart", "FlaxBartForCausalLM"), ("bert", "FlaxBertForCausalLM"), ("big_bird", "FlaxBigBirdForCausalLM"), ("electra", "FlaxElectraForCausalLM"), ("gpt-sw3", "FlaxGPT2LMHeadModel"), ("gpt2", "FlaxGPT2LMHeadModel"), ("gpt_neo", "FlaxGPTNeoForCausalLM"), ("gptj", "FlaxGPTJForCausalLM"), ("opt", "FlaxOPTForCausalLM"), ("roberta", "FlaxRobertaForCausalLM"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormForCausalLM"), ("xglm", "FlaxXGLMForCausalLM"), ("xlm-roberta", "FlaxXLMRobertaForCausalLM"), ] ) lowerCamelCase : Optional[int] = OrderedDict( [ # Model for Sequence Classification mapping ("albert", "FlaxAlbertForSequenceClassification"), ("bart", "FlaxBartForSequenceClassification"), ("bert", "FlaxBertForSequenceClassification"), ("big_bird", "FlaxBigBirdForSequenceClassification"), ("distilbert", "FlaxDistilBertForSequenceClassification"), ("electra", "FlaxElectraForSequenceClassification"), ("mbart", "FlaxMBartForSequenceClassification"), ("roberta", "FlaxRobertaForSequenceClassification"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormForSequenceClassification"), ("roformer", "FlaxRoFormerForSequenceClassification"), ("xlm-roberta", "FlaxXLMRobertaForSequenceClassification"), ] ) lowerCamelCase : Optional[Any] = OrderedDict( [ # Model for Question Answering mapping ("albert", "FlaxAlbertForQuestionAnswering"), ("bart", "FlaxBartForQuestionAnswering"), ("bert", "FlaxBertForQuestionAnswering"), ("big_bird", "FlaxBigBirdForQuestionAnswering"), ("distilbert", "FlaxDistilBertForQuestionAnswering"), ("electra", "FlaxElectraForQuestionAnswering"), ("mbart", "FlaxMBartForQuestionAnswering"), ("roberta", "FlaxRobertaForQuestionAnswering"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormForQuestionAnswering"), ("roformer", "FlaxRoFormerForQuestionAnswering"), ("xlm-roberta", "FlaxXLMRobertaForQuestionAnswering"), ] ) lowerCamelCase : List[str] = OrderedDict( [ # Model for Token Classification mapping ("albert", "FlaxAlbertForTokenClassification"), ("bert", "FlaxBertForTokenClassification"), ("big_bird", "FlaxBigBirdForTokenClassification"), ("distilbert", "FlaxDistilBertForTokenClassification"), ("electra", "FlaxElectraForTokenClassification"), ("roberta", "FlaxRobertaForTokenClassification"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormForTokenClassification"), ("roformer", "FlaxRoFormerForTokenClassification"), ("xlm-roberta", "FlaxXLMRobertaForTokenClassification"), ] ) lowerCamelCase : Optional[int] = OrderedDict( [ # Model for Multiple Choice mapping ("albert", "FlaxAlbertForMultipleChoice"), ("bert", "FlaxBertForMultipleChoice"), ("big_bird", "FlaxBigBirdForMultipleChoice"), ("distilbert", "FlaxDistilBertForMultipleChoice"), ("electra", "FlaxElectraForMultipleChoice"), ("roberta", "FlaxRobertaForMultipleChoice"), ("roberta-prelayernorm", "FlaxRobertaPreLayerNormForMultipleChoice"), ("roformer", "FlaxRoFormerForMultipleChoice"), ("xlm-roberta", "FlaxXLMRobertaForMultipleChoice"), ] ) lowerCamelCase : str = OrderedDict( [ ("bert", "FlaxBertForNextSentencePrediction"), ] ) lowerCamelCase : Dict = OrderedDict( [ ("speech-encoder-decoder", "FlaxSpeechEncoderDecoderModel"), ("whisper", "FlaxWhisperForConditionalGeneration"), ] ) lowerCamelCase : Optional[int] = OrderedDict( [ ("whisper", "FlaxWhisperForAudioClassification"), ] ) lowerCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) lowerCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) lowerCamelCase : Union[str, Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) lowerCamelCase : List[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) lowerCamelCase : Any = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) lowerCamelCase : Union[str, Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) lowerCamelCase : Optional[int] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) lowerCamelCase : Dict = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) lowerCamelCase : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) lowerCamelCase : Optional[int] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) lowerCamelCase : Optional[Any] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) lowerCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) lowerCamelCase : Optional[int] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) lowerCamelCase : List[str] = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class A__ ( _BaseAutoModelClass ): A__ = FLAX_MODEL_MAPPING lowerCamelCase : int = auto_class_update(FlaxAutoModel) class A__ ( _BaseAutoModelClass ): A__ = FLAX_MODEL_FOR_PRETRAINING_MAPPING lowerCamelCase : Any = auto_class_update(FlaxAutoModelForPreTraining, head_doc="pretraining") class A__ ( _BaseAutoModelClass ): A__ = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING lowerCamelCase : Union[str, Any] = auto_class_update(FlaxAutoModelForCausalLM, head_doc="causal language modeling") class A__ ( _BaseAutoModelClass ): A__ = FLAX_MODEL_FOR_MASKED_LM_MAPPING lowerCamelCase : List[Any] = auto_class_update(FlaxAutoModelForMaskedLM, head_doc="masked language modeling") class A__ ( _BaseAutoModelClass ): A__ = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING lowerCamelCase : Any = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc="sequence-to-sequence language modeling", checkpoint_for_example="t5-base" ) class A__ ( _BaseAutoModelClass ): A__ = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING lowerCamelCase : Tuple = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc="sequence classification" ) class A__ ( _BaseAutoModelClass ): A__ = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING lowerCamelCase : Optional[int] = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc="question answering") class A__ ( _BaseAutoModelClass ): A__ = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING lowerCamelCase : str = auto_class_update( FlaxAutoModelForTokenClassification, head_doc="token classification" ) class A__ ( _BaseAutoModelClass ): A__ = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING lowerCamelCase : Tuple = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc="multiple choice") class A__ ( _BaseAutoModelClass ): A__ = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING lowerCamelCase : List[str] = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc="next sentence prediction" ) class A__ ( _BaseAutoModelClass ): A__ = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING lowerCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForImageClassification, head_doc="image classification" ) class A__ ( _BaseAutoModelClass ): A__ = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING lowerCamelCase : List[str] = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc="vision-to-text modeling") class A__ ( _BaseAutoModelClass ): A__ = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING lowerCamelCase : Optional[Any] = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc="sequence-to-sequence speech-to-text modeling" )
47
'''simple docstring''' class A__ : def __init__( self : Union[str, Any] , _a : int ) -> None: '''simple docstring''' _SCREAMING_SNAKE_CASE =size _SCREAMING_SNAKE_CASE =[0] * size _SCREAMING_SNAKE_CASE =[0] * size @staticmethod def A ( _a : int ) -> int: '''simple docstring''' return index | (index + 1) @staticmethod def A ( _a : int ) -> int: '''simple docstring''' return (index & (index + 1)) - 1 def A ( self : Tuple , _a : int , _a : int ) -> None: '''simple docstring''' _SCREAMING_SNAKE_CASE =value while index < self.size: _SCREAMING_SNAKE_CASE =self.get_prev(_a ) + 1 if current_left_border == index: _SCREAMING_SNAKE_CASE =value else: _SCREAMING_SNAKE_CASE =max(_a , _a , _a ) _SCREAMING_SNAKE_CASE =self.get_next(_a ) def A ( self : int , _a : int , _a : int ) -> int: '''simple docstring''' right -= 1 # Because of right is exclusive _SCREAMING_SNAKE_CASE =0 while left <= right: _SCREAMING_SNAKE_CASE =self.get_prev(_a ) if left <= current_left: _SCREAMING_SNAKE_CASE =max(_a , self.tree[right] ) _SCREAMING_SNAKE_CASE =current_left else: _SCREAMING_SNAKE_CASE =max(_a , self.arr[right] ) right -= 1 return result if __name__ == "__main__": import doctest doctest.testmod()
47
1
import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class lowercase : '''simple docstring''' def __init__(self , __a , __a=3 , __a=32 , __a=3 , __a=10 , __a=[8, 16, 32, 64] , __a=[1, 1, 2, 1] , __a=True , __a=True , __a="relu" , __a=3 , __a=None , __a=["stage2", "stage3", "stage4"] , __a=[2, 3, 4] , __a=1 , ) -> int: """simple docstring""" UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = image_size UpperCAmelCase__ = num_channels UpperCAmelCase__ = embeddings_size UpperCAmelCase__ = hidden_sizes UpperCAmelCase__ = depths UpperCAmelCase__ = is_training UpperCAmelCase__ = use_labels UpperCAmelCase__ = hidden_act UpperCAmelCase__ = num_labels UpperCAmelCase__ = scope UpperCAmelCase__ = len(__a ) UpperCAmelCase__ = out_features UpperCAmelCase__ = out_indices UpperCAmelCase__ = num_groups def UpperCamelCase__ (self ) -> Dict: """simple docstring""" UpperCAmelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase__ = None if self.use_labels: UpperCAmelCase__ = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase__ = self.get_config() return config, pixel_values, labels def UpperCamelCase__ (self ) -> int: """simple docstring""" return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def UpperCamelCase__ (self , __a , __a , __a ) -> Dict: """simple docstring""" UpperCAmelCase__ = BitModel(config=__a ) model.to(__a ) model.eval() UpperCAmelCase__ = model(__a ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCamelCase__ (self , __a , __a , __a ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = self.num_labels UpperCAmelCase__ = BitForImageClassification(__a ) model.to(__a ) model.eval() UpperCAmelCase__ = model(__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase__ (self , __a , __a , __a ) -> Dict: """simple docstring""" UpperCAmelCase__ = BitBackbone(config=__a ) model.to(__a ) model.eval() UpperCAmelCase__ = model(__a ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None UpperCAmelCase__ = None UpperCAmelCase__ = BitBackbone(config=__a ) model.to(__a ) model.eval() UpperCAmelCase__ = model(__a ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def UpperCamelCase__ (self ) -> str: """simple docstring""" UpperCAmelCase__ = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = config_and_inputs UpperCAmelCase__ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class lowercase ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): '''simple docstring''' __SCREAMING_SNAKE_CASE = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () __SCREAMING_SNAKE_CASE = ( {"""feature-extraction""": BitModel, """image-classification""": BitForImageClassification} if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def UpperCamelCase__ (self ) -> str: """simple docstring""" UpperCAmelCase__ = BitModelTester(self ) UpperCAmelCase__ = ConfigTester(self , config_class=__a , has_text_modality=__a ) def UpperCamelCase__ (self ) -> Optional[int]: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase__ (self ) -> int: """simple docstring""" return @unittest.skip(reason='Bit does not output attentions' ) def UpperCamelCase__ (self ) -> int: """simple docstring""" pass @unittest.skip(reason='Bit does not use inputs_embeds' ) def UpperCamelCase__ (self ) -> Dict: """simple docstring""" pass @unittest.skip(reason='Bit does not support input and output embeddings' ) def UpperCamelCase__ (self ) -> int: """simple docstring""" pass def UpperCamelCase__ (self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(__a ) UpperCAmelCase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ = [*signature.parameters.keys()] UpperCAmelCase__ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __a ) def UpperCamelCase__ (self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a ) def UpperCamelCase__ (self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*__a ) def UpperCamelCase__ (self ) -> str: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(config=__a ) for name, module in model.named_modules(): if isinstance(__a , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F"Parameter {name} of model {model_class} seems not properly initialized" , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F"Parameter {name} of model {model_class} seems not properly initialized" , ) def UpperCamelCase__ (self ) -> str: """simple docstring""" def check_hidden_states_output(__a , __a , __a ): UpperCAmelCase__ = model_class(__a ) model.to(__a ) model.eval() with torch.no_grad(): UpperCAmelCase__ = model(**self._prepare_for_class(__a , __a ) ) UpperCAmelCase__ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCAmelCase__ = self.model_tester.num_stages self.assertEqual(len(__a ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ = ['preactivation', 'bottleneck'] for model_class in self.all_model_classes: for layer_type in layers_type: UpperCAmelCase__ = layer_type UpperCAmelCase__ = True check_hidden_states_output(__a , __a , __a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase__ = True check_hidden_states_output(__a , __a , __a ) @unittest.skip(reason='Bit does not use feedforward chunking' ) def UpperCamelCase__ (self ) -> Any: """simple docstring""" pass def UpperCamelCase__ (self ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__a ) @slow def UpperCamelCase__ (self ) -> int: """simple docstring""" for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ = BitModel.from_pretrained(__a ) self.assertIsNotNone(__a ) def UpperCamelCase_( ) -> Dict: UpperCAmelCase__ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class lowercase ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase__ (self ) -> Union[str, Any]: """simple docstring""" return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCamelCase__ (self ) -> Tuple: """simple docstring""" UpperCAmelCase__ = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(__a ) UpperCAmelCase__ = self.default_image_processor UpperCAmelCase__ = prepare_img() UpperCAmelCase__ = image_processor(images=__a , return_tensors='pt' ).to(__a ) # forward pass with torch.no_grad(): UpperCAmelCase__ = model(**__a ) # verify the logits UpperCAmelCase__ = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , __a ) UpperCAmelCase__ = torch.tensor([[-0.65_26, -0.52_63, -1.43_98]] ).to(__a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __a , atol=1E-4 ) ) @require_torch class lowercase ( _UpperCamelCase , unittest.TestCase ): '''simple docstring''' __SCREAMING_SNAKE_CASE = (BitBackbone,) if is_torch_available() else () __SCREAMING_SNAKE_CASE = BitConfig __SCREAMING_SNAKE_CASE = False def UpperCamelCase__ (self ) -> Any: """simple docstring""" UpperCAmelCase__ = BitModelTester(self )
368
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from transformers import TensorFlowBenchmark, TensorFlowBenchmarkArguments @require_tf class lowercase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ (self , __a ) -> List[Any]: """simple docstring""" for model_result in results.values(): for batch_size, sequence_length in zip(model_result['bs'] , model_result['ss'] ): UpperCAmelCase__ = model_result['result'][batch_size][sequence_length] self.assertIsNotNone(__a ) def UpperCamelCase__ (self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = 'sshleifer/tiny-gpt2' UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=__a , multi_process=__a , ) UpperCAmelCase__ = TensorFlowBenchmark(__a ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def UpperCamelCase__ (self ) -> Tuple: """simple docstring""" UpperCAmelCase__ = 'sgugger/tiny-distilbert-classification' UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__a , only_pretrain_model=__a , ) UpperCAmelCase__ = TensorFlowBenchmark(__a ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def UpperCamelCase__ (self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = 'sshleifer/tiny-gpt2' UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__a , ) UpperCAmelCase__ = TensorFlowBenchmark(__a ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def UpperCamelCase__ (self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = 'sshleifer/tiny-gpt2' UpperCAmelCase__ = AutoConfig.from_pretrained(__a ) UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , eager_mode=__a , multi_process=__a , ) UpperCAmelCase__ = TensorFlowBenchmark(__a , [config] ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def UpperCamelCase__ (self ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = 'sshleifer/tiny-gpt2' UpperCAmelCase__ = AutoConfig.from_pretrained(__a ) UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__a , ) UpperCAmelCase__ = TensorFlowBenchmark(__a , [config] ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def UpperCamelCase__ (self ) -> Dict: """simple docstring""" UpperCAmelCase__ = 'sshleifer/tiny-gpt2' UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__a , ) UpperCAmelCase__ = TensorFlowBenchmark(__a ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def UpperCamelCase__ (self ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = 'sshleifer/tiny-gpt2' UpperCAmelCase__ = AutoConfig.from_pretrained(__a ) UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__a , ) UpperCAmelCase__ = TensorFlowBenchmark(__a , [config] ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def UpperCamelCase__ (self ) -> Tuple: """simple docstring""" UpperCAmelCase__ = 'patrickvonplaten/t5-tiny-random' UpperCAmelCase__ = AutoConfig.from_pretrained(__a ) UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , multi_process=__a , ) UpperCAmelCase__ = TensorFlowBenchmark(__a , configs=[config] ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(is_tf_available() and len(tf.config.list_physical_devices('GPU' ) ) == 0 , 'Cannot do xla on CPU.' ) def UpperCamelCase__ (self ) -> List[str]: """simple docstring""" UpperCAmelCase__ = 'sshleifer/tiny-gpt2' UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , training=__a , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , use_xla=__a , multi_process=__a , ) UpperCAmelCase__ = TensorFlowBenchmark(__a ) UpperCAmelCase__ = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def UpperCamelCase__ (self ) -> Tuple: """simple docstring""" UpperCAmelCase__ = 'sshleifer/tiny-gpt2' with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=__a , save_to_csv=__a , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(__a , 'inf_time.csv' ) , inference_memory_csv_file=os.path.join(__a , 'inf_mem.csv' ) , env_info_csv_file=os.path.join(__a , 'env.csv' ) , multi_process=__a , ) UpperCAmelCase__ = TensorFlowBenchmark(__a ) benchmark.run() self.assertTrue(Path(os.path.join(__a , 'inf_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(__a , 'inf_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(__a , 'env.csv' ) ).exists() ) def UpperCamelCase__ (self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = 'sshleifer/tiny-gpt2' def _check_summary_is_not_empty(__a ): self.assertTrue(hasattr(__a , 'sequential' ) ) self.assertTrue(hasattr(__a , 'cumulative' ) ) self.assertTrue(hasattr(__a , 'current' ) ) self.assertTrue(hasattr(__a , 'total' ) ) with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase__ = TensorFlowBenchmarkArguments( models=[MODEL_ID] , inference=__a , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(__a , 'log.txt' ) , log_print=__a , trace_memory_line_by_line=__a , eager_mode=__a , multi_process=__a , ) UpperCAmelCase__ = TensorFlowBenchmark(__a ) UpperCAmelCase__ = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) self.assertTrue(Path(os.path.join(__a , 'log.txt' ) ).exists() )
335
0
'''simple docstring''' from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class lowercase__ : '''simple docstring''' A_ : Dict = XGLMConfig A_ : List[Any] = {} A_ : Optional[int] = """gelu""" def __init__( self , __snake_case , __snake_case=14 , __snake_case=7 , __snake_case=True , __snake_case=True , __snake_case=True , __snake_case=99 , __snake_case=32 , __snake_case=2 , __snake_case=4 , __snake_case=37 , __snake_case="gelu" , __snake_case=0.1 , __snake_case=0.1 , __snake_case=512 , __snake_case=0.02 , ): _SCREAMING_SNAKE_CASE : Dict = parent _SCREAMING_SNAKE_CASE : Dict = batch_size _SCREAMING_SNAKE_CASE : Tuple = seq_length _SCREAMING_SNAKE_CASE : int = is_training _SCREAMING_SNAKE_CASE : int = use_input_mask _SCREAMING_SNAKE_CASE : List[str] = use_labels _SCREAMING_SNAKE_CASE : Union[str, Any] = vocab_size _SCREAMING_SNAKE_CASE : Union[str, Any] = d_model _SCREAMING_SNAKE_CASE : Tuple = num_hidden_layers _SCREAMING_SNAKE_CASE : Optional[int] = num_attention_heads _SCREAMING_SNAKE_CASE : Optional[Any] = ffn_dim _SCREAMING_SNAKE_CASE : Any = activation_function _SCREAMING_SNAKE_CASE : Tuple = activation_dropout _SCREAMING_SNAKE_CASE : Union[str, Any] = attention_dropout _SCREAMING_SNAKE_CASE : int = max_position_embeddings _SCREAMING_SNAKE_CASE : List[str] = initializer_range _SCREAMING_SNAKE_CASE : Any = None _SCREAMING_SNAKE_CASE : Optional[int] = 0 _SCREAMING_SNAKE_CASE : int = 2 _SCREAMING_SNAKE_CASE : int = 1 def UpperCAmelCase_ ( self ): return XGLMConfig.from_pretrained("""facebook/xglm-564M""" ) def UpperCAmelCase_ ( self ): _SCREAMING_SNAKE_CASE : Union[str, Any] = tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) , clip_value_min=0 , clip_value_max=3 ) _SCREAMING_SNAKE_CASE : Any = None if self.use_input_mask: _SCREAMING_SNAKE_CASE : Tuple = random_attention_mask([self.batch_size, self.seq_length] ) _SCREAMING_SNAKE_CASE : List[Any] = self.get_config() _SCREAMING_SNAKE_CASE : Tuple = floats_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, ) def UpperCAmelCase_ ( self ): return XGLMConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , num_layers=self.num_hidden_layers , attention_heads=self.num_attention_heads , ffn_dim=self.ffn_dim , activation_function=self.activation_function , activation_dropout=self.activation_dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , use_cache=__snake_case , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , return_dict=__snake_case , ) def UpperCAmelCase_ ( self ): _SCREAMING_SNAKE_CASE : Dict = self.prepare_config_and_inputs() ( ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ( _SCREAMING_SNAKE_CASE ) , ) : Union[str, Any] = config_and_inputs _SCREAMING_SNAKE_CASE : Dict = { """input_ids""": input_ids, """head_mask""": head_mask, } return config, inputs_dict @require_tf class lowercase__ ( _snake_case , _snake_case , unittest.TestCase ): '''simple docstring''' A_ : List[Any] = (TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () A_ : Union[str, Any] = (TFXGLMForCausalLM,) if is_tf_available() else () A_ : List[str] = ( {"""feature-extraction""": TFXGLMModel, """text-generation""": TFXGLMForCausalLM} if is_tf_available() else {} ) A_ : Union[str, Any] = False A_ : Optional[Any] = False A_ : Union[str, Any] = False def UpperCAmelCase_ ( self ): _SCREAMING_SNAKE_CASE : Optional[Any] = TFXGLMModelTester(self ) _SCREAMING_SNAKE_CASE : Any = ConfigTester(self , config_class=__snake_case , n_embd=37 ) def UpperCAmelCase_ ( self ): self.config_tester.run_common_tests() @slow def UpperCAmelCase_ ( self ): for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _SCREAMING_SNAKE_CASE : Union[str, Any] = TFXGLMModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) @unittest.skip(reason="""Currently, model embeddings are going to undergo a major refactor.""" ) def UpperCAmelCase_ ( self ): super().test_resize_token_embeddings() @require_tf class lowercase__ ( unittest.TestCase ): '''simple docstring''' @slow def UpperCAmelCase_ ( self , __snake_case=True ): _SCREAMING_SNAKE_CASE : str = TFXGLMForCausalLM.from_pretrained("""facebook/xglm-564M""" ) _SCREAMING_SNAKE_CASE : Dict = tf.convert_to_tensor([[2, 268, 9865]] , dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off _SCREAMING_SNAKE_CASE : Dict = [2, 268, 9865, 67, 11, 1988, 5_7252, 9865, 5, 984, 67, 1988, 21_3838, 1658, 53, 7_0446, 33, 6657, 278, 1581] # fmt: on _SCREAMING_SNAKE_CASE : int = model.generate(__snake_case , do_sample=__snake_case , num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist() , __snake_case ) @slow def UpperCAmelCase_ ( self ): _SCREAMING_SNAKE_CASE : Tuple = XGLMTokenizer.from_pretrained("""facebook/xglm-564M""" ) _SCREAMING_SNAKE_CASE : Optional[int] = TFXGLMForCausalLM.from_pretrained("""facebook/xglm-564M""" ) tf.random.set_seed(0 ) _SCREAMING_SNAKE_CASE : int = tokenizer("""Today is a nice day and""" , return_tensors="""tf""" ) _SCREAMING_SNAKE_CASE : Optional[Any] = tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(""":/CPU:0""" ): _SCREAMING_SNAKE_CASE : Dict = model.generate(__snake_case , do_sample=__snake_case , seed=[7, 0] ) _SCREAMING_SNAKE_CASE : Optional[int] = tokenizer.decode(output_ids[0] , skip_special_tokens=__snake_case ) _SCREAMING_SNAKE_CASE : Optional[Any] = ( """Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due""" ) self.assertEqual(__snake_case , __snake_case ) @slow def UpperCAmelCase_ ( self ): _SCREAMING_SNAKE_CASE : str = TFXGLMForCausalLM.from_pretrained("""facebook/xglm-564M""" ) _SCREAMING_SNAKE_CASE : Any = XGLMTokenizer.from_pretrained("""facebook/xglm-564M""" ) _SCREAMING_SNAKE_CASE : Dict = """left""" # use different length sentences to test batching _SCREAMING_SNAKE_CASE : int = [ """This is an extremelly long sentence that only exists to test the ability of the model to cope with """ """left-padding, such as in batched generation. The output for the sequence below should be the same """ """regardless of whether left padding is applied or not. When""", """Hello, my dog is a little""", ] _SCREAMING_SNAKE_CASE : str = tokenizer(__snake_case , return_tensors="""tf""" , padding=__snake_case ) _SCREAMING_SNAKE_CASE : List[str] = inputs["""input_ids"""] _SCREAMING_SNAKE_CASE : Union[str, Any] = model.generate(input_ids=__snake_case , attention_mask=inputs["""attention_mask"""] , max_new_tokens=12 ) _SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer(sentences[0] , return_tensors="""tf""" ).input_ids _SCREAMING_SNAKE_CASE : List[Any] = model.generate(input_ids=__snake_case , max_new_tokens=12 ) _SCREAMING_SNAKE_CASE : int = tokenizer(sentences[1] , return_tensors="""tf""" ).input_ids _SCREAMING_SNAKE_CASE : Dict = model.generate(input_ids=__snake_case , max_new_tokens=12 ) _SCREAMING_SNAKE_CASE : str = tokenizer.batch_decode(__snake_case , skip_special_tokens=__snake_case ) _SCREAMING_SNAKE_CASE : Optional[int] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__snake_case ) _SCREAMING_SNAKE_CASE : List[str] = tokenizer.decode(output_padded[0] , skip_special_tokens=__snake_case ) _SCREAMING_SNAKE_CASE : Optional[int] = [ """This is an extremelly long sentence that only exists to test the ability of the model to cope with """ """left-padding, such as in batched generation. The output for the sequence below should be the same """ """regardless of whether left padding is applied or not. When left padding is applied, the sequence will be """ """a single""", """Hello, my dog is a little bit of a shy one, but he is very friendly""", ] self.assertListEqual(__snake_case , __snake_case ) self.assertListEqual(__snake_case , [non_padded_sentence, padded_sentence] )
200
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ : Union[str, Any] = logging.get_logger(__name__) UpperCAmelCase_ : Optional[int] = { 'google/bigbird-roberta-base': 'https://huggingface.co/google/bigbird-roberta-base/resolve/main/config.json', 'google/bigbird-roberta-large': 'https://huggingface.co/google/bigbird-roberta-large/resolve/main/config.json', 'google/bigbird-base-trivia-itc': 'https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/config.json', # See all BigBird models at https://huggingface.co/models?filter=big_bird } class lowercase__ ( _snake_case ): '''simple docstring''' A_ : str = """big_bird""" def __init__( self , __snake_case=5_0358 , __snake_case=768 , __snake_case=12 , __snake_case=12 , __snake_case=3072 , __snake_case="gelu_new" , __snake_case=0.1 , __snake_case=0.1 , __snake_case=4096 , __snake_case=2 , __snake_case=0.02 , __snake_case=1e-12 , __snake_case=True , __snake_case=0 , __snake_case=1 , __snake_case=2 , __snake_case=66 , __snake_case="block_sparse" , __snake_case=True , __snake_case=False , __snake_case=64 , __snake_case=3 , __snake_case=None , **__snake_case , ): super().__init__( pad_token_id=__snake_case , bos_token_id=__snake_case , eos_token_id=__snake_case , sep_token_id=__snake_case , **__snake_case , ) _SCREAMING_SNAKE_CASE : str = vocab_size _SCREAMING_SNAKE_CASE : Union[str, Any] = max_position_embeddings _SCREAMING_SNAKE_CASE : List[str] = hidden_size _SCREAMING_SNAKE_CASE : Union[str, Any] = num_hidden_layers _SCREAMING_SNAKE_CASE : Any = num_attention_heads _SCREAMING_SNAKE_CASE : Optional[int] = intermediate_size _SCREAMING_SNAKE_CASE : List[Any] = hidden_act _SCREAMING_SNAKE_CASE : Optional[int] = hidden_dropout_prob _SCREAMING_SNAKE_CASE : Optional[int] = attention_probs_dropout_prob _SCREAMING_SNAKE_CASE : Tuple = initializer_range _SCREAMING_SNAKE_CASE : Any = type_vocab_size _SCREAMING_SNAKE_CASE : Union[str, Any] = layer_norm_eps _SCREAMING_SNAKE_CASE : List[Any] = use_cache _SCREAMING_SNAKE_CASE : List[Any] = rescale_embeddings _SCREAMING_SNAKE_CASE : Union[str, Any] = attention_type _SCREAMING_SNAKE_CASE : Union[str, Any] = use_bias _SCREAMING_SNAKE_CASE : int = block_size _SCREAMING_SNAKE_CASE : Any = num_random_blocks _SCREAMING_SNAKE_CASE : List[str] = classifier_dropout class lowercase__ ( _snake_case ): '''simple docstring''' @property def UpperCAmelCase_ ( self ): if self.task == "multiple-choice": _SCREAMING_SNAKE_CASE : int = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _SCREAMING_SNAKE_CASE : Optional[int] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
200
1
from ...configuration_utils import PretrainedConfig from ...utils import logging A_ :List[Any] = logging.get_logger(__name__) A_ :Any = { '''facebook/nllb-moe-54B''': '''https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json''', } class __A ( a ): """simple docstring""" UpperCamelCase__ : Any ="""nllb-moe""" UpperCamelCase__ : Any =["""past_key_values"""] UpperCamelCase__ : Optional[Any] ={"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , lowerCamelCase__=128112 , lowerCamelCase__=1024 , lowerCamelCase__=12 , lowerCamelCase__=4096 , lowerCamelCase__=16 , lowerCamelCase__=12 , lowerCamelCase__=4096 , lowerCamelCase__=16 , lowerCamelCase__=0.05 , lowerCamelCase__=0.05 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__="relu" , lowerCamelCase__=1024 , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=0.0 , lowerCamelCase__=0.02 , lowerCamelCase__=2 , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__="float32" , lowerCamelCase__=False , lowerCamelCase__=128 , lowerCamelCase__=64 , lowerCamelCase__=4 , lowerCamelCase__=4 , lowerCamelCase__=0.001 , lowerCamelCase__=0.001 , lowerCamelCase__="all" , lowerCamelCase__=False , lowerCamelCase__=False , lowerCamelCase__=1.0 , lowerCamelCase__=0.2 , lowerCamelCase__=1 , lowerCamelCase__=0 , lowerCamelCase__=2 , lowerCamelCase__=False , **lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Optional[int] =vocab_size __UpperCamelCase : Dict =max_position_embeddings __UpperCamelCase : Tuple =d_model __UpperCamelCase : Optional[int] =encoder_ffn_dim __UpperCamelCase : str =encoder_layers __UpperCamelCase : Any =encoder_attention_heads __UpperCamelCase : List[Any] =decoder_ffn_dim __UpperCamelCase : Any =decoder_layers __UpperCamelCase : Dict =decoder_attention_heads __UpperCamelCase : Optional[int] =dropout __UpperCamelCase : Tuple =attention_dropout __UpperCamelCase : Optional[int] =activation_dropout __UpperCamelCase : int =activation_function __UpperCamelCase : Union[str, Any] =init_std __UpperCamelCase : Any =encoder_layerdrop __UpperCamelCase : Union[str, Any] =decoder_layerdrop __UpperCamelCase : str =use_cache __UpperCamelCase : int =encoder_layers __UpperCamelCase : List[str] =scale_embedding # scale factor will be sqrt(d_model) if True __UpperCamelCase : Union[str, Any] =router_z_loss_coef __UpperCamelCase : List[Any] =router_aux_loss_coef __UpperCamelCase : Union[str, Any] =decoder_sparse_step __UpperCamelCase : List[Any] =encoder_sparse_step __UpperCamelCase : int =num_experts __UpperCamelCase : Union[str, Any] =expert_capacity __UpperCamelCase : Union[str, Any] =router_bias if router_dtype not in ["float32", "float16", "bfloat16"]: raise ValueError(f'`router_dtype` must be one of \'float32\', \'float16\' or \'bfloat16\', got {router_dtype}' ) __UpperCamelCase : Tuple =router_dtype __UpperCamelCase : Tuple =router_ignore_padding_tokens __UpperCamelCase : Any =batch_prioritized_routing __UpperCamelCase : List[Any] =second_expert_policy __UpperCamelCase : List[Any] =normalize_router_prob_before_dropping __UpperCamelCase : Dict =moe_eval_capacity_token_fraction __UpperCamelCase : Optional[Any] =moe_token_dropout __UpperCamelCase : Any =output_router_logits super().__init__( pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , is_encoder_decoder=lowerCamelCase__ , decoder_start_token_id=lowerCamelCase__ , **lowerCamelCase__ , )
245
import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def A ( a_ ,a_ ,a_ ,a_ ,a_ ) -> Optional[int]: # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file __UpperCamelCase : Optional[int] =TapasConfig.from_json_file(a_ ) # set absolute/relative position embeddings parameter __UpperCamelCase : str =reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": __UpperCamelCase : Optional[Any] =TapasForQuestionAnswering(config=a_ ) elif task == "WTQ": # run_task_main.py hparams __UpperCamelCase : Optional[int] =4 __UpperCamelCase : Optional[Any] =True # hparam_utils.py hparams __UpperCamelCase : int =0.664_694 __UpperCamelCase : Any =0.207_951 __UpperCamelCase : Tuple =0.121_194 __UpperCamelCase : List[str] =True __UpperCamelCase : Dict =True __UpperCamelCase : Optional[Any] =False __UpperCamelCase : Optional[int] =0.0_352_513 __UpperCamelCase : Optional[Any] =TapasForQuestionAnswering(config=a_ ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams __UpperCamelCase : List[Any] =4 __UpperCamelCase : List[str] =False # hparam_utils.py hparams __UpperCamelCase : List[str] =36.4_519 __UpperCamelCase : Dict =0.903_421 __UpperCamelCase : List[Any] =222.088 __UpperCamelCase : Optional[Any] =True __UpperCamelCase : Optional[int] =True __UpperCamelCase : Dict =True __UpperCamelCase : Dict =0.763_141 __UpperCamelCase : Union[str, Any] =TapasForQuestionAnswering(config=a_ ) elif task == "TABFACT": __UpperCamelCase : List[Any] =TapasForSequenceClassification(config=a_ ) elif task == "MLM": __UpperCamelCase : Optional[Any] =TapasForMaskedLM(config=a_ ) elif task == "INTERMEDIATE_PRETRAINING": __UpperCamelCase : Optional[Any] =TapasModel(config=a_ ) else: raise ValueError(F'Task {task} not supported.' ) print(F'Building PyTorch model from configuration: {config}' ) # Load weights from tf checkpoint load_tf_weights_in_tapas(a_ ,a_ ,a_ ) # Save pytorch-model (weights and configuration) print(F'Save PyTorch model to {pytorch_dump_path}' ) model.save_pretrained(a_ ) # Save tokenizer files print(F'Save tokenizer files to {pytorch_dump_path}' ) __UpperCamelCase : Optional[Any] =TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + 'vocab.txt' ,model_max_length=512 ) tokenizer.save_pretrained(a_ ) print('Used relative position embeddings:' ,model.config.reset_position_index_per_cell ) if __name__ == "__main__": A_ :Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--task''', default='''SQA''', type=str, help='''Model task for which to convert a checkpoint. Defaults to SQA.''' ) parser.add_argument( '''--reset_position_index_per_cell''', default=False, action='''store_true''', help='''Whether to use relative position embeddings or not. Defaults to True.''', ) parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--tapas_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained TAPAS model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) A_ :Union[str, Any] = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
245
1
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bert import BertTokenizer A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''bert-base-uncased''': '''https://huggingface.co/bert-base-uncased/resolve/main/vocab.txt''', '''bert-large-uncased''': '''https://huggingface.co/bert-large-uncased/resolve/main/vocab.txt''', '''bert-base-cased''': '''https://huggingface.co/bert-base-cased/resolve/main/vocab.txt''', '''bert-large-cased''': '''https://huggingface.co/bert-large-cased/resolve/main/vocab.txt''', '''bert-base-multilingual-uncased''': ( '''https://huggingface.co/bert-base-multilingual-uncased/resolve/main/vocab.txt''' ), '''bert-base-multilingual-cased''': '''https://huggingface.co/bert-base-multilingual-cased/resolve/main/vocab.txt''', '''bert-base-chinese''': '''https://huggingface.co/bert-base-chinese/resolve/main/vocab.txt''', '''bert-base-german-cased''': '''https://huggingface.co/bert-base-german-cased/resolve/main/vocab.txt''', '''bert-large-uncased-whole-word-masking''': ( '''https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/vocab.txt''' ), '''bert-large-cased-whole-word-masking''': ( '''https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/vocab.txt''' ), '''bert-large-uncased-whole-word-masking-finetuned-squad''': ( '''https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt''' ), '''bert-large-cased-whole-word-masking-finetuned-squad''': ( '''https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt''' ), '''bert-base-cased-finetuned-mrpc''': ( '''https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/vocab.txt''' ), '''bert-base-german-dbmdz-cased''': '''https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/vocab.txt''', '''bert-base-german-dbmdz-uncased''': ( '''https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/vocab.txt''' ), '''TurkuNLP/bert-base-finnish-cased-v1''': ( '''https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/vocab.txt''' ), '''TurkuNLP/bert-base-finnish-uncased-v1''': ( '''https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/vocab.txt''' ), '''wietsedv/bert-base-dutch-cased''': ( '''https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''bert-base-uncased''': '''https://huggingface.co/bert-base-uncased/resolve/main/tokenizer.json''', '''bert-large-uncased''': '''https://huggingface.co/bert-large-uncased/resolve/main/tokenizer.json''', '''bert-base-cased''': '''https://huggingface.co/bert-base-cased/resolve/main/tokenizer.json''', '''bert-large-cased''': '''https://huggingface.co/bert-large-cased/resolve/main/tokenizer.json''', '''bert-base-multilingual-uncased''': ( '''https://huggingface.co/bert-base-multilingual-uncased/resolve/main/tokenizer.json''' ), '''bert-base-multilingual-cased''': ( '''https://huggingface.co/bert-base-multilingual-cased/resolve/main/tokenizer.json''' ), '''bert-base-chinese''': '''https://huggingface.co/bert-base-chinese/resolve/main/tokenizer.json''', '''bert-base-german-cased''': '''https://huggingface.co/bert-base-german-cased/resolve/main/tokenizer.json''', '''bert-large-uncased-whole-word-masking''': ( '''https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/tokenizer.json''' ), '''bert-large-cased-whole-word-masking''': ( '''https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/tokenizer.json''' ), '''bert-large-uncased-whole-word-masking-finetuned-squad''': ( '''https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json''' ), '''bert-large-cased-whole-word-masking-finetuned-squad''': ( '''https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json''' ), '''bert-base-cased-finetuned-mrpc''': ( '''https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/tokenizer.json''' ), '''bert-base-german-dbmdz-cased''': ( '''https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/tokenizer.json''' ), '''bert-base-german-dbmdz-uncased''': ( '''https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/tokenizer.json''' ), '''TurkuNLP/bert-base-finnish-cased-v1''': ( '''https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/tokenizer.json''' ), '''TurkuNLP/bert-base-finnish-uncased-v1''': ( '''https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/tokenizer.json''' ), '''wietsedv/bert-base-dutch-cased''': ( '''https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/tokenizer.json''' ), }, } A_ = { '''bert-base-uncased''': 5_12, '''bert-large-uncased''': 5_12, '''bert-base-cased''': 5_12, '''bert-large-cased''': 5_12, '''bert-base-multilingual-uncased''': 5_12, '''bert-base-multilingual-cased''': 5_12, '''bert-base-chinese''': 5_12, '''bert-base-german-cased''': 5_12, '''bert-large-uncased-whole-word-masking''': 5_12, '''bert-large-cased-whole-word-masking''': 5_12, '''bert-large-uncased-whole-word-masking-finetuned-squad''': 5_12, '''bert-large-cased-whole-word-masking-finetuned-squad''': 5_12, '''bert-base-cased-finetuned-mrpc''': 5_12, '''bert-base-german-dbmdz-cased''': 5_12, '''bert-base-german-dbmdz-uncased''': 5_12, '''TurkuNLP/bert-base-finnish-cased-v1''': 5_12, '''TurkuNLP/bert-base-finnish-uncased-v1''': 5_12, '''wietsedv/bert-base-dutch-cased''': 5_12, } A_ = { '''bert-base-uncased''': {'''do_lower_case''': True}, '''bert-large-uncased''': {'''do_lower_case''': True}, '''bert-base-cased''': {'''do_lower_case''': False}, '''bert-large-cased''': {'''do_lower_case''': False}, '''bert-base-multilingual-uncased''': {'''do_lower_case''': True}, '''bert-base-multilingual-cased''': {'''do_lower_case''': False}, '''bert-base-chinese''': {'''do_lower_case''': False}, '''bert-base-german-cased''': {'''do_lower_case''': False}, '''bert-large-uncased-whole-word-masking''': {'''do_lower_case''': True}, '''bert-large-cased-whole-word-masking''': {'''do_lower_case''': False}, '''bert-large-uncased-whole-word-masking-finetuned-squad''': {'''do_lower_case''': True}, '''bert-large-cased-whole-word-masking-finetuned-squad''': {'''do_lower_case''': False}, '''bert-base-cased-finetuned-mrpc''': {'''do_lower_case''': False}, '''bert-base-german-dbmdz-cased''': {'''do_lower_case''': False}, '''bert-base-german-dbmdz-uncased''': {'''do_lower_case''': True}, '''TurkuNLP/bert-base-finnish-cased-v1''': {'''do_lower_case''': False}, '''TurkuNLP/bert-base-finnish-uncased-v1''': {'''do_lower_case''': True}, '''wietsedv/bert-base-dutch-cased''': {'''do_lower_case''': False}, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_INIT_CONFIGURATION lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = BertTokenizer def __init__( self: int, a_: Any=None, a_: Tuple=None, a_: Tuple=True, a_: Union[str, Any]="[UNK]", a_: List[Any]="[SEP]", a_: Union[str, Any]="[PAD]", a_: List[str]="[CLS]", a_: Union[str, Any]="[MASK]", a_: List[str]=True, a_: List[str]=None, **a_: List[Any], ): '''simple docstring''' super().__init__( a_, tokenizer_file=a_, do_lower_case=a_, unk_token=a_, sep_token=a_, pad_token=a_, cls_token=a_, mask_token=a_, tokenize_chinese_chars=a_, strip_accents=a_, **a_, ) _snake_case : Optional[int] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""", a_ ) != do_lower_case or normalizer_state.get("""strip_accents""", a_ ) != strip_accents or normalizer_state.get("""handle_chinese_chars""", a_ ) != tokenize_chinese_chars ): _snake_case : Optional[int] = getattr(a_, normalizer_state.pop("""type""" ) ) _snake_case : Optional[Any] = do_lower_case _snake_case : Optional[int] = strip_accents _snake_case : int = tokenize_chinese_chars _snake_case : Union[str, Any] = normalizer_class(**a_ ) _snake_case : str = do_lower_case def UpperCamelCase_ ( self: Dict, a_: Dict, a_: str=None ): '''simple docstring''' _snake_case : List[str] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCamelCase_ ( self: Any, a_: List[int], a_: Optional[List[int]] = None ): '''simple docstring''' _snake_case : Dict = [self.sep_token_id] _snake_case : str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self: Union[str, Any], a_: str, a_: Optional[str] = None ): '''simple docstring''' _snake_case : List[Any] = self._tokenizer.model.save(a_, name=a_ ) return tuple(a_ )
64
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ....tokenization_utils_fast import PreTrainedTokenizerFast from ....utils import logging from .tokenization_retribert import RetriBertTokenizer A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/tokenizer.json''' ), }, } A_ = { '''yjernite/retribert-base-uncased''': 5_12, } A_ = { '''yjernite/retribert-base-uncased''': {'''do_lower_case''': True}, } class lowercase( __a ): '''simple docstring''' lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = PRETRAINED_INIT_CONFIGURATION lowercase__ = RetriBertTokenizer lowercase__ = ["input_ids", "attention_mask"] def __init__( self: int, a_: int=None, a_: Dict=None, a_: Any=True, a_: int="[UNK]", a_: Any="[SEP]", a_: List[Any]="[PAD]", a_: List[Any]="[CLS]", a_: str="[MASK]", a_: Dict=True, a_: Optional[int]=None, **a_: Tuple, ): '''simple docstring''' super().__init__( a_, tokenizer_file=a_, do_lower_case=a_, unk_token=a_, sep_token=a_, pad_token=a_, cls_token=a_, mask_token=a_, tokenize_chinese_chars=a_, strip_accents=a_, **a_, ) _snake_case : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""", a_ ) != do_lower_case or normalizer_state.get("""strip_accents""", a_ ) != strip_accents or normalizer_state.get("""handle_chinese_chars""", a_ ) != tokenize_chinese_chars ): _snake_case : Dict = getattr(a_, normalizer_state.pop("""type""" ) ) _snake_case : List[Any] = do_lower_case _snake_case : List[str] = strip_accents _snake_case : Tuple = tokenize_chinese_chars _snake_case : Tuple = normalizer_class(**a_ ) _snake_case : List[str] = do_lower_case def UpperCamelCase_ ( self: Any, a_: str, a_: Optional[int]=None ): '''simple docstring''' _snake_case : Optional[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCamelCase_ ( self: List[str], a_: List[int], a_: Optional[List[int]] = None ): '''simple docstring''' _snake_case : Union[str, Any] = [self.sep_token_id] _snake_case : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase_ ( self: Dict, a_: str, a_: Optional[str] = None ): '''simple docstring''' _snake_case : Union[str, Any] = self._tokenizer.model.save(a_, name=a_ ) return tuple(a_ )
64
1
import unittest from transformers import RoFormerTokenizer, RoFormerTokenizerFast from transformers.testing_utils import require_rjieba, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_rjieba @require_tokenizers class __lowercase (UpperCamelCase__ , unittest.TestCase ): """simple docstring""" _snake_case = RoFormerTokenizer _snake_case = RoFormerTokenizerFast _snake_case = True _snake_case = True def UpperCAmelCase ( self ) -> Any: super().setUp() def UpperCAmelCase ( self , **A ) -> List[Any]: return self.tokenizer_class.from_pretrained("""junnyu/roformer_chinese_base""" , **A ) def UpperCAmelCase ( self , **A ) -> Union[str, Any]: return self.rust_tokenizer_class.from_pretrained("""junnyu/roformer_chinese_base""" , **A ) def UpperCAmelCase ( self ) -> Dict: snake_case : int = """永和服装饰品有限公司,今天天气非常好""" snake_case : List[Any] = """永和 服装 饰品 有限公司 , 今 天 天 气 非常 好""" return input_text, output_text def UpperCAmelCase ( self ) -> Tuple: snake_case : Dict = self.get_tokenizer() snake_case , snake_case : Any = self.get_chinese_input_output_texts() snake_case : Optional[Any] = tokenizer.tokenize(A ) self.assertListEqual(A , output_text.split() ) snake_case : Optional[Any] = tokens + [tokenizer.unk_token] snake_case : List[str] = [2_2_9_4_3, 2_1_3_3_2, 3_4_4_3_1, 4_5_9_0_4, 1_1_7, 3_0_6, 1_2_3_1, 1_2_3_1, 2_6_5_3, 3_3_9_9_4, 1_2_6_6, 1_0_0] self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) , A ) def UpperCAmelCase ( self ) -> Optional[int]: snake_case : List[str] = self.get_rust_tokenizer() snake_case , snake_case : str = self.get_chinese_input_output_texts() snake_case : int = tokenizer.tokenize(A ) self.assertListEqual(A , output_text.split() ) snake_case : int = tokens + [tokenizer.unk_token] snake_case : str = [2_2_9_4_3, 2_1_3_3_2, 3_4_4_3_1, 4_5_9_0_4, 1_1_7, 3_0_6, 1_2_3_1, 1_2_3_1, 2_6_5_3, 3_3_9_9_4, 1_2_6_6, 1_0_0] self.assertListEqual(tokenizer.convert_tokens_to_ids(A ) , A ) def UpperCAmelCase ( self ) -> Dict: pass def UpperCAmelCase ( self ) -> Union[str, Any]: pass def UpperCAmelCase ( self ) -> Tuple: pass
176
import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging lowerCamelCase : Optional[Any] = logging.get_logger(__name__) lowerCamelCase : Optional[int] = { 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/config.json', # See all BART models at https://huggingface.co/models?filter=bart } class __lowercase (UpperCamelCase__ ): """simple docstring""" _snake_case = """bart""" _snake_case = ["""past_key_values"""] _snake_case = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , A=5_0_2_6_5 , A=1_0_2_4 , A=1_2 , A=4_0_9_6 , A=1_6 , A=1_2 , A=4_0_9_6 , A=1_6 , A=0.0 , A=0.0 , A="gelu" , A=1_0_2_4 , A=0.1 , A=0.0 , A=0.0 , A=0.02 , A=0.0 , A=False , A=True , A=3 , A=1 , A=0 , A=2 , A=True , A=2 , A=2 , **A , ) -> Any: snake_case : Optional[int] = vocab_size snake_case : Union[str, Any] = max_position_embeddings snake_case : List[str] = d_model snake_case : List[Any] = encoder_ffn_dim snake_case : Optional[Any] = encoder_layers snake_case : Union[str, Any] = encoder_attention_heads snake_case : str = decoder_ffn_dim snake_case : Union[str, Any] = decoder_layers snake_case : Any = decoder_attention_heads snake_case : Union[str, Any] = dropout snake_case : List[str] = attention_dropout snake_case : List[Any] = activation_dropout snake_case : Optional[int] = activation_function snake_case : Union[str, Any] = init_std snake_case : List[str] = encoder_layerdrop snake_case : int = decoder_layerdrop snake_case : str = classifier_dropout snake_case : List[str] = use_cache snake_case : Tuple = encoder_layers snake_case : int = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=A , pad_token_id=A , bos_token_id=A , eos_token_id=A , is_encoder_decoder=A , decoder_start_token_id=A , forced_eos_token_id=A , **A , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get("""force_bos_token_to_be_generated""" , A ): snake_case : Any = self.bos_token_id warnings.warn( f"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ """The config can simply be saved and uploaded again to be fixed.""" ) class __lowercase (UpperCamelCase__ ): """simple docstring""" @property def UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: snake_case : Optional[Any] = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: snake_case : Tuple = {0: """batch"""} snake_case : List[Any] = {0: """batch""", 1: """past_decoder_sequence + sequence"""} else: snake_case : Union[str, Any] = {0: """batch""", 1: """decoder_sequence"""} snake_case : Any = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(A , direction="""inputs""" ) elif self.task == "causal-lm": # TODO: figure this case out. snake_case : Union[str, Any] = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ] ) if self.use_past: snake_case , snake_case : List[Any] = self.num_layers for i in range(A ): snake_case : List[Any] = {0: """batch""", 2: """past_sequence + sequence"""} snake_case : Optional[int] = {0: """batch""", 2: """past_sequence + sequence"""} else: snake_case : Union[str, Any] = OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """encoder_sequence"""}), ("""attention_mask""", {0: """batch""", 1: """encoder_sequence"""}), ("""decoder_input_ids""", {0: """batch""", 1: """decoder_sequence"""}), ("""decoder_attention_mask""", {0: """batch""", 1: """decoder_sequence"""}), ] ) return common_inputs @property def UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: snake_case : Any = super().outputs else: snake_case : Any = super(A , self ).outputs if self.use_past: snake_case , snake_case : Any = self.num_layers for i in range(A ): snake_case : Any = {0: """batch""", 2: """past_sequence + sequence"""} snake_case : Union[str, Any] = {0: """batch""", 2: """past_sequence + sequence"""} return common_outputs def UpperCAmelCase ( self , A , A = -1 , A = -1 , A = False , A = None , ) -> Mapping[str, Any]: snake_case : Optional[Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A , A , A , A , A ) # Generate decoder inputs snake_case : Any = seq_length if not self.use_past else 1 snake_case : List[str] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A , A , A , A , A ) snake_case : Optional[int] = {f"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} snake_case : List[str] = dict(**A , **A ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch snake_case , snake_case : Optional[int] = common_inputs["""input_ids"""].shape snake_case : Any = common_inputs["""decoder_input_ids"""].shape[1] snake_case , snake_case : Optional[Any] = self.num_attention_heads snake_case : Optional[int] = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) snake_case : Any = decoder_seq_length + 3 snake_case : List[Any] = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) snake_case : str = torch.cat( [common_inputs["""decoder_attention_mask"""], torch.ones(A , A )] , dim=1 ) snake_case : str = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered snake_case , snake_case : Any = self.num_layers snake_case : List[str] = min(A , A ) snake_case : Dict = max(A , A ) - min_num_layers snake_case : List[str] = """encoder""" if num_encoder_layers > num_decoder_layers else """decoder""" for _ in range(A ): common_inputs["past_key_values"].append( ( torch.zeros(A ), torch.zeros(A ), torch.zeros(A ), torch.zeros(A ), ) ) # TODO: test this. snake_case : Tuple = encoder_shape if remaining_side_name == """encoder""" else decoder_shape for _ in range(A , A ): common_inputs["past_key_values"].append((torch.zeros(A ), torch.zeros(A )) ) return common_inputs def UpperCAmelCase ( self , A , A = -1 , A = -1 , A = False , A = None , ) -> Mapping[str, Any]: snake_case : str = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A , A , A , A , A ) if self.use_past: if not is_torch_available(): raise ValueError("""Cannot generate dummy past_keys inputs without PyTorch installed.""" ) else: import torch snake_case , snake_case : str = common_inputs["""input_ids"""].shape # Not using the same length for past_key_values snake_case : Optional[int] = seqlen + 2 snake_case , snake_case : Tuple = self.num_layers snake_case , snake_case : Optional[Any] = self.num_attention_heads snake_case : Union[str, Any] = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) snake_case : Optional[Any] = common_inputs["""attention_mask"""].dtype snake_case : int = torch.cat( [common_inputs["""attention_mask"""], torch.ones(A , A , dtype=A )] , dim=1 ) snake_case : Union[str, Any] = [ (torch.zeros(A ), torch.zeros(A )) for _ in range(A ) ] return common_inputs def UpperCAmelCase ( self , A , A = -1 , A = -1 , A = False , A = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX snake_case : int = compute_effective_axis_dimension( A , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX snake_case : int = tokenizer.num_special_tokens_to_add(A ) snake_case : Tuple = compute_effective_axis_dimension( A , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=A ) # Generate dummy inputs according to compute batch and sequence snake_case : int = [""" """.join([tokenizer.unk_token] ) * seq_length] * batch_size snake_case : str = dict(tokenizer(A , return_tensors=A ) ) return common_inputs def UpperCAmelCase ( self , A , A = -1 , A = -1 , A = False , A = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: snake_case : Optional[Any] = self._generate_dummy_inputs_for_default_and_seqaseq_lm( A , batch_size=A , seq_length=A , is_pair=A , framework=A ) elif self.task == "causal-lm": snake_case : Optional[int] = self._generate_dummy_inputs_for_causal_lm( A , batch_size=A , seq_length=A , is_pair=A , framework=A ) else: snake_case : List[Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( A , batch_size=A , seq_length=A , is_pair=A , framework=A ) return common_inputs def UpperCAmelCase ( self , A , A , A , A ) -> Union[str, Any]: if self.task in ["default", "seq2seq-lm"]: snake_case : Optional[Any] = super()._flatten_past_key_values_(A , A , A , A ) else: snake_case : Union[str, Any] = super(A , self )._flatten_past_key_values_( A , A , A , A )
176
1
from queue import PriorityQueue from typing import Any import numpy as np def lowerCAmelCase_ ( _lowercase : dict , _lowercase : str , _lowercase : set , _lowercase : set , _lowercase : dict , _lowercase : dict , _lowercase : PriorityQueue , _lowercase : dict , _lowercase : float | int , ) -> float | int: """simple docstring""" for nxt, d in graph[v]: if nxt in visited_forward: continue a__ : Tuple = cst_fwd.get(_lowercase , np.inf) a__ : List[str] = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt)) a__ : int = new_cost_f a__ : List[Any] = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: a__ : Union[str, Any] = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def lowerCAmelCase_ ( _lowercase : str , _lowercase : str , _lowercase : dict , _lowercase : dict) -> int: """simple docstring""" a__ : Union[str, Any] = -1 a__ : List[str] = set() a__ : Any = set() a__ : int = {source: 0} a__ : Any = {destination: 0} a__ : Optional[int] = {source: None} a__ : int = {destination: None} a__ : PriorityQueue[Any] = PriorityQueue() a__ : PriorityQueue[Any] = PriorityQueue() a__ : Tuple = np.inf queue_forward.put((0, source)) queue_backward.put((0, destination)) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): a__ , a__ : int = queue_forward.get() visited_forward.add(_lowercase) a__ , a__ : List[Any] = queue_backward.get() visited_backward.add(_lowercase) a__ : int = pass_and_relaxation( _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , ) a__ : List[Any] = pass_and_relaxation( _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: a__ : Any = shortest_distance return shortest_path_distance _lowercase : List[str] ={ "B": [["C", 1]], "C": [["D", 1]], "D": [["F", 1]], "E": [["B", 1], ["G", 2]], "F": [], "G": [["F", 1]], } _lowercase : List[Any] ={ "B": [["E", 1]], "C": [["B", 1]], "D": [["C", 1]], "F": [["D", 1], ["G", 1]], "E": [[None, np.inf]], "G": [["E", 2]], } if __name__ == "__main__": import doctest doctest.testmod()
170
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase : Optional[int] =logging.get_logger(__name__) _lowercase : Tuple ={ "microsoft/swinv2-tiny-patch4-window8-256": ( "https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json" ), } class snake_case__ (A__ ): """simple docstring""" __lowerCAmelCase :List[Any] = "swinv2" __lowerCAmelCase :List[str] = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self , __lowercase=2_2_4 , __lowercase=4 , __lowercase=3 , __lowercase=9_6 , __lowercase=[2, 2, 6, 2] , __lowercase=[3, 6, 1_2, 2_4] , __lowercase=7 , __lowercase=4.0 , __lowercase=True , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.1 , __lowercase="gelu" , __lowercase=False , __lowercase=0.0_2 , __lowercase=1E-5 , __lowercase=3_2 , **__lowercase , ) -> Any: """simple docstring""" super().__init__(**__lowercase ) a__ : Optional[Any] = image_size a__ : Union[str, Any] = patch_size a__ : List[Any] = num_channels a__ : Union[str, Any] = embed_dim a__ : Any = depths a__ : List[str] = len(__lowercase ) a__ : Optional[Any] = num_heads a__ : Union[str, Any] = window_size a__ : Optional[int] = mlp_ratio a__ : List[str] = qkv_bias a__ : Dict = hidden_dropout_prob a__ : str = attention_probs_dropout_prob a__ : List[Any] = drop_path_rate a__ : Tuple = hidden_act a__ : Dict = use_absolute_embeddings a__ : Tuple = layer_norm_eps a__ : Tuple = initializer_range a__ : Union[str, Any] = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model a__ : int = int(embed_dim * 2 ** (len(__lowercase ) - 1) ) a__ : Dict = (0, 0, 0, 0)
170
1
'''simple docstring''' import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class __snake_case( _lowerCAmelCase ): '''simple docstring''' @require_torch def __snake_case ( self ) -> Union[str, Any]: # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched lowerCAmelCase = """ from transformers import BertConfig, BertModel, BertTokenizer, pipeline """ lowerCAmelCase = """ mname = \"hf-internal-testing/tiny-random-bert\" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task=\"fill-mask\", model=mname) print(\"success\") """ lowerCAmelCase = """ import socket def offline_socket(*args, **kwargs): raise RuntimeError(\"Offline mode is enabled, we shouldn't access internet\") socket.socket = offline_socket """ # Force fetching the files so that we can use the cache lowerCAmelCase = """hf-internal-testing/tiny-random-bert""" BertConfig.from_pretrained(A_ ) BertModel.from_pretrained(A_ ) BertTokenizer.from_pretrained(A_ ) pipeline(task="""fill-mask""" , model=A_ ) # baseline - just load from_pretrained with normal network lowerCAmelCase = [sys.executable, """-c""", """\n""".join([load, run, mock] )] # should succeed lowerCAmelCase = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files lowerCAmelCase = """1""" lowerCAmelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) @require_torch def __snake_case ( self ) -> Optional[Any]: # python one-liner segments # this must be loaded before socket.socket is monkey-patched lowerCAmelCase = """ from transformers import BertConfig, BertModel, BertTokenizer, pipeline """ lowerCAmelCase = """ mname = \"hf-internal-testing/tiny-random-bert\" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task=\"fill-mask\", model=mname) print(\"success\") """ lowerCAmelCase = """ import socket def offline_socket(*args, **kwargs): raise socket.error(\"Faking flaky internet\") socket.socket = offline_socket """ # Force fetching the files so that we can use the cache lowerCAmelCase = """hf-internal-testing/tiny-random-bert""" BertConfig.from_pretrained(A_ ) BertModel.from_pretrained(A_ ) BertTokenizer.from_pretrained(A_ ) pipeline(task="""fill-mask""" , model=A_ ) # baseline - just load from_pretrained with normal network lowerCAmelCase = [sys.executable, """-c""", """\n""".join([load, run, mock] )] # should succeed lowerCAmelCase = self.get_env() lowerCAmelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) @require_torch def __snake_case ( self ) -> List[Any]: # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched lowerCAmelCase = """ from transformers import BertConfig, BertModel, BertTokenizer """ lowerCAmelCase = """ mname = \"hf-internal-testing/tiny-random-bert-sharded\" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) print(\"success\") """ lowerCAmelCase = """ import socket def offline_socket(*args, **kwargs): raise ValueError(\"Offline mode is enabled\") socket.socket = offline_socket """ # baseline - just load from_pretrained with normal network lowerCAmelCase = [sys.executable, """-c""", """\n""".join([load, run] )] # should succeed lowerCAmelCase = self.get_env() lowerCAmelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) # next emulate no network lowerCAmelCase = [sys.executable, """-c""", """\n""".join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files lowerCAmelCase = """1""" lowerCAmelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) @require_torch def __snake_case ( self ) -> Tuple: lowerCAmelCase = """ from transformers import pipeline """ lowerCAmelCase = """ mname = \"hf-internal-testing/tiny-random-bert\" pipe = pipeline(model=mname) """ lowerCAmelCase = """ import socket def offline_socket(*args, **kwargs): raise socket.error(\"Offline mode is enabled\") socket.socket = offline_socket """ lowerCAmelCase = self.get_env() lowerCAmelCase = """1""" lowerCAmelCase = [sys.executable, """-c""", """\n""".join([load, mock, run] )] lowerCAmelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( """You cannot infer task automatically within `pipeline` when using offline mode""" , result.stderr.decode().replace("""\n""" , """""" ) , ) @require_torch def __snake_case ( self ) -> Optional[int]: lowerCAmelCase = """ from transformers import AutoModel """ lowerCAmelCase = """ mname = \"hf-internal-testing/test_dynamic_model\" AutoModel.from_pretrained(mname, trust_remote_code=True) print(\"success\") """ # baseline - just load from_pretrained with normal network lowerCAmelCase = [sys.executable, """-c""", """\n""".join([load, run] )] # should succeed lowerCAmelCase = self.get_env() lowerCAmelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files lowerCAmelCase = """1""" lowerCAmelCase = subprocess.run(A_ , env=A_ , check=A_ , capture_output=A_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("""success""" , result.stdout.decode() )
187
'''simple docstring''' from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class __snake_case: '''simple docstring''' def __init__( self , A_ , A_=13 , A_=7 , A_=True , A_=True , A_=True , A_=True , A_=99 , A_=32 , A_=2 , A_=4 , A_=37 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=16 , A_=2 , A_=0.0_2 , A_=3 , A_=4 , A_=None , ) -> Dict: lowerCAmelCase = parent lowerCAmelCase = 13 lowerCAmelCase = 7 lowerCAmelCase = True lowerCAmelCase = True lowerCAmelCase = True lowerCAmelCase = True lowerCAmelCase = 99 lowerCAmelCase = 384 lowerCAmelCase = 2 lowerCAmelCase = 4 lowerCAmelCase = 37 lowerCAmelCase = """gelu""" lowerCAmelCase = 0.1 lowerCAmelCase = 0.1 lowerCAmelCase = 512 lowerCAmelCase = 16 lowerCAmelCase = 2 lowerCAmelCase = 0.0_2 lowerCAmelCase = 3 lowerCAmelCase = 4 lowerCAmelCase = 128 lowerCAmelCase = 2 lowerCAmelCase = 9 lowerCAmelCase = 1 lowerCAmelCase = None def __snake_case ( self ) -> Optional[int]: lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase = None if self.use_input_mask: lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase = None if self.use_token_type_ids: lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = None if self.use_labels: lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=A_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __snake_case ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) -> int: lowerCAmelCase = TFConvBertModel(config=A_ ) lowerCAmelCase = {"""input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids} lowerCAmelCase = [input_ids, input_mask] lowerCAmelCase = model(A_ ) lowerCAmelCase = model(A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) -> List[Any]: lowerCAmelCase = TFConvBertForMaskedLM(config=A_ ) lowerCAmelCase = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } lowerCAmelCase = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __snake_case ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) -> Optional[int]: lowerCAmelCase = self.num_labels lowerCAmelCase = TFConvBertForSequenceClassification(config=A_ ) lowerCAmelCase = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } lowerCAmelCase = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) -> Any: lowerCAmelCase = self.num_choices lowerCAmelCase = TFConvBertForMultipleChoice(config=A_ ) lowerCAmelCase = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) lowerCAmelCase = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) lowerCAmelCase = tf.tile(tf.expand_dims(A_ , 1 ) , (1, self.num_choices, 1) ) lowerCAmelCase = { """input_ids""": multiple_choice_inputs_ids, """attention_mask""": multiple_choice_input_mask, """token_type_ids""": multiple_choice_token_type_ids, } lowerCAmelCase = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __snake_case ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) -> Union[str, Any]: lowerCAmelCase = self.num_labels lowerCAmelCase = TFConvBertForTokenClassification(config=A_ ) lowerCAmelCase = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } lowerCAmelCase = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __snake_case ( self , A_ , A_ , A_ , A_ , A_ , A_ , A_ ) -> Optional[int]: lowerCAmelCase = TFConvBertForQuestionAnswering(config=A_ ) lowerCAmelCase = { """input_ids""": input_ids, """attention_mask""": input_mask, """token_type_ids""": token_type_ids, } lowerCAmelCase = model(A_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __snake_case ( self ) -> Any: lowerCAmelCase = self.prepare_config_and_inputs() ( ( lowerCAmelCase ), ( lowerCAmelCase ), ( lowerCAmelCase ), ( lowerCAmelCase ), ( lowerCAmelCase ), ( lowerCAmelCase ), ( lowerCAmelCase ), ) = config_and_inputs lowerCAmelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class __snake_case( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase : Optional[int] = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) UpperCAmelCase : Union[str, Any] = ( { "feature-extraction": TFConvBertModel, "fill-mask": TFConvBertForMaskedLM, "question-answering": TFConvBertForQuestionAnswering, "text-classification": TFConvBertForSequenceClassification, "token-classification": TFConvBertForTokenClassification, "zero-shot": TFConvBertForSequenceClassification, } if is_tf_available() else {} ) UpperCAmelCase : Union[str, Any] = False UpperCAmelCase : Optional[int] = False UpperCAmelCase : Dict = False def __snake_case ( self ) -> Optional[int]: lowerCAmelCase = TFConvBertModelTester(self ) lowerCAmelCase = ConfigTester(self , config_class=A_ , hidden_size=37 ) def __snake_case ( self ) -> Tuple: self.config_tester.run_common_tests() def __snake_case ( self ) -> Union[str, Any]: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def __snake_case ( self ) -> Tuple: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A_ ) def __snake_case ( self ) -> Optional[int]: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*A_ ) def __snake_case ( self ) -> List[str]: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A_ ) def __snake_case ( self ) -> str: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A_ ) def __snake_case ( self ) -> Tuple: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A_ ) @slow def __snake_case ( self ) -> Any: lowerCAmelCase, lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase = True lowerCAmelCase = True if hasattr(A_ , """use_cache""" ): lowerCAmelCase = True lowerCAmelCase = getattr(self.model_tester , """encoder_seq_length""" , self.model_tester.seq_length ) lowerCAmelCase = getattr(self.model_tester , """key_length""" , A_ ) for model_class in self.all_model_classes: lowerCAmelCase = self._prepare_for_class(A_ , A_ ) lowerCAmelCase = model_class(A_ ) lowerCAmelCase = len(model(A_ ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(A_ , saved_model=A_ ) lowerCAmelCase = os.path.join(A_ , """saved_model""" , """1""" ) lowerCAmelCase = tf.keras.models.load_model(A_ ) lowerCAmelCase = model(A_ ) if self.is_encoder_decoder: lowerCAmelCase = outputs["""encoder_hidden_states"""] lowerCAmelCase = outputs["""encoder_attentions"""] else: lowerCAmelCase = outputs["""hidden_states"""] lowerCAmelCase = outputs["""attentions"""] self.assertEqual(len(A_ ) , A_ ) lowerCAmelCase = getattr( self.model_tester , """expected_num_hidden_layers""" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(A_ ) , A_ ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def __snake_case ( self ) -> Optional[Any]: lowerCAmelCase = TFConvBertModel.from_pretrained("""YituTech/conv-bert-base""" ) self.assertIsNotNone(A_ ) def __snake_case ( self ) -> str: lowerCAmelCase, lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase = True lowerCAmelCase = getattr(self.model_tester , """decoder_seq_length""" , self.model_tester.seq_length ) lowerCAmelCase = getattr(self.model_tester , """encoder_seq_length""" , self.model_tester.seq_length ) lowerCAmelCase = getattr(self.model_tester , """key_length""" , A_ ) lowerCAmelCase = getattr(self.model_tester , """key_length""" , A_ ) def check_decoder_attentions_output(A_ ): lowerCAmelCase = len(A_ ) self.assertEqual(out_len % 2 , 0 ) lowerCAmelCase = outputs.decoder_attentions self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(A_ ): lowerCAmelCase = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(A_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: lowerCAmelCase = True lowerCAmelCase = False lowerCAmelCase = model_class(A_ ) lowerCAmelCase = model(self._prepare_for_class(A_ , A_ ) ) lowerCAmelCase = len(A_ ) self.assertEqual(config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) if self.is_encoder_decoder: lowerCAmelCase = model_class(A_ ) lowerCAmelCase = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(config.output_hidden_states , A_ ) check_decoder_attentions_output(A_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] lowerCAmelCase = True lowerCAmelCase = model_class(A_ ) lowerCAmelCase = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) # Check attention is always last and order is fine lowerCAmelCase = True lowerCAmelCase = True lowerCAmelCase = model_class(A_ ) lowerCAmelCase = model(self._prepare_for_class(A_ , A_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(A_ ) ) self.assertEqual(model.config.output_hidden_states , A_ ) check_encoder_attentions_output(A_ ) @require_tf class __snake_case( unittest.TestCase ): '''simple docstring''' @slow def __snake_case ( self ) -> Any: lowerCAmelCase = TFConvBertModel.from_pretrained("""YituTech/conv-bert-base""" ) lowerCAmelCase = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowerCAmelCase = model(A_ )[0] lowerCAmelCase = [1, 6, 768] self.assertEqual(output.shape , A_ ) lowerCAmelCase = tf.constant( [ [ [-0.0_3_4_7_5_4_9_3, -0.4_6_8_6_0_3_4, -0.3_0_6_3_8_8_3_2], [0.2_2_6_3_7_2_4_8, -0.2_6_9_8_8_6_4_6, -0.7_4_2_3_4_2_4], [0.1_0_3_2_4_8_6_8, -0.4_5_0_1_3_5_0_8, -0.5_8_2_8_0_7_8_4], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , A_ , atol=1e-4 )
187
1
from __future__ import annotations def a__ ( UpperCAmelCase : list[int] , UpperCAmelCase : list[int] , UpperCAmelCase : int ) -> tuple[float, list[float]]: UpperCAmelCase : List[Any] = list(range(len(UpperCAmelCase ) ) ) UpperCAmelCase : Dict = [v / w for v, w in zip(UpperCAmelCase , UpperCAmelCase )] index.sort(key=lambda UpperCAmelCase : ratio[i] , reverse=UpperCAmelCase ) UpperCAmelCase : List[Any] = 0 UpperCAmelCase : Optional[int] = [0] * len(UpperCAmelCase ) for i in index: if weight[i] <= capacity: UpperCAmelCase : str = 1 max_value += value[i] capacity -= weight[i] else: UpperCAmelCase : Optional[Any] = capacity / weight[i] max_value += value[i] * capacity / weight[i] break return max_value, fractions if __name__ == "__main__": import doctest doctest.testmod()
336
'''simple docstring''' import argparse import json import logging import os import sys from unittest.mock import patch from transformers.testing_utils import TestCasePlus, get_gpu_count, slow lowercase__ : List[str] = [ os.path.join(os.path.dirname(__file__), dirname) for dirname in [ 'text-classification', 'language-modeling', 'summarization', 'token-classification', 'question-answering', ] ] sys.path.extend(SRC_DIRS) if SRC_DIRS is not None: import run_clm_flax import run_flax_glue import run_flax_ner import run_mlm_flax import run_qa import run_summarization_flax import run_ta_mlm_flax logging.basicConfig(level=logging.DEBUG) lowercase__ : Dict = logging.getLogger() def a__ ( ) -> Optional[int]: """simple docstring""" _UpperCamelCase = argparse.ArgumentParser() parser.add_argument('''-f''' ) _UpperCamelCase = parser.parse_args() return args.f def a__ ( lowercase : Tuple, lowercase : Dict="eval" ) -> int: """simple docstring""" _UpperCamelCase = os.path.join(lowercase, F"""{split}_results.json""" ) if os.path.exists(lowercase ): with open(lowercase, '''r''' ) as f: return json.load(lowercase ) raise ValueError(F"""can't find {path}""" ) lowercase__ : int = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" def snake_case__ ( self : Any ) -> str: '''simple docstring''' _UpperCamelCase = self.get_auto_remove_tmp_dir() _UpperCamelCase = f""" run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --eval_steps=2 --warmup_steps=2 --seed=42 --max_seq_length=128 """.split() with patch.object(lowerCAmelCase__ , '''argv''' , lowerCAmelCase__ ): run_flax_glue.main() _UpperCamelCase = get_results(lowerCAmelCase__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.75 ) @slow def snake_case__ ( self : Tuple ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.get_auto_remove_tmp_dir() _UpperCamelCase = f""" run_clm_flax.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --block_size 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(lowerCAmelCase__ , '''argv''' , lowerCAmelCase__ ): run_clm_flax.main() _UpperCamelCase = get_results(lowerCAmelCase__ ) self.assertLess(result['''eval_perplexity'''] , 100 ) @slow def snake_case__ ( self : Tuple ) -> str: '''simple docstring''' _UpperCamelCase = self.get_auto_remove_tmp_dir() _UpperCamelCase = f""" run_summarization.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --test_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=8 --do_train --do_eval --do_predict --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --predict_with_generate """.split() with patch.object(lowerCAmelCase__ , '''argv''' , lowerCAmelCase__ ): run_summarization_flax.main() _UpperCamelCase = get_results(lowerCAmelCase__ , split='''test''' ) self.assertGreaterEqual(result['''test_rouge1'''] , 10 ) self.assertGreaterEqual(result['''test_rouge2'''] , 2 ) self.assertGreaterEqual(result['''test_rougeL'''] , 7 ) self.assertGreaterEqual(result['''test_rougeLsum'''] , 7 ) @slow def snake_case__ ( self : Tuple ) -> Any: '''simple docstring''' _UpperCamelCase = self.get_auto_remove_tmp_dir() _UpperCamelCase = f""" run_mlm.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --overwrite_output_dir --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --logging_steps 2 --eval_steps 2 --do_train --do_eval --num_train_epochs=1 """.split() with patch.object(lowerCAmelCase__ , '''argv''' , lowerCAmelCase__ ): run_mlm_flax.main() _UpperCamelCase = get_results(lowerCAmelCase__ ) self.assertLess(result['''eval_perplexity'''] , 42 ) @slow def snake_case__ ( self : str ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.get_auto_remove_tmp_dir() _UpperCamelCase = f""" run_t5_mlm_flax.py --model_name_or_path t5-small --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(lowerCAmelCase__ , '''argv''' , lowerCAmelCase__ ): run_ta_mlm_flax.main() _UpperCamelCase = get_results(lowerCAmelCase__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.42 ) @slow def snake_case__ ( self : List[Any] ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = 7 if get_gpu_count() > 1 else 2 _UpperCamelCase = self.get_auto_remove_tmp_dir() _UpperCamelCase = f""" run_flax_ner.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --overwrite_output_dir --do_train --do_eval --warmup_steps=2 --learning_rate=2e-4 --logging_steps 2 --eval_steps 2 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 """.split() with patch.object(lowerCAmelCase__ , '''argv''' , lowerCAmelCase__ ): run_flax_ner.main() _UpperCamelCase = get_results(lowerCAmelCase__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.75 ) self.assertGreaterEqual(result['''eval_f1'''] , 0.3 ) @slow def snake_case__ ( self : str ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = self.get_auto_remove_tmp_dir() _UpperCamelCase = f""" run_qa.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=2 --do_train --do_eval --logging_steps 2 --eval_steps 2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 """.split() with patch.object(lowerCAmelCase__ , '''argv''' , lowerCAmelCase__ ): run_qa.main() _UpperCamelCase = get_results(lowerCAmelCase__ ) self.assertGreaterEqual(result['''eval_f1'''] , 30 ) self.assertGreaterEqual(result['''eval_exact'''] , 30 )
324
0
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch _lowercase : Optional[Any] =random.Random() def lowerCAmelCase_ ( _lowercase : str , _lowercase : List[str]=1.0 , _lowercase : Any=None , _lowercase : Optional[Any]=None) -> Optional[Any]: """simple docstring""" if rng is None: a__ : Any = global_rng a__ : int = [] for batch_idx in range(shape[0]): values.append([]) for _ in range(shape[1]): values[-1].append(rng.random() * scale) return values @require_torch class snake_case__ (unittest.TestCase ): """simple docstring""" def __init__( self , __lowercase , __lowercase=7 , __lowercase=4_0_0 , __lowercase=2_0_0_0 , __lowercase=1 , __lowercase=0.0 , __lowercase=1_6_0_0_0 , __lowercase=True , __lowercase=8_0 , __lowercase=1_6 , __lowercase=6_4 , __lowercase="hann_window" , __lowercase=8_0 , __lowercase=7_6_0_0 , __lowercase=1E-10 , __lowercase=True , ) -> List[str]: """simple docstring""" a__ : List[Any] = parent a__ : Any = batch_size a__ : List[Any] = min_seq_length a__ : List[str] = max_seq_length a__ : Union[str, Any] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) a__ : Tuple = feature_size a__ : int = padding_value a__ : Dict = sampling_rate a__ : Tuple = do_normalize a__ : List[Any] = num_mel_bins a__ : Dict = hop_length a__ : Optional[Any] = win_length a__ : Union[str, Any] = win_function a__ : Optional[int] = fmin a__ : str = fmax a__ : Optional[int] = mel_floor a__ : Any = return_attention_mask def SCREAMING_SNAKE_CASE__( self ) -> int: """simple docstring""" return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def SCREAMING_SNAKE_CASE__( self , __lowercase=False , __lowercase=False ) -> Optional[Any]: """simple docstring""" def _flatten(__lowercase ): return list(itertools.chain(*SCREAMING_SNAKE_CASE_ ) ) if equal_length: a__ : Tuple = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size a__ : Any = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a__ : Tuple = [np.asarray(SCREAMING_SNAKE_CASE_ ) for x in speech_inputs] return speech_inputs def SCREAMING_SNAKE_CASE__( self , __lowercase=False , __lowercase=False ) -> Any: """simple docstring""" if equal_length: a__ : List[str] = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size a__ : List[Any] = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: a__ : Tuple = [np.asarray(SCREAMING_SNAKE_CASE_ ) for x in speech_inputs] return speech_inputs @require_torch class snake_case__ (_UpperCAmelCase , unittest.TestCase ): """simple docstring""" __lowerCAmelCase :List[Any] = SpeechTaFeatureExtractor def SCREAMING_SNAKE_CASE__( self ) -> Any: """simple docstring""" a__ : Optional[Any] = SpeechTaFeatureExtractionTester(self ) def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> int: """simple docstring""" self.assertTrue(np.all(np.mean(SCREAMING_SNAKE_CASE_ , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(SCREAMING_SNAKE_CASE_ , axis=0 ) - 1 ) < 1E-3 ) ) def SCREAMING_SNAKE_CASE__( self ) -> Union[str, Any]: """simple docstring""" a__ : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a__ : Dict = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] a__ : Optional[int] = [np.asarray(SCREAMING_SNAKE_CASE_ ) for speech_input in speech_inputs] # Test not batched input a__ : Optional[Any] = feat_extract(speech_inputs[0] , return_tensors="""np""" ).input_values a__ : Dict = feat_extract(np_speech_inputs[0] , return_tensors="""np""" ).input_values self.assertTrue(np.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , atol=1E-3 ) ) # Test batched a__ : Optional[Any] = feat_extract(SCREAMING_SNAKE_CASE_ , return_tensors="""np""" ).input_values a__ : Dict = feat_extract(SCREAMING_SNAKE_CASE_ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): self.assertTrue(np.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , atol=1E-3 ) ) def SCREAMING_SNAKE_CASE__( self ) -> Dict: """simple docstring""" a__ : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a__ : Any = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] a__ : List[Any] = ["""longest""", """max_length""", """do_not_pad"""] a__ : Union[str, Any] = [None, 1_6_0_0, None] for max_length, padding in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): a__ : Optional[int] = feat_extract(SCREAMING_SNAKE_CASE_ , padding=SCREAMING_SNAKE_CASE_ , max_length=SCREAMING_SNAKE_CASE_ , return_tensors="""np""" ) a__ : Dict = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_0_0] ) self.assertTrue(input_values[0][8_0_0:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1_0_0_0] ) self.assertTrue(input_values[0][1_0_0_0:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1_2_0_0] ) def SCREAMING_SNAKE_CASE__( self ) -> str: """simple docstring""" a__ : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a__ : Optional[Any] = range(8_0_0 , 1_4_0_0 , 2_0_0 ) a__ : Any = [floats_list((1, x) )[0] for x in lengths] a__ : int = ["""longest""", """max_length""", """do_not_pad"""] a__ : str = [None, 1_6_0_0, None] for max_length, padding in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): a__ : Any = feat_extract(SCREAMING_SNAKE_CASE_ , max_length=SCREAMING_SNAKE_CASE_ , padding=SCREAMING_SNAKE_CASE_ ) a__ : Optional[Any] = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:8_0_0] ) self._check_zero_mean_unit_variance(input_values[1][:1_0_0_0] ) self._check_zero_mean_unit_variance(input_values[2][:1_2_0_0] ) def SCREAMING_SNAKE_CASE__( self ) -> Dict: """simple docstring""" a__ : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a__ : Union[str, Any] = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] a__ : str = feat_extract( SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ , max_length=1_0_0_0 , padding="""max_length""" , return_tensors="""np""" ) a__ : int = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_0_0] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def SCREAMING_SNAKE_CASE__( self ) -> Any: """simple docstring""" a__ : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a__ : int = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] a__ : str = feat_extract( SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ , max_length=1_0_0_0 , padding="""longest""" , return_tensors="""np""" ) a__ : Optional[int] = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_0_0] ) self._check_zero_mean_unit_variance(input_values[1, :1_0_0_0] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1_0_0_0) ) a__ : Any = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] a__ : Dict = feat_extract( SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ , max_length=2_0_0_0 , padding="""longest""" , return_tensors="""np""" ) a__ : Any = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :8_0_0] ) self._check_zero_mean_unit_variance(input_values[1, :1_0_0_0] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1_2_0_0) ) def SCREAMING_SNAKE_CASE__( self ) -> int: """simple docstring""" a__ : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) a__ : Dict = np.random.rand(1_0_0 ).astype(np.floataa ) a__ : Any = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: a__ : Any = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) a__ : int = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def SCREAMING_SNAKE_CASE__( self ) -> Union[str, Any]: """simple docstring""" a__ : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 a__ : Any = [floats_list((1, x) )[0] for x in range(8_0_0 , 1_4_0_0 , 2_0_0 )] a__ : Tuple = [np.asarray(SCREAMING_SNAKE_CASE_ ) for speech_input in speech_inputs] # Test feature size a__ : Optional[Any] = feature_extractor(audio_target=SCREAMING_SNAKE_CASE_ , padding=SCREAMING_SNAKE_CASE_ , return_tensors="""np""" ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input a__ : List[str] = feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_values a__ : List[str] = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_values self.assertTrue(np.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , atol=1E-3 ) ) # Test batched a__ : int = feature_extractor(SCREAMING_SNAKE_CASE_ , return_tensors="""np""" ).input_values a__ : Tuple = feature_extractor(SCREAMING_SNAKE_CASE_ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): self.assertTrue(np.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. a__ : Optional[Any] = [floats_list((1, x) )[0] for x in (8_0_0, 8_0_0, 8_0_0)] a__ : int = np.asarray(SCREAMING_SNAKE_CASE_ ) a__ : List[Any] = feature_extractor(SCREAMING_SNAKE_CASE_ , return_tensors="""np""" ).input_values a__ : List[str] = feature_extractor(SCREAMING_SNAKE_CASE_ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): self.assertTrue(np.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , atol=1E-3 ) ) def SCREAMING_SNAKE_CASE__( self ) -> Optional[Any]: """simple docstring""" a__ : Dict = self.feat_extract_tester.prepare_inputs_for_target() a__ : Any = self.feature_extraction_class(**self.feat_extract_dict ) a__ : List[str] = feat_extract.model_input_names[0] a__ : Optional[int] = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(SCREAMING_SNAKE_CASE_ ) == len(SCREAMING_SNAKE_CASE_ ) for x, y in zip(SCREAMING_SNAKE_CASE_ , processed_features[input_name] ) ) ) a__ : str = self.feat_extract_tester.prepare_inputs_for_target(equal_length=SCREAMING_SNAKE_CASE_ ) a__ : List[str] = BatchFeature({input_name: speech_inputs} , tensor_type="""np""" ) a__ : str = processed_features[input_name] if len(batch_features_input.shape ) < 3: a__ : Union[str, Any] = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def SCREAMING_SNAKE_CASE__( self ) -> Union[str, Any]: """simple docstring""" a__ : Optional[int] = self.feat_extract_tester.prepare_inputs_for_target(equal_length=SCREAMING_SNAKE_CASE_ ) a__ : List[Any] = self.feature_extraction_class(**self.feat_extract_dict ) a__ : List[Any] = feat_extract.model_input_names[0] a__ : Any = BatchFeature({input_name: speech_inputs} , tensor_type="""pt""" ) a__ : Dict = processed_features[input_name] if len(batch_features_input.shape ) < 3: a__ : str = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def SCREAMING_SNAKE_CASE__( self ) -> Tuple: """simple docstring""" a__ : Any = self.feature_extraction_class(**self.feat_extract_dict ) a__ : int = self.feat_extract_tester.prepare_inputs_for_target() a__ : str = feat_extract.model_input_names[0] a__ : str = BatchFeature({input_name: speech_inputs} ) a__ : Optional[int] = feat_extract.num_mel_bins # hack! a__ : int = feat_extract.pad(SCREAMING_SNAKE_CASE_ , padding="""longest""" , return_tensors="""np""" )[input_name] a__ : Dict = feat_extract.pad(SCREAMING_SNAKE_CASE_ , padding="""longest""" , return_tensors="""pt""" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def SCREAMING_SNAKE_CASE__( self ) -> str: """simple docstring""" a__ : List[Any] = self.feat_extract_dict a__ : List[Any] = True a__ : Optional[Any] = self.feature_extraction_class(**SCREAMING_SNAKE_CASE_ ) a__ : Optional[Any] = self.feat_extract_tester.prepare_inputs_for_target() a__ : str = [len(SCREAMING_SNAKE_CASE_ ) for x in speech_inputs] a__ : str = feat_extract.model_input_names[0] a__ : List[Any] = BatchFeature({input_name: speech_inputs} ) a__ : Tuple = feat_extract.num_mel_bins # hack! a__ : Optional[int] = feat_extract.pad(SCREAMING_SNAKE_CASE_ , padding="""longest""" , return_tensors="""np""" ) self.assertIn("""attention_mask""" , SCREAMING_SNAKE_CASE_ ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , SCREAMING_SNAKE_CASE_ ) def SCREAMING_SNAKE_CASE__( self ) -> Optional[Any]: """simple docstring""" a__ : Optional[Any] = self.feat_extract_dict a__ : Union[str, Any] = True a__ : List[str] = self.feature_extraction_class(**SCREAMING_SNAKE_CASE_ ) a__ : Dict = self.feat_extract_tester.prepare_inputs_for_target() a__ : Any = [len(SCREAMING_SNAKE_CASE_ ) for x in speech_inputs] a__ : int = feat_extract.model_input_names[0] a__ : Any = BatchFeature({input_name: speech_inputs} ) a__ : str = min(SCREAMING_SNAKE_CASE_ ) a__ : Tuple = feat_extract.num_mel_bins # hack! a__ : Optional[Any] = feat_extract.pad( SCREAMING_SNAKE_CASE_ , padding="""max_length""" , max_length=SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ , return_tensors="""np""" ) self.assertIn("""attention_mask""" , SCREAMING_SNAKE_CASE_ ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] ) def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> Dict: """simple docstring""" from datasets import load_dataset a__ : List[str] = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech a__ : Union[str, Any] = ds.sort("""id""" ).select(range(SCREAMING_SNAKE_CASE_ ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def SCREAMING_SNAKE_CASE__( self ) -> Optional[Any]: """simple docstring""" a__ : int = torch.tensor( [2.3804E-03, 2.0752E-03, 1.9836E-03, 2.1057E-03, 1.6174E-03, 3.0518E-04, 9.1553E-05, 3.3569E-04, 9.7656E-04, 1.8311E-03, 2.0142E-03, 2.1057E-03, 1.7395E-03, 4.5776E-04, -3.9673E-04, 4.5776E-04, 1.0071E-03, 9.1553E-05, 4.8828E-04, 1.1597E-03, 7.3242E-04, 9.4604E-04, 1.8005E-03, 1.8311E-03, 8.8501E-04, 4.2725E-04, 4.8828E-04, 7.3242E-04, 1.0986E-03, 2.1057E-03] ) # fmt: on a__ : List[Any] = self._load_datasamples(1 ) a__ : int = SpeechTaFeatureExtractor() a__ : str = feature_extractor(SCREAMING_SNAKE_CASE_ , return_tensors="""pt""" ).input_values self.assertEquals(input_values.shape , (1, 9_3_6_8_0) ) self.assertTrue(torch.allclose(input_values[0, :3_0] , SCREAMING_SNAKE_CASE_ , atol=1E-6 ) ) def SCREAMING_SNAKE_CASE__( self ) -> Dict: """simple docstring""" a__ : Dict = torch.tensor( [-2.6_8_7_0, -3.0_1_0_4, -3.1_3_5_6, -3.5_3_5_2, -3.0_0_4_4, -3.0_3_5_3, -3.4_7_1_9, -3.6_7_7_7, -3.1_5_2_0, -2.9_4_3_5, -2.6_5_5_3, -2.8_7_9_5, -2.9_9_4_4, -2.5_9_2_1, -3.0_2_7_9, -3.0_3_8_6, -3.0_8_6_4, -3.1_2_9_1, -3.2_3_5_3, -2.7_4_4_4, -2.6_8_3_1, -2.7_2_8_7, -3.1_7_6_1, -3.1_5_7_1, -3.2_7_2_6, -3.0_5_8_2, -3.1_0_0_7, -3.4_5_3_3, -3.4_6_9_5, -3.0_9_9_8] ) # fmt: on a__ : Tuple = self._load_datasamples(1 ) a__ : Optional[Any] = SpeechTaFeatureExtractor() a__ : Optional[int] = feature_extractor(audio_target=SCREAMING_SNAKE_CASE_ , return_tensors="""pt""" ).input_values self.assertEquals(input_values.shape , (1, 3_6_6, 8_0) ) self.assertTrue(torch.allclose(input_values[0, 0, :3_0] , SCREAMING_SNAKE_CASE_ , atol=1E-4 ) )
355
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _lowercase : int ={ "configuration_longt5": ["LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP", "LongT5Config", "LongT5OnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Optional[Any] =[ "LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST", "LongT5EncoderModel", "LongT5ForConditionalGeneration", "LongT5Model", "LongT5PreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : List[Any] =[ "FlaxLongT5ForConditionalGeneration", "FlaxLongT5Model", "FlaxLongT5PreTrainedModel", ] if TYPE_CHECKING: from .configuration_longta import LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP, LongTaConfig, LongTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longta import ( LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST, LongTaEncoderModel, LongTaForConditionalGeneration, LongTaModel, LongTaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_longta import ( FlaxLongTaForConditionalGeneration, FlaxLongTaModel, FlaxLongTaPreTrainedModel, ) else: import sys _lowercase : Any =_LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
266
0
import json import os import shutil import tempfile import unittest import numpy as np from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES, BertTokenizer from transformers.testing_utils import require_tokenizers, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor, ViTImageProcessor @require_tokenizers @require_vision class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> int: __UpperCamelCase :List[str] = tempfile.mkdtemp() # fmt: off __UpperCamelCase :Any = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing''', ''',''', '''low''', '''lowest'''] # fmt: on __UpperCamelCase :Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file''']) with open(self.vocab_file , '''w''' , encoding='''utf-8''') as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens])) __UpperCamelCase :List[Any] = { '''do_resize''': True, '''size''': {'''height''': 18, '''width''': 18}, '''do_normalize''': True, '''image_mean''': [0.5, 0.5, 0.5], '''image_std''': [0.5, 0.5, 0.5], } __UpperCamelCase :List[str] = os.path.join(self.tmpdirname , __lowercase) with open(self.image_processor_file , '''w''' , encoding='''utf-8''') as fp: json.dump(__lowercase , __lowercase) def UpperCamelCase__ ( self , **__lowercase) -> List[Any]: return BertTokenizer.from_pretrained(self.tmpdirname , **__lowercase) def UpperCamelCase__ ( self , **__lowercase) -> Any: return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowercase) def UpperCamelCase__ ( self) -> Dict: shutil.rmtree(self.tmpdirname) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :List[Any] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta)] __UpperCamelCase :Any = [Image.fromarray(np.moveaxis(__lowercase , 0 , -1)) for x in image_inputs] return image_inputs def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :str = self.get_tokenizer() __UpperCamelCase :List[str] = self.get_image_processor() __UpperCamelCase :List[Any] = VisionTextDualEncoderProcessor(tokenizer=__lowercase , image_processor=__lowercase) processor.save_pretrained(self.tmpdirname) __UpperCamelCase :Any = VisionTextDualEncoderProcessor.from_pretrained(self.tmpdirname) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab()) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast)) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string()) self.assertIsInstance(processor.image_processor , __lowercase) def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :List[Any] = VisionTextDualEncoderProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) __UpperCamelCase :List[Any] = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''') __UpperCamelCase :Any = self.get_image_processor(do_normalize=__lowercase , padding_value=1.0) __UpperCamelCase :List[Any] = VisionTextDualEncoderProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=__lowercase , padding_value=1.0) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab()) self.assertIsInstance(processor.tokenizer , (BertTokenizer, BertTokenizerFast)) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , __lowercase) def UpperCamelCase__ ( self) -> Optional[Any]: __UpperCamelCase :Optional[Any] = self.get_image_processor() __UpperCamelCase :Optional[Any] = self.get_tokenizer() __UpperCamelCase :str = VisionTextDualEncoderProcessor(tokenizer=__lowercase , image_processor=__lowercase) __UpperCamelCase :str = self.prepare_image_inputs() __UpperCamelCase :Optional[int] = image_processor(__lowercase , return_tensors='''np''') __UpperCamelCase :str = processor(images=__lowercase , return_tensors='''np''') for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2) def UpperCamelCase__ ( self) -> str: __UpperCamelCase :List[Any] = self.get_image_processor() __UpperCamelCase :List[Any] = self.get_tokenizer() __UpperCamelCase :Any = VisionTextDualEncoderProcessor(tokenizer=__lowercase , image_processor=__lowercase) __UpperCamelCase :Optional[Any] = '''lower newer''' __UpperCamelCase :Optional[int] = processor(text=__lowercase) __UpperCamelCase :Any = tokenizer(__lowercase) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key]) def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :str = self.get_image_processor() __UpperCamelCase :int = self.get_tokenizer() __UpperCamelCase :Any = VisionTextDualEncoderProcessor(tokenizer=__lowercase , image_processor=__lowercase) __UpperCamelCase :Tuple = '''lower newer''' __UpperCamelCase :Tuple = self.prepare_image_inputs() __UpperCamelCase :Dict = processor(text=__lowercase , images=__lowercase) self.assertListEqual(list(inputs.keys()) , ['''input_ids''', '''token_type_ids''', '''attention_mask''', '''pixel_values''']) # test if it raises when no input is passed with self.assertRaises(__lowercase): processor() def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Dict = self.get_image_processor() __UpperCamelCase :Dict = self.get_tokenizer() __UpperCamelCase :Optional[int] = VisionTextDualEncoderProcessor(tokenizer=__lowercase , image_processor=__lowercase) __UpperCamelCase :Union[str, Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __UpperCamelCase :Union[str, Any] = processor.batch_decode(__lowercase) __UpperCamelCase :str = tokenizer.batch_decode(__lowercase) self.assertListEqual(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :int = self.get_image_processor() __UpperCamelCase :Tuple = self.get_tokenizer() __UpperCamelCase :Optional[int] = VisionTextDualEncoderProcessor(tokenizer=__lowercase , image_processor=__lowercase) __UpperCamelCase :Tuple = '''lower newer''' __UpperCamelCase :Optional[int] = self.prepare_image_inputs() __UpperCamelCase :Tuple = processor(text=__lowercase , images=__lowercase) self.assertListEqual(list(inputs.keys()) , processor.model_input_names)
43
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModel, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableUnCLIPImgaImgPipeline, UNetaDConditionModel from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, load_image, load_numpy, require_torch_gpu, skip_mps, slow, torch_device, ) from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class lowerCamelCase_ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : int = StableUnCLIPImgaImgPipeline a__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS a__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS a__ : Optional[Any] = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess a__ : int = frozenset([] ) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Tuple = 32 __UpperCamelCase :Optional[int] = embedder_hidden_size # image encoding components __UpperCamelCase :Union[str, Any] = CLIPImageProcessor(crop_size=32 , size=32) torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPVisionModelWithProjection( CLIPVisionConfig( hidden_size=__lowercase , projection_dim=__lowercase , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , )) # regular denoising components torch.manual_seed(0) __UpperCamelCase :str = StableUnCLIPImageNormalizer(embedding_dim=__lowercase) __UpperCamelCase :Optional[int] = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''') torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') torch.manual_seed(0) __UpperCamelCase :Dict = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__lowercase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , )) torch.manual_seed(0) __UpperCamelCase :List[Any] = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=__lowercase , layers_per_block=1 , upcast_attention=__lowercase , use_linear_projection=__lowercase , ) torch.manual_seed(0) __UpperCamelCase :Tuple = DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.0_00_85 , beta_end=0.0_12 , prediction_type='''v_prediction''' , set_alpha_to_one=__lowercase , steps_offset=1 , ) torch.manual_seed(0) __UpperCamelCase :List[str] = AutoencoderKL() __UpperCamelCase :Tuple = { # image encoding components '''feature_extractor''': feature_extractor, '''image_encoder''': image_encoder.eval(), # image noising components '''image_normalizer''': image_normalizer.eval(), '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder.eval(), '''unet''': unet.eval(), '''scheduler''': scheduler, '''vae''': vae.eval(), } return components def UpperCamelCase__ ( self , __lowercase , __lowercase=0 , __lowercase=True) -> str: if str(__lowercase).startswith('''mps'''): __UpperCamelCase :Union[str, Any] = torch.manual_seed(__lowercase) else: __UpperCamelCase :int = torch.Generator(device=__lowercase).manual_seed(__lowercase) __UpperCamelCase :int = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowercase)).to(__lowercase) if pil_image: __UpperCamelCase :List[Any] = input_image * 0.5 + 0.5 __UpperCamelCase :Optional[Any] = input_image.clamp(0 , 1) __UpperCamelCase :int = input_image.cpu().permute(0 , 2 , 3 , 1).float().numpy() __UpperCamelCase :Optional[Any] = DiffusionPipeline.numpy_to_pil(__lowercase)[0] return { "prompt": "An anime racoon running a marathon", "image": input_image, "generator": generator, "num_inference_steps": 2, "output_type": "np", } @skip_mps def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :Dict = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Tuple = self.get_dummy_components() __UpperCamelCase :Any = StableUnCLIPImgaImgPipeline(**__lowercase) __UpperCamelCase :Optional[Any] = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :List[Any] = self.get_dummy_inputs(__lowercase) inputs.update({'''image_embeds''': None}) __UpperCamelCase :Any = sd_pipe(**__lowercase).images __UpperCamelCase :List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCamelCase :List[Any] = np.array([0.38_72, 0.72_24, 0.56_01, 0.47_41, 0.68_72, 0.58_14, 0.46_36, 0.38_67, 0.50_78]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Optional[Any] = torch_device in ['''cpu''', '''mps'''] self._test_attention_slicing_forward_pass(test_max_difference=__lowercase) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Optional[Any] = torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=__lowercase) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase__ ( self) -> Union[str, Any]: self._test_xformers_attention_forwardGenerator_pass(test_max_difference=__lowercase) @slow @require_torch_gpu class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Union[str, Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :int = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') __UpperCamelCase :Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_img2img_anime_turtle_fp16.npy''') __UpperCamelCase :List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-l-img2img''' , torch_dtype=torch.floataa) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :Dict = pipe(__lowercase , '''anime turle''' , generator=__lowercase , output_type='''np''') __UpperCamelCase :Dict = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Optional[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') __UpperCamelCase :Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_h_img2img_anime_turtle_fp16.npy''') __UpperCamelCase :Any = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :Optional[int] = pipe(__lowercase , '''anime turle''' , generator=__lowercase , output_type='''np''') __UpperCamelCase :List[Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __UpperCamelCase :List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa) __UpperCamelCase :Union[str, Any] = pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :Optional[Any] = pipe( __lowercase , '''anime turtle''' , num_inference_steps=2 , output_type='''np''' , ) __UpperCamelCase :int = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
43
1
"""simple docstring""" import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint __lowerCamelCase = { "169M": 12, "430M": 24, "1B5": 24, "3B": 32, "7B": 32, "14B": 40, } __lowerCamelCase = { "169M": 7_68, "430M": 10_24, "1B5": 20_48, "3B": 25_60, "7B": 40_96, "14B": 51_20, } def UpperCAmelCase ( UpperCamelCase__ ): """simple docstring""" A__ = list(state_dict.keys() ) for name in state_dict_keys: A__ = state_dict.pop(UpperCamelCase__ ) # emb -> embedding if name.startswith('emb.' ): A__ = name.replace('emb.' , 'embeddings.' ) # ln_0 -> pre_ln (only present at block 0) if name.startswith('blocks.0.ln0' ): A__ = name.replace('blocks.0.ln0' , 'blocks.0.pre_ln' ) # att -> attention A__ = re.sub(r'blocks\.(\d+)\.att' , r'blocks.\1.attention' , UpperCamelCase__ ) # ffn -> feed_forward A__ = re.sub(r'blocks\.(\d+)\.ffn' , r'blocks.\1.feed_forward' , UpperCamelCase__ ) # time_mix_k -> time_mix_key and reshape if name.endswith('.time_mix_k' ): A__ = name.replace('.time_mix_k' , '.time_mix_key' ) # time_mix_v -> time_mix_value and reshape if name.endswith('.time_mix_v' ): A__ = name.replace('.time_mix_v' , '.time_mix_value' ) # time_mix_r -> time_mix_key and reshape if name.endswith('.time_mix_r' ): A__ = name.replace('.time_mix_r' , '.time_mix_receptance' ) if name != "head.weight": A__ = 'rwkv.' + name A__ = weight return state_dict def UpperCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=False , UpperCamelCase__=None ): """simple docstring""" if tokenizer_file is None: print('No `--tokenizer_file` provided, we will use the default tokenizer.' ) A__ = 50_277 A__ = AutoTokenizer.from_pretrained('EleutherAI/gpt-neox-20b' ) else: A__ = PreTrainedTokenizerFast(tokenizer_file=UpperCamelCase__ ) A__ = len(UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) # 2. Build the config A__ = list(NUM_HIDDEN_LAYERS_MAPPING.keys() ) if size is None: # Try to infer size from the checkpoint name for candidate in possible_sizes: if candidate in checkpoint_file: A__ = candidate break if size is None: raise ValueError('Could not infer the size, please provide it with the `--size` argument.' ) if size not in possible_sizes: raise ValueError(F'''`size` should be one of {possible_sizes}, got {size}.''' ) A__ = RwkvConfig( vocab_size=UpperCamelCase__ , num_hidden_layers=NUM_HIDDEN_LAYERS_MAPPING[size] , hidden_size=HIDEN_SIZE_MAPPING[size] , ) config.save_pretrained(UpperCamelCase__ ) # 3. Download model file then convert state_dict A__ = hf_hub_download(UpperCamelCase__ , UpperCamelCase__ ) A__ = torch.load(UpperCamelCase__ , map_location='cpu' ) A__ = convert_state_dict(UpperCamelCase__ ) # 4. Split in shards and save A__ , A__ = shard_checkpoint(UpperCamelCase__ ) for shard_file, shard in shards.items(): torch.save(UpperCamelCase__ , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) if index is not None: A__ = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) # Save the index as well with open(UpperCamelCase__ , 'w' , encoding='utf-8' ) as f: A__ = json.dumps(UpperCamelCase__ , indent=2 , sort_keys=UpperCamelCase__ ) + '\n' f.write(UpperCamelCase__ ) # 5. Clean up shards (for some reason the file PyTorch saves take the same space as the whole state_dict print( 'Cleaning up shards. This may error with an OOM error, it this is the case don\'t worry you still have converted the model.' ) A__ = list(shards.keys() ) del state_dict del shards gc.collect() for shard_file in shard_files: A__ = torch.load(os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) torch.save({k: v.cpu().clone() for k, v in state_dict.items()} , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) del state_dict gc.collect() if push_to_hub: if model_name is None: raise ValueError('Please provide a `model_name` to push the model to the Hub.' ) A__ = AutoModelForCausalLM.from_pretrained(UpperCamelCase__ ) model.push_to_hub(UpperCamelCase__ , max_shard_size='2GB' ) tokenizer.push_to_hub(UpperCamelCase__ ) if __name__ == "__main__": __lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--repo_id", default=None, type=str, required=True, help="Repo ID from which to pull the checkpoint." ) parser.add_argument( "--checkpoint_file", default=None, type=str, required=True, help="Name of the checkpoint file in the repo." ) parser.add_argument( "--output_dir", default=None, type=str, required=True, help="Where to save the converted model." ) parser.add_argument( "--tokenizer_file", default=None, type=str, help="Path to the tokenizer file to use (if not provided, only the model is converted).", ) parser.add_argument( "--size", default=None, type=str, help="Size of the model. Will be inferred from the `checkpoint_file` if not passed.", ) parser.add_argument( "--push_to_hub", action="store_true", help="Push to the Hub the converted model.", ) parser.add_argument( "--model_name", default=None, type=str, help="Name of the pushed model on the Hub, including the username / organization.", ) __lowerCamelCase = parser.parse_args() convert_rmkv_checkpoint_to_hf_format( args.repo_id, args.checkpoint_file, args.output_dir, size=args.size, tokenizer_file=args.tokenizer_file, push_to_hub=args.push_to_hub, model_name=args.model_name, )
154
"""simple docstring""" def UpperCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" if not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise ValueError('iterations must be defined as integers' ) if not isinstance(UpperCamelCase__ , UpperCamelCase__ ) or not number >= 1: raise ValueError( 'starting number must be\n and integer and be more than 0' ) if not iterations >= 1: raise ValueError('Iterations must be done more than 0 times to play FizzBuzz' ) A__ = '' while number <= iterations: if number % 3 == 0: out += "Fizz" if number % 5 == 0: out += "Buzz" if 0 not in (number % 3, number % 5): out += str(UpperCamelCase__ ) # print(out) number += 1 out += " " return out if __name__ == "__main__": import doctest doctest.testmod()
154
1
'''simple docstring''' import os import numpy import onnx def a ( lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : Dict = a.name A_ : Union[str, Any] = b.name A_ : Any = """""" A_ : Optional[Any] = """""" A_ : str = a == b A_ : List[str] = name_a A_ : Dict = name_b return res def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' for i, input_name in enumerate(node_proto.input ): if input_name == name: node_proto.input.insert(lowerCamelCase__ , lowerCamelCase__ ) node_proto.input.pop(i + 1 ) if node_proto.op_type == "If": _graph_replace_input_with(node_proto.attribute[0].g , lowerCamelCase__ , lowerCamelCase__ ) _graph_replace_input_with(node_proto.attribute[1].g , lowerCamelCase__ , lowerCamelCase__ ) if node_proto.op_type == "Loop": _graph_replace_input_with(node_proto.attribute[0].g , lowerCamelCase__ , lowerCamelCase__ ) def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' for n in graph_proto.node: _node_replace_input_with(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): '''simple docstring''' A_ : Any = list(model.graph.initializer ) A_ : Optional[int] = list(model_without_ext.graph.initializer ) for i, ref_i in ind_to_replace: assert inits_with_data[i].name == inits[i].name assert inits_with_data[ref_i].name == inits[ref_i].name assert i > ref_i A_ : Optional[Any] = inits[i].name A_ : Any = inits[ref_i].name model_without_ext.graph.initializer.remove(inits[i] ) # for n in model.graph.node: _graph_replace_input_with(model_without_ext.graph , lowerCamelCase__ , lowerCamelCase__ ) def a ( lowerCamelCase__ ): '''simple docstring''' A_ : Optional[Any] = os.path.dirname(lowerCamelCase__ ) A_ : Any = os.path.basename(lowerCamelCase__ ) A_ : Union[str, Any] = onnx.load(os.path.join(lowerCamelCase__ , lowerCamelCase__ ) ) A_ : Tuple = list(model.graph.initializer ) A_ : Dict = set() A_ : str = {} A_ : Optional[Any] = [] A_ : Optional[Any] = 0 for i in range(len(lowerCamelCase__ ) ): if i in dup_set: continue for j in range(i + 1 , len(lowerCamelCase__ ) ): if j in dup_set: continue if _is_equal_tensor_proto(inits[i] , inits[j] ): dup_set.add(lowerCamelCase__ ) dup_set.add(lowerCamelCase__ ) A_ : Union[str, Any] = inits[j].data_type A_ : Dict = numpy.prod(inits[j].dims ) if dtype == 1: mem_size *= 4 elif dtype == 6: mem_size *= 4 elif dtype == 7 or dtype == 11: mem_size *= 8 else: print("""unexpected data type: """ , lowerCamelCase__ ) total_reduced_size += mem_size A_ : List[Any] = inits[i].name A_ : Union[str, Any] = inits[j].name if name_i in dup_map: dup_map[name_i].append(lowerCamelCase__ ) else: A_ : Union[str, Any] = [name_j] ind_to_replace.append((j, i) ) print("""total reduced size: """ , total_reduced_size / 10_24 / 10_24 / 10_24 , """GB""" ) A_ : Union[str, Any] = sorted(lowerCamelCase__ ) _remove_dup_initializers_from_model(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) A_ : Optional[int] = """optimized_""" + model_file_name A_ : Optional[Any] = os.path.join(lowerCamelCase__ , lowerCamelCase__ ) onnx.save(lowerCamelCase__ , lowerCamelCase__ ) return new_model
206
'''simple docstring''' # Author: OMKAR PATHAK, Nwachukwu Chidiebere # Use a Python dictionary to construct the graph. from __future__ import annotations from pprint import pformat from typing import Generic, TypeVar lowerCamelCase :str = TypeVar('''T''') class _lowerCAmelCase ( Generic[T] ): def __init__(self , lowercase = True ): A_ : dict[T, list[T]] = {} # dictionary of lists A_ : Any = directed def _a (self , lowercase , lowercase ): if not self.directed: # For undirected graphs # if both source vertex and destination vertex are both present in the # adjacency list, add destination vertex to source vertex list of adjacent # vertices and add source vertex to destination vertex list of adjacent # vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase ) self.adj_list[destination_vertex].append(lowercase ) # if only source vertex is present in adjacency list, add destination vertex # to source vertex list of adjacent vertices, then create a new vertex with # destination vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase ) A_ : Dict = [source_vertex] # if only destination vertex is present in adjacency list, add source vertex # to destination vertex list of adjacent vertices, then create a new vertex # with source vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif destination_vertex in self.adj_list: self.adj_list[destination_vertex].append(lowercase ) A_ : int = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and assign a list # containing the destination vertex as it's first adjacent vertex also # create a new vertex with destination vertex as key and assign a list # containing the source vertex as it's first adjacent vertex. else: A_ : Optional[Any] = [destination_vertex] A_ : Tuple = [source_vertex] else: # For directed graphs # if both source vertex and destination vertex are present in adjacency # list, add destination vertex to source vertex list of adjacent vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase ) # if only source vertex is present in adjacency list, add destination # vertex to source vertex list of adjacent vertices and create a new vertex # with destination vertex as key, which has no adjacent vertex elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(lowercase ) A_ : Tuple = [] # if only destination vertex is present in adjacency list, create a new # vertex with source vertex as key and assign a list containing destination # vertex as first adjacent vertex elif destination_vertex in self.adj_list: A_ : Tuple = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and a list containing # destination vertex as it's first adjacent vertex. Then create a new vertex # with destination vertex as key, which has no adjacent vertex else: A_ : int = [destination_vertex] A_ : List[str] = [] return self def __repr__(self ): return pformat(self.adj_list )
206
1
'''simple docstring''' import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging __UpperCAmelCase :int = logging.get_logger(__name__) class a ( _a ): """simple docstring""" SCREAMING_SNAKE_CASE : str = ["input_features"] def __init__( self : List[str] , snake_case : Union[str, Any]=80 , snake_case : Optional[int]=1_6000 , snake_case : Dict=160 , snake_case : Optional[Any]=30 , snake_case : List[str]=400 , snake_case : Optional[int]=0.0 , snake_case : Any=False , **snake_case : List[Any] , ) -> Optional[Any]: super().__init__( feature_size=snake_case , sampling_rate=snake_case , padding_value=snake_case , return_attention_mask=snake_case , **snake_case , ) __UpperCAmelCase : Optional[Any] = n_fft __UpperCAmelCase : Union[str, Any] = hop_length __UpperCAmelCase : Union[str, Any] = chunk_length __UpperCAmelCase : Tuple = chunk_length * sampling_rate __UpperCAmelCase : Any = self.n_samples // hop_length __UpperCAmelCase : Any = sampling_rate __UpperCAmelCase : Optional[Any] = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=snake_case , min_frequency=0.0 , max_frequency=8_000.0 , sampling_rate=snake_case , norm='''slaney''' , mel_scale='''slaney''' , ) def lowerCamelCase__ ( self : List[Any] , snake_case : np.array ) -> np.ndarray: __UpperCAmelCase : Dict = spectrogram( snake_case , window_function(self.n_fft , '''hann''' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel='''log10''' , ) __UpperCAmelCase : Tuple = log_spec[:, :-1] __UpperCAmelCase : str = np.maximum(snake_case , log_spec.max() - 8.0 ) __UpperCAmelCase : Any = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def lowerCamelCase__ ( snake_case : List[np.ndarray] , snake_case : List[np.ndarray] , snake_case : float = 0.0 ) -> List[np.ndarray]: if attention_mask is not None: __UpperCAmelCase : List[str] = np.array(snake_case , np.intaa ) __UpperCAmelCase : Optional[Any] = [] for vector, length in zip(snake_case , attention_mask.sum(-1 ) ): __UpperCAmelCase : List[str] = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: __UpperCAmelCase : Optional[Any] = padding_value normed_input_values.append(snake_case ) else: __UpperCAmelCase : Tuple = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def __call__( self : str , snake_case : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , snake_case : bool = True , snake_case : Optional[int] = None , snake_case : Optional[Union[str, TensorType]] = None , snake_case : Optional[bool] = None , snake_case : Optional[str] = "max_length" , snake_case : Optional[int] = None , snake_case : Optional[int] = None , snake_case : Optional[bool] = None , **snake_case : Any , ) -> BatchFeature: if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a' f' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input' f' was sampled with {self.sampling_rate} and not {sampling_rate}.' ) else: logger.warning( '''It is strongly recommended to pass the `sampling_rate` argument to this function. ''' '''Failing to do so can result in silent errors that might be hard to debug.''' ) __UpperCAmelCase : List[str] = isinstance(snake_case , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'Only mono-channel audio is supported for input to {self}' ) __UpperCAmelCase : int = is_batched_numpy or ( isinstance(snake_case , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __UpperCAmelCase : Any = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(snake_case , np.ndarray ): __UpperCAmelCase : Dict = np.asarray(snake_case , dtype=np.floataa ) elif isinstance(snake_case , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __UpperCAmelCase : str = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __UpperCAmelCase : Tuple = [np.asarray([raw_speech] ).T] __UpperCAmelCase : Tuple = BatchFeature({'''input_features''': raw_speech} ) # convert into correct format for padding __UpperCAmelCase : Optional[int] = self.pad( snake_case , padding=snake_case , max_length=max_length if max_length else self.n_samples , truncation=snake_case , pad_to_multiple_of=snake_case , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: __UpperCAmelCase : List[str] = self.zero_mean_unit_var_norm( padded_inputs['''input_features'''] , attention_mask=padded_inputs['''attention_mask'''] , padding_value=self.padding_value , ) __UpperCAmelCase : Dict = np.stack(padded_inputs['''input_features'''] , axis=0 ) # make sure list is in array format __UpperCAmelCase : List[Any] = padded_inputs.get('''input_features''' ).transpose(2 , 0 , 1 ) __UpperCAmelCase : Optional[int] = [self._np_extract_fbank_features(snake_case ) for waveform in input_features[0]] if isinstance(input_features[0] , snake_case ): __UpperCAmelCase : str = [np.asarray(snake_case , dtype=np.floataa ) for feature in input_features] else: __UpperCAmelCase : int = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) __UpperCAmelCase : Any = padded_inputs['''attention_mask'''][:, :: self.hop_length] if return_tensors is not None: __UpperCAmelCase : Tuple = padded_inputs.convert_to_tensors(snake_case ) return padded_inputs def lowerCamelCase__ ( self : Any ) -> Dict[str, Any]: __UpperCAmelCase : int = copy.deepcopy(self.__dict__ ) __UpperCAmelCase : Any = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
240
'''simple docstring''' from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class a : """simple docstring""" SCREAMING_SNAKE_CASE : torch.Tensor # [batch_size x 3] SCREAMING_SNAKE_CASE : torch.Tensor # [batch_size x 3] SCREAMING_SNAKE_CASE : torch.Tensor # [batch_size x 3] SCREAMING_SNAKE_CASE : torch.Tensor # [batch_size x 3] SCREAMING_SNAKE_CASE : int SCREAMING_SNAKE_CASE : int SCREAMING_SNAKE_CASE : float SCREAMING_SNAKE_CASE : float SCREAMING_SNAKE_CASE : Tuple[int] def lowerCamelCase__ ( self : Any ) -> int: assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def lowerCamelCase__ ( self : Union[str, Any] ) -> str: return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def lowerCamelCase__ ( self : Any ) -> Optional[Any]: return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def lowerCamelCase__ ( self : Any ) -> torch.Tensor: __UpperCAmelCase : Dict = torch.arange(self.height * self.width ) __UpperCAmelCase : Dict = torch.stack( [ pixel_indices % self.width, torch.div(snake_case , self.width , rounding_mode='''trunc''' ), ] , axis=1 , ) return coords @property def lowerCamelCase__ ( self : Any ) -> int: __UpperCAmelCase , *__UpperCAmelCase : str = self.shape __UpperCAmelCase : Dict = int(np.prod(snake_case ) ) __UpperCAmelCase : Tuple = self.get_image_coords() __UpperCAmelCase : List[Any] = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) __UpperCAmelCase : Any = self.get_camera_rays(snake_case ) __UpperCAmelCase : List[str] = rays.view(snake_case , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def lowerCamelCase__ ( self : Union[str, Any] , snake_case : torch.Tensor ) -> torch.Tensor: __UpperCAmelCase , *__UpperCAmelCase , __UpperCAmelCase : List[str] = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] __UpperCAmelCase : List[str] = coords.view(snake_case , -1 , 2 ) __UpperCAmelCase : Optional[Any] = self.resolution() __UpperCAmelCase : Tuple = self.fov() __UpperCAmelCase : Optional[int] = (flat.float() / (res - 1)) * 2 - 1 __UpperCAmelCase : Union[str, Any] = fracs * torch.tan(fov / 2 ) __UpperCAmelCase : str = fracs.view(snake_case , -1 , 2 ) __UpperCAmelCase : Any = ( self.z.view(snake_case , 1 , 3 ) + self.x.view(snake_case , 1 , 3 ) * fracs[:, :, :1] + self.y.view(snake_case , 1 , 3 ) * fracs[:, :, 1:] ) __UpperCAmelCase : Union[str, Any] = directions / directions.norm(dim=-1 , keepdim=snake_case ) __UpperCAmelCase : Union[str, Any] = torch.stack( [ torch.broadcast_to(self.origin.view(snake_case , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(snake_case , *snake_case , 2 , 3 ) def lowerCamelCase__ ( self : Any , snake_case : int , snake_case : int ) -> "DifferentiableProjectiveCamera": assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=snake_case , height=snake_case , x_fov=self.x_fov , y_fov=self.y_fov , ) def _a ( _lowercase : int ): '''simple docstring''' __UpperCAmelCase : str = [] __UpperCAmelCase : Optional[int] = [] __UpperCAmelCase : Union[str, Any] = [] __UpperCAmelCase : List[Any] = [] for theta in np.linspace(0 , 2 * np.pi , num=20 ): __UpperCAmelCase : Dict = np.array([np.sin(_lowercase ), np.cos(_lowercase ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) __UpperCAmelCase : Any = -z * 4 __UpperCAmelCase : Dict = np.array([np.cos(_lowercase ), -np.sin(_lowercase ), 0.0] ) __UpperCAmelCase : List[str] = np.cross(_lowercase , _lowercase ) origins.append(_lowercase ) xs.append(_lowercase ) ys.append(_lowercase ) zs.append(_lowercase ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(_lowercase , axis=0 ) ).float() , x=torch.from_numpy(np.stack(_lowercase , axis=0 ) ).float() , y=torch.from_numpy(np.stack(_lowercase , axis=0 ) ).float() , z=torch.from_numpy(np.stack(_lowercase , axis=0 ) ).float() , width=_lowercase , height=_lowercase , x_fov=0.7 , y_fov=0.7 , shape=(1, len(_lowercase )) , )
240
1
from collections.abc import Iterable from typing import Any class __magic_name__ : '''simple docstring''' def __init__( self, lowercase_ = None ) -> Optional[Any]: """simple docstring""" a__ =value a__ =None # Added in order to delete a node easier a__ =None a__ =None def __repr__( self ) -> str: """simple docstring""" from pprint import pformat if self.left is None and self.right is None: return str(self.value ) return pformat({F"""{self.value}""": (self.left, self.right)}, indent=1 ) class __magic_name__ : '''simple docstring''' def __init__( self, lowercase_ = None ) -> str: """simple docstring""" a__ =root def __str__( self ) -> str: """simple docstring""" return str(self.root ) def _UpperCAmelCase ( self, lowercase_, lowercase_ ) -> None: """simple docstring""" if new_children is not None: # reset its kids a__ =node.parent if node.parent is not None: # reset its parent if self.is_right(lowercase_ ): # If it is the right children a__ =new_children else: a__ =new_children else: a__ =new_children def _UpperCAmelCase ( self, lowercase_ ) -> bool: """simple docstring""" if node.parent and node.parent.right: return node == node.parent.right return False def _UpperCAmelCase ( self ) -> bool: """simple docstring""" return self.root is None def _UpperCAmelCase ( self, lowercase_ ) -> None: """simple docstring""" a__ =Node(lowercase_ ) # create a new Node if self.empty(): # if Tree is empty a__ =new_node # set its root else: # Tree is not empty a__ =self.root # from root if parent_node is None: return while True: # While we don't get to a leaf if value < parent_node.value: # We go left if parent_node.left is None: a__ =new_node # We insert the new node in a leaf break else: a__ =parent_node.left else: if parent_node.right is None: a__ =new_node break else: a__ =parent_node.right a__ =parent_node def _UpperCAmelCase ( self, *lowercase_ ) -> None: """simple docstring""" for value in values: self.__insert(lowercase_ ) def _UpperCAmelCase ( self, lowercase_ ) -> Node | None: """simple docstring""" if self.empty(): raise IndexError('''Warning: Tree is empty! please use another.''' ) else: a__ =self.root # use lazy evaluation here to avoid NoneType Attribute error while node is not None and node.value is not value: a__ =node.left if value < node.value else node.right return node def _UpperCAmelCase ( self, lowercase_ = None ) -> Node | None: """simple docstring""" if node is None: if self.root is None: return None a__ =self.root if not self.empty(): while node.right is not None: a__ =node.right return node def _UpperCAmelCase ( self, lowercase_ = None ) -> Node | None: """simple docstring""" if node is None: a__ =self.root if self.root is None: return None if not self.empty(): a__ =self.root while node.left is not None: a__ =node.left return node def _UpperCAmelCase ( self, lowercase_ ) -> None: """simple docstring""" a__ =self.search(lowercase_ ) # Look for the node with that label if node is not None: if node.left is None and node.right is None: # If it has no children self.__reassign_nodes(lowercase_, lowercase_ ) elif node.left is None: # Has only right children self.__reassign_nodes(lowercase_, node.right ) elif node.right is None: # Has only left children self.__reassign_nodes(lowercase_, node.left ) else: a__ =self.get_max( node.left ) # Gets the max value of the left branch self.remove(tmp_node.value ) # type: ignore a__ =( tmp_node.value # type: ignore ) # Assigns the value to the node to delete and keep tree structure def _UpperCAmelCase ( self, lowercase_ ) -> Iterable: """simple docstring""" if node is not None: yield node # Preorder Traversal yield from self.preorder_traverse(node.left ) yield from self.preorder_traverse(node.right ) def _UpperCAmelCase ( self, lowercase_=None ) -> Any: """simple docstring""" if traversal_function is None: return self.preorder_traverse(self.root ) else: return traversal_function(self.root ) def _UpperCAmelCase ( self, lowercase_, lowercase_ ) -> None: """simple docstring""" if node: self.inorder(lowercase_, node.left ) arr.append(node.value ) self.inorder(lowercase_, node.right ) def _UpperCAmelCase ( self, lowercase_, lowercase_ ) -> int: """simple docstring""" a__ =[] self.inorder(lowercase_, lowercase_ ) # append all values to list using inorder traversal return arr[k - 1] def UpperCAmelCase__ ( _A : Node | None ): '''simple docstring''' a__ =[] if curr_node is not None: a__ =postorder(curr_node.left ) + postorder(curr_node.right ) + [curr_node] return node_list def UpperCAmelCase__ ( ): '''simple docstring''' a__ =(8, 3, 6, 1, 10, 14, 13, 4, 7) a__ =BinarySearchTree() for i in testlist: t.insert(_A ) # Prints all the elements of the list in order traversal print(_A ) if t.search(6 ) is not None: print('''The value 6 exists''' ) else: print('''The value 6 doesn\'t exist''' ) if t.search(-1 ) is not None: print('''The value -1 exists''' ) else: print('''The value -1 doesn\'t exist''' ) if not t.empty(): print('''Max Value: ''' , t.get_max().value ) # type: ignore print('''Min Value: ''' , t.get_min().value ) # type: ignore for i in testlist: t.remove(_A ) print(_A ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
188
import os import string import sys lowerCamelCase = 1 << 8 lowerCamelCase = { '''tab''': ord('''\t'''), '''newline''': ord('''\r'''), '''esc''': 27, '''up''': 65 + ARROW_KEY_FLAG, '''down''': 66 + ARROW_KEY_FLAG, '''right''': 67 + ARROW_KEY_FLAG, '''left''': 68 + ARROW_KEY_FLAG, '''mod_int''': 91, '''undefined''': sys.maxsize, '''interrupt''': 3, '''insert''': 50, '''delete''': 51, '''pg_up''': 53, '''pg_down''': 54, } lowerCamelCase = KEYMAP['''up'''] lowerCamelCase = KEYMAP['''left'''] if sys.platform == "win32": lowerCamelCase = [] lowerCamelCase = { b'''\xe0H''': KEYMAP['''up'''] - ARROW_KEY_FLAG, b'''\x00H''': KEYMAP['''up'''] - ARROW_KEY_FLAG, b'''\xe0P''': KEYMAP['''down'''] - ARROW_KEY_FLAG, b'''\x00P''': KEYMAP['''down'''] - ARROW_KEY_FLAG, b'''\xe0M''': KEYMAP['''right'''] - ARROW_KEY_FLAG, b'''\x00M''': KEYMAP['''right'''] - ARROW_KEY_FLAG, b'''\xe0K''': KEYMAP['''left'''] - ARROW_KEY_FLAG, b'''\x00K''': KEYMAP['''left'''] - ARROW_KEY_FLAG, } for i in range(10): lowerCamelCase = ord(str(i)) def UpperCAmelCase__ ( ): '''simple docstring''' if os.name == "nt": import msvcrt a__ ='''mbcs''' # Flush the keyboard buffer while msvcrt.kbhit(): msvcrt.getch() if len(_A ) == 0: # Read the keystroke a__ =msvcrt.getch() # If it is a prefix char, get second part if ch in (b"\x00", b"\xe0"): a__ =ch + msvcrt.getch() # Translate actual Win chars to bullet char types try: a__ =chr(WIN_KEYMAP[cha] ) WIN_CH_BUFFER.append(chr(KEYMAP['''mod_int'''] ) ) WIN_CH_BUFFER.append(_A ) if ord(_A ) in ( KEYMAP["insert"] - 1 << 9, KEYMAP["delete"] - 1 << 9, KEYMAP["pg_up"] - 1 << 9, KEYMAP["pg_down"] - 1 << 9, ): WIN_CH_BUFFER.append(chr(1_26 ) ) a__ =chr(KEYMAP['''esc'''] ) except KeyError: a__ =cha[1] else: a__ =ch.decode(_A ) else: a__ =WIN_CH_BUFFER.pop(0 ) elif os.name == "posix": import termios import tty a__ =sys.stdin.fileno() a__ =termios.tcgetattr(_A ) try: tty.setraw(_A ) a__ =sys.stdin.read(1 ) finally: termios.tcsetattr(_A , termios.TCSADRAIN , _A ) return ch def UpperCAmelCase__ ( ): '''simple docstring''' a__ =get_raw_chars() if ord(_A ) in [KEYMAP["interrupt"], KEYMAP["newline"]]: return char elif ord(_A ) == KEYMAP["esc"]: a__ =get_raw_chars() if ord(_A ) == KEYMAP["mod_int"]: a__ =get_raw_chars() if ord(_A ) >= KEYMAP["arrow_begin"] - ARROW_KEY_FLAG and ord(_A ) <= KEYMAP["arrow_end"] - ARROW_KEY_FLAG: return chr(ord(_A ) + ARROW_KEY_FLAG ) else: return KEYMAP["undefined"] else: return get_raw_chars() else: if char in string.printable: return char else: return KEYMAP["undefined"]
188
1
"""simple docstring""" from maths.is_square_free import is_square_free from maths.prime_factors import prime_factors def lowercase ( _SCREAMING_SNAKE_CASE : int ): '''simple docstring''' _UpperCAmelCase = prime_factors(UpperCAmelCase__ ) if is_square_free(UpperCAmelCase__ ): return -1 if len(UpperCAmelCase__ ) % 2 else 1 return 0 if __name__ == "__main__": import doctest doctest.testmod()
371
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available __A : List[Any] = { "configuration_gpt_neo": ["GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoConfig", "GPTNeoOnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : str = [ "GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST", "GPTNeoForCausalLM", "GPTNeoForQuestionAnswering", "GPTNeoForSequenceClassification", "GPTNeoForTokenClassification", "GPTNeoModel", "GPTNeoPreTrainedModel", "load_tf_weights_in_gpt_neo", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : List[Any] = [ "FlaxGPTNeoForCausalLM", "FlaxGPTNeoModel", "FlaxGPTNeoPreTrainedModel", ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys __A : Dict = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
326
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mobilebert import MobileBertTokenizer __A : List[Any] = logging.get_logger(__name__) __A : List[Any] = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} __A : Optional[Any] = { '''vocab_file''': {'''mobilebert-uncased''': '''https://huggingface.co/google/mobilebert-uncased/resolve/main/vocab.txt'''}, '''tokenizer_file''': { '''mobilebert-uncased''': '''https://huggingface.co/google/mobilebert-uncased/resolve/main/tokenizer.json''' }, } __A : Optional[int] = {'''mobilebert-uncased''': 512} __A : Tuple = {} class __A ( lowerCAmelCase ): lowerCAmelCase_ : str = VOCAB_FILES_NAMES lowerCAmelCase_ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ : str = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase_ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase_ : Dict = MobileBertTokenizer def __init__( self : str , UpperCAmelCase_ : int=None , UpperCAmelCase_ : str=None , UpperCAmelCase_ : str=True , UpperCAmelCase_ : str="[UNK]" , UpperCAmelCase_ : List[Any]="[SEP]" , UpperCAmelCase_ : str="[PAD]" , UpperCAmelCase_ : str="[CLS]" , UpperCAmelCase_ : int="[MASK]" , UpperCAmelCase_ : List[Any]=True , UpperCAmelCase_ : Optional[Any]=None , **UpperCAmelCase_ : Tuple , ): super().__init__( UpperCAmelCase_ , tokenizer_file=UpperCAmelCase_ , do_lower_case=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , tokenize_chinese_chars=UpperCAmelCase_ , strip_accents=UpperCAmelCase_ , **UpperCAmelCase_ , ) lowerCAmelCase : Dict = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , UpperCAmelCase_ ) != do_lower_case or normalizer_state.get('strip_accents' , UpperCAmelCase_ ) != strip_accents or normalizer_state.get('handle_chinese_chars' , UpperCAmelCase_ ) != tokenize_chinese_chars ): lowerCAmelCase : Dict = getattr(UpperCAmelCase_ , normalizer_state.pop('type' ) ) lowerCAmelCase : Optional[Any] = do_lower_case lowerCAmelCase : List[Any] = strip_accents lowerCAmelCase : int = tokenize_chinese_chars lowerCAmelCase : str = normalizer_class(**UpperCAmelCase_ ) lowerCAmelCase : Optional[int] = do_lower_case def lowercase__ ( self : Any , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Dict=None ): lowerCAmelCase : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowercase__ ( self : int , UpperCAmelCase_ : List[int] , UpperCAmelCase_ : Optional[List[int]] = None ): lowerCAmelCase : Any = [self.sep_token_id] lowerCAmelCase : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowercase__ ( self : List[Any] , UpperCAmelCase_ : str , UpperCAmelCase_ : Optional[str] = None ): lowerCAmelCase : Dict = self._tokenizer.model.save(UpperCAmelCase_ , name=UpperCAmelCase_ ) return tuple(UpperCAmelCase_ )
138
def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> str: '''simple docstring''' return " ".join( ''.join(word[::-1] ) if len(_UpperCAmelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words('''Hey wollef sroirraw'''))
138
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging snake_case_ = logging.get_logger(__name__) if is_vision_available(): import PIL class SCREAMING_SNAKE_CASE__ (__snake_case ): __lowerCamelCase : Optional[int] = ["""pixel_values"""] def __init__( self , a = True , a = None , a = PILImageResampling.BICUBIC , a = True , a = None , a = True , a = 1 / 255 , a = True , a = None , a = None , a = True , **a , ): super().__init__(**a) lowercase__ : int = size if size is not None else {'shortest_edge': 224} lowercase__ : int = get_size_dict(a , default_to_square=a) lowercase__ : int = crop_size if crop_size is not None else {'height': 224, 'width': 224} lowercase__ : Any = get_size_dict(a , default_to_square=a , param_name='crop_size') lowercase__ : int = do_resize lowercase__ : List[str] = size lowercase__ : int = resample lowercase__ : Optional[Any] = do_center_crop lowercase__ : Tuple = crop_size lowercase__ : Any = do_rescale lowercase__ : Union[str, Any] = rescale_factor lowercase__ : Dict = do_normalize lowercase__ : List[Any] = image_mean if image_mean is not None else OPENAI_CLIP_MEAN lowercase__ : List[Any] = image_std if image_std is not None else OPENAI_CLIP_STD lowercase__ : str = do_convert_rgb def snake_case_ ( self , a , a , a = PILImageResampling.BICUBIC , a = None , **a , ): lowercase__ : Union[str, Any] = get_size_dict(a , default_to_square=a) if "shortest_edge" not in size: raise ValueError(f"""The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}""") lowercase__ : str = get_resize_output_image_size(a , size=size['shortest_edge'] , default_to_square=a) return resize(a , size=a , resample=a , data_format=a , **a) def snake_case_ ( self , a , a , a = None , **a , ): lowercase__ : Union[str, Any] = get_size_dict(a) if "height" not in size or "width" not in size: raise ValueError(f"""The `size` parameter must contain the keys (height, width). Got {size.keys()}""") return center_crop(a , size=(size['height'], size['width']) , data_format=a , **a) def snake_case_ ( self , a , a , a = None , **a , ): return rescale(a , scale=a , data_format=a , **a) def snake_case_ ( self , a , a , a , a = None , **a , ): return normalize(a , mean=a , std=a , data_format=a , **a) def snake_case_ ( self , a , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = ChannelDimension.FIRST , **a , ): lowercase__ : int = do_resize if do_resize is not None else self.do_resize lowercase__ : Optional[Any] = size if size is not None else self.size lowercase__ : str = get_size_dict(a , param_name='size' , default_to_square=a) lowercase__ : Optional[Any] = resample if resample is not None else self.resample lowercase__ : List[str] = do_center_crop if do_center_crop is not None else self.do_center_crop lowercase__ : Tuple = crop_size if crop_size is not None else self.crop_size lowercase__ : List[str] = get_size_dict(a , param_name='crop_size' , default_to_square=a) lowercase__ : Any = do_rescale if do_rescale is not None else self.do_rescale lowercase__ : Union[str, Any] = rescale_factor if rescale_factor is not None else self.rescale_factor lowercase__ : Tuple = do_normalize if do_normalize is not None else self.do_normalize lowercase__ : Dict = image_mean if image_mean is not None else self.image_mean lowercase__ : List[str] = image_std if image_std is not None else self.image_std lowercase__ : Dict = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb lowercase__ : Tuple = make_list_of_images(a) if not valid_images(a): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.') if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.') if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.') if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.') if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.') # PIL RGBA images are converted to RGB if do_convert_rgb: lowercase__ : Any = [convert_to_rgb(a) for image in images] # All transformations expect numpy arrays. lowercase__ : int = [to_numpy_array(a) for image in images] if do_resize: lowercase__ : Union[str, Any] = [self.resize(image=a , size=a , resample=a) for image in images] if do_center_crop: lowercase__ : List[Any] = [self.center_crop(image=a , size=a) for image in images] if do_rescale: lowercase__ : Optional[Any] = [self.rescale(image=a , scale=a) for image in images] if do_normalize: lowercase__ : Tuple = [self.normalize(image=a , mean=a , std=a) for image in images] lowercase__ : Tuple = [to_channel_dimension_format(a , a) for image in images] lowercase__ : Union[str, Any] = {'pixel_values': images} return BatchFeature(data=a , tensor_type=a)
357
from __future__ import annotations from collections.abc import Callable def snake_case__ ( SCREAMING_SNAKE_CASE_ : Callable[[int | float], int | float] , SCREAMING_SNAKE_CASE_ : int | float , SCREAMING_SNAKE_CASE_ : int | float , SCREAMING_SNAKE_CASE_ : int = 100 , ): '''simple docstring''' lowercase__ : Tuple = x_start lowercase__ : Tuple = fnc(SCREAMING_SNAKE_CASE_ ) lowercase__ : List[Any] = 0.0 for _ in range(SCREAMING_SNAKE_CASE_ ): # Approximates small segments of curve as linear and solve # for trapezoidal area lowercase__ : Any = (x_end - x_start) / steps + xa lowercase__ : Optional[Any] = fnc(SCREAMING_SNAKE_CASE_ ) area += abs(fxa + fxa ) * (xa - xa) / 2 # Increment step lowercase__ : Any = xa lowercase__ : str = fxa return area if __name__ == "__main__": def snake_case__ ( SCREAMING_SNAKE_CASE_ : Tuple ): '''simple docstring''' return x**3 + x**2 print('''f(x) = x^3 + x^2''') print('''The area between the curve, x = -5, x = 5 and the x axis is:''') snake_case_ = 10 while i <= 100_000: print(F'''with {i} steps: {trapezoidal_area(f, -5, 5, i)}''') i *= 10
216
0
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __lowerCAmelCase ( __magic_name__ , unittest.TestCase ): """simple docstring""" snake_case_ = KandinskyInpaintPipeline snake_case_ = ['''prompt''', '''image_embeds''', '''negative_image_embeds''', '''image''', '''mask_image'''] snake_case_ = [ '''prompt''', '''negative_prompt''', '''image_embeds''', '''negative_image_embeds''', '''image''', '''mask_image''', ] snake_case_ = [ '''generator''', '''height''', '''width''', '''latents''', '''guidance_scale''', '''negative_prompt''', '''num_inference_steps''', '''return_dict''', '''guidance_scale''', '''num_images_per_prompt''', '''output_type''', '''return_dict''', ] snake_case_ = False @property def lowercase_ ( self ) -> Any: '''simple docstring''' return 32 @property def lowercase_ ( self ) -> str: '''simple docstring''' return 32 @property def lowercase_ ( self ) -> Any: '''simple docstring''' return self.time_input_dim @property def lowercase_ ( self ) -> Tuple: '''simple docstring''' return self.time_input_dim * 4 @property def lowercase_ ( self ) -> List[Any]: '''simple docstring''' return 100 @property def lowercase_ ( self ) -> Optional[Any]: '''simple docstring''' __lowerCamelCase = XLMRobertaTokenizerFast.from_pretrained('YiYiXu/tiny-random-mclip-base' ) return tokenizer @property def lowercase_ ( self ) -> Any: '''simple docstring''' torch.manual_seed(0 ) __lowerCamelCase = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=1_005 , ) __lowerCamelCase = MultilingualCLIP(lowerCamelCase__ ) __lowerCamelCase = text_encoder.eval() return text_encoder @property def lowercase_ ( self ) -> Any: '''simple docstring''' torch.manual_seed(0 ) __lowerCamelCase = { 'in_channels': 9, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'text_image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'text_image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } __lowerCamelCase = UNetaDConditionModel(**lowerCamelCase__ ) return model @property def lowercase_ ( self ) -> List[Any]: '''simple docstring''' return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def lowercase_ ( self ) -> Dict: '''simple docstring''' torch.manual_seed(0 ) __lowerCamelCase = VQModel(**self.dummy_movq_kwargs ) return model def lowercase_ ( self ) -> Tuple: '''simple docstring''' __lowerCamelCase = self.dummy_text_encoder __lowerCamelCase = self.dummy_tokenizer __lowerCamelCase = self.dummy_unet __lowerCamelCase = self.dummy_movq __lowerCamelCase = DDIMScheduler( num_train_timesteps=1_000 , beta_schedule='linear' , beta_start=0.0_00_85 , beta_end=0.0_12 , clip_sample=lowerCamelCase__ , set_alpha_to_one=lowerCamelCase__ , steps_offset=1 , prediction_type='epsilon' , thresholding=lowerCamelCase__ , ) __lowerCamelCase = { 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__=0 ) -> List[str]: '''simple docstring''' __lowerCamelCase = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __lowerCamelCase = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1 ) ).to(lowerCamelCase__ ) # create init_image __lowerCamelCase = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __lowerCamelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCamelCase = Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('RGB' ).resize((256, 256) ) # create mask __lowerCamelCase = np.ones((64, 64) , dtype=np.floataa ) __lowerCamelCase = 0 if str(lowerCamelCase__ ).startswith('mps' ): __lowerCamelCase = torch.manual_seed(lowerCamelCase__ ) else: __lowerCamelCase = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __lowerCamelCase = { 'prompt': 'horse', 'image': init_image, 'mask_image': mask, 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 64, 'width': 64, 'num_inference_steps': 2, 'guidance_scale': 4.0, 'output_type': 'np', } return inputs def lowercase_ ( self ) -> int: '''simple docstring''' __lowerCamelCase = 'cpu' __lowerCamelCase = self.get_dummy_components() __lowerCamelCase = self.pipeline_class(**lowerCamelCase__ ) __lowerCamelCase = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __lowerCamelCase = pipe(**self.get_dummy_inputs(lowerCamelCase__ ) ) __lowerCamelCase = output.images __lowerCamelCase = pipe( **self.get_dummy_inputs(lowerCamelCase__ ) , return_dict=lowerCamelCase__ , )[0] __lowerCamelCase = image[0, -3:, -3:, -1] __lowerCamelCase = image_from_tuple[0, -3:, -3:, -1] print(f"""image.shape {image.shape}""" ) assert image.shape == (1, 64, 64, 3) __lowerCamelCase = np.array( [0.8_32_69_19, 0.73_79_04_67, 0.20_91_85_81, 0.9_30_96_12, 0.5_51_17_91, 0.43_71_33_28, 0.5_51_33_21, 0.49_92_29_34, 0.59_49_77_86] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f""" expected_slice {expected_slice}, but got {image_slice.flatten()}""" assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f""" expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}""" def lowercase_ ( self ) -> Optional[int]: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def lowercase_ ( self ) -> List[str]: '''simple docstring''' # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ ( self ) -> Dict: '''simple docstring''' __lowerCamelCase = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/kandinsky_inpaint_cat_with_hat_fp16.npy' ) __lowerCamelCase = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinsky/cat.png' ) __lowerCamelCase = np.ones((768, 768) , dtype=np.floataa ) __lowerCamelCase = 0 __lowerCamelCase = 'a hat' __lowerCamelCase = KandinskyPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-prior' , torch_dtype=torch.floataa ) pipe_prior.to(lowerCamelCase__ ) __lowerCamelCase = KandinskyInpaintPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-1-inpaint' , torch_dtype=torch.floataa ) __lowerCamelCase = pipeline.to(lowerCamelCase__ ) pipeline.set_progress_bar_config(disable=lowerCamelCase__ ) __lowerCamelCase = torch.Generator(device='cpu' ).manual_seed(0 ) __lowerCamelCase , __lowerCamelCase = pipe_prior( lowerCamelCase__ , generator=lowerCamelCase__ , num_inference_steps=5 , negative_prompt='' , ).to_tuple() __lowerCamelCase = pipeline( lowerCamelCase__ , image=lowerCamelCase__ , mask_image=lowerCamelCase__ , image_embeds=lowerCamelCase__ , negative_image_embeds=lowerCamelCase__ , generator=lowerCamelCase__ , num_inference_steps=100 , height=768 , width=768 , output_type='np' , ) __lowerCamelCase = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(lowerCamelCase__ , lowerCamelCase__ )
90
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class __a : '''simple docstring''' def __init__( self , _a , _a=13 , _a=2 , _a=24 , _a=16 , _a=True , _a=True , _a=32 , _a=5 , _a=4 , _a=37 , _a="gelu" , _a=0.1 , _a=0.1 , _a=10 , _a=0.02 , _a=None , _a=2 , _a=2 , ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = parent SCREAMING_SNAKE_CASE__ : Union[str, Any] = batch_size SCREAMING_SNAKE_CASE__ : Optional[Any] = patch_size SCREAMING_SNAKE_CASE__ : str = max_length SCREAMING_SNAKE_CASE__ : Optional[Any] = num_mel_bins SCREAMING_SNAKE_CASE__ : Optional[Any] = is_training SCREAMING_SNAKE_CASE__ : Union[str, Any] = use_labels SCREAMING_SNAKE_CASE__ : Optional[int] = hidden_size SCREAMING_SNAKE_CASE__ : Tuple = num_hidden_layers SCREAMING_SNAKE_CASE__ : List[str] = num_attention_heads SCREAMING_SNAKE_CASE__ : int = intermediate_size SCREAMING_SNAKE_CASE__ : Tuple = hidden_act SCREAMING_SNAKE_CASE__ : Optional[Any] = hidden_dropout_prob SCREAMING_SNAKE_CASE__ : List[str] = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ : List[str] = type_sequence_label_size SCREAMING_SNAKE_CASE__ : Tuple = initializer_range SCREAMING_SNAKE_CASE__ : List[Any] = scope SCREAMING_SNAKE_CASE__ : List[str] = frequency_stride SCREAMING_SNAKE_CASE__ : Union[str, Any] = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) SCREAMING_SNAKE_CASE__ : Optional[int] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 SCREAMING_SNAKE_CASE__ : Any = (self.max_length - self.patch_size) // self.time_stride + 1 SCREAMING_SNAKE_CASE__ : Union[str, Any] = frequency_out_dimension * time_out_dimension SCREAMING_SNAKE_CASE__ : Any = num_patches + 2 def _a ( self ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) SCREAMING_SNAKE_CASE__ : int = None if self.use_labels: SCREAMING_SNAKE_CASE__ : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.get_config() return config, input_values, labels def _a ( self ) -> Union[str, Any]: """simple docstring""" return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_a , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def _a ( self , _a , _a , _a ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = ASTModel(config=_a ) model.to(_a ) model.eval() SCREAMING_SNAKE_CASE__ : Union[str, Any] = model(_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = self.prepare_config_and_inputs() ( ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ) : List[str] = config_and_inputs SCREAMING_SNAKE_CASE__ : Union[str, Any] = {"""input_values""": input_values} return config, inputs_dict @require_torch class __a (UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase): '''simple docstring''' _SCREAMING_SNAKE_CASE :Optional[Any] = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) _SCREAMING_SNAKE_CASE :Dict = ( {"""audio-classification""": ASTForAudioClassification, """feature-extraction""": ASTModel} if is_torch_available() else {} ) _SCREAMING_SNAKE_CASE :Union[str, Any] = False _SCREAMING_SNAKE_CASE :Any = False _SCREAMING_SNAKE_CASE :Union[str, Any] = False _SCREAMING_SNAKE_CASE :Tuple = False def _a ( self , _a , _a , _a , _a , _a ) -> Dict: """simple docstring""" if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def _a ( self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = ASTModelTester(self ) SCREAMING_SNAKE_CASE__ : str = ConfigTester(self , config_class=_a , has_text_modality=_a , hidden_size=37 ) def _a ( self ) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""AST does not use inputs_embeds""" ) def _a ( self ) -> List[str]: """simple docstring""" pass def _a ( self ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ : str = model_class(_a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) SCREAMING_SNAKE_CASE__ : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_a , nn.Linear ) ) def _a ( self ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ : Any = model_class(_a ) SCREAMING_SNAKE_CASE__ : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE__ : Dict = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE__ : Dict = ["""input_values"""] self.assertListEqual(arg_names[:1] , _a ) def _a ( self ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) @slow def _a ( self ) -> Union[str, Any]: """simple docstring""" for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE__ : Optional[Any] = ASTModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def _lowercase ( ) -> int: SCREAMING_SNAKE_CASE__ : List[Any] = hf_hub_download( repo_id="""nielsr/audio-spectogram-transformer-checkpoint""" , filename="""sample_audio.flac""" , repo_type="""dataset""" ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Tuple = torchaudio.load(__lowerCAmelCase ) return audio, sampling_rate @require_torch @require_torchaudio class __a (unittest.TestCase): '''simple docstring''' @cached_property def _a ( self ) -> int: """simple docstring""" return ( ASTFeatureExtractor.from_pretrained("""MIT/ast-finetuned-audioset-10-10-0.4593""" ) if is_torchaudio_available() else None ) @slow def _a ( self ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = self.default_feature_extractor SCREAMING_SNAKE_CASE__ : Optional[Any] = ASTForAudioClassification.from_pretrained("""MIT/ast-finetuned-audioset-10-10-0.4593""" ).to(_a ) SCREAMING_SNAKE_CASE__ : Dict = self.default_feature_extractor SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Union[str, Any] = prepare_audio() SCREAMING_SNAKE_CASE__ : List[str] = audio.squeeze().numpy() SCREAMING_SNAKE_CASE__ : List[str] = feature_extractor(_a , sampling_rate=_a , return_tensors="""pt""" ).to(_a ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE__ : List[Any] = model(**_a ) # verify the logits SCREAMING_SNAKE_CASE__ : List[Any] = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , _a ) SCREAMING_SNAKE_CASE__ : Tuple = torch.tensor([-0.8_760, -7.0_042, -8.6_602] ).to(_a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _a , atol=1E-4 ) )
132
0
"""simple docstring""" from __future__ import annotations from math import pi, sqrt def _snake_case ( lowercase__ , lowercase__ ): if inductance <= 0: raise ValueError('Inductance cannot be 0 or negative' ) elif capacitance <= 0: raise ValueError('Capacitance cannot be 0 or negative' ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
12
"""simple docstring""" def _snake_case ( lowercase__ ): # if the collection is empty, returns empty if collection == []: return [] # get some information about the collection _lowerCamelCase : List[str] = len(lowercase__ ) _lowerCamelCase : List[str] = max(lowercase__ ) _lowerCamelCase : List[str] = min(lowercase__ ) # create the counting array _lowerCamelCase : List[Any] = coll_max + 1 - coll_min _lowerCamelCase : List[Any] = [0] * counting_arr_length # count how much a number appears in the collection for number in collection: counting_arr[number - coll_min] += 1 # sum each position with it's predecessors. now, counting_arr[i] tells # us how many elements <= i has in the collection for i in range(1 , lowercase__ ): _lowerCamelCase : Optional[int] = counting_arr[i] + counting_arr[i - 1] # create the output collection _lowerCamelCase : Dict = [0] * coll_len # place the elements in the output, respecting the original order (stable # sort) from end to begin, updating counting_arr for i in reversed(range(0 , lowercase__ ) ): _lowerCamelCase : Any = collection[i] counting_arr[collection[i] - coll_min] -= 1 return ordered def _snake_case ( lowercase__ ): return "".join([chr(lowercase__ ) for i in counting_sort([ord(lowercase__ ) for c in string] )] ) if __name__ == "__main__": # Test string sort assert counting_sort_string("""thisisthestring""") == "eghhiiinrsssttt" lowercase__ = input("""Enter numbers separated by a comma:\n""").strip() lowercase__ = [int(item) for item in user_input.split(""",""")] print(counting_sort(unsorted))
12
1
import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowercase__ : Dict = "▁" lowercase__ : List[Any] = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class UpperCAmelCase ( UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowerCAmelCase_ = BigBirdTokenizer lowerCAmelCase_ = BigBirdTokenizerFast lowerCAmelCase_ = True lowerCAmelCase_ = True def snake_case__ ( self : List[Any] ): """simple docstring""" super().setUp() snake_case_ = self.tokenizer_class(__lowercase , keep_accents=__lowercase ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case__ ( self : Tuple ): """simple docstring""" snake_case_ = "<s>" snake_case_ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowercase ) , __lowercase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowercase ) , __lowercase ) def snake_case__ ( self : List[Any] ): """simple docstring""" snake_case_ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "[MASK]" ) self.assertEqual(len(__lowercase ) , 10_04 ) def snake_case__ ( self : List[Any] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 10_00 ) def snake_case__ ( self : List[Any] ): """simple docstring""" if not self.test_rust_tokenizer: return snake_case_ = self.get_tokenizer() snake_case_ = self.get_rust_tokenizer() snake_case_ = "I was born in 92000, and this is falsé." snake_case_ = tokenizer.tokenize(__lowercase ) snake_case_ = rust_tokenizer.tokenize(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) snake_case_ = tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) snake_case_ = rust_tokenizer.encode(__lowercase , add_special_tokens=__lowercase ) self.assertListEqual(__lowercase , __lowercase ) snake_case_ = self.get_rust_tokenizer() snake_case_ = tokenizer.encode(__lowercase ) snake_case_ = rust_tokenizer.encode(__lowercase ) self.assertListEqual(__lowercase , __lowercase ) def snake_case__ ( self : Optional[Any] ): """simple docstring""" snake_case_ = BigBirdTokenizer(__lowercase , keep_accents=__lowercase ) snake_case_ = tokenizer.tokenize("This is a test" ) self.assertListEqual(__lowercase , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowercase ) , [2_85, 46, 10, 1_70, 3_82] , ) snake_case_ = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( __lowercase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) snake_case_ = tokenizer.convert_tokens_to_ids(__lowercase ) self.assertListEqual( __lowercase , [8, 21, 84, 55, 24, 19, 7, 0, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 0, 4] , ) snake_case_ = tokenizer.convert_ids_to_tokens(__lowercase ) self.assertListEqual( __lowercase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) @cached_property def snake_case__ ( self : Optional[int] ): """simple docstring""" return BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) @slow def snake_case__ ( self : Optional[Any] ): """simple docstring""" snake_case_ = "Hello World!" snake_case_ = [65, 1_85_36, 22_60, 1_01, 66] self.assertListEqual(__lowercase , self.big_tokenizer.encode(__lowercase ) ) @slow def snake_case__ ( self : str ): """simple docstring""" snake_case_ = ( "This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) \" [ ] ! : - . Also we will" " add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth" ) # fmt: off snake_case_ = [65, 8_71, 4_19, 3_58, 9_46, 9_91, 25_21, 4_52, 3_58, 13_57, 3_87, 77_51, 35_36, 1_12, 9_85, 4_56, 1_26, 8_65, 9_38, 54_00, 57_34, 4_58, 13_68, 4_67, 7_86, 24_62, 52_46, 11_59, 6_33, 8_65, 45_19, 4_57, 5_82, 8_52, 25_57, 4_27, 9_16, 5_08, 4_05, 3_43_24, 4_97, 3_91, 4_08, 1_13_42, 12_44, 3_85, 1_00, 9_38, 9_85, 4_56, 5_74, 3_62, 1_25_97, 32_00, 31_29, 11_72, 66] # noqa: E231 # fmt: on self.assertListEqual(__lowercase , self.big_tokenizer.encode(__lowercase ) ) @require_torch @slow def snake_case__ ( self : Optional[int] ): """simple docstring""" import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence snake_case_ = list(self.big_tokenizer.get_vocab().keys() )[:10] snake_case_ = " ".join(__lowercase ) snake_case_ = self.big_tokenizer.encode_plus(__lowercase , return_tensors="pt" , return_token_type_ids=__lowercase ) snake_case_ = self.big_tokenizer.batch_encode_plus( [sequence + " " + sequence] , return_tensors="pt" , return_token_type_ids=__lowercase ) snake_case_ = BigBirdConfig(attention_type="original_full" ) snake_case_ = BigBirdModel(__lowercase ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**__lowercase ) model(**__lowercase ) @slow def snake_case__ ( self : int ): """simple docstring""" snake_case_ = BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base" ) snake_case_ = tokenizer.decode(tokenizer("Paris is the [MASK]." ).input_ids ) self.assertTrue(decoded_text == "[CLS] Paris is the[MASK].[SEP]" ) @slow def snake_case__ ( self : Any ): """simple docstring""" snake_case_ = {"input_ids": [[65, 3_92_86, 4_58, 3_63_35, 20_01, 4_56, 1_30_73, 1_32_66, 4_55, 1_13, 77_46, 17_41, 1_11_57, 3_91, 1_30_73, 1_32_66, 4_55, 1_13, 39_67, 3_54_12, 1_13, 49_36, 1_09, 38_70, 23_77, 1_13, 3_00_84, 4_57_20, 4_58, 1_34, 1_74_96, 1_12, 5_03, 1_16_72, 1_13, 1_18, 1_12, 56_65, 1_33_47, 3_86_87, 1_12, 14_96, 3_13_89, 1_12, 32_68, 4_72_64, 1_34, 9_62, 1_12, 1_63_77, 80_35, 2_31_30, 4_30, 1_21_69, 1_55_18, 2_85_92, 4_58, 1_46, 4_16_97, 1_09, 3_91, 1_21_69, 1_55_18, 1_66_89, 4_58, 1_46, 4_13_58, 1_09, 4_52, 7_26, 40_34, 1_11, 7_63, 3_54_12, 50_82, 3_88, 19_03, 1_11, 90_51, 3_91, 28_70, 4_89_18, 19_00, 11_23, 5_50, 9_98, 1_12, 95_86, 1_59_85, 4_55, 3_91, 4_10, 2_29_55, 3_76_36, 1_14, 66], [65, 4_48, 1_74_96, 4_19, 36_63, 3_85, 7_63, 1_13, 2_75_33, 28_70, 32_83, 1_30_43, 16_39, 2_47_13, 5_23, 6_56, 2_40_13, 1_85_50, 25_21, 5_17, 2_70_14, 2_12_44, 4_20, 12_12, 14_65, 3_91, 9_27, 48_33, 3_88, 5_78, 1_17_86, 1_14, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 4_84, 21_69, 76_87, 2_19_32, 1_81_46, 7_26, 3_63, 1_70_32, 33_91, 1_14, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowercase , model_name="google/bigbird-roberta-base" , revision="215c99f1600e06f83acce68422f2035b2b5c3510" , )
187
def lowerCamelCase__ ( _A , _A ): '''simple docstring''' _enforce_args(_A , _A ) if n == 0: return 0 snake_case_ = float("-inf" ) for i in range(1 , n + 1 ): snake_case_ = max( _A , prices[i - 1] + naive_cut_rod_recursive(n - i , _A ) ) return max_revue def lowerCamelCase__ ( _A , _A ): '''simple docstring''' _enforce_args(_A , _A ) snake_case_ = [float("-inf" ) for _ in range(n + 1 )] return _top_down_cut_rod_recursive(_A , _A , _A ) def lowerCamelCase__ ( _A , _A , _A ): '''simple docstring''' if max_rev[n] >= 0: return max_rev[n] elif n == 0: return 0 else: snake_case_ = float("-inf" ) for i in range(1 , n + 1 ): snake_case_ = max( _A , prices[i - 1] + _top_down_cut_rod_recursive(n - i , _A , _A ) , ) snake_case_ = max_revenue return max_rev[n] def lowerCamelCase__ ( _A , _A ): '''simple docstring''' _enforce_args(_A , _A ) # length(max_rev) = n + 1, to accommodate for the revenue obtainable from a rod of # length 0. snake_case_ = [float("-inf" ) for _ in range(n + 1 )] snake_case_ = 0 for i in range(1 , n + 1 ): snake_case_ = max_rev[i] for j in range(1 , i + 1 ): snake_case_ = max(_A , prices[j - 1] + max_rev[i - j] ) snake_case_ = max_revenue_i return max_rev[n] def lowerCamelCase__ ( _A , _A ): '''simple docstring''' if n < 0: snake_case_ = f"n must be greater than or equal to 0. Got n = {n}" raise ValueError(_A ) if n > len(_A ): snake_case_ = ( "Each integral piece of rod must have a corresponding price. " f"Got n = {n} but length of prices = {len(_A )}" ) raise ValueError(_A ) def lowerCamelCase__ ( ): '''simple docstring''' snake_case_ = [6, 10, 12, 15, 20, 23] snake_case_ = len(_A ) # the best revenue comes from cutting the rod into 6 pieces, each # of length 1 resulting in a revenue of 6 * 6 = 36. snake_case_ = 36 snake_case_ = top_down_cut_rod(_A , _A ) snake_case_ = bottom_up_cut_rod(_A , _A ) snake_case_ = naive_cut_rod_recursive(_A , _A ) assert expected_max_revenue == max_rev_top_down assert max_rev_top_down == max_rev_bottom_up assert max_rev_bottom_up == max_rev_naive if __name__ == "__main__": main()
187
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __magic_name__ = logging.get_logger(__name__) __magic_name__ = { "camembert-base": "https://huggingface.co/camembert-base/resolve/main/config.json", "umberto-commoncrawl-cased-v1": ( "https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json" ), "umberto-wikipedia-uncased-v1": ( "https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json" ), } class lowercase ( A__ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = """camembert""" def __init__( self , _snake_case=3_0522 , _snake_case=768 , _snake_case=12 , _snake_case=12 , _snake_case=3072 , _snake_case="gelu" , _snake_case=0.1 , _snake_case=0.1 , _snake_case=512 , _snake_case=2 , _snake_case=0.02 , _snake_case=1e-12 , _snake_case=1 , _snake_case=0 , _snake_case=2 , _snake_case="absolute" , _snake_case=True , _snake_case=None , **_snake_case , ) -> Any: """simple docstring""" super().__init__(pad_token_id=_snake_case , bos_token_id=_snake_case , eos_token_id=_snake_case , **_snake_case ) UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_act UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps UpperCAmelCase = position_embedding_type UpperCAmelCase = use_cache UpperCAmelCase = classifier_dropout class lowercase ( A__ ): '''simple docstring''' @property def snake_case_ ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": UpperCAmelCase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: UpperCAmelCase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
152
import argparse import torch from ...utils import logging from . import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert logging.set_verbosity_info() def _lowerCAmelCase ( A__: str , A__: List[str] , A__: str ): '''simple docstring''' UpperCAmelCase = AlbertConfig.from_json_file(A__ ) print(F"""Building PyTorch model from configuration: {config}""" ) UpperCAmelCase = AlbertForPreTraining(A__ ) # Load weights from tf checkpoint load_tf_weights_in_albert(A__ , A__ , A__ ) # Save pytorch-model print(F"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , A__ ) if __name__ == "__main__": __magic_name__ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--albert_config_file", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained ALBERT model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) __magic_name__ = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.albert_config_file, args.pytorch_dump_path)
152
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { """shi-labs/dinat-mini-in1k-224""": """https://huggingface.co/shi-labs/dinat-mini-in1k-224/resolve/main/config.json""", # See all Dinat models at https://huggingface.co/models?filter=dinat } class lowerCAmelCase_ ( UpperCAmelCase__ ,UpperCAmelCase__ ): __lowerCamelCase : Optional[int] = "dinat" __lowerCamelCase : Tuple = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self , _lowerCAmelCase=4 , _lowerCAmelCase=3 , _lowerCAmelCase=64 , _lowerCAmelCase=[3, 4, 6, 5] , _lowerCAmelCase=[2, 4, 8, 16] , _lowerCAmelCase=7 , _lowerCAmelCase=[[1, 8, 1], [1, 4, 1, 4], [1, 2, 1, 2, 1, 2], [1, 1, 1, 1, 1]] , _lowerCAmelCase=3.0 , _lowerCAmelCase=True , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.1 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.02 , _lowerCAmelCase=1E-5 , _lowerCAmelCase=0.0 , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase , ) -> Any: super().__init__(**_a ) _lowerCAmelCase = patch_size _lowerCAmelCase = num_channels _lowerCAmelCase = embed_dim _lowerCAmelCase = depths _lowerCAmelCase = len(_a ) _lowerCAmelCase = num_heads _lowerCAmelCase = kernel_size _lowerCAmelCase = dilations _lowerCAmelCase = mlp_ratio _lowerCAmelCase = qkv_bias _lowerCAmelCase = hidden_dropout_prob _lowerCAmelCase = attention_probs_dropout_prob _lowerCAmelCase = drop_path_rate _lowerCAmelCase = hidden_act _lowerCAmelCase = layer_norm_eps _lowerCAmelCase = initializer_range # we set the hidden_size attribute in order to make Dinat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model _lowerCAmelCase = int(embed_dim * 2 ** (len(_a ) - 1) ) _lowerCAmelCase = layer_scale_init_value _lowerCAmelCase = ["stem"] + [f'''stage{idx}''' for idx in range(1 , len(_a ) + 1 )] _lowerCAmelCase , _lowerCAmelCase = get_aligned_output_features_output_indices( out_features=_a , out_indices=_a , stage_names=self.stage_names )
158
"""simple docstring""" def a__ ( snake_case__ , snake_case__ = False ) -> str: if not isinstance(snake_case__ , snake_case__ ): lowerCamelCase = F'Expected string as input, found {type(snake_case__ )}' raise ValueError(snake_case__ ) if not isinstance(snake_case__ , snake_case__ ): lowerCamelCase = F'Expected boolean as use_pascal parameter, found {type(snake_case__ )}' raise ValueError(snake_case__ ) lowerCamelCase = input_str.split("""_""" ) lowerCamelCase = 0 if use_pascal else 1 lowerCamelCase = words[start_index:] lowerCamelCase = [word[0].upper() + word[1:] for word in words_to_capitalize] lowerCamelCase = """""" if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
291
0
import os import tempfile import unittest from transformers import NezhaConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, ) from transformers.models.nezha.modeling_nezha import NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST class __UpperCAmelCase : '''simple docstring''' def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=13 , _SCREAMING_SNAKE_CASE=7 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=99 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=5 , _SCREAMING_SNAKE_CASE=4 , _SCREAMING_SNAKE_CASE=37 , _SCREAMING_SNAKE_CASE="gelu" , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=128 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=16 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=4 , _SCREAMING_SNAKE_CASE=None , ) -> str: A_ = parent A_ = batch_size A_ = seq_length A_ = is_training A_ = use_input_mask A_ = use_token_type_ids A_ = use_labels A_ = vocab_size A_ = hidden_size A_ = num_hidden_layers A_ = num_attention_heads A_ = intermediate_size A_ = hidden_act A_ = hidden_dropout_prob A_ = attention_probs_dropout_prob A_ = max_position_embeddings A_ = type_vocab_size A_ = type_sequence_label_size A_ = initializer_range A_ = num_labels A_ = num_choices A_ = scope def __A ( self ) -> int: A_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A_ = None if self.use_input_mask: A_ = random_attention_mask([self.batch_size, self.seq_length] ) A_ = None if self.use_token_type_ids: A_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A_ = None A_ = None A_ = None if self.use_labels: A_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A_ = ids_tensor([self.batch_size] , self.num_choices ) A_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __A ( self ) -> Dict: return NezhaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , ) def __A ( self ) -> Tuple: ( ( A_ ) ,( A_ ) ,( A_ ) ,( A_ ) ,( A_ ) ,( A_ ) ,( A_ ) , ) = self.prepare_config_and_inputs() A_ = True A_ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) A_ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def __A ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Optional[int]: A_ = NezhaModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() A_ = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE ) A_ = model(_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE ) A_ = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def __A ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) -> Tuple: A_ = True A_ = NezhaModel(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() A_ = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , encoder_hidden_states=_SCREAMING_SNAKE_CASE , encoder_attention_mask=_SCREAMING_SNAKE_CASE , ) A_ = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , encoder_hidden_states=_SCREAMING_SNAKE_CASE , ) A_ = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def __A ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[str]: A_ = NezhaForMaskedLM(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() A_ = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __A ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Tuple: A_ = NezhaForNextSentencePrediction(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() A_ = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def __A ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> int: A_ = NezhaForPreTraining(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() A_ = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE , next_sentence_label=_SCREAMING_SNAKE_CASE , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def __A ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Dict: A_ = NezhaForQuestionAnswering(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() A_ = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , start_positions=_SCREAMING_SNAKE_CASE , end_positions=_SCREAMING_SNAKE_CASE , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __A ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> int: A_ = self.num_labels A_ = NezhaForSequenceClassification(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() A_ = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __A ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Dict: A_ = self.num_labels A_ = NezhaForTokenClassification(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() A_ = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __A ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> Tuple: A_ = self.num_choices A_ = NezhaForMultipleChoice(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() A_ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A_ = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A_ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A_ = model( _SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , token_type_ids=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __A ( self ) -> List[str]: A_ = self.prepare_config_and_inputs() ( ( A_ ) ,( A_ ) ,( A_ ) ,( A_ ) ,( A_ ) ,( A_ ) ,( A_ ) , ) = config_and_inputs A_ = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __UpperCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' __lowercase : Optional[int] = ( ( NezhaModel, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, ) if is_torch_available() else () ) __lowercase : int = ( { "feature-extraction": NezhaModel, "fill-mask": NezhaForMaskedLM, "question-answering": NezhaForQuestionAnswering, "text-classification": NezhaForSequenceClassification, "token-classification": NezhaForTokenClassification, "zero-shot": NezhaForSequenceClassification, } if is_torch_available() else {} ) __lowercase : Tuple = True def __A ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ) -> Optional[Any]: A_ = super()._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_labels=_SCREAMING_SNAKE_CASE ) if return_labels: if model_class in get_values(_SCREAMING_SNAKE_CASE ): A_ = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=_SCREAMING_SNAKE_CASE ) A_ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_SCREAMING_SNAKE_CASE ) return inputs_dict def __A ( self ) -> Dict: A_ = NezhaModelTester(self ) A_ = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , hidden_size=37 ) def __A ( self ) -> List[str]: self.config_tester.run_common_tests() def __A ( self ) -> str: A_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) def __A ( self ) -> Union[str, Any]: A_ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*_SCREAMING_SNAKE_CASE ) def __A ( self ) -> List[Any]: # This regression test was failing with PyTorch < 1.3 ( ( A_ ) ,( A_ ) ,( A_ ) ,( A_ ) ,( A_ ) ,( A_ ) ,( A_ ) ,( A_ ) ,( A_ ) , ) = self.model_tester.prepare_config_and_inputs_for_decoder() A_ = None self.model_tester.create_and_check_model_as_decoder( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , ) def __A ( self ) -> Dict: A_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_SCREAMING_SNAKE_CASE ) def __A ( self ) -> List[Any]: A_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*_SCREAMING_SNAKE_CASE ) def __A ( self ) -> List[Any]: A_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_next_sequence_prediction(*_SCREAMING_SNAKE_CASE ) def __A ( self ) -> Optional[Any]: A_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*_SCREAMING_SNAKE_CASE ) def __A ( self ) -> Union[str, Any]: A_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_SCREAMING_SNAKE_CASE ) def __A ( self ) -> Tuple: A_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_SCREAMING_SNAKE_CASE ) def __A ( self ) -> Any: A_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_SCREAMING_SNAKE_CASE ) @slow def __A ( self ) -> Union[str, Any]: for model_name in NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ = NezhaModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) @slow @require_torch_gpu def __A ( self ) -> Any: A_ ,A_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # NezhaForMultipleChoice behaves incorrectly in JIT environments. if model_class == NezhaForMultipleChoice: return A_ = True A_ = model_class(config=_SCREAMING_SNAKE_CASE ) A_ = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) A_ = torch.jit.trace( _SCREAMING_SNAKE_CASE , (inputs_dict['''input_ids'''].to('''cpu''' ), inputs_dict['''attention_mask'''].to('''cpu''' )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(_SCREAMING_SNAKE_CASE , os.path.join(_SCREAMING_SNAKE_CASE , '''bert.pt''' ) ) A_ = torch.jit.load(os.path.join(_SCREAMING_SNAKE_CASE , '''bert.pt''' ) , map_location=_SCREAMING_SNAKE_CASE ) loaded(inputs_dict['''input_ids'''].to(_SCREAMING_SNAKE_CASE ) , inputs_dict['''attention_mask'''].to(_SCREAMING_SNAKE_CASE ) ) @require_torch class __UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def __A ( self ) -> Dict: A_ = NezhaModel.from_pretrained('''sijunhe/nezha-cn-base''' ) A_ = torch.tensor([[0, 1, 2, 3, 4, 5]] ) A_ = torch.tensor([[0, 1, 1, 1, 1, 1]] ) with torch.no_grad(): A_ = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE )[0] A_ = torch.Size((1, 6, 768) ) self.assertEqual(output.shape , _SCREAMING_SNAKE_CASE ) A_ = torch.tensor([[[0.0_685, 0.2_441, 0.1_102], [0.0_600, 0.1_906, 0.1_349], [0.0_221, 0.0_819, 0.0_586]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _SCREAMING_SNAKE_CASE , atol=1E-4 ) ) @slow def __A ( self ) -> Optional[Any]: A_ = NezhaForMaskedLM.from_pretrained('''sijunhe/nezha-cn-base''' ) A_ = torch.tensor([[0, 1, 2, 3, 4, 5]] ) A_ = torch.tensor([[1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): A_ = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE )[0] A_ = torch.Size((1, 6, 2_1128) ) self.assertEqual(output.shape , _SCREAMING_SNAKE_CASE ) A_ = torch.tensor( [[-2.7_939, -1.7_902, -2.2_189], [-2.8_585, -1.8_908, -2.3_723], [-2.6_499, -1.7_750, -2.2_558]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _SCREAMING_SNAKE_CASE , atol=1E-4 ) )
354
'''simple docstring''' import math def _UpperCAmelCase ( _UpperCamelCase : float, _UpperCamelCase : float ) -> float: if initial_intensity < 0: raise ValueError('''The value of intensity cannot be negative''' ) # handling of negative values of initial intensity if angle < 0 or angle > 3_60: raise ValueError('''In Malus Law, the angle is in the range 0-360 degrees''' ) # handling of values out of allowed range return initial_intensity * (math.cos(math.radians(_UpperCamelCase ) ) ** 2) if __name__ == "__main__": import doctest doctest.testmod(name='malus_law')
18
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = { '''google/vivit-b-16x2-kinetics400''': ( '''https://huggingface.co/google/vivit-b-16x2-kinetics400/resolve/main/config.json''' ), # See all Vivit models at https://huggingface.co/models?filter=vivit } class snake_case_ ( __A ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = "vivit" def __init__( self : int , _UpperCamelCase : Union[str, Any]=2_2_4 , _UpperCamelCase : List[str]=3_2 , _UpperCamelCase : Optional[Any]=[2, 1_6, 1_6] , _UpperCamelCase : Dict=3 , _UpperCamelCase : int=7_6_8 , _UpperCamelCase : Any=1_2 , _UpperCamelCase : Dict=1_2 , _UpperCamelCase : Union[str, Any]=3_0_7_2 , _UpperCamelCase : List[Any]="gelu_fast" , _UpperCamelCase : str=0.0 , _UpperCamelCase : Dict=0.0 , _UpperCamelCase : str=0.02 , _UpperCamelCase : Optional[int]=1e-06 , _UpperCamelCase : Any=True , **_UpperCamelCase : List[Any] , ) ->int: snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = initializer_range snake_case_ = layer_norm_eps snake_case_ = image_size snake_case_ = num_frames snake_case_ = tubelet_size snake_case_ = num_channels snake_case_ = qkv_bias super().__init__(**_UpperCamelCase )
8
def __UpperCamelCase ( _A : int ) ->int: """simple docstring""" assert ( isinstance(_A , _A ) and number_of_steps > 0 ), f'number_of_steps needs to be positive integer, your input {number_of_steps}' if number_of_steps == 1: return 1 lowerCamelCase_ , lowerCamelCase_ =1, 1 for _ in range(number_of_steps - 1 ): lowerCamelCase_ , lowerCamelCase_ =current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
154
0
"""simple docstring""" import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore snake_case = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" snake_case = [file for file in filepaths if file != file.lower()] if upper_files: print(F"""{len(upper_files)} files contain uppercase characters:""") print("""\n""".join(upper_files) + """\n""") snake_case = [file for file in filepaths if """ """ in file] if space_files: print(F"""{len(space_files)} files contain space characters:""") print("""\n""".join(space_files) + """\n""") snake_case = [file for file in filepaths if """-""" in file] if hyphen_files: print(F"""{len(hyphen_files)} files contain hyphen characters:""") print("""\n""".join(hyphen_files) + """\n""") snake_case = [file for file in filepaths if os.sep not in file] if nodir_files: print(F"""{len(nodir_files)} files are not in a directory:""") print("""\n""".join(nodir_files) + """\n""") snake_case = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
352
def lowerCamelCase__ ( lowercase ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = n ** (1 / 3) return (val * val * val) == n if __name__ == "__main__": print(perfect_cube(27)) print(perfect_cube(4))
319
0
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' @slow def _snake_case ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' A: int = XLMRobertaModel.from_pretrained('''xlm-roberta-base''' ) A: Union[str, Any] = torch.tensor([[0, 5_81, 1_02_69, 83, 9_99_42, 1_36, 6_07_42, 23, 70, 8_05_83, 1_82_76, 2]] ) # The dog is cute and lives in the garden house A: List[Any] = torch.Size((1, 12, 7_68) ) # batch_size, sequence_length, embedding_vector_dim A: Tuple = torch.tensor( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): A: Tuple = model(SCREAMING_SNAKE_CASE_ )['''last_hidden_state'''].detach() self.assertEqual(output.shape , SCREAMING_SNAKE_CASE_ ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , SCREAMING_SNAKE_CASE_ , atol=1E-3 ) ) @slow def _snake_case ( self : Dict ) -> Any: '''simple docstring''' A: Dict = XLMRobertaModel.from_pretrained('''xlm-roberta-large''' ) A: Any = torch.tensor([[0, 5_81, 1_02_69, 83, 9_99_42, 1_36, 6_07_42, 23, 70, 8_05_83, 1_82_76, 2]] ) # The dog is cute and lives in the garden house A: int = torch.Size((1, 12, 10_24) ) # batch_size, sequence_length, embedding_vector_dim A: str = torch.tensor( [[-0.0699, -0.0318, 0.0705, -0.1241, 0.0999, -0.0520, 0.1004, -0.1838, -0.4704, 0.1437, 0.0821, 0.0126]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.large') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): A: List[Any] = model(SCREAMING_SNAKE_CASE_ )['''last_hidden_state'''].detach() self.assertEqual(output.shape , SCREAMING_SNAKE_CASE_ ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , SCREAMING_SNAKE_CASE_ , atol=1E-3 ) )
319
'''simple docstring''' from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = Dict[str, Any] UpperCamelCase = List[Prediction] @add_end_docstrings(UpperCAmelCase_ ) class lowerCAmelCase_ ( UpperCAmelCase_ ): '''simple docstring''' def __init__( self : Union[str, Any] , *SCREAMING_SNAKE_CASE_ : Union[str, Any] , **SCREAMING_SNAKE_CASE_ : List[str] ) -> int: '''simple docstring''' super().__init__(*SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) if self.framework == "tf": raise ValueError(f"""The {self.__class__} is only available in PyTorch.""" ) requires_backends(self , '''vision''' ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def _snake_case ( self : int , **SCREAMING_SNAKE_CASE_ : Optional[int] ) -> Optional[Any]: '''simple docstring''' A: Any = {} if "threshold" in kwargs: A: List[Any] = kwargs['''threshold'''] return {}, {}, postprocess_kwargs def __call__( self : str , *SCREAMING_SNAKE_CASE_ : str , **SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> Union[Predictions, List[Prediction]]: '''simple docstring''' return super().__call__(*SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def _snake_case ( self : Optional[int] , SCREAMING_SNAKE_CASE_ : Optional[int] ) -> Union[str, Any]: '''simple docstring''' A: int = load_image(SCREAMING_SNAKE_CASE_ ) A: Optional[Any] = torch.IntTensor([[image.height, image.width]] ) A: Union[str, Any] = self.image_processor(images=[image] , return_tensors='''pt''' ) if self.tokenizer is not None: A: int = self.tokenizer(text=inputs['''words'''] , boxes=inputs['''boxes'''] , return_tensors='''pt''' ) A: Any = target_size return inputs def _snake_case ( self : int , SCREAMING_SNAKE_CASE_ : str ) -> List[Any]: '''simple docstring''' A: Tuple = model_inputs.pop('''target_size''' ) A: Tuple = self.model(**SCREAMING_SNAKE_CASE_ ) A: List[str] = outputs.__class__({'''target_size''': target_size, **outputs} ) if self.tokenizer is not None: A: Dict = model_inputs['''bbox'''] return model_outputs def _snake_case ( self : Any , SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : str=0.9 ) -> Union[str, Any]: '''simple docstring''' A: List[Any] = model_outputs['''target_size'''] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. A , A: Union[str, Any] = target_size[0].tolist() def unnormalize(SCREAMING_SNAKE_CASE_ : str ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 10_00), (height * bbox[1] / 10_00), (width * bbox[2] / 10_00), (height * bbox[3] / 10_00), ] ) ) A , A: Dict = model_outputs['''logits'''].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) A: List[str] = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] A: List[str] = [unnormalize(SCREAMING_SNAKE_CASE_ ) for bbox in model_outputs['''bbox'''].squeeze(0 )] A: Dict = ['''score''', '''label''', '''box'''] A: Optional[int] = [dict(zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) for vals in zip(scores.tolist() , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel A: Any = self.image_processor.post_process_object_detection(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) A: List[str] = raw_annotations[0] A: List[Any] = raw_annotation['''scores'''] A: List[Any] = raw_annotation['''labels'''] A: int = raw_annotation['''boxes'''] A: Any = scores.tolist() A: List[Any] = [self.model.config.idalabel[label.item()] for label in labels] A: List[Any] = [self._get_bounding_box(SCREAMING_SNAKE_CASE_ ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] A: Tuple = ['''score''', '''label''', '''box'''] A: str = [ dict(zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) for vals in zip(raw_annotation['''scores'''] , raw_annotation['''labels'''] , raw_annotation['''boxes'''] ) ] return annotation def _snake_case ( self : Tuple , SCREAMING_SNAKE_CASE_ : "torch.Tensor" ) -> Dict[str, int]: '''simple docstring''' if self.framework != "pt": raise ValueError('''The ObjectDetectionPipeline is only available in PyTorch.''' ) A , A , A , A: str = box.int().tolist() A: str = { '''xmin''': xmin, '''ymin''': ymin, '''xmax''': xmax, '''ymax''': ymax, } return bbox
319
1
'''simple docstring''' import argparse import glob import logging import os from argparse import Namespace from importlib import import_module import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch.nn import CrossEntropyLoss from torch.utils.data import DataLoader, TensorDataset from utils_ner import TokenClassificationTask __lowerCamelCase = logging.getLogger(__name__) class A__ ( _snake_case ): lowercase = "token-classification" def __init__( self , UpperCamelCase__ ) -> Any: '''simple docstring''' if type(UpperCamelCase__ ) == dict: A_ = Namespace(**UpperCamelCase__ ) A_ = import_module("""tasks""" ) try: A_ = getattr(UpperCamelCase__ , hparams.task_type ) A_ = token_classification_task_clazz() except AttributeError: raise ValueError( f'''Task {hparams.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. ''' f'''Available tasks classes are: {TokenClassificationTask.__subclasses__()}''' ) A_ = self.token_classification_task.get_labels(hparams.labels ) A_ = CrossEntropyLoss().ignore_index super().__init__(UpperCamelCase__ , len(self.labels ) , self.mode ) def snake_case_ ( self , **UpperCamelCase__ ) -> int: '''simple docstring''' return self.model(**UpperCamelCase__ ) def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ ) -> List[Any]: '''simple docstring''' A_ = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type != "distilbert": A_ = ( batch[2] if self.config.model_type in ["""bert""", """xlnet"""] else None ) # XLM and RoBERTa don"t use token_type_ids A_ = self(**UpperCamelCase__ ) A_ = outputs[0] # tensorboard_logs = {"loss": loss, "rate": self.lr_scheduler.get_last_lr()[-1]} return {"loss": loss} def snake_case_ ( self ) -> Any: '''simple docstring''' A_ = self.hparams for mode in ["train", "dev", "test"]: A_ = self._feature_file(UpperCamelCase__ ) if os.path.exists(UpperCamelCase__ ) and not args.overwrite_cache: logger.info("""Loading features from cached file %s""" , UpperCamelCase__ ) A_ = torch.load(UpperCamelCase__ ) else: logger.info("""Creating features from dataset file at %s""" , args.data_dir ) A_ = self.token_classification_task.read_examples_from_file(args.data_dir , UpperCamelCase__ ) A_ = self.token_classification_task.convert_examples_to_features( UpperCamelCase__ , self.labels , args.max_seq_length , self.tokenizer , cls_token_at_end=bool(self.config.model_type in ["""xlnet"""] ) , cls_token=self.tokenizer.cls_token , cls_token_segment_id=2 if self.config.model_type in ["""xlnet"""] else 0 , sep_token=self.tokenizer.sep_token , sep_token_extra=UpperCamelCase__ , pad_on_left=bool(self.config.model_type in ["""xlnet"""] ) , pad_token=self.tokenizer.pad_token_id , pad_token_segment_id=self.tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info("""Saving features into cached file %s""" , UpperCamelCase__ ) torch.save(UpperCamelCase__ , UpperCamelCase__ ) def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = False ) -> DataLoader: '''simple docstring''' A_ = self._feature_file(UpperCamelCase__ ) logger.info("""Loading features from cached file %s""" , UpperCamelCase__ ) A_ = torch.load(UpperCamelCase__ ) A_ = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) A_ = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) if features[0].token_type_ids is not None: A_ = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) else: A_ = torch.tensor([0 for f in features] , dtype=torch.long ) # HACK(we will not use this anymore soon) A_ = torch.tensor([f.label_ids for f in features] , dtype=torch.long ) return DataLoader( TensorDataset(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) , batch_size=UpperCamelCase__ ) def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ ) -> List[Any]: '''simple docstring''' """Compute validation""" "" A_ = {"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type != "distilbert": A_ = ( batch[2] if self.config.model_type in ["""bert""", """xlnet"""] else None ) # XLM and RoBERTa don"t use token_type_ids A_ = self(**UpperCamelCase__ ) A_ , A_ = outputs[:2] A_ = logits.detach().cpu().numpy() A_ = inputs["""labels"""].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def snake_case_ ( self , UpperCamelCase__ ) -> int: '''simple docstring''' A_ = torch.stack([x["""val_loss"""] for x in outputs] ).mean() A_ = np.concatenate([x["""pred"""] for x in outputs] , axis=0 ) A_ = np.argmax(UpperCamelCase__ , axis=2 ) A_ = np.concatenate([x["""target"""] for x in outputs] , axis=0 ) A_ = dict(enumerate(self.labels ) ) A_ = [[] for _ in range(out_label_ids.shape[0] )] A_ = [[] for _ in range(out_label_ids.shape[0] )] for i in range(out_label_ids.shape[0] ): for j in range(out_label_ids.shape[1] ): if out_label_ids[i, j] != self.pad_token_label_id: out_label_list[i].append(label_map[out_label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) A_ = { """val_loss""": val_loss_mean, """accuracy_score""": accuracy_score(UpperCamelCase__ , UpperCamelCase__ ), """precision""": precision_score(UpperCamelCase__ , UpperCamelCase__ ), """recall""": recall_score(UpperCamelCase__ , UpperCamelCase__ ), """f1""": fa_score(UpperCamelCase__ , UpperCamelCase__ ), } A_ = dict(results.items() ) A_ = results return ret, preds_list, out_label_list def snake_case_ ( self , UpperCamelCase__ ) -> int: '''simple docstring''' # when stable A_ , A_ , A_ = self._eval_end(UpperCamelCase__ ) A_ = ret["""log"""] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def snake_case_ ( self , UpperCamelCase__ ) -> str: '''simple docstring''' # updating to test_epoch_end instead of deprecated test_end A_ , A_ , A_ = self._eval_end(UpperCamelCase__ ) # Converting to the dict required by pl # https://github.com/PyTorchLightning/pytorch-lightning/blob/master/\ # pytorch_lightning/trainer/logging.py#L139 A_ = ret["""log"""] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def snake_case_ ( UpperCamelCase__ , UpperCamelCase__ ) -> List[Any]: '''simple docstring''' # Add NER specific options BaseTransformer.add_model_specific_args(UpperCamelCase__ , UpperCamelCase__ ) parser.add_argument( """--task_type""" , default="""NER""" , type=UpperCamelCase__ , help="""Task type to fine tune in training (e.g. NER, POS, etc)""" ) parser.add_argument( """--max_seq_length""" , default=128 , type=UpperCamelCase__ , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument( """--labels""" , default="""""" , type=UpperCamelCase__ , help="""Path to a file containing all labels. If not specified, CoNLL-2003 labels are used.""" , ) parser.add_argument( """--gpus""" , default=0 , type=UpperCamelCase__ , help="""The number of GPUs allocated for this, it is by default 0 meaning none""" , ) parser.add_argument( """--overwrite_cache""" , action="""store_true""" , help="""Overwrite the cached training and evaluation sets""" ) return parser if __name__ == "__main__": __lowerCamelCase = argparse.ArgumentParser() add_generic_args(parser, os.getcwd()) __lowerCamelCase = NERTransformer.add_model_specific_args(parser, os.getcwd()) __lowerCamelCase = parser.parse_args() __lowerCamelCase = NERTransformer(args) __lowerCamelCase = generic_train(model, args) if args.do_predict: # See https://github.com/huggingface/transformers/issues/3159 # pl use this default format to create a checkpoint: # https://github.com/PyTorchLightning/pytorch-lightning/blob/master\ # /pytorch_lightning/callbacks/model_checkpoint.py#L322 __lowerCamelCase = sorted(glob.glob(os.path.join(args.output_dir, '''checkpoint-epoch=*.ckpt'''), recursive=True)) __lowerCamelCase = model.load_from_checkpoint(checkpoints[-1]) trainer.test(model)
101
'''simple docstring''' def UpperCAmelCase__ ( UpperCAmelCase__ = 1_00 ) -> int: A_ = n * (n + 1) * (2 * n + 1) / 6 A_ = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(f"""{solution() = }""")
101
1
"""simple docstring""" import random import sys import numpy as np from matplotlib import pyplot as plt from matplotlib.colors import ListedColormap lowercase__ : Tuple = """Usage of script: script_name <size_of_canvas:int>""" lowercase__ : Any = [0] * 1_0_0 + [1] * 1_0 random.shuffle(choice) def UpperCamelCase_ ( lowerCAmelCase__ : Optional[Any] ) -> list[list[bool]]: """simple docstring""" lowerCAmelCase_ : Dict = [[False for i in range(_lowerCAmelCase )] for j in range(_lowerCAmelCase )] return canvas def UpperCamelCase_ ( lowerCAmelCase__ : List[str] ) -> None: """simple docstring""" for i, row in enumerate(_lowerCAmelCase ): for j, _ in enumerate(_lowerCAmelCase ): lowerCAmelCase_ : Optional[int] = bool(random.getrandbits(1 ) ) def UpperCamelCase_ ( lowerCAmelCase__ : Optional[Any] ) -> list[list[bool]]: """simple docstring""" lowerCAmelCase_ : Dict = np.array(_lowerCAmelCase ) lowerCAmelCase_ : Dict = np.array(create_canvas(current_canvas.shape[0] ) ) for r, row in enumerate(_lowerCAmelCase ): for c, pt in enumerate(_lowerCAmelCase ): lowerCAmelCase_ : int = __judge_point( _lowerCAmelCase , current_canvas[r - 1 : r + 2, c - 1 : c + 2] ) lowerCAmelCase_ : List[Any] = next_gen_canvas del next_gen_canvas # cleaning memory as we move on. lowerCAmelCase_ : list[list[bool]] = current_canvas.tolist() return return_canvas def UpperCamelCase_ ( lowerCAmelCase__ : Tuple , lowerCAmelCase__ : int ) -> bool: """simple docstring""" lowerCAmelCase_ : Dict = 0 lowerCAmelCase_ : int = 0 # finding dead or alive neighbours count. for i in neighbours: for status in i: if status: alive += 1 else: dead += 1 # handling duplicate entry for focus pt. if pt: alive -= 1 else: dead -= 1 # running the rules of game here. lowerCAmelCase_ : int = pt if pt: if alive < 2: lowerCAmelCase_ : Tuple = False elif alive == 2 or alive == 3: lowerCAmelCase_ : int = True elif alive > 3: lowerCAmelCase_ : Dict = False else: if alive == 3: lowerCAmelCase_ : Optional[Any] = True return state if __name__ == "__main__": if len(sys.argv) != 2: raise Exception(usage_doc) lowercase__ : Tuple = int(sys.argv[1]) # main working structure of this module. lowercase__ : Any = create_canvas(canvas_size) seed(c) lowercase__ , lowercase__ : Optional[int] = plt.subplots() fig.show() lowercase__ : List[str] = ListedColormap(["""w""", """k"""]) try: while True: lowercase__ : Tuple = run(c) ax.matshow(c, cmap=cmap) fig.canvas.draw() ax.cla() except KeyboardInterrupt: # do nothing. pass
224
'''simple docstring''' from __future__ import annotations import math from collections.abc import Callable def a_ ( _lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase ,_lowerCAmelCase = 100 ,) -> float: __lowerCamelCase : Dict = x_start __lowerCamelCase : int = fnc(_lowerCAmelCase ) __lowerCamelCase : Dict = 0.0 for _ in range(_lowerCAmelCase ): # Approximates curve as a sequence of linear lines and sums their length __lowerCamelCase : List[str] = (x_end - x_start) / steps + xa __lowerCamelCase : List[Any] = fnc(_lowerCAmelCase ) length += math.hypot(xa - xa ,fxa - fxa ) # Increment step __lowerCamelCase : Any = xa __lowerCamelCase : Tuple = fxa return length if __name__ == "__main__": def a_ ( _lowerCAmelCase ) -> Dict: return math.sin(10 * x ) print('f(x) = sin(10 * x)') print('The length of the curve from x = -10 to x = 10 is:') _UpperCamelCase = 10 while i <= 100000: print(f'''With {i} steps: {line_length(f, -10, 10, i)}''') i *= 10
208
0
def A ( _UpperCAmelCase : Any , _UpperCAmelCase : Optional[Any] ) -> float: '''simple docstring''' def get_matched_characters(_UpperCAmelCase : List[str] , _UpperCAmelCase : List[str] ) -> str: _UpperCAmelCase = [] _UpperCAmelCase = min(len(_stra ) , len(_stra ) ) // 2 for i, l in enumerate(_stra ): _UpperCAmelCase = int(max(0 , i - limit ) ) _UpperCAmelCase = int(min(i + limit + 1 , len(_stra ) ) ) if l in _stra[left:right]: matched.append(lowerCAmelCase_ ) _UpperCAmelCase = F"{_stra[0:_stra.index(lowerCAmelCase_ )]} {_stra[_stra.index(lowerCAmelCase_ ) + 1:]}" return "".join(lowerCAmelCase_ ) # matching characters _UpperCAmelCase = get_matched_characters(lowerCAmelCase_ , lowerCAmelCase_ ) _UpperCAmelCase = get_matched_characters(lowerCAmelCase_ , lowerCAmelCase_ ) _UpperCAmelCase = len(lowerCAmelCase_ ) # transposition _UpperCAmelCase = ( len([(ca, ca) for ca, ca in zip(lowerCAmelCase_ , lowerCAmelCase_ ) if ca != ca] ) // 2 ) if not match_count: _UpperCAmelCase = 0.0 else: _UpperCAmelCase = ( 1 / 3 * ( match_count / len(lowerCAmelCase_ ) + match_count / len(lowerCAmelCase_ ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters _UpperCAmelCase = 0 for ca, ca in zip(stra[:4] , stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler("hello", "world"))
369
from __future__ import annotations import unittest from transformers import DebertaVaConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, TFDebertaVaModel, ) class __lowerCAmelCase : def __init__( self : str , A : str , A : Dict=13 , A : int=7 , A : Tuple=True , A : Union[str, Any]=True , A : Any=True , A : Dict=True , A : Dict=99 , A : Tuple=32 , A : Any=2 , A : Any=4 , A : Any=37 , A : Optional[Any]="gelu" , A : List[Any]=0.1 , A : Tuple=0.1 , A : Optional[Any]=5_12 , A : Tuple=16 , A : int=2 , A : List[str]=0.0_2 , A : int=False , A : List[Any]=True , A : Optional[Any]="None" , A : Union[str, Any]=3 , A : List[str]=4 , A : List[Any]=None , ) -> int: """simple docstring""" _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = seq_length _UpperCAmelCase = is_training _UpperCAmelCase = use_input_mask _UpperCAmelCase = use_token_type_ids _UpperCAmelCase = use_labels _UpperCAmelCase = vocab_size _UpperCAmelCase = hidden_size _UpperCAmelCase = num_hidden_layers _UpperCAmelCase = num_attention_heads _UpperCAmelCase = intermediate_size _UpperCAmelCase = hidden_act _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = type_vocab_size _UpperCAmelCase = type_sequence_label_size _UpperCAmelCase = initializer_range _UpperCAmelCase = num_labels _UpperCAmelCase = num_choices _UpperCAmelCase = relative_attention _UpperCAmelCase = position_biased_input _UpperCAmelCase = pos_att_type _UpperCAmelCase = scope def _lowerCamelCase ( self : Any) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) _UpperCAmelCase = None if self.use_input_mask: _UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length]) _UpperCAmelCase = None if self.use_token_type_ids: _UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) _UpperCAmelCase = None _UpperCAmelCase = None _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size) _UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) _UpperCAmelCase = DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , initializer_range=self.initializer_range , return_dict=A , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowerCamelCase ( self : Union[str, Any] , A : Optional[int] , A : Tuple , A : int , A : Any , A : List[str] , A : List[str] , A : int) -> Tuple: """simple docstring""" _UpperCAmelCase = TFDebertaVaModel(config=A) _UpperCAmelCase = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} _UpperCAmelCase = [input_ids, input_mask] _UpperCAmelCase = model(A) _UpperCAmelCase = model(A) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def _lowerCamelCase ( self : str , A : Tuple , A : Tuple , A : Optional[int] , A : List[str] , A : Any , A : List[str] , A : List[str]) -> Optional[int]: """simple docstring""" _UpperCAmelCase = TFDebertaVaForMaskedLM(config=A) _UpperCAmelCase = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } _UpperCAmelCase = model(A) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def _lowerCamelCase ( self : List[Any] , A : Tuple , A : Tuple , A : Optional[int] , A : Optional[int] , A : List[Any] , A : Any , A : Optional[int]) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase = self.num_labels _UpperCAmelCase = TFDebertaVaForSequenceClassification(config=A) _UpperCAmelCase = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } _UpperCAmelCase = model(A) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def _lowerCamelCase ( self : Union[str, Any] , A : List[Any] , A : List[Any] , A : List[str] , A : Optional[Any] , A : int , A : Any , A : int) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase = self.num_labels _UpperCAmelCase = TFDebertaVaForTokenClassification(config=A) _UpperCAmelCase = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } _UpperCAmelCase = model(A) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def _lowerCamelCase ( self : List[Any] , A : List[Any] , A : List[str] , A : Dict , A : Dict , A : Any , A : Tuple , A : List[Any]) -> List[str]: """simple docstring""" _UpperCAmelCase = TFDebertaVaForQuestionAnswering(config=A) _UpperCAmelCase = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } _UpperCAmelCase = model(A) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def _lowerCamelCase ( self : Union[str, Any]) -> List[Any]: """simple docstring""" _UpperCAmelCase = self.prepare_config_and_inputs() ( ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ) = config_and_inputs _UpperCAmelCase = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class __lowerCAmelCase ( A , A , unittest.TestCase ): UpperCamelCase = ( ( TFDebertaVaModel, TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, ) if is_tf_available() else () ) UpperCamelCase = ( { '''feature-extraction''': TFDebertaVaModel, '''fill-mask''': TFDebertaVaForMaskedLM, '''question-answering''': TFDebertaVaForQuestionAnswering, '''text-classification''': TFDebertaVaForSequenceClassification, '''token-classification''': TFDebertaVaForTokenClassification, '''zero-shot''': TFDebertaVaForSequenceClassification, } if is_tf_available() else {} ) UpperCamelCase = False UpperCamelCase = False def _lowerCamelCase ( self : int) -> Optional[int]: """simple docstring""" _UpperCAmelCase = TFDebertaVaModelTester(self) _UpperCAmelCase = ConfigTester(self , config_class=A , hidden_size=37) def _lowerCamelCase ( self : Optional[int]) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() def _lowerCamelCase ( self : Tuple) -> Optional[int]: """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A) def _lowerCamelCase ( self : Tuple) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A) def _lowerCamelCase ( self : Tuple) -> str: """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A) def _lowerCamelCase ( self : int) -> List[Any]: """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A) def _lowerCamelCase ( self : Dict) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A) @slow def _lowerCamelCase ( self : List[Any]) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase = TFDebertaVaModel.from_pretrained('kamalkraj/deberta-v2-xlarge') self.assertIsNotNone(A) @require_tf class __lowerCAmelCase ( unittest.TestCase ): @unittest.skip(reason='Model not available yet') def _lowerCamelCase ( self : Optional[int]) -> Dict: """simple docstring""" pass @slow def _lowerCamelCase ( self : List[str]) -> Optional[int]: """simple docstring""" _UpperCAmelCase = TFDebertaVaModel.from_pretrained('kamalkraj/deberta-v2-xlarge') _UpperCAmelCase = tf.constant([[0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2]]) _UpperCAmelCase = tf.constant([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]) _UpperCAmelCase = model(A , attention_mask=A)[0] _UpperCAmelCase = tf.constant( [[[0.2_3_5_6, 0.1_9_4_8, 0.0_3_6_9], [-0.1_0_6_3, 0.3_5_8_6, -0.5_1_5_2], [-0.6_3_9_9, -0.0_2_5_9, -0.2_5_2_5]]]) tf.debugging.assert_near(output[:, 1:4, 1:4] , A , atol=1E-4)
290
0
'''simple docstring''' def __lowercase ( __lowercase ) -> str: '''simple docstring''' if isinstance(__lowercase , __lowercase ): raise TypeError("'float' object cannot be interpreted as an integer" ) if isinstance(__lowercase , __lowercase ): raise TypeError("'str' object cannot be interpreted as an integer" ) if num == 0: return "0b0" _A = False if num < 0: _A = True _A = -num _A = [] while num > 0: binary.insert(0 , num % 2 ) num >>= 1 if negative: return "-0b" + "".join(str(__lowercase ) for e in binary ) return "0b" + "".join(str(__lowercase ) for e in binary ) if __name__ == "__main__": import doctest doctest.testmod()
79
import argparse import torch from transformers import YosoConfig, YosoForMaskedLM def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> Union[str, Any]: '''simple docstring''' if "model" in orig_key: UpperCAmelCase = orig_key.replace('''model.''' , '''''' ) if "norm1" in orig_key: UpperCAmelCase = orig_key.replace('''norm1''' , '''attention.output.LayerNorm''' ) if "norm2" in orig_key: UpperCAmelCase = orig_key.replace('''norm2''' , '''output.LayerNorm''' ) if "norm" in orig_key: UpperCAmelCase = orig_key.replace('''norm''' , '''LayerNorm''' ) if "transformer" in orig_key: UpperCAmelCase = orig_key.split('''.''' )[0].split('''_''' )[-1] UpperCAmelCase = orig_key.replace(F"""transformer_{layer_num}""" , F"""encoder.layer.{layer_num}""" ) if "mha.attn" in orig_key: UpperCAmelCase = orig_key.replace('''mha.attn''' , '''attention.self''' ) if "mha" in orig_key: UpperCAmelCase = orig_key.replace('''mha''' , '''attention''' ) if "W_q" in orig_key: UpperCAmelCase = orig_key.replace('''W_q''' , '''self.query''' ) if "W_k" in orig_key: UpperCAmelCase = orig_key.replace('''W_k''' , '''self.key''' ) if "W_v" in orig_key: UpperCAmelCase = orig_key.replace('''W_v''' , '''self.value''' ) if "ff1" in orig_key: UpperCAmelCase = orig_key.replace('''ff1''' , '''intermediate.dense''' ) if "ff2" in orig_key: UpperCAmelCase = orig_key.replace('''ff2''' , '''output.dense''' ) if "ff" in orig_key: UpperCAmelCase = orig_key.replace('''ff''' , '''output.dense''' ) if "mlm_class" in orig_key: UpperCAmelCase = orig_key.replace('''mlm.mlm_class''' , '''cls.predictions.decoder''' ) if "mlm" in orig_key: UpperCAmelCase = orig_key.replace('''mlm''' , '''cls.predictions.transform''' ) if "cls" not in orig_key: UpperCAmelCase = '''yoso.''' + orig_key return orig_key def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ ) -> Dict: '''simple docstring''' for key in orig_state_dict.copy().keys(): UpperCAmelCase = orig_state_dict.pop(UpperCamelCase__ ) if ("pooler" in key) or ("sen_class" in key): continue else: UpperCAmelCase = val UpperCAmelCase = orig_state_dict['''cls.predictions.decoder.bias'''] UpperCAmelCase = torch.arange(UpperCamelCase__ ).expand((1, -1) ) + 2 return orig_state_dict def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> int: '''simple docstring''' UpperCAmelCase = torch.load(UpperCamelCase__ , map_location='''cpu''' )['''model_state_dict'''] UpperCAmelCase = YosoConfig.from_json_file(UpperCamelCase__ ) UpperCAmelCase = YosoForMaskedLM(UpperCamelCase__ ) UpperCAmelCase = convert_checkpoint_helper(config.max_position_embeddings , UpperCamelCase__ ) print(model.load_state_dict(UpperCamelCase__ ) ) model.eval() model.save_pretrained(UpperCamelCase__ ) print(F"""Checkpoint successfuly converted. Model saved at {pytorch_dump_path}""" ) if __name__ == "__main__": __A : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--pytorch_model_path", default=None, type=str, required=True, help="Path to YOSO pytorch checkpoint." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The json file for YOSO model config.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) __A : List[str] = parser.parse_args() convert_yoso_checkpoint(args.pytorch_model_path, args.config_file, args.pytorch_dump_path)
273
0
'''simple docstring''' import argparse import json import os import re import torch from transformers import BloomConfig, BloomModel from transformers.file_utils import CONFIG_NAME, WEIGHTS_NAME from transformers.utils import logging logging.set_verbosity_info() _lowercase : Optional[int] = [ "word_embeddings_layernorm.weight", "word_embeddings_layernorm.bias", "input_layernorm.weight", "input_layernorm.bias", "post_attention_layernorm.weight", "post_attention_layernorm.bias", "self_attention.dense.bias", "mlp.dense_4h_to_h.bias", "ln_f.weight", "ln_f.bias", ] _lowercase : Optional[int] = [ "mlp.dense_4h_to_h.weight", "self_attention.dense.weight", ] def snake_case_ ( __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : List[str] ): """simple docstring""" lowercase_ : List[str] = { '''word_embeddings.weight''': '''word_embeddings.weight''', '''word_embeddings.norm.weight''': '''word_embeddings_layernorm.weight''', '''word_embeddings.norm.bias''': '''word_embeddings_layernorm.bias''', '''weight''': '''ln_f.weight''', '''bias''': '''ln_f.bias''', } if key in layer_rename_map: return layer_rename_map[key] # Handle transformer blocks lowercase_ : Any = int(re.match(R'''.*layer_(\d*).*''' , __SCREAMING_SNAKE_CASE )[1] ) layer_number -= 3 return F'''h.{layer_number}.''' + key def snake_case_ ( __SCREAMING_SNAKE_CASE : str ): """simple docstring""" if dtype == torch.bool: return 1 / 8 lowercase_ : List[Any] = re.search(R'''[^\d](\d+)$''' , str(__SCREAMING_SNAKE_CASE ) ) if bit_search is None: raise ValueError(F'''`dtype` is not a valid dtype: {dtype}.''' ) lowercase_ : Tuple = int(bit_search.groups()[0] ) return bit_size // 8 def snake_case_ ( __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : List[Any] ): """simple docstring""" if bloom_config_file == "": lowercase_ : str = BloomConfig() else: lowercase_ : Optional[int] = BloomConfig.from_json_file(__SCREAMING_SNAKE_CASE ) if shard_model: lowercase_ : Tuple = os.listdir(__SCREAMING_SNAKE_CASE ) lowercase_ : int = sorted(filter(lambda __SCREAMING_SNAKE_CASE : s.startswith('''layer''' ) and "model_00" in s , __SCREAMING_SNAKE_CASE ) ) lowercase_ : str = {'''weight_map''': {}, '''metadata''': {}} lowercase_ : List[Any] = 0 lowercase_ : str = None lowercase_ : Union[str, Any] = BloomConfig() for j, file in enumerate(__SCREAMING_SNAKE_CASE ): print('''Processing file: {}'''.format(__SCREAMING_SNAKE_CASE ) ) lowercase_ : Tuple = None for i in range(__SCREAMING_SNAKE_CASE ): # load all TP files lowercase_ : Optional[Any] = file.replace('''model_00''' , F'''model_0{i}''' ) lowercase_ : int = torch.load(os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) , map_location='''cpu''' ) # Rename keys in the transformers names lowercase_ : int = list(temp.keys() ) for key in keys: lowercase_ : Dict = temp.pop(__SCREAMING_SNAKE_CASE ) if tensors is None: lowercase_ : Union[str, Any] = temp else: for key in tensors.keys(): if any(key.endswith(__SCREAMING_SNAKE_CASE ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel lowercase_ : Tuple = 1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks lowercase_ : Union[str, Any] = torch.cat([tensors[key], temp[key]] , dim=__SCREAMING_SNAKE_CASE ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(__SCREAMING_SNAKE_CASE ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): lowercase_ : str = tensors[key] / pretraining_tp torch.save( __SCREAMING_SNAKE_CASE , os.path.join( __SCREAMING_SNAKE_CASE , '''pytorch_model_{}-of-{}.bin'''.format(str(j + 1 ).zfill(5 ) , str(len(__SCREAMING_SNAKE_CASE ) ).zfill(5 ) ) , ) , ) for key in tensors.keys(): lowercase_ : int = tensors[key] total_size += value.numel() * get_dtype_size(value.dtype ) if key not in index_dict["weight_map"]: lowercase_ : Any = '''pytorch_model_{}-of-{}.bin'''.format( str(j + 1 ).zfill(5 ) , str(len(__SCREAMING_SNAKE_CASE ) ).zfill(5 ) ) lowercase_ : Dict = BloomConfig() lowercase_ : str = pytorch_dump_folder_path + '''/''' + CONFIG_NAME lowercase_ : Tuple = total_size with open(__SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) with open(os.path.join(__SCREAMING_SNAKE_CASE , WEIGHTS_NAME + '''.index.json''' ) , '''w''' , encoding='''utf-8''' ) as f: lowercase_ : Any = json.dumps(__SCREAMING_SNAKE_CASE , indent=2 , sort_keys=__SCREAMING_SNAKE_CASE ) + '''\n''' f.write(__SCREAMING_SNAKE_CASE ) else: lowercase_ : Any = BloomModel(__SCREAMING_SNAKE_CASE ) lowercase_ : Dict = os.listdir(__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[int] = sorted(filter(lambda __SCREAMING_SNAKE_CASE : s.startswith('''layer''' ) and "model_00" in s , __SCREAMING_SNAKE_CASE ) ) lowercase_ : Optional[int] = None for i, file in enumerate(__SCREAMING_SNAKE_CASE ): lowercase_ : List[str] = None for i in range(__SCREAMING_SNAKE_CASE ): # load all TP files lowercase_ : Tuple = file.replace('''model_00''' , F'''model_0{i}''' ) lowercase_ : Dict = torch.load(os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) , map_location='''cpu''' ) # Rename keys in the transformers names lowercase_ : List[Any] = list(temp.keys() ) for key in keys: lowercase_ : Optional[int] = temp.pop(__SCREAMING_SNAKE_CASE ) if tensors is None: lowercase_ : int = temp else: for key in tensors.keys(): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) if any(key.endswith(__SCREAMING_SNAKE_CASE ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel lowercase_ : List[str] = 1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks lowercase_ : List[str] = torch.cat([tensors[key], temp[key]] , dim=__SCREAMING_SNAKE_CASE ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(__SCREAMING_SNAKE_CASE ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): lowercase_ : Tuple = tensors[key] / pretraining_tp lowercase_ : Union[str, Any] = model.load_state_dict(__SCREAMING_SNAKE_CASE , strict=__SCREAMING_SNAKE_CASE ) assert not other_keys.unexpected_keys, F'''The keys {other_keys.unexpected_keys} are unexpected''' if missing_keys is None: lowercase_ : Any = set(other_keys.missing_keys ) else: lowercase_ : Optional[int] = missing_keys.intersection(set(other_keys.missing_keys ) ) assert not missing_keys, F'''The keys {missing_keys} are missing''' # Save pytorch-model os.makedirs(__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE ) lowercase_ : List[Any] = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME lowercase_ : Any = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(F'''Save PyTorch model to {pytorch_weights_dump_path} with dtype {config.torch_dtype}''' ) if config.torch_dtype is not None: lowercase_ : Union[str, Any] = model.to(config.torch_dtype ) torch.save(model.state_dict() , __SCREAMING_SNAKE_CASE ) print(F'''Save configuration file to {pytorch_config_dump_path}''' ) with open(__SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _lowercase : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--bloom_checkpoint_path", default=None, type=str, required=True, help="Path to the Megatron-LM checkpoint path.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--bloom_config_file", default="", type=str, help=( "An optional config json file corresponding to the pre-trained model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--shard_model", action="store_true", help="An optional setting to shard the output model \nThis enables sharding the converted checkpoint", ) parser.add_argument( "--pretraining_tp", default=4, type=int, help="Pretraining TP rank that has been used when training the model in Megatron-LM \n", ) _lowercase : Union[str, Any] = parser.parse_args() convert_bloom_checkpoint_to_pytorch( args.bloom_checkpoint_path, args.bloom_config_file, args.pytorch_dump_folder_path, args.shard_model, args.pretraining_tp, )
264
'''simple docstring''' import logging from dataclasses import dataclass, field from pathlib import Path from typing import Optional, Union from .generation.configuration_utils import GenerationConfig from .training_args import TrainingArguments from .utils import add_start_docstrings _lowercase : Tuple = logging.getLogger(__name__) @dataclass @add_start_docstrings(TrainingArguments.__doc__ ) class lowerCAmelCase__ ( lowerCamelCase_ ): lowerCAmelCase_ = field(default=lowerCamelCase_ , metadata={'''help''': '''Whether to use SortishSampler or not.'''} ) lowerCAmelCase_ = field( default=lowerCamelCase_ , metadata={'''help''': '''Whether to use generate to calculate generative metrics (ROUGE, BLEU).'''} ) lowerCAmelCase_ = field( default=lowerCamelCase_ , metadata={ '''help''': ( '''The `max_length` to use on each evaluation loop when `predict_with_generate=True`. Will default ''' '''to the `max_length` value of the model configuration.''' ) } , ) lowerCAmelCase_ = field( default=lowerCamelCase_ , metadata={ '''help''': ( '''The `num_beams` to use on each evaluation loop when `predict_with_generate=True`. Will default ''' '''to the `num_beams` value of the model configuration.''' ) } , ) lowerCAmelCase_ = field( default=lowerCamelCase_ , metadata={ '''help''': '''Model id, file path or url pointing to a GenerationConfig json file, to use during prediction.''' } , ) def _snake_case ( self ): """simple docstring""" lowercase_ : Dict = super().to_dict() for k, v in d.items(): if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase_ : str = v.to_dict() return d
264
1
from __future__ import annotations from math import pi, sqrt def lowerCamelCase__ ( A__ : float , A__ : float ): '''simple docstring''' if inductance <= 0: raise ValueError("""Inductance cannot be 0 or negative""" ) elif capacitance <= 0: raise ValueError("""Capacitance cannot be 0 or negative""" ) else: return ( "Resonant frequency", float(1 / (2 * pi * (sqrt(inductance * capacitance ))) ), ) if __name__ == "__main__": import doctest doctest.testmod()
12
import os import torch from ..logging import get_logger from .constants import FSDP_PYTORCH_VERSION, MODEL_NAME, OPTIMIZER_NAME from .versions import is_torch_version if is_torch_version('>=', FSDP_PYTORCH_VERSION): import torch.distributed.checkpoint as dist_cp from torch.distributed.checkpoint.default_planner import DefaultLoadPlanner, DefaultSavePlanner from torch.distributed.checkpoint.optimizer import load_sharded_optimizer_state_dict from torch.distributed.fsdp.fully_sharded_data_parallel import FullyShardedDataParallel as FSDP from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType UpperCAmelCase_ = get_logger(__name__) def lowerCamelCase__ ( A__ : Union[str, Any] , A__ : str , A__ : Any , A__ : Dict , A__ : Any=0 ): '''simple docstring''' os.makedirs(A__ , exist_ok=A__ ) with FSDP.state_dict_type( A__ , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): __lowerCamelCase = model.state_dict() if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: __lowerCamelCase = f'{MODEL_NAME}.bin' if model_index == 0 else f'{MODEL_NAME}_{model_index}.bin' __lowerCamelCase = os.path.join(A__ , A__ ) if accelerator.process_index == 0: logger.info(f'Saving model to {output_model_file}' ) torch.save(A__ , A__ ) logger.info(f'Model saved to {output_model_file}' ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: __lowerCamelCase = ( f'{MODEL_NAME}_rank{accelerator.process_index}.bin' if model_index == 0 else f'{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin' ) __lowerCamelCase = os.path.join(A__ , A__ ) logger.info(f'Saving model to {output_model_file}' ) torch.save(A__ , A__ ) logger.info(f'Model saved to {output_model_file}' ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: __lowerCamelCase = os.path.join(A__ , f'{MODEL_NAME}_{model_index}' ) os.makedirs(A__ , exist_ok=A__ ) logger.info(f'Saving model to {ckpt_dir}' ) __lowerCamelCase = {"""model""": state_dict} dist_cp.save_state_dict( state_dict=A__ , storage_writer=dist_cp.FileSystemWriter(A__ ) , planner=DefaultSavePlanner() , ) logger.info(f'Model saved to {ckpt_dir}' ) def lowerCamelCase__ ( A__ : int , A__ : Dict , A__ : int , A__ : List[str] , A__ : Any=0 ): '''simple docstring''' accelerator.wait_for_everyone() with FSDP.state_dict_type( A__ , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if type(A__ ) != FSDP and accelerator.process_index != 0: if not fsdp_plugin.sync_module_states: raise ValueError( """Set the `sync_module_states` flag to `True` so that model states are synced across processes when """ """initializing FSDP object""" ) return __lowerCamelCase = f'{MODEL_NAME}.bin' if model_index == 0 else f'{MODEL_NAME}_{model_index}.bin' __lowerCamelCase = os.path.join(A__ , A__ ) logger.info(f'Loading model from {input_model_file}' ) __lowerCamelCase = torch.load(A__ ) logger.info(f'Model loaded from {input_model_file}' ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: __lowerCamelCase = ( f'{MODEL_NAME}_rank{accelerator.process_index}.bin' if model_index == 0 else f'{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin' ) __lowerCamelCase = os.path.join(A__ , A__ ) logger.info(f'Loading model from {input_model_file}' ) __lowerCamelCase = torch.load(A__ ) logger.info(f'Model loaded from {input_model_file}' ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: __lowerCamelCase = ( os.path.join(A__ , f'{MODEL_NAME}_{model_index}' ) if f'{MODEL_NAME}' not in input_dir else input_dir ) logger.info(f'Loading model from {ckpt_dir}' ) __lowerCamelCase = {"""model""": model.state_dict()} dist_cp.load_state_dict( state_dict=A__ , storage_reader=dist_cp.FileSystemReader(A__ ) , planner=DefaultLoadPlanner() , ) __lowerCamelCase = state_dict["""model"""] logger.info(f'Model loaded from {ckpt_dir}' ) model.load_state_dict(A__ ) def lowerCamelCase__ ( A__ : List[str] , A__ : List[str] , A__ : str , A__ : Dict , A__ : Optional[Any] , A__ : Optional[int]=0 ): '''simple docstring''' os.makedirs(A__ , exist_ok=A__ ) with FSDP.state_dict_type( A__ , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): __lowerCamelCase = FSDP.optim_state_dict(A__ , A__ ) if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if accelerator.process_index == 0: __lowerCamelCase = ( f'{OPTIMIZER_NAME}.bin' if optimizer_index == 0 else f'{OPTIMIZER_NAME}_{optimizer_index}.bin' ) __lowerCamelCase = os.path.join(A__ , A__ ) logger.info(f'Saving Optimizer state to {output_optimizer_file}' ) torch.save(A__ , A__ ) logger.info(f'Optimizer state saved in {output_optimizer_file}' ) else: __lowerCamelCase = os.path.join(A__ , f'{OPTIMIZER_NAME}_{optimizer_index}' ) os.makedirs(A__ , exist_ok=A__ ) logger.info(f'Saving Optimizer state to {ckpt_dir}' ) dist_cp.save_state_dict( state_dict={"""optimizer""": optim_state} , storage_writer=dist_cp.FileSystemWriter(A__ ) , planner=DefaultSavePlanner() , ) logger.info(f'Optimizer state saved in {ckpt_dir}' ) def lowerCamelCase__ ( A__ : int , A__ : List[str] , A__ : int , A__ : Any , A__ : Union[str, Any] , A__ : List[Any]=0 ): '''simple docstring''' accelerator.wait_for_everyone() with FSDP.state_dict_type( A__ , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: __lowerCamelCase = None # below check should work but currently it isn't working (mostly opytorch issue), # in the meantime disabling it at the cost of excess memory usage # if accelerator.process_index == 0 or not fsdp_plugin.optim_state_dict_config.rank0_only: __lowerCamelCase = ( f'{OPTIMIZER_NAME}.bin' if optimizer_index == 0 else f'{OPTIMIZER_NAME}_{optimizer_index}.bin' ) __lowerCamelCase = os.path.join(A__ , A__ ) logger.info(f'Loading Optimizer state from {input_optimizer_file}' ) __lowerCamelCase = torch.load(A__ ) logger.info(f'Optimizer state loaded from {input_optimizer_file}' ) else: __lowerCamelCase = ( os.path.join(A__ , f'{OPTIMIZER_NAME}_{optimizer_index}' ) if f'{OPTIMIZER_NAME}' not in input_dir else input_dir ) logger.info(f'Loading Optimizer from {ckpt_dir}' ) __lowerCamelCase = load_sharded_optimizer_state_dict( model_state_dict=model.state_dict() , optimizer_key="""optimizer""" , storage_reader=dist_cp.FileSystemReader(A__ ) , ) __lowerCamelCase = optim_state["""optimizer"""] logger.info(f'Optimizer loaded from {ckpt_dir}' ) __lowerCamelCase = FSDP.optim_state_dict_to_load(A__ , A__ , A__ ) optimizer.load_state_dict(A__ )
12
1
import random import unittest import torch from diffusers import IFInpaintingPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _A ( __UpperCAmelCase ,__UpperCAmelCase ,unittest.TestCase ): UpperCamelCase__ : List[Any] = IFInpaintingPipeline UpperCamelCase__ : List[Any] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} UpperCamelCase__ : int = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS UpperCamelCase__ : int = PipelineTesterMixin.required_optional_params - {'''latents'''} def _lowerCamelCase ( self : str): '''simple docstring''' return self._get_dummy_components() def _lowerCamelCase ( self : int , __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : Dict=0): '''simple docstring''' if str(__SCREAMING_SNAKE_CASE).startswith('''mps'''): __a = torch.manual_seed(__SCREAMING_SNAKE_CASE) else: __a = torch.Generator(device=__SCREAMING_SNAKE_CASE).manual_seed(__SCREAMING_SNAKE_CASE) __a = floats_tensor((1, 3, 32, 32) , rng=random.Random(__SCREAMING_SNAKE_CASE)).to(__SCREAMING_SNAKE_CASE) __a = floats_tensor((1, 3, 32, 32) , rng=random.Random(__SCREAMING_SNAKE_CASE)).to(__SCREAMING_SNAKE_CASE) __a = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def _lowerCamelCase ( self : List[Any]): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3) def _lowerCamelCase ( self : Dict): '''simple docstring''' self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''') def _lowerCamelCase ( self : Any): '''simple docstring''' super().test_save_load_floataa(expected_max_diff=1E-1) def _lowerCamelCase ( self : str): '''simple docstring''' self._test_attention_slicing_forward_pass(expected_max_diff=1E-2) def _lowerCamelCase ( self : Optional[Any]): '''simple docstring''' self._test_save_load_local() def _lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
131
from __future__ import annotations from collections import namedtuple from dataclasses import dataclass @dataclass class _A : UpperCamelCase__ : int UpperCamelCase__ : TreeNode | None = None UpperCamelCase__ : TreeNode | None = None __snake_case :Optional[Any] = namedtuple('''CoinsDistribResult''', '''moves excess''') def __snake_case ( _UpperCAmelCase ): if root is None: return 0 # Validation def count_nodes(_UpperCAmelCase ) -> int: if node is None: return 0 return count_nodes(node.left ) + count_nodes(node.right ) + 1 def count_coins(_UpperCAmelCase ) -> int: if node is None: return 0 return count_coins(node.left ) + count_coins(node.right ) + node.data if count_nodes(_UpperCAmelCase ) != count_coins(_UpperCAmelCase ): raise ValueError('''The nodes number should be same as the number of coins''' ) # Main calculation def get_distrib(_UpperCAmelCase ) -> CoinsDistribResult: if node is None: return CoinsDistribResult(0 , 1 ) __a , __a = get_distrib(node.left ) __a , __a = get_distrib(node.right ) __a = 1 - left_distrib_excess __a = 1 - right_distrib_excess __a = ( left_distrib_moves + right_distrib_moves + abs(_UpperCAmelCase ) + abs(_UpperCAmelCase ) ) __a = node.data - coins_to_left - coins_to_right return CoinsDistribResult(_UpperCAmelCase , _UpperCAmelCase ) return get_distrib(_UpperCAmelCase )[0] if __name__ == "__main__": import doctest doctest.testmod()
131
1
from __future__ import annotations def __A ( __lowerCamelCase , __lowerCamelCase ) -> int: if len(__lowerCamelCase ) < k or k < 0: raise ValueError("""Invalid Input""" ) a = a = sum(array[:k] ) for i in range(len(__lowerCamelCase ) - k ): a = current_sum - array[i] + array[i + k] a = max(__lowerCamelCase , __lowerCamelCase ) return max_sum if __name__ == "__main__": from doctest import testmod from random import randint testmod() __UpperCamelCase : Union[str, Any] = [randint(-1_000, 1_000) for i in range(100)] __UpperCamelCase : Optional[int] = randint(0, 110) print(F'The maximum sum of {k} consecutive elements is {max_sum_in_array(array,k)}')
228
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def __A ( ) -> Any: a = ArgumentParser( description=( """PyTorch TPU distributed training launch """ """helper utility that will spawn up """ """multiple distributed processes""" ) ) # Optional arguments for the launch helper parser.add_argument("""--num_cores""" , type=__lowerCamelCase , default=1 , help="""Number of TPU cores to use (1 or 8).""" ) # positional parser.add_argument( """training_script""" , type=__lowerCamelCase , help=( """The full path to the single TPU training """ """program/script to be launched in parallel, """ """followed by all the arguments for the """ """training script""" ) , ) # rest from the training program parser.add_argument("""training_script_args""" , nargs=__lowerCamelCase ) return parser.parse_args() def __A ( ) -> Union[str, Any]: a = parse_args() # Import training_script as a module. a = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) a = script_fpath.stem a = importlib.import_module(__lowerCamelCase ) # Patch sys.argv a = [args.training_script] + args.training_script_args + ["""--tpu_num_cores""", str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
228
1
'''simple docstring''' class _UpperCamelCase ( lowerCamelCase__ ): '''simple docstring''' pass class _UpperCamelCase ( lowerCamelCase__ ): '''simple docstring''' pass class _UpperCamelCase : '''simple docstring''' def __init__( self : Optional[int]): '''simple docstring''' __lowercase =[ [], [], [], ] def __lowerCamelCase ( self : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : int): '''simple docstring''' try: if len(self.queues[priority]) >= 1_0_0: raise OverflowError('Maximum queue size is 100') self.queues[priority].append(lowercase__) except IndexError: raise ValueError('Valid priorities are 0, 1, and 2') def __lowerCamelCase ( self : Dict): '''simple docstring''' for queue in self.queues: if queue: return queue.pop(0) raise UnderFlowError('All queues are empty') def __str__( self : Optional[Any]): '''simple docstring''' return "\n".join(f"""Priority {i}: {q}""" for i, q in enumerate(self.queues)) class _UpperCamelCase : '''simple docstring''' def __init__( self : int): '''simple docstring''' __lowercase =[] def __lowerCamelCase ( self : int , _lowerCAmelCase : int): '''simple docstring''' if len(self.queue) == 1_0_0: raise OverFlowError('Maximum queue size is 100') self.queue.append(lowercase__) def __lowerCamelCase ( self : List[str]): '''simple docstring''' if not self.queue: raise UnderFlowError('The queue is empty') else: __lowercase =min(self.queue) self.queue.remove(lowercase__) return data def __str__( self : List[Any]): '''simple docstring''' return str(self.queue) def _A ( ): """simple docstring""" __lowercase =FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 100 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 128 ) print(A__ ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(A__ ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def _A ( ): """simple docstring""" __lowercase =ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(100 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(128 ) print(A__ ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(A__ ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
357
'''simple docstring''' import argparse import requests import torch from PIL import Image from transformers import SwinConfig, SwinForMaskedImageModeling, ViTImageProcessor def _A ( _lowerCAmelCase ): """simple docstring""" __lowercase =SwinConfig(image_size=192 ) if "base" in model_name: __lowercase =6 __lowercase =128 __lowercase =(2, 2, 18, 2) __lowercase =(4, 8, 16, 32) elif "large" in model_name: __lowercase =12 __lowercase =192 __lowercase =(2, 2, 18, 2) __lowercase =(6, 12, 24, 48) else: raise ValueError('Model not supported, only supports base and large variants' ) __lowercase =window_size __lowercase =embed_dim __lowercase =depths __lowercase =num_heads return config def _A ( _lowerCAmelCase ): """simple docstring""" if "encoder.mask_token" in name: __lowercase =name.replace('encoder.mask_token' , 'embeddings.mask_token' ) if "encoder.patch_embed.proj" in name: __lowercase =name.replace('encoder.patch_embed.proj' , 'embeddings.patch_embeddings.projection' ) if "encoder.patch_embed.norm" in name: __lowercase =name.replace('encoder.patch_embed.norm' , 'embeddings.norm' ) if "attn.proj" in name: __lowercase =name.replace('attn.proj' , 'attention.output.dense' ) if "attn" in name: __lowercase =name.replace('attn' , 'attention.self' ) if "norm1" in name: __lowercase =name.replace('norm1' , 'layernorm_before' ) if "norm2" in name: __lowercase =name.replace('norm2' , 'layernorm_after' ) if "mlp.fc1" in name: __lowercase =name.replace('mlp.fc1' , 'intermediate.dense' ) if "mlp.fc2" in name: __lowercase =name.replace('mlp.fc2' , 'output.dense' ) if name == "encoder.norm.weight": __lowercase ='layernorm.weight' if name == "encoder.norm.bias": __lowercase ='layernorm.bias' if "decoder" in name: pass else: __lowercase ='swin.' + name return name def _A ( _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" for key in orig_state_dict.copy().keys(): __lowercase =orig_state_dict.pop(_lowerCAmelCase ) if "attn_mask" in key: pass elif "qkv" in key: __lowercase =key.split('.' ) __lowercase =int(key_split[2] ) __lowercase =int(key_split[4] ) __lowercase =model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: __lowercase =val[:dim, :] __lowercase =val[ dim : dim * 2, : ] __lowercase =val[-dim:, :] else: __lowercase =val[ :dim ] __lowercase =val[ dim : dim * 2 ] __lowercase =val[ -dim: ] else: __lowercase =val return orig_state_dict def _A ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =torch.load(_lowerCAmelCase , map_location='cpu' )['model'] __lowercase =get_swin_config(_lowerCAmelCase ) __lowercase =SwinForMaskedImageModeling(_lowerCAmelCase ) model.eval() __lowercase =convert_state_dict(_lowerCAmelCase , _lowerCAmelCase ) model.load_state_dict(_lowerCAmelCase ) __lowercase ='http://images.cocodataset.org/val2017/000000039769.jpg' __lowercase =ViTImageProcessor(size={'height': 192, 'width': 192} ) __lowercase =Image.open(requests.get(_lowerCAmelCase , stream=_lowerCAmelCase ).raw ) __lowercase =image_processor(images=_lowerCAmelCase , return_tensors='pt' ) with torch.no_grad(): __lowercase =model(**_lowerCAmelCase ).logits print(outputs.keys() ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(f"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_lowerCAmelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_lowerCAmelCase ) if push_to_hub: print(f"""Pushing model and image processor for {model_name} to hub""" ) model.push_to_hub(f"""microsoft/{model_name}""" ) image_processor.push_to_hub(f"""microsoft/{model_name}""" ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""swin-base-simmim-window6-192""", type=str, choices=["""swin-base-simmim-window6-192""", """swin-large-simmim-window12-192"""], help="""Name of the Swin SimMIM model you'd like to convert.""", ) parser.add_argument( """--checkpoint_path""", default="""/Users/nielsrogge/Documents/SwinSimMIM/simmim_pretrain__swin_base__img192_window6__100ep.pth""", type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) lowerCamelCase = parser.parse_args() convert_swin_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
48
0
import tempfile import unittest import numpy as np from diffusers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionPipeline, PNDMScheduler, ) from diffusers.utils.testing_utils import is_onnx_available, nightly, require_onnxruntime, require_torch_gpu from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class UpperCamelCase__ (lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' lowerCamelCase_ : Union[str, Any] = """hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline""" def _lowercase ( self , UpperCamelCase__=0 ) -> Dict: lowerCamelCase : List[Any] = np.random.RandomState(UpperCamelCase__ ) lowerCamelCase : Optional[Any] = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def _lowercase ( self ) -> Union[str, Any]: lowerCamelCase : List[str] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase : Dict = self.get_dummy_inputs() lowerCamelCase : Optional[Any] = pipe(**UpperCamelCase__ ).images lowerCamelCase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) lowerCamelCase : str = np.array([0.65072, 0.58492, 0.48219, 0.55521, 0.53180, 0.55939, 0.50697, 0.39800, 0.46455] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self ) -> Optional[int]: lowerCamelCase : Union[str, Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) lowerCamelCase : Any = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase : Any = self.get_dummy_inputs() lowerCamelCase : int = pipe(**UpperCamelCase__ ).images lowerCamelCase : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) lowerCamelCase : Any = np.array([0.65863, 0.59425, 0.49326, 0.56313, 0.53875, 0.56627, 0.51065, 0.39777, 0.46330] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self ) -> List[Any]: lowerCamelCase : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) lowerCamelCase : Union[str, Any] = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase : Dict = self.get_dummy_inputs() lowerCamelCase : Any = pipe(**UpperCamelCase__ ).images lowerCamelCase : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) lowerCamelCase : List[Any] = np.array([0.53755, 0.60786, 0.47402, 0.49488, 0.51869, 0.49819, 0.47985, 0.38957, 0.44279] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self ) -> int: lowerCamelCase : Union[str, Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) lowerCamelCase : List[str] = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase : List[str] = self.get_dummy_inputs() lowerCamelCase : str = pipe(**UpperCamelCase__ ).images lowerCamelCase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) lowerCamelCase : Optional[Any] = np.array([0.53755, 0.60786, 0.47402, 0.49488, 0.51869, 0.49819, 0.47985, 0.38957, 0.44279] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self ) -> int: lowerCamelCase : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) lowerCamelCase : Union[str, Any] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase : List[str] = self.get_dummy_inputs() lowerCamelCase : str = pipe(**UpperCamelCase__ ).images lowerCamelCase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) lowerCamelCase : Optional[Any] = np.array([0.53817, 0.60812, 0.47384, 0.49530, 0.51894, 0.49814, 0.47984, 0.38958, 0.44271] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self ) -> Union[str, Any]: lowerCamelCase : Dict = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) lowerCamelCase : List[str] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase : int = self.get_dummy_inputs() lowerCamelCase : Union[str, Any] = pipe(**UpperCamelCase__ ).images lowerCamelCase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) lowerCamelCase : str = np.array([0.53895, 0.60808, 0.47933, 0.49608, 0.51886, 0.49950, 0.48053, 0.38957, 0.44200] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _lowercase ( self ) -> List[Any]: lowerCamelCase : List[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase : Any = self.get_dummy_inputs() lowerCamelCase : Optional[int] = 3 * [inputs["prompt"]] # forward lowerCamelCase : str = pipe(**UpperCamelCase__ ) lowerCamelCase : Optional[Any] = output.images[0, -3:, -3:, -1] lowerCamelCase : Optional[Any] = self.get_dummy_inputs() lowerCamelCase : Optional[Any] = 3 * [inputs.pop("prompt" )] lowerCamelCase : Any = pipe.tokenizer( UpperCamelCase__ , padding="max_length" , max_length=pipe.tokenizer.model_max_length , truncation=UpperCamelCase__ , return_tensors="np" , ) lowerCamelCase : Optional[int] = text_inputs["input_ids"] lowerCamelCase : Dict = pipe.text_encoder(input_ids=text_inputs.astype(np.intaa ) )[0] lowerCamelCase : str = prompt_embeds # forward lowerCamelCase : Tuple = pipe(**UpperCamelCase__ ) lowerCamelCase : Tuple = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 def _lowercase ( self ) -> Tuple: lowerCamelCase : List[Any] = OnnxStableDiffusionPipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase : Optional[int] = self.get_dummy_inputs() lowerCamelCase : Any = 3 * ["this is a negative prompt"] lowerCamelCase : str = negative_prompt lowerCamelCase : Tuple = 3 * [inputs["prompt"]] # forward lowerCamelCase : Dict = pipe(**UpperCamelCase__ ) lowerCamelCase : int = output.images[0, -3:, -3:, -1] lowerCamelCase : List[Any] = self.get_dummy_inputs() lowerCamelCase : Tuple = 3 * [inputs.pop("prompt" )] lowerCamelCase : Tuple = [] for p in [prompt, negative_prompt]: lowerCamelCase : int = pipe.tokenizer( UpperCamelCase__ , padding="max_length" , max_length=pipe.tokenizer.model_max_length , truncation=UpperCamelCase__ , return_tensors="np" , ) lowerCamelCase : Optional[int] = text_inputs["input_ids"] embeds.append(pipe.text_encoder(input_ids=text_inputs.astype(np.intaa ) )[0] ) lowerCamelCase , lowerCamelCase : List[str] = embeds # forward lowerCamelCase : Optional[int] = pipe(**UpperCamelCase__ ) lowerCamelCase : Optional[Any] = output.images[0, -3:, -3:, -1] assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @nightly @require_onnxruntime @require_torch_gpu class UpperCamelCase__ (unittest.TestCase ): '''simple docstring''' @property def _lowercase ( self ) -> int: return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def _lowercase ( self ) -> Optional[int]: lowerCamelCase : str = ort.SessionOptions() lowerCamelCase : str = False return options def _lowercase ( self ) -> Optional[Any]: # using the PNDM scheduler by default lowerCamelCase : Optional[int] = OnnxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="onnx" , safety_checker=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase : Any = "A painting of a squirrel eating a burger" np.random.seed(0 ) lowerCamelCase : Tuple = sd_pipe([prompt] , guidance_scale=6.0 , num_inference_steps=10 , output_type="np" ) lowerCamelCase : str = output.images lowerCamelCase : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) lowerCamelCase : Tuple = np.array([0.0452, 0.0390, 0.0087, 0.0350, 0.0617, 0.0364, 0.0544, 0.0523, 0.0720] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def _lowercase ( self ) -> Optional[int]: lowerCamelCase : List[Any] = DDIMScheduler.from_pretrained( "runwayml/stable-diffusion-v1-5" , subfolder="scheduler" , revision="onnx" ) lowerCamelCase : Optional[Any] = OnnxStableDiffusionPipeline.from_pretrained( "runwayml/stable-diffusion-v1-5" , revision="onnx" , scheduler=UpperCamelCase__ , safety_checker=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase : Union[str, Any] = "open neural network exchange" lowerCamelCase : Tuple = np.random.RandomState(0 ) lowerCamelCase : str = sd_pipe([prompt] , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCamelCase__ , output_type="np" ) lowerCamelCase : Tuple = output.images lowerCamelCase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) lowerCamelCase : Dict = np.array([0.2867, 0.1974, 0.1481, 0.7294, 0.7251, 0.6667, 0.4194, 0.5642, 0.6486] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def _lowercase ( self ) -> Any: lowerCamelCase : Any = LMSDiscreteScheduler.from_pretrained( "runwayml/stable-diffusion-v1-5" , subfolder="scheduler" , revision="onnx" ) lowerCamelCase : Any = OnnxStableDiffusionPipeline.from_pretrained( "runwayml/stable-diffusion-v1-5" , revision="onnx" , scheduler=UpperCamelCase__ , safety_checker=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) sd_pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase : int = "open neural network exchange" lowerCamelCase : Optional[Any] = np.random.RandomState(0 ) lowerCamelCase : int = sd_pipe([prompt] , guidance_scale=7.5 , num_inference_steps=10 , generator=UpperCamelCase__ , output_type="np" ) lowerCamelCase : Optional[Any] = output.images lowerCamelCase : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) lowerCamelCase : List[Any] = np.array([0.2306, 0.1959, 0.1593, 0.6549, 0.6394, 0.5408, 0.5065, 0.6010, 0.6161] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def _lowercase ( self ) -> int: lowerCamelCase : List[str] = 0 def test_callback_fn(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> None: lowerCamelCase : Dict = True nonlocal number_of_steps number_of_steps += 1 if step == 0: assert latents.shape == (1, 4, 64, 64) lowerCamelCase : int = latents[0, -3:, -3:, -1] lowerCamelCase : Optional[int] = np.array( [-0.6772, -0.3835, -1.2456, 0.1905, -1.0974, 0.6967, -1.9353, 0.0178, 1.0167] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 1e-3 elif step == 5: assert latents.shape == (1, 4, 64, 64) lowerCamelCase : Union[str, Any] = latents[0, -3:, -3:, -1] lowerCamelCase : List[str] = np.array( [-0.3351, 0.2241, -0.1837, -0.2325, -0.6577, 0.3393, -0.0241, 0.5899, 1.3875] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 1e-3 lowerCamelCase : List[str] = False lowerCamelCase : str = OnnxStableDiffusionPipeline.from_pretrained( "runwayml/stable-diffusion-v1-5" , revision="onnx" , safety_checker=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCamelCase : Union[str, Any] = "Andromeda galaxy in a bottle" lowerCamelCase : List[str] = np.random.RandomState(0 ) pipe( prompt=UpperCamelCase__ , num_inference_steps=5 , guidance_scale=7.5 , generator=UpperCamelCase__ , callback=UpperCamelCase__ , callback_steps=1 , ) assert test_callback_fn.has_been_called assert number_of_steps == 6 def _lowercase ( self ) -> str: lowerCamelCase : int = OnnxStableDiffusionPipeline.from_pretrained( "runwayml/stable-diffusion-v1-5" , revision="onnx" , safety_checker=UpperCamelCase__ , feature_extractor=UpperCamelCase__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) assert isinstance(UpperCamelCase__ , UpperCamelCase__ ) assert pipe.safety_checker is None lowerCamelCase : Optional[int] = pipe("example prompt" , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(UpperCamelCase__ ) lowerCamelCase : int = OnnxStableDiffusionPipeline.from_pretrained(UpperCamelCase__ ) # sanity check that the pipeline still works assert pipe.safety_checker is None lowerCamelCase : Tuple = pipe("example prompt" , num_inference_steps=2 ).images[0] assert image is not None
48
from functools import lru_cache @lru_cache def _snake_case ( lowerCAmelCase : int ): """simple docstring""" if num < 0: raise ValueError("Number should not be negative." ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
18
0
'''simple docstring''' import math import unittest def _A (lowerCAmelCase__ :int ) -> bool: '''simple docstring''' assert isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(lowerCAmelCase__ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True class a ( unittest.TestCase ): def __UpperCAmelCase ( self ) -> Optional[int]: self.assertTrue(is_prime(2 ) ) self.assertTrue(is_prime(3 ) ) self.assertTrue(is_prime(5 ) ) self.assertTrue(is_prime(7 ) ) self.assertTrue(is_prime(11 ) ) self.assertTrue(is_prime(13 ) ) self.assertTrue(is_prime(17 ) ) self.assertTrue(is_prime(19 ) ) self.assertTrue(is_prime(23 ) ) self.assertTrue(is_prime(29 ) ) def __UpperCAmelCase ( self ) -> Tuple: with self.assertRaises(__magic_name__ ): is_prime(-19 ) self.assertFalse( is_prime(0 ) , 'Zero doesn\'t have any positive factors, primes must have exactly two.' , ) self.assertFalse( is_prime(1 ) , 'One only has 1 positive factor, primes must have exactly two.' , ) self.assertFalse(is_prime(2 * 2 ) ) self.assertFalse(is_prime(2 * 3 ) ) self.assertFalse(is_prime(3 * 3 ) ) self.assertFalse(is_prime(3 * 5 ) ) self.assertFalse(is_prime(3 * 5 * 7 ) ) if __name__ == "__main__": unittest.main()
104
'''simple docstring''' import colorsys from PIL import Image # type: ignore def _A (lowerCAmelCase__ :float , lowerCAmelCase__ :float , lowerCAmelCase__ :int ) -> float: '''simple docstring''' _a = x _a = y for step in range(lowerCAmelCase__ ): # noqa: B007 _a = a * a - b * b + x _a = 2 * a * b + y _a = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def _A (lowerCAmelCase__ :float ) -> tuple: '''simple docstring''' if distance == 1: return (0, 0, 0) else: return (2_55, 2_55, 2_55) def _A (lowerCAmelCase__ :float ) -> tuple: '''simple docstring''' if distance == 1: return (0, 0, 0) else: return tuple(round(i * 2_55 ) for i in colorsys.hsv_to_rgb(lowerCAmelCase__ , 1 , 1 ) ) def _A (lowerCAmelCase__ :int = 8_00 , lowerCAmelCase__ :int = 6_00 , lowerCAmelCase__ :float = -0.6 , lowerCAmelCase__ :float = 0 , lowerCAmelCase__ :float = 3.2 , lowerCAmelCase__ :int = 50 , lowerCAmelCase__ :bool = True , ) -> Image.Image: '''simple docstring''' _a = Image.new('RGB' , (image_width, image_height) ) _a = img.load() # loop through the image-coordinates for image_x in range(lowerCAmelCase__ ): for image_y in range(lowerCAmelCase__ ): # determine the figure-coordinates based on the image-coordinates _a = figure_width / image_width * image_height _a = figure_center_x + (image_x / image_width - 0.5) * figure_width _a = figure_center_y + (image_y / image_height - 0.5) * figure_height _a = get_distance(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: _a = get_color_coded_rgb(lowerCAmelCase__ ) else: _a = get_black_and_white_rgb(lowerCAmelCase__ ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure a_ : Optional[Any] = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
104
1
'''simple docstring''' import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home _lowerCAmelCase = HUGGINGFACE_HUB_CACHE _lowerCAmelCase = '''config.json''' _lowerCAmelCase = '''diffusion_pytorch_model.bin''' _lowerCAmelCase = '''diffusion_flax_model.msgpack''' _lowerCAmelCase = '''model.onnx''' _lowerCAmelCase = '''diffusion_pytorch_model.safetensors''' _lowerCAmelCase = '''weights.pb''' _lowerCAmelCase = '''https://huggingface.co''' _lowerCAmelCase = default_cache_path _lowerCAmelCase = '''diffusers_modules''' _lowerCAmelCase = os.getenv('''HF_MODULES_CACHE''', os.path.join(hf_cache_home, '''modules''')) _lowerCAmelCase = ['''fp16''', '''non-ema'''] _lowerCAmelCase = '''.self_attn'''
298
'''simple docstring''' def __lowerCAmelCase ( snake_case__ , snake_case__ , snake_case__ ): def count_of_possible_combinations(snake_case__ ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(snake_case__ ) def __lowerCAmelCase ( snake_case__ , snake_case__ , snake_case__ ): def count_of_possible_combinations_with_dp_array( snake_case__ , snake_case__ ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] __UpperCamelCase : Any = sum( count_of_possible_combinations_with_dp_array(target - item , snake_case__ ) for item in array ) __UpperCamelCase : List[str] = answer return answer __UpperCamelCase : Optional[int] = [-1] * (target + 1) return count_of_possible_combinations_with_dp_array(snake_case__ , snake_case__ ) def __lowerCAmelCase ( snake_case__ , snake_case__ , snake_case__ ): __UpperCamelCase : Optional[int] = [0] * (target + 1) __UpperCamelCase : Tuple = 1 for i in range(1 , target + 1 ): for j in range(snake_case__ ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() _lowerCAmelCase = 3 _lowerCAmelCase = 5 _lowerCAmelCase = [1, 2, 5] print(combination_sum_iv(n, array, target))
298
1
import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ : List[str] =logging.get_logger(__name__) lowerCAmelCase__ : str ={'''vocab_file''': '''vocab.txt'''} lowerCAmelCase__ : List[Any] ={ '''vocab_file''': { '''openbmb/cpm-ant-10b''': '''https://huggingface.co/openbmb/cpm-ant-10b/blob/main/vocab.txt''', }, } lowerCAmelCase__ : Optional[Any] ={ '''openbmb/cpm-ant-10b''': 1024, } def __lowercase ( a__ ) -> Optional[Any]: __SCREAMING_SNAKE_CASE = collections.OrderedDict() with open(a__ , 'r' , encoding='utf-8' ) as reader: __SCREAMING_SNAKE_CASE = reader.readlines() for index, token in enumerate(a__ ): __SCREAMING_SNAKE_CASE = token.rstrip('\n' ) __SCREAMING_SNAKE_CASE = index return vocab class UpperCAmelCase_ ( UpperCamelCase_ ): '''simple docstring''' def __init__( self , _A , _A="<unk>" , _A=200 ): '''simple docstring''' __SCREAMING_SNAKE_CASE = vocab __SCREAMING_SNAKE_CASE = unk_token __SCREAMING_SNAKE_CASE = max_input_chars_per_word def _A ( self , _A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = list(_A ) if len(_A ) > self.max_input_chars_per_word: return [self.unk_token] __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = [] while start < len(_A ): __SCREAMING_SNAKE_CASE = len(_A ) __SCREAMING_SNAKE_CASE = None while start < end: __SCREAMING_SNAKE_CASE = ''.join(chars[start:end] ) if substr in self.vocab: __SCREAMING_SNAKE_CASE = substr break end -= 1 if cur_substr is None: sub_tokens.append(self.unk_token ) start += 1 else: sub_tokens.append(_A ) __SCREAMING_SNAKE_CASE = end return sub_tokens class UpperCAmelCase_ ( UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ : Tuple = VOCAB_FILES_NAMES UpperCamelCase__ : Any = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : Any = ['''input_ids''', '''attention_mask'''] UpperCamelCase__ : Dict = False def __init__( self , _A , _A="<d>" , _A="</d>" , _A="<s>" , _A="</s>" , _A="<pad>" , _A="<unk>" , _A="</n>" , _A="</_>" , _A="left" , **_A , ): '''simple docstring''' requires_backends(self , ['jieba'] ) super().__init__( bod_token=_A , eod_token=_A , bos_token=_A , eos_token=_A , pad_token=_A , unk_token=_A , line_token=_A , space_token=_A , padding_side=_A , **_A , ) __SCREAMING_SNAKE_CASE = bod_token __SCREAMING_SNAKE_CASE = eod_token __SCREAMING_SNAKE_CASE = load_vocab(_A ) __SCREAMING_SNAKE_CASE = self.encoder[space_token] __SCREAMING_SNAKE_CASE = self.encoder[line_token] del self.encoder[space_token] del self.encoder[line_token] __SCREAMING_SNAKE_CASE = collections.OrderedDict(sorted(self.encoder.items() , key=lambda _A : x[1] ) ) __SCREAMING_SNAKE_CASE = {v: k for k, v in self.encoder.items()} __SCREAMING_SNAKE_CASE = WordpieceTokenizer(vocab=self.encoder , unk_token=self.unk_token ) @property def _A ( self ): '''simple docstring''' return self.encoder[self.bod_token] @property def _A ( self ): '''simple docstring''' return self.encoder[self.eod_token] @property def _A ( self ): '''simple docstring''' return self.encoder["\n"] @property def _A ( self ): '''simple docstring''' return len(self.encoder ) def _A ( self ): '''simple docstring''' return dict(self.encoder , **self.added_tokens_encoder ) def _A ( self , _A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = [] for x in jieba.cut(_A , cut_all=_A ): output_tokens.extend(self.wordpiece_tokenizer.tokenize(_A ) ) return output_tokens def _A ( self , _A , **_A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = [i for i in token_ids if i >= 0] __SCREAMING_SNAKE_CASE = [ x for x in token_ids if x != self.pad_token_id and x != self.eos_token_id and x != self.bos_token_id ] return super()._decode(_A , **_A ) def _A ( self , _A ): '''simple docstring''' return token in self.encoder def _A ( self , _A ): '''simple docstring''' return "".join(_A ) def _A ( self , _A ): '''simple docstring''' return self.encoder.get(_A , self.encoder.get(self.unk_token ) ) def _A ( self , _A ): '''simple docstring''' return self.decoder.get(_A , self.unk_token ) def _A ( self , _A , _A = None ): '''simple docstring''' if os.path.isdir(_A ): __SCREAMING_SNAKE_CASE = os.path.join( _A , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) else: __SCREAMING_SNAKE_CASE = (filename_prefix + '-' if filename_prefix else '') + save_directory __SCREAMING_SNAKE_CASE = 0 if " " in self.encoder: __SCREAMING_SNAKE_CASE = self.encoder[' '] del self.encoder[" "] if "\n" in self.encoder: __SCREAMING_SNAKE_CASE = self.encoder['\n'] del self.encoder["\n"] __SCREAMING_SNAKE_CASE = collections.OrderedDict(sorted(self.encoder.items() , key=lambda _A : x[1] ) ) with open(_A , 'w' , encoding='utf-8' ) as writer: for token, token_index in self.encoder.items(): if index != token_index: logger.warning( f"""Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.""" ' Please check that the vocabulary is not corrupted!' ) __SCREAMING_SNAKE_CASE = token_index writer.write(token + '\n' ) index += 1 return (vocab_file,) def _A ( self , _A , _A = None ): '''simple docstring''' if token_ids_a is None: return [self.bos_token_id] + token_ids_a return [self.bos_token_id] + token_ids_a + [self.bos_token_id] + token_ids_a def _A ( self , _A , _A = None , _A = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_A , token_ids_a=_A , already_has_special_tokens=_A ) if token_ids_a is not None: return [1] + ([0] * len(_A )) + [1] + ([0] * len(_A )) return [1] + ([0] * len(_A ))
118
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase__ : List[str] ={ '''configuration_timesformer''': ['''TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TimesformerConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : List[str] =[ '''TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TimesformerModel''', '''TimesformerForVideoClassification''', '''TimesformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys lowerCAmelCase__ : Union[str, Any] =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
118
1
import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": lowercase__ :Any = pd.read_csv("sample_data.csv", header=None) lowercase__ :Optional[Any] = df.shape[:1][0] # If you're using some other dataset input the target column lowercase__ :Any = df.iloc[:, 1:2] lowercase__ :Tuple = actual_data.values.reshape(len_data, 1) lowercase__ :int = MinMaxScaler().fit_transform(actual_data) lowercase__ :Optional[Any] = 10 lowercase__ :Union[str, Any] = 5 lowercase__ :str = 20 lowercase__ :Dict = len_data - periods * look_back lowercase__ :Optional[int] = actual_data[:division] lowercase__ :Any = actual_data[division - look_back :] lowercase__ , lowercase__ :Dict = [], [] lowercase__ , lowercase__ :List[Any] = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) lowercase__ :Optional[Any] = np.array(train_x) lowercase__ :Optional[Any] = np.array(test_x) lowercase__ :int = np.array([list(i.ravel()) for i in train_y]) lowercase__ :Union[str, Any] = np.array([list(i.ravel()) for i in test_y]) lowercase__ :Union[str, Any] = Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss="mean_squared_error", optimizer="adam") lowercase__ :str = model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) lowercase__ :Dict = model.predict(x_test)
101
import random import unittest import torch from diffusers import IFInpaintingPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class lowercase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): lowercase_ : List[Any] =IFInpaintingPipeline lowercase_ : Optional[int] =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} lowercase_ : Any =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS lowercase_ : str =PipelineTesterMixin.required_optional_params - {'''latents'''} def A__ ( self): return self._get_dummy_components() def A__ ( self ,A__ ,A__=0): if str(A__).startswith('''mps'''): lowercase = torch.manual_seed(A__) else: lowercase = torch.Generator(device=A__).manual_seed(A__) lowercase = floats_tensor((1, 3, 3_2, 3_2) ,rng=random.Random(A__)).to(A__) lowercase = floats_tensor((1, 3, 3_2, 3_2) ,rng=random.Random(A__)).to(A__) lowercase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() ,reason='''XFormers attention is only available with CUDA and `xformers` installed''' ,) def A__ ( self): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3) def A__ ( self): self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' ,reason='''float16 requires CUDA''') def A__ ( self): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1) def A__ ( self): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2) def A__ ( self): self._test_save_load_local() def A__ ( self): self._test_inference_batch_single_identical( expected_max_diff=1E-2 ,)
101
1
'''simple docstring''' import pytest from datasets.splits import SplitDict, SplitInfo from datasets.utils.py_utils import asdict @pytest.mark.parametrize( '''split_dict''' , [ SplitDict(), SplitDict({'''train''': SplitInfo(name='''train''' , num_bytes=1_3_3_7 , num_examples=4_2 , dataset_name='''my_dataset''' )} ), SplitDict({'''train''': SplitInfo(name='''train''' , num_bytes=1_3_3_7 , num_examples=4_2 )} ), SplitDict({'''train''': SplitInfo()} ), ] , ) def _lowerCAmelCase ( lowerCamelCase_ : SplitDict ): __lowercase = split_dict._to_yaml_list() assert len(__lowerCAmelCase ) == len(__lowerCAmelCase ) __lowercase = SplitDict._from_yaml_list(__lowerCAmelCase ) for split_name, split_info in split_dict.items(): # dataset_name field is deprecated, and is therefore not part of the YAML dump __lowercase = None # the split name of split_dict takes over the name of the split info object __lowercase = split_name assert split_dict == reloaded @pytest.mark.parametrize( '''split_info''' , [SplitInfo(), SplitInfo(dataset_name=__lowerCAmelCase ), SplitInfo(dataset_name='''my_dataset''' )] ) def _lowerCAmelCase ( lowerCamelCase_ : int ): # For backward compatibility, we need asdict(split_dict) to return split info dictrionaries with the "dataset_name" # field even if it's deprecated. This way old versionso of `datasets` can still reload dataset_infos.json files __lowercase = asdict(SplitDict({'''train''': split_info} ) ) assert "dataset_name" in split_dict_asdict["train"] assert split_dict_asdict["train"]["dataset_name"] == split_info.dataset_name
353
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { '''s-JoL/Open-Llama-V1''': '''https://huggingface.co/s-JoL/Open-Llama-V1/blob/main/config.json''', } class __lowercase ( lowerCAmelCase__ ): '''simple docstring''' a : Dict = "open-llama" def __init__(self ,_lowerCamelCase=100000 ,_lowerCamelCase=4096 ,_lowerCamelCase=11008 ,_lowerCamelCase=32 ,_lowerCamelCase=32 ,_lowerCamelCase="silu" ,_lowerCamelCase=2048 ,_lowerCamelCase=0.0_2 ,_lowerCamelCase=1E-6 ,_lowerCamelCase=True ,_lowerCamelCase=0 ,_lowerCamelCase=1 ,_lowerCamelCase=2 ,_lowerCamelCase=False ,_lowerCamelCase=True ,_lowerCamelCase=0.1 ,_lowerCamelCase=0.1 ,_lowerCamelCase=True ,_lowerCamelCase=True ,_lowerCamelCase=None ,**_lowerCamelCase ,) -> Union[str, Any]: '''simple docstring''' __lowercase = vocab_size __lowercase = max_position_embeddings __lowercase = hidden_size __lowercase = intermediate_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = initializer_range __lowercase = rms_norm_eps __lowercase = use_cache __lowercase = kwargs.pop( '''use_memorry_efficient_attention''' ,_lowerCamelCase ) __lowercase = hidden_dropout_prob __lowercase = attention_dropout_prob __lowercase = use_stable_embedding __lowercase = shared_input_output_embedding __lowercase = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=_lowerCamelCase ,bos_token_id=_lowerCamelCase ,eos_token_id=_lowerCamelCase ,tie_word_embeddings=_lowerCamelCase ,**_lowerCamelCase ,) def _UpperCAmelCase (self ) -> List[Any]: '''simple docstring''' if self.rope_scaling is None: return if not isinstance(self.rope_scaling ,_lowerCamelCase ) or len(self.rope_scaling ) != 2: raise ValueError( '''`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ''' f"got {self.rope_scaling}" ) __lowercase = self.rope_scaling.get('''type''' ,_lowerCamelCase ) __lowercase = self.rope_scaling.get('''factor''' ,_lowerCamelCase ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}" ) if rope_scaling_factor is None or not isinstance(_lowerCamelCase ,_lowerCamelCase ) or rope_scaling_factor <= 1.0: raise ValueError(f"`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}" )
217
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class a__ ( SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__, unittest.TestCase ): _lowerCamelCase = CycleDiffusionPipeline _lowerCamelCase = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { 'negative_prompt', 'height', 'width', 'negative_prompt_embeds', } _lowerCamelCase = PipelineTesterMixin.required_optional_params - {'latents'} _lowerCamelCase = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'source_prompt'} ) _lowerCamelCase = IMAGE_TO_IMAGE_IMAGE_PARAMS _lowerCamelCase = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase ( self : Tuple ) -> List[str]: torch.manual_seed(0 ) lowercase : Tuple = UNetaDConditionModel( block_out_channels=(32, 64), layers_per_block=2, sample_size=32, in_channels=4, out_channels=4, down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D'), up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D'), cross_attention_dim=32, ) lowercase : Optional[Any] = DDIMScheduler( beta_start=0.0_0085, beta_end=0.012, beta_schedule='scaled_linear', num_train_timesteps=1000, clip_sample=lowerCAmelCase, set_alpha_to_one=lowerCAmelCase, ) torch.manual_seed(0 ) lowercase : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64], in_channels=3, out_channels=3, down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'], up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'], latent_channels=4, ) torch.manual_seed(0 ) lowercase : Any = CLIPTextConfig( bos_token_id=0, eos_token_id=2, hidden_size=32, intermediate_size=37, layer_norm_eps=1e-05, num_attention_heads=4, num_hidden_layers=5, pad_token_id=1, vocab_size=1000, ) lowercase : Optional[int] = CLIPTextModel(lowerCAmelCase ) lowercase : Dict = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) lowercase : Union[str, Any] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowercase ( self : Tuple, lowerCAmelCase : List[Any], lowerCAmelCase : Optional[int]=0 ) -> Any: lowercase : List[Any] = floats_tensor((1, 3, 32, 32), rng=random.Random(lowerCAmelCase ) ).to(lowerCAmelCase ) lowercase : int = image / 2 + 0.5 if str(lowerCAmelCase ).startswith('mps' ): lowercase : int = torch.manual_seed(lowerCAmelCase ) else: lowercase : Union[str, Any] = torch.Generator(device=lowerCAmelCase ).manual_seed(lowerCAmelCase ) lowercase : Optional[Any] = { 'prompt': 'An astronaut riding an elephant', 'source_prompt': 'An astronaut riding a horse', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'eta': 0.1, 'strength': 0.8, 'guidance_scale': 3, 'source_guidance_scale': 1, 'output_type': 'numpy', } return inputs def lowercase ( self : Optional[Any] ) -> Dict: lowercase : List[str] = 'cpu' # ensure determinism for the device-dependent torch.Generator lowercase : List[str] = self.get_dummy_components() lowercase : Tuple = CycleDiffusionPipeline(**lowerCAmelCase ) lowercase : Union[str, Any] = pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) lowercase : Optional[int] = self.get_dummy_inputs(lowerCAmelCase ) lowercase : Optional[int] = pipe(**lowerCAmelCase ) lowercase : int = output.images lowercase : Union[str, Any] = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) lowercase : Tuple = np.array([0.4459, 0.4943, 0.4544, 0.6643, 0.5474, 0.4327, 0.5701, 0.5959, 0.5179] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf(torch_device != 'cuda', 'This test requires a GPU' ) def lowercase ( self : List[str] ) -> List[Any]: lowercase : Optional[int] = self.get_dummy_components() for name, module in components.items(): if hasattr(lowerCAmelCase, 'half' ): lowercase : List[Any] = module.half() lowercase : str = CycleDiffusionPipeline(**lowerCAmelCase ) lowercase : Tuple = pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) lowercase : int = self.get_dummy_inputs(lowerCAmelCase ) lowercase : Tuple = pipe(**lowerCAmelCase ) lowercase : Optional[int] = output.images lowercase : List[str] = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) lowercase : Any = np.array([0.3506, 0.4543, 0.446, 0.4575, 0.5195, 0.4155, 0.5273, 0.518, 0.4116] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def lowercase ( self : Optional[int] ) -> str: return super().test_save_load_local() @unittest.skip('non-deterministic pipeline' ) def lowercase ( self : Union[str, Any] ) -> str: return super().test_inference_batch_single_identical() @skip_mps def lowercase ( self : Any ) -> List[Any]: return super().test_dict_tuple_outputs_equivalent() @skip_mps def lowercase ( self : List[str] ) -> Union[str, Any]: return super().test_save_load_optional_components() @skip_mps def lowercase ( self : Union[str, Any] ) -> Dict: return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class a__ ( unittest.TestCase ): def lowercase ( self : int ) -> Optional[int]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase ( self : Optional[int] ) -> int: lowercase : Any = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/cycle-diffusion/black_colored_car.png' ) lowercase : Optional[Any] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy' ) lowercase : int = init_image.resize((512, 512) ) lowercase : Tuple = 'CompVis/stable-diffusion-v1-4' lowercase : Optional[int] = DDIMScheduler.from_pretrained(lowerCAmelCase, subfolder='scheduler' ) lowercase : Optional[Any] = CycleDiffusionPipeline.from_pretrained( lowerCAmelCase, scheduler=lowerCAmelCase, safety_checker=lowerCAmelCase, torch_dtype=torch.floataa, revision='fp16' ) pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) pipe.enable_attention_slicing() lowercase : Tuple = 'A black colored car' lowercase : Optional[Any] = 'A blue colored car' lowercase : List[Any] = torch.manual_seed(0 ) lowercase : Tuple = pipe( prompt=lowerCAmelCase, source_prompt=lowerCAmelCase, image=lowerCAmelCase, num_inference_steps=100, eta=0.1, strength=0.85, guidance_scale=3, source_guidance_scale=1, generator=lowerCAmelCase, output_type='np', ) lowercase : List[str] = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5e-1 def lowercase ( self : Dict ) -> Any: lowercase : int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/cycle-diffusion/black_colored_car.png' ) lowercase : str = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy' ) lowercase : str = init_image.resize((512, 512) ) lowercase : Any = 'CompVis/stable-diffusion-v1-4' lowercase : Optional[Any] = DDIMScheduler.from_pretrained(lowerCAmelCase, subfolder='scheduler' ) lowercase : Optional[int] = CycleDiffusionPipeline.from_pretrained(lowerCAmelCase, scheduler=lowerCAmelCase, safety_checker=lowerCAmelCase ) pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) pipe.enable_attention_slicing() lowercase : Union[str, Any] = 'A black colored car' lowercase : Any = 'A blue colored car' lowercase : Optional[int] = torch.manual_seed(0 ) lowercase : Tuple = pipe( prompt=lowerCAmelCase, source_prompt=lowerCAmelCase, image=lowerCAmelCase, num_inference_steps=100, eta=0.1, strength=0.85, guidance_scale=3, source_guidance_scale=1, generator=lowerCAmelCase, output_type='np', ) lowercase : List[str] = output.images assert np.abs(image - expected_image ).max() < 2e-2
255
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _UpperCamelCase: List[str] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase: Union[str, Any] = ['NllbTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase: Optional[int] = ['NllbTokenizerFast'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys _UpperCamelCase: Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
255
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCamelCase_ = { "configuration_altclip": [ "ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "AltCLIPConfig", "AltCLIPTextConfig", "AltCLIPVisionConfig", ], "processing_altclip": ["AltCLIPProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ = [ "ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "AltCLIPPreTrainedModel", "AltCLIPModel", "AltCLIPTextModel", "AltCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys lowerCamelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
355
'''simple docstring''' from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def __lowercase ( __lowercase ) -> str: '''simple docstring''' return {key.lstrip("-" ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def __lowercase ( ) -> Tuple: '''simple docstring''' _A = ArgumentParser( "HuggingFace Datasets CLI tool" , usage="datasets-cli <command> [<args>]" , allow_abbrev=__lowercase ) _A = parser.add_subparsers(help="datasets-cli command helpers" ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(__lowercase ) EnvironmentCommand.register_subcommand(__lowercase ) TestCommand.register_subcommand(__lowercase ) RunBeamCommand.register_subcommand(__lowercase ) DummyDataCommand.register_subcommand(__lowercase ) # Parse args _A , _A = parser.parse_known_args() if not hasattr(__lowercase , "func" ): parser.print_help() exit(1 ) _A = parse_unknown_args(__lowercase ) # Run _A = args.func(__lowercase , **__lowercase ) service.run() if __name__ == "__main__": main()
174
0