code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" import re def _lowerCamelCase ( _UpperCamelCase ): '''simple docstring''' return [char.split() for char in re.split(R"[^ a-z A-Z 0-9 \s]" , str_ )] def _lowerCamelCase ( _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = split_input(str_ ) return "".join( ["".join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' try: __lowerCAmelCase = split_input(snake_case_ ) if upper: __lowerCAmelCase = "".join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: __lowerCAmelCase = "".join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def _lowerCamelCase ( _UpperCamelCase ): '''simple docstring''' return to_simple_case(snake_case_ ) def _lowerCamelCase ( _UpperCamelCase ): '''simple docstring''' try: __lowerCAmelCase = to_simple_case(snake_case_ ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' return to_complex_case(snake_case_ , snake_case_ , "_" ) def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' return to_complex_case(snake_case_ , snake_case_ , "-" ) if __name__ == "__main__": __import__("doctest").testmod()
57
import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase , unittest.TestCase ): A_ : Tuple = 'hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline' def a (self : int , a__ : List[Any]=0 ): """simple docstring""" __snake_case = floats_tensor((1, 3, 128, 128) , rng=random.Random(a__ ) ) __snake_case = np.random.RandomState(a__ ) __snake_case = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 3, '''strength''': 0.7_5, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def a (self : List[Any] ): """simple docstring""" __snake_case = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) pipe.set_progress_bar_config(disable=a__ ) __snake_case = self.get_dummy_inputs() __snake_case = pipe(**a__ ).images __snake_case = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 128, 128, 3) __snake_case = np.array([0.6_9_6_4_3, 0.5_8_4_8_4, 0.5_0_3_1_4, 0.5_8_7_6_0, 0.5_5_3_6_8, 0.5_9_6_4_3, 0.5_1_5_2_9, 0.4_1_2_1_7, 0.4_9_0_8_7] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def a (self : Dict ): """simple docstring""" __snake_case = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) __snake_case = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=a__ ) pipe.set_progress_bar_config(disable=a__ ) __snake_case = self.get_dummy_inputs() __snake_case = pipe(**a__ ).images __snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __snake_case = np.array([0.6_1_7_3_7, 0.5_4_6_4_2, 0.5_3_1_8_3, 0.5_4_4_6_5, 0.5_2_7_4_2, 0.6_0_5_2_5, 0.4_9_9_6_9, 0.4_0_6_5_5, 0.4_8_1_5_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def a (self : List[str] ): """simple docstring""" __snake_case = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) __snake_case = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=a__ ) # warmup pass to apply optimizations __snake_case = pipe(**self.get_dummy_inputs() ) __snake_case = self.get_dummy_inputs() __snake_case = pipe(**a__ ).images __snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __snake_case = np.array([0.5_2_7_6_1, 0.5_9_9_7_7, 0.4_9_0_3_3, 0.4_9_6_1_9, 0.5_4_2_8_2, 0.5_0_3_1_1, 0.4_7_6_0_0, 0.4_0_9_1_8, 0.4_5_2_0_3] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def a (self : Any ): """simple docstring""" __snake_case = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) __snake_case = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=a__ ) __snake_case = self.get_dummy_inputs() __snake_case = pipe(**a__ ).images __snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __snake_case = np.array([0.5_2_9_1_1, 0.6_0_0_0_4, 0.4_9_2_2_9, 0.4_9_8_0_5, 0.5_4_5_0_2, 0.5_0_6_8_0, 0.4_7_7_7_7, 0.4_1_0_2_8, 0.4_5_3_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def a (self : Dict ): """simple docstring""" __snake_case = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) __snake_case = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=a__ ) __snake_case = self.get_dummy_inputs() __snake_case = pipe(**a__ ).images __snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __snake_case = np.array([0.5_2_9_1_1, 0.6_0_0_0_4, 0.4_9_2_2_9, 0.4_9_8_0_5, 0.5_4_5_0_2, 0.5_0_6_8_0, 0.4_7_7_7_7, 0.4_1_0_2_8, 0.4_5_3_0_4] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def a (self : List[str] ): """simple docstring""" __snake_case = OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider='''CPUExecutionProvider''' ) __snake_case = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=a__ ) __snake_case = self.get_dummy_inputs() __snake_case = pipe(**a__ ).images __snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __snake_case = np.array([0.6_5_3_3_1, 0.5_8_2_7_7, 0.4_8_2_0_4, 0.5_6_0_5_9, 0.5_3_6_6_5, 0.5_6_2_3_5, 0.5_0_9_6_9, 0.4_0_0_0_9, 0.4_6_5_5_2] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @property def a (self : List[str] ): """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def a (self : Optional[Any] ): """simple docstring""" __snake_case = ort.SessionOptions() __snake_case = False return options def a (self : Optional[Any] ): """simple docstring""" __snake_case = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) __snake_case = init_image.resize((768, 512) ) # using the PNDM scheduler by default __snake_case = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''onnx''' , safety_checker=a__ , feature_extractor=a__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=a__ ) __snake_case = '''A fantasy landscape, trending on artstation''' __snake_case = np.random.RandomState(0 ) __snake_case = pipe( prompt=a__ , image=a__ , strength=0.7_5 , guidance_scale=7.5 , num_inference_steps=10 , generator=a__ , output_type='''np''' , ) __snake_case = output.images __snake_case = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) __snake_case = np.array([0.4_9_0_9, 0.5_0_5_9, 0.5_3_7_2, 0.4_6_2_3, 0.4_8_7_6, 0.5_0_4_9, 0.4_8_2_0, 0.4_9_5_6, 0.5_0_1_9] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def a (self : Dict ): """simple docstring""" __snake_case = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/img2img/sketch-mountains-input.jpg''' ) __snake_case = init_image.resize((768, 512) ) __snake_case = LMSDiscreteScheduler.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , subfolder='''scheduler''' , revision='''onnx''' ) __snake_case = OnnxStableDiffusionImgaImgPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , revision='''onnx''' , scheduler=a__ , safety_checker=a__ , feature_extractor=a__ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=a__ ) __snake_case = '''A fantasy landscape, trending on artstation''' __snake_case = np.random.RandomState(0 ) __snake_case = pipe( prompt=a__ , image=a__ , strength=0.7_5 , guidance_scale=7.5 , num_inference_steps=20 , generator=a__ , output_type='''np''' , ) __snake_case = output.images __snake_case = images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) __snake_case = np.array([0.8_0_4_3, 0.9_2_6, 0.9_5_8_1, 0.8_1_1_9, 0.8_9_5_4, 0.9_1_3, 0.7_2_0_9, 0.7_4_6_3, 0.7_4_3_1] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
24
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCAmelCase : Tuple = { """configuration_xlm_roberta_xl""": [ """XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XLMRobertaXLConfig""", """XLMRobertaXLOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = [ """XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST""", """XLMRobertaXLForCausalLM""", """XLMRobertaXLForMaskedLM""", """XLMRobertaXLForMultipleChoice""", """XLMRobertaXLForQuestionAnswering""", """XLMRobertaXLForSequenceClassification""", """XLMRobertaXLForTokenClassification""", """XLMRobertaXLModel""", """XLMRobertaXLPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaXLConfig, XLMRobertaXLOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaXLForCausalLM, XLMRobertaXLForMaskedLM, XLMRobertaXLForMultipleChoice, XLMRobertaXLForQuestionAnswering, XLMRobertaXLForSequenceClassification, XLMRobertaXLForTokenClassification, XLMRobertaXLModel, XLMRobertaXLPreTrainedModel, ) else: import sys _UpperCAmelCase : Optional[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
352
def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' global f # a global dp table for knapsack if f[i][j] < 0: if j < wt[i - 1]: snake_case_ = mf_knapsack(i - 1 , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) else: snake_case_ = max( mf_knapsack(i - 1 , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) , mf_knapsack(i - 1 , UpperCamelCase__ , UpperCamelCase__ , j - wt[i - 1] ) + val[i - 1] , ) snake_case_ = val return f[i][j] def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' snake_case_ = [[0] * (w + 1) for _ in range(n + 1 )] for i in range(1 , n + 1 ): for w_ in range(1 , w + 1 ): if wt[i - 1] <= w_: snake_case_ = max(val[i - 1] + dp[i - 1][w_ - wt[i - 1]] , dp[i - 1][w_] ) else: snake_case_ = dp[i - 1][w_] return dp[n][w_], dp def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' if not (isinstance(UpperCamelCase__ , (list, tuple) ) and isinstance(UpperCamelCase__ , (list, tuple) )): raise ValueError( 'Both the weights and values vectors must be either lists or tuples' ) snake_case_ = len(UpperCamelCase__ ) if num_items != len(UpperCamelCase__ ): snake_case_ = ( 'The number of weights must be the same as the number of values.\n' F'''But got {num_items} weights and {len(UpperCamelCase__ )} values''' ) raise ValueError(UpperCamelCase__ ) for i in range(UpperCamelCase__ ): if not isinstance(wt[i] , UpperCamelCase__ ): snake_case_ = ( 'All weights must be integers but got weight of ' F'''type {type(wt[i] )} at index {i}''' ) raise TypeError(UpperCamelCase__ ) snake_case_ , snake_case_ = knapsack(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) snake_case_ = set() _construct_solution(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) return optimal_val, example_optional_set def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' if i > 0 and j > 0: if dp[i - 1][j] == dp[i][j]: _construct_solution(UpperCamelCase__ , UpperCamelCase__ , i - 1 , UpperCamelCase__ , UpperCamelCase__ ) else: optimal_set.add(UpperCamelCase__ ) _construct_solution(UpperCamelCase__ , UpperCamelCase__ , i - 1 , j - wt[i - 1] , UpperCamelCase__ ) if __name__ == "__main__": _UpperCAmelCase : int = [3, 2, 4, 4] _UpperCAmelCase : Tuple = [4, 3, 2, 3] _UpperCAmelCase : Dict = 4 _UpperCAmelCase : int = 6 _UpperCAmelCase : Union[str, Any] = [[0] * (w + 1)] + [[0] + [-1] * (w + 1) for _ in range(n + 1)] _UpperCAmelCase , _UpperCAmelCase : Optional[int] = knapsack(w, wt, val, n) print(optimal_solution) print(mf_knapsack(n, wt, val, w)) # switched the n and w # testing the dynamic programming problem with example # the optimal subset for the above example are items 3 and 4 _UpperCAmelCase , _UpperCAmelCase : Optional[int] = knapsack_with_example_solution(w, wt, val) assert optimal_solution == 8 assert optimal_subset == {3, 4} print("""optimal_value = """, optimal_solution) print("""An optimal subset corresponding to the optimal value""", optimal_subset)
200
0
'''simple docstring''' import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> np.array: '''simple docstring''' snake_case_ = F"{sampling_rate}" snake_case_ = '''1''' snake_case_ = '''f32le''' snake_case_ = [ '''ffmpeg''', '''-i''', '''pipe:0''', '''-ac''', ac, '''-ar''', ar, '''-f''', format_for_conversion, '''-hide_banner''', '''-loglevel''', '''quiet''', '''pipe:1''', ] try: with subprocess.Popen(__UpperCAmelCase, stdin=subprocess.PIPE, stdout=subprocess.PIPE ) as ffmpeg_process: snake_case_ = ffmpeg_process.communicate(__UpperCAmelCase ) except FileNotFoundError as error: raise ValueError('''ffmpeg was not found but is required to load audio files from filename''' ) from error snake_case_ = output_stream[0] snake_case_ = np.frombuffer(__UpperCAmelCase, np.floataa ) if audio.shape[0] == 0: raise ValueError('''Malformed soundfile''' ) return audio def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase = "f32le", ) -> List[Any]: '''simple docstring''' snake_case_ = F"{sampling_rate}" snake_case_ = '''1''' if format_for_conversion == "s16le": snake_case_ = 2 elif format_for_conversion == "f32le": snake_case_ = 4 else: raise ValueError(F"Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`" ) snake_case_ = platform.system() if system == "Linux": snake_case_ = '''alsa''' snake_case_ = '''default''' elif system == "Darwin": snake_case_ = '''avfoundation''' snake_case_ = ''':0''' elif system == "Windows": snake_case_ = '''dshow''' snake_case_ = '''default''' snake_case_ = [ '''ffmpeg''', '''-f''', format_, '''-i''', input_, '''-ac''', ac, '''-ar''', ar, '''-f''', format_for_conversion, '''-fflags''', '''nobuffer''', '''-hide_banner''', '''-loglevel''', '''quiet''', '''pipe:1''', ] snake_case_ = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample snake_case_ = _ffmpeg_stream(__UpperCAmelCase, __UpperCAmelCase ) for item in iterator: yield item def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase = None, __UpperCAmelCase = None, __UpperCAmelCase = "f32le", ) -> Dict: '''simple docstring''' if stream_chunk_s is not None: snake_case_ = stream_chunk_s else: snake_case_ = chunk_length_s snake_case_ = ffmpeg_microphone(__UpperCAmelCase, __UpperCAmelCase, format_for_conversion=__UpperCAmelCase ) if format_for_conversion == "s16le": snake_case_ = np.intaa snake_case_ = 2 elif format_for_conversion == "f32le": snake_case_ = np.floataa snake_case_ = 4 else: raise ValueError(F"Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`" ) if stride_length_s is None: snake_case_ = chunk_length_s / 6 snake_case_ = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(__UpperCAmelCase, (int, float) ): snake_case_ = [stride_length_s, stride_length_s] snake_case_ = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample snake_case_ = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample snake_case_ = datetime.datetime.now() snake_case_ = datetime.timedelta(seconds=__UpperCAmelCase ) for item in chunk_bytes_iter(__UpperCAmelCase, __UpperCAmelCase, stride=(stride_left, stride_right), stream=__UpperCAmelCase ): # Put everything back in numpy scale snake_case_ = np.frombuffer(item['''raw'''], dtype=__UpperCAmelCase ) snake_case_ = ( item['''stride'''][0] // size_of_sample, item['''stride'''][1] // size_of_sample, ) snake_case_ = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase = False ) -> Any: '''simple docstring''' snake_case_ = b'''''' snake_case_ ,snake_case_ = stride if stride_left + stride_right >= chunk_len: raise ValueError( F"Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}" ) snake_case_ = 0 for raw in iterator: acc += raw if stream and len(__UpperCAmelCase ) < chunk_len: snake_case_ = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(__UpperCAmelCase ) >= chunk_len: # We are flushing the accumulator snake_case_ = (_stride_left, stride_right) snake_case_ = {'''raw''': acc[:chunk_len], '''stride''': stride} if stream: snake_case_ = False yield item snake_case_ = stride_left snake_case_ = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(__UpperCAmelCase ) > stride_left: snake_case_ = {'''raw''': acc, '''stride''': (_stride_left, 0)} if stream: snake_case_ = False yield item def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' snake_case_ = 2**24 # 16Mo try: with subprocess.Popen(__UpperCAmelCase, stdout=subprocess.PIPE, bufsize=__UpperCAmelCase ) as ffmpeg_process: while True: snake_case_ = ffmpeg_process.stdout.read(__UpperCAmelCase ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('''ffmpeg was not found but is required to stream audio files from filename''' ) from error
56
'''simple docstring''' from collections import defaultdict def __magic_name__ ( __UpperCAmelCase ) -> int: '''simple docstring''' snake_case_ = 1 snake_case_ = True for v in tree[start]: if v not in visited: ret += dfs(__UpperCAmelCase ) if ret % 2 == 0: cuts.append(__UpperCAmelCase ) return ret def __magic_name__ ( ) -> Union[str, Any]: '''simple docstring''' dfs(1 ) if __name__ == "__main__": a ,a : Dict = 10, 9 a : Dict = defaultdict(list) a : dict[int, bool] = {} a : list[int] = [] a : Tuple = 0 a : str = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
56
1
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer __A : Union[str, Any] = logging.get_logger(__name__) __A : Optional[int] = {"vocab_file": "vocab.txt"} __A : Any = { "vocab_file": { "YituTech/conv-bert-base": "https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt", "YituTech/conv-bert-medium-small": ( "https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt" ), "YituTech/conv-bert-small": "https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt", } } __A : List[str] = { "YituTech/conv-bert-base": 512, "YituTech/conv-bert-medium-small": 512, "YituTech/conv-bert-small": 512, } __A : Optional[Any] = { "YituTech/conv-bert-base": {"do_lower_case": True}, "YituTech/conv-bert-medium-small": {"do_lower_case": True}, "YituTech/conv-bert-small": {"do_lower_case": True}, } class __snake_case ( _SCREAMING_SNAKE_CASE): """simple docstring""" lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_INIT_CONFIGURATION lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = ConvBertTokenizer def __init__( self : int , lowerCamelCase : Any=None , lowerCamelCase : List[Any]=None , lowerCamelCase : List[Any]=True , lowerCamelCase : str="[UNK]" , lowerCamelCase : Tuple="[SEP]" , lowerCamelCase : Optional[int]="[PAD]" , lowerCamelCase : Tuple="[CLS]" , lowerCamelCase : Any="[MASK]" , lowerCamelCase : Any=True , lowerCamelCase : List[str]=None , **lowerCamelCase : List[Any] , ) -> Union[str, Any]: super().__init__( lowerCamelCase , tokenizer_file=lowerCamelCase , do_lower_case=lowerCamelCase , unk_token=lowerCamelCase , sep_token=lowerCamelCase , pad_token=lowerCamelCase , cls_token=lowerCamelCase , mask_token=lowerCamelCase , tokenize_chinese_chars=lowerCamelCase , strip_accents=lowerCamelCase , **lowerCamelCase , ) lowerCAmelCase_ : List[str] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , lowerCamelCase ) != do_lower_case or normalizer_state.get("""strip_accents""" , lowerCamelCase ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , lowerCamelCase ) != tokenize_chinese_chars ): lowerCAmelCase_ : Optional[Any] = getattr(lowerCamelCase , normalizer_state.pop("""type""" ) ) lowerCAmelCase_ : Union[str, Any] = do_lower_case lowerCAmelCase_ : Optional[Any] = strip_accents lowerCAmelCase_ : Optional[Any] = tokenize_chinese_chars lowerCAmelCase_ : List[str] = normalizer_class(**lowerCamelCase ) lowerCAmelCase_ : str = do_lower_case def __lowercase ( self : Tuple , lowerCamelCase : List[Any] , lowerCamelCase : Tuple=None ) -> Dict: lowerCAmelCase_ : Union[str, Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __lowercase ( self : int , lowerCamelCase : List[int] , lowerCamelCase : Optional[List[int]] = None ) -> List[int]: lowerCAmelCase_ : Dict = [self.sep_token_id] lowerCAmelCase_ : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __lowercase ( self : Optional[Any] , lowerCamelCase : str , lowerCamelCase : Optional[str] = None ) -> Tuple[str]: lowerCAmelCase_ : int = self._tokenizer.model.save(lowerCamelCase , name=lowerCamelCase ) return tuple(lowerCamelCase )
89
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __A : List[str] = { "configuration_bigbird_pegasus": [ "BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP", "BigBirdPegasusConfig", "BigBirdPegasusOnnxConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Union[str, Any] = [ "BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST", "BigBirdPegasusForCausalLM", "BigBirdPegasusForConditionalGeneration", "BigBirdPegasusForQuestionAnswering", "BigBirdPegasusForSequenceClassification", "BigBirdPegasusModel", "BigBirdPegasusPreTrainedModel", ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys __A : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
89
1
'''simple docstring''' import argparse import os import gluonnlp as nlp import mxnet as mx import numpy as np import torch from gluonnlp.base import get_home_dir from gluonnlp.model.bert import BERTEncoder from gluonnlp.model.utils import _load_vocab from gluonnlp.vocab import Vocab from packaging import version from torch import nn from transformers import BertConfig, BertForMaskedLM, BertModel, RobertaTokenizer from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging if version.parse(nlp.__version__) != version.parse("""0.8.3"""): raise Exception("""requires gluonnlp == 0.8.3""") if version.parse(mx.__version__) != version.parse("""1.5.0"""): raise Exception("""requires mxnet == 1.5.0""") logging.set_verbosity_info() lowercase_ = logging.get_logger(__name__) lowercase_ = """The Nymphenburg Palace is a beautiful palace in Munich!""" def lowerCamelCase ( __lowerCamelCase : str , __lowerCamelCase : str ) ->List[str]: _SCREAMING_SNAKE_CASE = { """attention_cell""": """multi_head""", """num_layers""": 4, """units""": 1024, """hidden_size""": 768, """max_length""": 512, """num_heads""": 8, """scaled""": True, """dropout""": 0.1, """use_residual""": True, """embed_size""": 1024, """embed_dropout""": 0.1, """word_embed""": None, """layer_norm_eps""": 1e-5, """token_type_vocab_size""": 2, } _SCREAMING_SNAKE_CASE = bort_4_8_768_1024_hparams # Let's construct the original Bort model here # Taken from official BERT implementation, see: # https://github.com/alexa/bort/blob/master/bort/bort.py _SCREAMING_SNAKE_CASE = BERTEncoder( attention_cell=predefined_args["""attention_cell"""] , num_layers=predefined_args["""num_layers"""] , units=predefined_args["""units"""] , hidden_size=predefined_args["""hidden_size"""] , max_length=predefined_args["""max_length"""] , num_heads=predefined_args["""num_heads"""] , scaled=predefined_args["""scaled"""] , dropout=predefined_args["""dropout"""] , output_attention=UpperCamelCase__ , output_all_encodings=UpperCamelCase__ , use_residual=predefined_args["""use_residual"""] , activation=predefined_args.get("""activation""" , """gelu""" ) , layer_norm_eps=predefined_args.get("""layer_norm_eps""" , UpperCamelCase__ ) , ) # Vocab information needs to be fetched first # It's the same as RoBERTa, so RobertaTokenizer can be used later _SCREAMING_SNAKE_CASE = """openwebtext_ccnews_stories_books_cased""" # Specify download folder to Gluonnlp's vocab _SCREAMING_SNAKE_CASE = os.path.join(get_home_dir() , """models""" ) _SCREAMING_SNAKE_CASE = _load_vocab(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , cls=UpperCamelCase__ ) _SCREAMING_SNAKE_CASE = nlp.model.BERTModel( UpperCamelCase__ , len(UpperCamelCase__ ) , units=predefined_args["""units"""] , embed_size=predefined_args["""embed_size"""] , embed_dropout=predefined_args["""embed_dropout"""] , word_embed=predefined_args["""word_embed"""] , use_pooler=UpperCamelCase__ , use_token_type_embed=UpperCamelCase__ , token_type_vocab_size=predefined_args["""token_type_vocab_size"""] , use_classifier=UpperCamelCase__ , use_decoder=UpperCamelCase__ , ) original_bort.load_parameters(UpperCamelCase__ , cast_dtype=UpperCamelCase__ , ignore_extra=UpperCamelCase__ ) _SCREAMING_SNAKE_CASE = original_bort._collect_params_with_prefix() # Build our config 🤗 _SCREAMING_SNAKE_CASE = { """architectures""": ["""BertForMaskedLM"""], """attention_probs_dropout_prob""": predefined_args["""dropout"""], """hidden_act""": """gelu""", """hidden_dropout_prob""": predefined_args["""dropout"""], """hidden_size""": predefined_args["""embed_size"""], """initializer_range""": 0.02, """intermediate_size""": predefined_args["""hidden_size"""], """layer_norm_eps""": predefined_args["""layer_norm_eps"""], """max_position_embeddings""": predefined_args["""max_length"""], """model_type""": """bort""", """num_attention_heads""": predefined_args["""num_heads"""], """num_hidden_layers""": predefined_args["""num_layers"""], """pad_token_id""": 1, # 2 = BERT, 1 = RoBERTa """type_vocab_size""": 1, # 2 = BERT, 1 = RoBERTa """vocab_size""": len(UpperCamelCase__ ), } _SCREAMING_SNAKE_CASE = BertConfig.from_dict(UpperCamelCase__ ) _SCREAMING_SNAKE_CASE = BertForMaskedLM(UpperCamelCase__ ) hf_bort_model.eval() # Parameter mapping table (Gluonnlp to Transformers) # * denotes layer index # # | Gluon Parameter | Transformers Parameter # | -------------------------------------------------------------- | ---------------------- # | `encoder.layer_norm.beta` | `bert.embeddings.LayerNorm.bias` # | `encoder.layer_norm.gamma` | `bert.embeddings.LayerNorm.weight` # | `encoder.position_weight` | `bert.embeddings.position_embeddings.weight` # | `word_embed.0.weight` | `bert.embeddings.word_embeddings.weight` # | `encoder.transformer_cells.*.attention_cell.proj_key.bias` | `bert.encoder.layer.*.attention.self.key.bias` # | `encoder.transformer_cells.*.attention_cell.proj_key.weight` | `bert.encoder.layer.*.attention.self.key.weight` # | `encoder.transformer_cells.*.attention_cell.proj_query.bias` | `bert.encoder.layer.*.attention.self.query.bias` # | `encoder.transformer_cells.*.attention_cell.proj_query.weight` | `bert.encoder.layer.*.attention.self.query.weight` # | `encoder.transformer_cells.*.attention_cell.proj_value.bias` | `bert.encoder.layer.*.attention.self.value.bias` # | `encoder.transformer_cells.*.attention_cell.proj_value.weight` | `bert.encoder.layer.*.attention.self.value.weight` # | `encoder.transformer_cells.*.ffn.ffn_2.bias` | `bert.encoder.layer.*.attention.output.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_2.weight` | `bert.encoder.layer.*.attention.output.dense.weight` # | `encoder.transformer_cells.*.layer_norm.beta` | `bert.encoder.layer.*.attention.output.LayerNorm.bias` # | `encoder.transformer_cells.*.layer_norm.gamma` | `bert.encoder.layer.*.attention.output.LayerNorm.weight` # | `encoder.transformer_cells.*.ffn.ffn_1.bias` | `bert.encoder.layer.*.intermediate.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_1.weight` | `bert.encoder.layer.*.intermediate.dense.weight` # | `encoder.transformer_cells.*.ffn.layer_norm.beta` | `bert.encoder.layer.*.output.LayerNorm.bias` # | `encoder.transformer_cells.*.ffn.layer_norm.gamma` | `bert.encoder.layer.*.output.LayerNorm.weight` # | `encoder.transformer_cells.*.proj.bias` | `bert.encoder.layer.*.output.dense.bias` # | `encoder.transformer_cells.*.proj.weight` | `bert.encoder.layer.*.output.dense.weight` # Helper function to convert MXNET Arrays to PyTorch def to_torch(__lowerCamelCase : int ) -> nn.Parameter: return nn.Parameter(torch.FloatTensor(mx_array.data().asnumpy() ) ) # Check param shapes and map new HF param back def check_and_map_params(__lowerCamelCase : Union[str, Any] , __lowerCamelCase : Dict ): _SCREAMING_SNAKE_CASE = hf_param.shape _SCREAMING_SNAKE_CASE = to_torch(params[gluon_param] ) _SCREAMING_SNAKE_CASE = gluon_param.shape assert ( shape_hf == shape_gluon ), F'The gluon parameter {gluon_param} has shape {shape_gluon}, but expects shape {shape_hf} for Transformers' return gluon_param _SCREAMING_SNAKE_CASE = check_and_map_params( hf_bort_model.bert.embeddings.word_embeddings.weight , """word_embed.0.weight""" ) _SCREAMING_SNAKE_CASE = check_and_map_params( hf_bort_model.bert.embeddings.position_embeddings.weight , """encoder.position_weight""" ) _SCREAMING_SNAKE_CASE = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.bias , """encoder.layer_norm.beta""" ) _SCREAMING_SNAKE_CASE = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.weight , """encoder.layer_norm.gamma""" ) # Inspired by RoBERTa conversion script, we just zero them out (Bort does not use them) _SCREAMING_SNAKE_CASE = torch.zeros_like( hf_bort_model.bert.embeddings.token_type_embeddings.weight.data ) for i in range(hf_bort_config.num_hidden_layers ): _SCREAMING_SNAKE_CASE = hf_bort_model.bert.encoder.layer[i] # self attention _SCREAMING_SNAKE_CASE = layer.attention.self _SCREAMING_SNAKE_CASE = check_and_map_params( self_attn.key.bias.data , F'encoder.transformer_cells.{i}.attention_cell.proj_key.bias' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_attn.key.weight.data , F'encoder.transformer_cells.{i}.attention_cell.proj_key.weight' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_attn.query.bias.data , F'encoder.transformer_cells.{i}.attention_cell.proj_query.bias' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_attn.query.weight.data , F'encoder.transformer_cells.{i}.attention_cell.proj_query.weight' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_attn.value.bias.data , F'encoder.transformer_cells.{i}.attention_cell.proj_value.bias' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_attn.value.weight.data , F'encoder.transformer_cells.{i}.attention_cell.proj_value.weight' ) # self attention output _SCREAMING_SNAKE_CASE = layer.attention.output _SCREAMING_SNAKE_CASE = check_and_map_params( self_output.dense.bias , F'encoder.transformer_cells.{i}.proj.bias' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_output.dense.weight , F'encoder.transformer_cells.{i}.proj.weight' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_output.LayerNorm.bias , F'encoder.transformer_cells.{i}.layer_norm.beta' ) _SCREAMING_SNAKE_CASE = check_and_map_params( self_output.LayerNorm.weight , F'encoder.transformer_cells.{i}.layer_norm.gamma' ) # intermediate _SCREAMING_SNAKE_CASE = layer.intermediate _SCREAMING_SNAKE_CASE = check_and_map_params( intermediate.dense.bias , F'encoder.transformer_cells.{i}.ffn.ffn_1.bias' ) _SCREAMING_SNAKE_CASE = check_and_map_params( intermediate.dense.weight , F'encoder.transformer_cells.{i}.ffn.ffn_1.weight' ) # output _SCREAMING_SNAKE_CASE = layer.output _SCREAMING_SNAKE_CASE = check_and_map_params( bert_output.dense.bias , F'encoder.transformer_cells.{i}.ffn.ffn_2.bias' ) _SCREAMING_SNAKE_CASE = check_and_map_params( bert_output.dense.weight , F'encoder.transformer_cells.{i}.ffn.ffn_2.weight' ) _SCREAMING_SNAKE_CASE = check_and_map_params( bert_output.LayerNorm.bias , F'encoder.transformer_cells.{i}.ffn.layer_norm.beta' ) _SCREAMING_SNAKE_CASE = check_and_map_params( bert_output.LayerNorm.weight , F'encoder.transformer_cells.{i}.ffn.layer_norm.gamma' ) # Save space and energy 🎄 hf_bort_model.half() # Compare output of both models _SCREAMING_SNAKE_CASE = RobertaTokenizer.from_pretrained("""roberta-base""" ) _SCREAMING_SNAKE_CASE = tokenizer.encode_plus(UpperCamelCase__ )["""input_ids"""] # Get gluon output _SCREAMING_SNAKE_CASE = mx.nd.array([input_ids] ) _SCREAMING_SNAKE_CASE = original_bort(inputs=UpperCamelCase__ , token_types=[] ) # Get Transformer output (save and reload model again) hf_bort_model.save_pretrained(UpperCamelCase__ ) _SCREAMING_SNAKE_CASE = BertModel.from_pretrained(UpperCamelCase__ ) hf_bort_model.eval() _SCREAMING_SNAKE_CASE = tokenizer.encode_plus(UpperCamelCase__ , return_tensors="""pt""" ) _SCREAMING_SNAKE_CASE = hf_bort_model(**UpperCamelCase__ )[0] _SCREAMING_SNAKE_CASE = output_gluon[0].asnumpy() _SCREAMING_SNAKE_CASE = output_hf[0].detach().numpy() _SCREAMING_SNAKE_CASE = np.max(np.abs(hf_layer - gluon_layer ) ).item() _SCREAMING_SNAKE_CASE = np.allclose(UpperCamelCase__ , UpperCamelCase__ , atol=1e-3 ) if success: print("""✔️ Both model do output the same tensors""" ) else: print("""❌ Both model do **NOT** output the same tensors""" ) print("""Absolute difference is:""" , UpperCamelCase__ ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--bort_checkpoint_path""", default=None, type=str, required=True, help="""Path the official Bort params file.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) lowercase_ = parser.parse_args() convert_bort_checkpoint_to_pytorch(args.bort_checkpoint_path, args.pytorch_dump_folder_path)
58
from math import pi, sqrt, tan def lowerCamelCase_ ( UpperCamelCase__ : float ) -> float: """simple docstring""" if side_length < 0: raise ValueError('surface_area_cube() only accepts non-negative values' ) return 6 * side_length**2 def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if length < 0 or breadth < 0 or height < 0: raise ValueError('surface_area_cuboid() only accepts non-negative values' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def lowerCamelCase_ ( UpperCamelCase__ : float ) -> float: """simple docstring""" if radius < 0: raise ValueError('surface_area_sphere() only accepts non-negative values' ) return 4 * pi * radius**2 def lowerCamelCase_ ( UpperCamelCase__ : float ) -> float: """simple docstring""" if radius < 0: raise ValueError('surface_area_hemisphere() only accepts non-negative values' ) return 3 * pi * radius**2 def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if radius < 0 or height < 0: raise ValueError('surface_area_cone() only accepts non-negative values' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( 'surface_area_conical_frustum() only accepts non-negative values' ) __lowerCamelCase = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if radius < 0 or height < 0: raise ValueError('surface_area_cylinder() only accepts non-negative values' ) return 2 * pi * radius * (height + radius) def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if torus_radius < 0 or tube_radius < 0: raise ValueError('surface_area_torus() only accepts non-negative values' ) if torus_radius < tube_radius: raise ValueError( 'surface_area_torus() does not support spindle or self intersecting tori' ) return 4 * pow(UpperCamelCase__ , 2 ) * torus_radius * tube_radius def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if length < 0 or width < 0: raise ValueError('area_rectangle() only accepts non-negative values' ) return length * width def lowerCamelCase_ ( UpperCamelCase__ : float ) -> float: """simple docstring""" if side_length < 0: raise ValueError('area_square() only accepts non-negative values' ) return side_length**2 def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if base < 0 or height < 0: raise ValueError('area_triangle() only accepts non-negative values' ) return (base * height) / 2 def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('area_triangle_three_sides() only accepts non-negative values' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('Given three sides do not form a triangle' ) __lowerCamelCase = (sidea + sidea + sidea) / 2 __lowerCamelCase = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if base < 0 or height < 0: raise ValueError('area_parallelogram() only accepts non-negative values' ) return base * height def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if basea < 0 or basea < 0 or height < 0: raise ValueError('area_trapezium() only accepts non-negative values' ) return 1 / 2 * (basea + basea) * height def lowerCamelCase_ ( UpperCamelCase__ : float ) -> float: """simple docstring""" if radius < 0: raise ValueError('area_circle() only accepts non-negative values' ) return pi * radius**2 def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if radius_x < 0 or radius_y < 0: raise ValueError('area_ellipse() only accepts non-negative values' ) return pi * radius_x * radius_y def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if diagonal_a < 0 or diagonal_a < 0: raise ValueError('area_rhombus() only accepts non-negative values' ) return 1 / 2 * diagonal_a * diagonal_a def lowerCamelCase_ ( UpperCamelCase__ : int , UpperCamelCase__ : float ) -> float: """simple docstring""" if not isinstance(UpperCamelCase__ , UpperCamelCase__ ) or sides < 3: raise ValueError( 'area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides' ) elif length < 0: raise ValueError( 'area_reg_polygon() only accepts non-negative values as \ length of a side' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print("[DEMO] Areas of various geometric shapes: \n") print(f'''Rectangle: {area_rectangle(10, 20) = }''') print(f'''Square: {area_square(10) = }''') print(f'''Triangle: {area_triangle(10, 10) = }''') print(f'''Triangle: {area_triangle_three_sides(5, 12, 13) = }''') print(f'''Parallelogram: {area_parallelogram(10, 20) = }''') print(f'''Rhombus: {area_rhombus(10, 20) = }''') print(f'''Trapezium: {area_trapezium(10, 20, 30) = }''') print(f'''Circle: {area_circle(20) = }''') print(f'''Ellipse: {area_ellipse(10, 20) = }''') print("\nSurface Areas of various geometric shapes: \n") print(f'''Cube: {surface_area_cube(20) = }''') print(f'''Cuboid: {surface_area_cuboid(10, 20, 30) = }''') print(f'''Sphere: {surface_area_sphere(20) = }''') print(f'''Hemisphere: {surface_area_hemisphere(20) = }''') print(f'''Cone: {surface_area_cone(10, 20) = }''') print(f'''Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }''') print(f'''Cylinder: {surface_area_cylinder(10, 20) = }''') print(f'''Torus: {surface_area_torus(20, 10) = }''') print(f'''Equilateral Triangle: {area_reg_polygon(3, 10) = }''') print(f'''Square: {area_reg_polygon(4, 10) = }''') print(f'''Reqular Pentagon: {area_reg_polygon(5, 10) = }''')
90
0
import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def lowerCamelCase ( a_ , a_ ) -> List[Any]: lowerCAmelCase_ = checkpoint lowerCAmelCase_ = {} lowerCAmelCase_ = vae_state_dict['encoder.conv_in.weight'] lowerCAmelCase_ = vae_state_dict['encoder.conv_in.bias'] lowerCAmelCase_ = vae_state_dict['encoder.conv_out.weight'] lowerCAmelCase_ = vae_state_dict['encoder.conv_out.bias'] lowerCAmelCase_ = vae_state_dict['encoder.norm_out.weight'] lowerCAmelCase_ = vae_state_dict['encoder.norm_out.bias'] lowerCAmelCase_ = vae_state_dict['decoder.conv_in.weight'] lowerCAmelCase_ = vae_state_dict['decoder.conv_in.bias'] lowerCAmelCase_ = vae_state_dict['decoder.conv_out.weight'] lowerCAmelCase_ = vae_state_dict['decoder.conv_out.bias'] lowerCAmelCase_ = vae_state_dict['decoder.norm_out.weight'] lowerCAmelCase_ = vae_state_dict['decoder.norm_out.bias'] lowerCAmelCase_ = vae_state_dict['quant_conv.weight'] lowerCAmelCase_ = vae_state_dict['quant_conv.bias'] lowerCAmelCase_ = vae_state_dict['post_quant_conv.weight'] lowerCAmelCase_ = vae_state_dict['post_quant_conv.bias'] # Retrieves the keys for the encoder down blocks only lowerCAmelCase_ = len({'.'.join(layer.split('.' )[:3] ) for layer in vae_state_dict if 'encoder.down' in layer} ) lowerCAmelCase_ = { layer_id: [key for key in vae_state_dict if F'''down.{layer_id}''' in key] for layer_id in range(a_ ) } # Retrieves the keys for the decoder up blocks only lowerCAmelCase_ = len({'.'.join(layer.split('.' )[:3] ) for layer in vae_state_dict if 'decoder.up' in layer} ) lowerCAmelCase_ = { layer_id: [key for key in vae_state_dict if F'''up.{layer_id}''' in key] for layer_id in range(a_ ) } for i in range(a_ ): lowerCAmelCase_ = [key for key in down_blocks[i] if F'''down.{i}''' in key and F'''down.{i}.downsample''' not in key] if F'''encoder.down.{i}.downsample.conv.weight''' in vae_state_dict: lowerCAmelCase_ = vae_state_dict.pop( F'''encoder.down.{i}.downsample.conv.weight''' ) lowerCAmelCase_ = vae_state_dict.pop( F'''encoder.down.{i}.downsample.conv.bias''' ) lowerCAmelCase_ = renew_vae_resnet_paths(a_ ) lowerCAmelCase_ = {'old': F'''down.{i}.block''', 'new': F'''down_blocks.{i}.resnets'''} assign_to_checkpoint(a_ , a_ , a_ , additional_replacements=[meta_path] , config=a_ ) lowerCAmelCase_ = [key for key in vae_state_dict if 'encoder.mid.block' in key] lowerCAmelCase_ = 2 for i in range(1 , num_mid_res_blocks + 1 ): lowerCAmelCase_ = [key for key in mid_resnets if F'''encoder.mid.block_{i}''' in key] lowerCAmelCase_ = renew_vae_resnet_paths(a_ ) lowerCAmelCase_ = {'old': F'''mid.block_{i}''', 'new': F'''mid_block.resnets.{i - 1}'''} assign_to_checkpoint(a_ , a_ , a_ , additional_replacements=[meta_path] , config=a_ ) lowerCAmelCase_ = [key for key in vae_state_dict if 'encoder.mid.attn' in key] lowerCAmelCase_ = renew_vae_attention_paths(a_ ) lowerCAmelCase_ = {'old': 'mid.attn_1', 'new': 'mid_block.attentions.0'} assign_to_checkpoint(a_ , a_ , a_ , additional_replacements=[meta_path] , config=a_ ) conv_attn_to_linear(a_ ) for i in range(a_ ): lowerCAmelCase_ = num_up_blocks - 1 - i lowerCAmelCase_ = [ key for key in up_blocks[block_id] if F'''up.{block_id}''' in key and F'''up.{block_id}.upsample''' not in key ] if F'''decoder.up.{block_id}.upsample.conv.weight''' in vae_state_dict: lowerCAmelCase_ = vae_state_dict[ F'''decoder.up.{block_id}.upsample.conv.weight''' ] lowerCAmelCase_ = vae_state_dict[ F'''decoder.up.{block_id}.upsample.conv.bias''' ] lowerCAmelCase_ = renew_vae_resnet_paths(a_ ) lowerCAmelCase_ = {'old': F'''up.{block_id}.block''', 'new': F'''up_blocks.{i}.resnets'''} assign_to_checkpoint(a_ , a_ , a_ , additional_replacements=[meta_path] , config=a_ ) lowerCAmelCase_ = [key for key in vae_state_dict if 'decoder.mid.block' in key] lowerCAmelCase_ = 2 for i in range(1 , num_mid_res_blocks + 1 ): lowerCAmelCase_ = [key for key in mid_resnets if F'''decoder.mid.block_{i}''' in key] lowerCAmelCase_ = renew_vae_resnet_paths(a_ ) lowerCAmelCase_ = {'old': F'''mid.block_{i}''', 'new': F'''mid_block.resnets.{i - 1}'''} assign_to_checkpoint(a_ , a_ , a_ , additional_replacements=[meta_path] , config=a_ ) lowerCAmelCase_ = [key for key in vae_state_dict if 'decoder.mid.attn' in key] lowerCAmelCase_ = renew_vae_attention_paths(a_ ) lowerCAmelCase_ = {'old': 'mid.attn_1', 'new': 'mid_block.attentions.0'} assign_to_checkpoint(a_ , a_ , a_ , additional_replacements=[meta_path] , config=a_ ) conv_attn_to_linear(a_ ) return new_checkpoint def lowerCamelCase ( a_ , a_ , ) -> List[Any]: # Only support V1 lowerCAmelCase_ = requests.get( ' https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml' ) lowerCAmelCase_ = io.BytesIO(r.content ) lowerCAmelCase_ = OmegaConf.load(a_ ) lowerCAmelCase_ = 512 lowerCAmelCase_ = 'cuda' if torch.cuda.is_available() else 'cpu' if checkpoint_path.endswith('safetensors' ): from safetensors import safe_open lowerCAmelCase_ = {} with safe_open(a_ , framework='pt' , device='cpu' ) as f: for key in f.keys(): lowerCAmelCase_ = f.get_tensor(a_ ) else: lowerCAmelCase_ = torch.load(a_ , map_location=a_ )['state_dict'] # Convert the VAE model. lowerCAmelCase_ = create_vae_diffusers_config(a_ , image_size=a_ ) lowerCAmelCase_ = custom_convert_ldm_vae_checkpoint(a_ , a_ ) lowerCAmelCase_ = AutoencoderKL(**a_ ) vae.load_state_dict(a_ ) vae.save_pretrained(a_ ) if __name__ == "__main__": lowerCamelCase_ = argparse.ArgumentParser() parser.add_argument("""--vae_pt_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") lowerCamelCase_ = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
14
lowerCamelCase_ = 6_5_5_2_1 def lowerCamelCase ( a_ ) -> int: lowerCAmelCase_ = 1 lowerCAmelCase_ = 0 for plain_chr in plain_text: lowerCAmelCase_ = (a + ord(a_ )) % MOD_ADLER lowerCAmelCase_ = (b + a) % MOD_ADLER return (b << 16) | a
14
1
def _lowerCamelCase( lowercase__ ) -> float: '''simple docstring''' if edge <= 0 or not isinstance(lowercase__ , lowercase__ ): raise ValueError('Length must be a positive.' ) return 3 * ((2_5 + 1_0 * (5 ** (1 / 2))) ** (1 / 2)) * (edge**2) def _lowerCamelCase( lowercase__ ) -> float: '''simple docstring''' if edge <= 0 or not isinstance(lowercase__ , lowercase__ ): raise ValueError('Length must be a positive.' ) return ((1_5 + (7 * (5 ** (1 / 2)))) / 4) * (edge**3) if __name__ == "__main__": import doctest doctest.testmod()
295
import os import re import warnings from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_ta import TaTokenizer else: lowerCAmelCase = None lowerCAmelCase = logging.get_logger(__name__) lowerCAmelCase = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase = { '''vocab_file''': { '''t5-small''': '''https://huggingface.co/t5-small/resolve/main/spiece.model''', '''t5-base''': '''https://huggingface.co/t5-base/resolve/main/spiece.model''', '''t5-large''': '''https://huggingface.co/t5-large/resolve/main/spiece.model''', '''t5-3b''': '''https://huggingface.co/t5-3b/resolve/main/spiece.model''', '''t5-11b''': '''https://huggingface.co/t5-11b/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''t5-small''': '''https://huggingface.co/t5-small/resolve/main/tokenizer.json''', '''t5-base''': '''https://huggingface.co/t5-base/resolve/main/tokenizer.json''', '''t5-large''': '''https://huggingface.co/t5-large/resolve/main/tokenizer.json''', '''t5-3b''': '''https://huggingface.co/t5-3b/resolve/main/tokenizer.json''', '''t5-11b''': '''https://huggingface.co/t5-11b/resolve/main/tokenizer.json''', }, } # TODO(PVP) - this should be removed in Transformers v5 lowerCAmelCase = { '''t5-small''': 5_1_2, '''t5-base''': 5_1_2, '''t5-large''': 5_1_2, '''t5-3b''': 5_1_2, '''t5-11b''': 5_1_2, } class A ( A_ ): UpperCamelCase_ : Dict =VOCAB_FILES_NAMES UpperCamelCase_ : Dict =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ : str =['''input_ids''', '''attention_mask'''] UpperCamelCase_ : List[str] =TaTokenizer UpperCamelCase_ : List[int] =[] def __init__(self , lowerCAmelCase=None , lowerCAmelCase=None , lowerCAmelCase="</s>" , lowerCAmelCase="<unk>" , lowerCAmelCase="<pad>" , lowerCAmelCase=1_0_0 , lowerCAmelCase=None , **lowerCAmelCase , ): # Add extra_ids to the special token list if extra_ids > 0 and additional_special_tokens is None: __lowercase= [f'<extra_id_{i}>' for i in range(lowerCAmelCase )] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra special tokens __lowercase= len(set(filter(lambda lowerCAmelCase : bool('extra_id_' in str(lowerCAmelCase ) ) , lowerCAmelCase ) ) ) if extra_tokens != extra_ids: raise ValueError( f'Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are' ' provided to T5Tokenizer. In this case the additional_special_tokens must include the extra_ids' ' tokens' ) super().__init__( lowerCAmelCase , tokenizer_file=lowerCAmelCase , eos_token=lowerCAmelCase , unk_token=lowerCAmelCase , pad_token=lowerCAmelCase , extra_ids=lowerCAmelCase , additional_special_tokens=lowerCAmelCase , **lowerCAmelCase , ) __lowercase= vocab_file __lowercase= False if not self.vocab_file else True __lowercase= extra_ids @staticmethod def _A (lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): if pretrained_model_name_or_path in TaTokenizerFast.max_model_input_sizes: __lowercase= TaTokenizerFast.max_model_input_sizes[pretrained_model_name_or_path] if init_max_model_length is not None and init_max_model_length != max_model_length: return init_max_model_length elif init_max_model_length is None: warnings.warn( 'This tokenizer was incorrectly instantiated with a model max length of' f' {deprecated_max_model_length} which will be corrected in Transformers v5.\nFor now, this' ' behavior is kept to avoid breaking backwards compatibility when padding/encoding with' ' `truncation is True`.\n- Be aware that you SHOULD NOT rely on' f' {pretrained_model_name_or_path} automatically truncating your input to' f' {deprecated_max_model_length} when padding/encoding.\n- If you want to encode/pad to sequences' f' longer than {deprecated_max_model_length} you can either instantiate this tokenizer with' ' `model_max_length` or pass `max_length` when encoding/padding.\n- To avoid this warning, please' ' instantiate this tokenizer with `model_max_length` set to your preferred value.' , lowerCAmelCase , ) return max_model_length def _A (self , lowerCAmelCase , lowerCAmelCase = None ): if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(lowerCAmelCase ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return __lowercase= os.path.join( lowerCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase ): copyfile(self.vocab_file , lowerCAmelCase ) logger.info(f'Copy vocab file to {out_vocab_file}' ) return (out_vocab_file,) def _A (self , lowerCAmelCase , lowerCAmelCase = None ): __lowercase= token_ids_a + [self.eos_token_id] if token_ids_a is None: return self.prefix_tokens + token_ids_a else: __lowercase= token_ids_a + [self.eos_token_id] return self.prefix_tokens + token_ids_a + token_ids_a def _A (self , lowerCAmelCase , lowerCAmelCase = None ): __lowercase= [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos ) * [0] return len(token_ids_a + eos + token_ids_a + eos ) * [0] def _A (self ): return list( set(filter(lambda lowerCAmelCase : bool(re.search(r'<extra_id_\d+>' , lowerCAmelCase ) ) is not None , self.additional_special_tokens ) ) ) def _A (self ): return [self.convert_tokens_to_ids(lowerCAmelCase ) for token in self.get_sentinel_tokens()]
295
1
'''simple docstring''' from __future__ import annotations from collections.abc import Callable from typing import Generic, TypeVar lowerCAmelCase_ : List[str] = TypeVar('''T''') lowerCAmelCase_ : Union[str, Any] = TypeVar('''U''') class __lowerCAmelCase ( Generic[T, U] ): def __init__(self , lowerCAmelCase__ , lowerCAmelCase__ ): _UpperCAmelCase : Optional[Any] = key _UpperCAmelCase : Optional[int] = val _UpperCAmelCase : DoubleLinkedListNode[T, U] | None = None _UpperCAmelCase : DoubleLinkedListNode[T, U] | None = None def __repr__(self ): return ( F"Node: key: {self.key}, val: {self.val}, " F"has next: {bool(self.next )}, has prev: {bool(self.prev )}" ) class __lowerCAmelCase ( Generic[T, U] ): def __init__(self ): _UpperCAmelCase : DoubleLinkedListNode[T, U] = DoubleLinkedListNode(lowerCAmelCase__ , lowerCAmelCase__ ) _UpperCAmelCase : DoubleLinkedListNode[T, U] = DoubleLinkedListNode(lowerCAmelCase__ , lowerCAmelCase__ ) _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = self.rear, self.head def __repr__(self ): _UpperCAmelCase : Tuple = ["""DoubleLinkedList"""] _UpperCAmelCase : str = self.head while node.next is not None: rep.append(str(lowerCAmelCase__ ) ) _UpperCAmelCase : Tuple = node.next rep.append(str(self.rear ) ) return ",\n ".join(lowerCAmelCase__ ) def snake_case_ (self , lowerCAmelCase__ ): _UpperCAmelCase : Any = self.rear.prev # All nodes other than self.head are guaranteed to have non-None previous assert previous is not None _UpperCAmelCase : Union[str, Any] = node _UpperCAmelCase : str = previous _UpperCAmelCase : List[Any] = node _UpperCAmelCase : Any = self.rear def snake_case_ (self , lowerCAmelCase__ ): if node.prev is None or node.next is None: return None _UpperCAmelCase : Optional[int] = node.next _UpperCAmelCase : str = node.prev _UpperCAmelCase : Optional[Any] = None _UpperCAmelCase : str = None return node class __lowerCAmelCase ( Generic[T, U] ): snake_case : dict[Callable[[T], U], LRUCache[T, U]] = {} def __init__(self , lowerCAmelCase__ ): _UpperCAmelCase : DoubleLinkedList[T, U] = DoubleLinkedList() _UpperCAmelCase : List[Any] = capacity _UpperCAmelCase : Optional[Any] = 0 _UpperCAmelCase : Optional[Any] = 0 _UpperCAmelCase : Optional[Any] = 0 _UpperCAmelCase : dict[T, DoubleLinkedListNode[T, U]] = {} def __repr__(self ): return ( F"CacheInfo(hits={self.hits}, misses={self.miss}, " F"capacity={self.capacity}, current size={self.num_keys})" ) def __contains__(self , lowerCAmelCase__ ): return key in self.cache def snake_case_ (self , lowerCAmelCase__ ): # Note: pythonic interface would throw KeyError rather than return None if key in self.cache: self.hits += 1 _UpperCAmelCase : DoubleLinkedListNode[T, U] = self.cache[key] _UpperCAmelCase : Optional[Any] = self.list.remove(self.cache[key] ) assert node == value_node # node is guaranteed not None because it is in self.cache assert node is not None self.list.add(lowerCAmelCase__ ) return node.val self.miss += 1 return None def snake_case_ (self , lowerCAmelCase__ , lowerCAmelCase__ ): if key not in self.cache: if self.num_keys >= self.capacity: # delete first node (oldest) when over capacity _UpperCAmelCase : List[Any] = self.list.head.next # guaranteed to have a non-None first node when num_keys > 0 # explain to type checker via assertions assert first_node is not None assert first_node.key is not None assert ( self.list.remove(lowerCAmelCase__ ) is not None ) # node guaranteed to be in list assert node.key is not None del self.cache[first_node.key] self.num_keys -= 1 _UpperCAmelCase : Optional[int] = DoubleLinkedListNode(lowerCAmelCase__ , lowerCAmelCase__ ) self.list.add(self.cache[key] ) self.num_keys += 1 else: # bump node to the end of the list, update value _UpperCAmelCase : str = self.list.remove(self.cache[key] ) assert node is not None # node guaranteed to be in list _UpperCAmelCase : Dict = value self.list.add(lowerCAmelCase__ ) @classmethod def snake_case_ (cls , lowerCAmelCase__ = 1_2_8 ): def cache_decorator_inner(lowerCAmelCase__ ) -> Callable[..., U]: def cache_decorator_wrapper(*lowerCAmelCase__ ) -> U: if func not in cls.decorator_function_to_instance_map: _UpperCAmelCase : Union[str, Any] = LRUCache(lowerCAmelCase__ ) _UpperCAmelCase : List[str] = cls.decorator_function_to_instance_map[func].get(args[0] ) if result is None: _UpperCAmelCase : List[Any] = func(*lowerCAmelCase__ ) cls.decorator_function_to_instance_map[func].put(args[0] , lowerCAmelCase__ ) return result def cache_info() -> LRUCache[T, U]: return cls.decorator_function_to_instance_map[func] setattr(lowerCAmelCase__ , """cache_info""" , lowerCAmelCase__ ) # noqa: B010 return cache_decorator_wrapper return cache_decorator_inner if __name__ == "__main__": import doctest doctest.testmod()
170
'''simple docstring''' from __future__ import annotations # This is the precision for this function which can be altered. # It is recommended for users to keep this number greater than or equal to 10. lowerCAmelCase_ : Optional[Any] = 10 def __A ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): for i in range(lowerCAmelCase_ , lowerCAmelCase_ ): if array[i] == target: return i return -1 def __A ( lowerCAmelCase_ , lowerCAmelCase_ ): _UpperCAmelCase : Optional[int] = 0 _UpperCAmelCase : str = len(lowerCAmelCase_ ) while left <= right: if right - left < precision: return lin_search(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _UpperCAmelCase : Tuple = (left + right) // 3 + 1 _UpperCAmelCase : str = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: _UpperCAmelCase : List[Any] = one_third - 1 elif array[two_third] < target: _UpperCAmelCase : Optional[Any] = two_third + 1 else: _UpperCAmelCase : Dict = one_third + 1 _UpperCAmelCase : List[Any] = two_third - 1 else: return -1 def __A ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ): if left < right: if right - left < precision: return lin_search(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _UpperCAmelCase : List[Any] = (left + right) // 3 + 1 _UpperCAmelCase : int = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: return rec_ternary_search(lowerCAmelCase_ , one_third - 1 , lowerCAmelCase_ , lowerCAmelCase_ ) elif array[two_third] < target: return rec_ternary_search(two_third + 1 , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) else: return rec_ternary_search(one_third + 1 , two_third - 1 , lowerCAmelCase_ , lowerCAmelCase_ ) else: return -1 if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase_ : int = input('''Enter numbers separated by comma:\n''').strip() lowerCAmelCase_ : Tuple = [int(item.strip()) for item in user_input.split(''',''')] assert collection == sorted(collection), F"List must be ordered.\n{collection}." lowerCAmelCase_ : Any = int(input('''Enter the number to be found in the list:\n''').strip()) lowerCAmelCase_ : List[str] = ite_ternary_search(collection, target) lowerCAmelCase_ : int = rec_ternary_search(0, len(collection) - 1, collection, target) if resulta != -1: print(F"Iterative search: {target} found at positions: {resulta}") print(F"Recursive search: {target} found at positions: {resulta}") else: print('''Not found''')
170
1
A_ : Union[str, Any] = '''Alexander Joslin''' import operator as op from .stack import Stack def __a ( SCREAMING_SNAKE_CASE ) -> str: '''simple docstring''' __UpperCAmelCase = {'''*''': op.mul, '''/''': op.truediv, '''+''': op.add, '''-''': op.sub} __UpperCAmelCase = Stack() __UpperCAmelCase = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(_snake_case ) ) elif i in operators: # RULE 2 operator_stack.push(_snake_case ) elif i == ")": # RULE 4 __UpperCAmelCase = operator_stack.peek() operator_stack.pop() __UpperCAmelCase = operand_stack.peek() operand_stack.pop() __UpperCAmelCase = operand_stack.peek() operand_stack.pop() __UpperCAmelCase = operators[opr](_snake_case , _snake_case ) operand_stack.push(_snake_case ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": A_ : int = '''(5 + ((4 * 2) * (2 + 3)))''' # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
333
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() snake_case__ : Any = logging.get_logger(__name__) def _snake_case ( _snake_case : List[Any] , _snake_case : Tuple=False ): lowerCAmelCase : List[str] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'''blocks.{i}.norm1.weight''', f'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((f'''blocks.{i}.norm1.bias''', f'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((f'''blocks.{i}.attn.proj.weight''', f'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.attn.proj.bias''', f'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((f'''blocks.{i}.norm2.weight''', f'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((f'''blocks.{i}.norm2.bias''', f'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.weight''', f'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc1.bias''', f'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.weight''', f'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((f'''blocks.{i}.mlp.fc2.bias''', f'''vit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" lowerCAmelCase : Union[str, Any] = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def _snake_case ( _snake_case : Tuple , _snake_case : List[Any] , _snake_case : Tuple=False ): for i in range(config.num_hidden_layers ): if base_model: lowerCAmelCase : Optional[int] = '''''' else: lowerCAmelCase : Union[str, Any] = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCAmelCase : List[Any] = state_dict.pop(f'''blocks.{i}.attn.qkv.weight''' ) lowerCAmelCase : Tuple = state_dict.pop(f'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict lowerCAmelCase : Optional[Any] = in_proj_weight[ : config.hidden_size, : ] lowerCAmelCase : Tuple = in_proj_bias[: config.hidden_size] lowerCAmelCase : Tuple = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCAmelCase : Tuple = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCAmelCase : Union[str, Any] = in_proj_weight[ -config.hidden_size :, : ] lowerCAmelCase : List[Any] = in_proj_bias[-config.hidden_size :] def _snake_case ( _snake_case : Tuple ): lowerCAmelCase : List[Any] = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(_snake_case , _snake_case ) def _snake_case ( _snake_case : Union[str, Any] , _snake_case : Any , _snake_case : List[Any] ): lowerCAmelCase : Optional[int] = dct.pop(_snake_case ) lowerCAmelCase : Union[str, Any] = val def _snake_case ( ): lowerCAmelCase : Any = '''http://images.cocodataset.org/val2017/000000039769.jpg''' lowerCAmelCase : Any = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) return im @torch.no_grad() def _snake_case ( _snake_case : Optional[int] , _snake_case : Optional[Any] ): lowerCAmelCase : Any = ViTConfig() lowerCAmelCase : Any = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": lowerCAmelCase : List[str] = True lowerCAmelCase : int = int(vit_name[-12:-10] ) lowerCAmelCase : List[Any] = int(vit_name[-9:-6] ) else: lowerCAmelCase : str = 1000 lowerCAmelCase : Optional[int] = '''huggingface/label-files''' lowerCAmelCase : Any = '''imagenet-1k-id2label.json''' lowerCAmelCase : Optional[Any] = json.load(open(hf_hub_download(_snake_case , _snake_case , repo_type='''dataset''' ) , '''r''' ) ) lowerCAmelCase : Optional[Any] = {int(_snake_case ): v for k, v in idalabel.items()} lowerCAmelCase : Dict = idalabel lowerCAmelCase : List[Any] = {v: k for k, v in idalabel.items()} lowerCAmelCase : List[str] = int(vit_name[-6:-4] ) lowerCAmelCase : int = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith('''tiny''' ): lowerCAmelCase : str = 192 lowerCAmelCase : int = 768 lowerCAmelCase : List[str] = 12 lowerCAmelCase : str = 3 elif vit_name[9:].startswith('''small''' ): lowerCAmelCase : List[str] = 384 lowerCAmelCase : Optional[int] = 1536 lowerCAmelCase : int = 12 lowerCAmelCase : str = 6 else: pass else: if vit_name[4:].startswith('''small''' ): lowerCAmelCase : List[str] = 768 lowerCAmelCase : Dict = 2304 lowerCAmelCase : Dict = 8 lowerCAmelCase : Tuple = 8 elif vit_name[4:].startswith('''base''' ): pass elif vit_name[4:].startswith('''large''' ): lowerCAmelCase : Union[str, Any] = 1024 lowerCAmelCase : List[Any] = 4096 lowerCAmelCase : Union[str, Any] = 24 lowerCAmelCase : Any = 16 elif vit_name[4:].startswith('''huge''' ): lowerCAmelCase : Any = 1280 lowerCAmelCase : str = 5120 lowerCAmelCase : Tuple = 32 lowerCAmelCase : Tuple = 16 # load original model from timm lowerCAmelCase : Any = timm.create_model(_snake_case , pretrained=_snake_case ) timm_model.eval() # load state_dict of original model, remove and rename some keys lowerCAmelCase : int = timm_model.state_dict() if base_model: remove_classification_head_(_snake_case ) lowerCAmelCase : Optional[Any] = create_rename_keys(_snake_case , _snake_case ) for src, dest in rename_keys: rename_key(_snake_case , _snake_case , _snake_case ) read_in_q_k_v(_snake_case , _snake_case , _snake_case ) # load HuggingFace model if vit_name[-5:] == "in21k": lowerCAmelCase : Any = ViTModel(_snake_case ).eval() else: lowerCAmelCase : Any = ViTForImageClassification(_snake_case ).eval() model.load_state_dict(_snake_case ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: lowerCAmelCase : Dict = DeiTImageProcessor(size=config.image_size ) else: lowerCAmelCase : Union[str, Any] = ViTImageProcessor(size=config.image_size ) lowerCAmelCase : Union[str, Any] = image_processor(images=prepare_img() , return_tensors='''pt''' ) lowerCAmelCase : Dict = encoding['''pixel_values'''] lowerCAmelCase : List[Any] = model(_snake_case ) if base_model: lowerCAmelCase : Dict = timm_model.forward_features(_snake_case ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(_snake_case , outputs.pooler_output , atol=1E-3 ) else: lowerCAmelCase : Dict = timm_model(_snake_case ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_snake_case , outputs.logits , atol=1E-3 ) Path(_snake_case ).mkdir(exist_ok=_snake_case ) print(f'''Saving model {vit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(_snake_case ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(_snake_case ) if __name__ == "__main__": snake_case__ : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--vit_name''', default='''vit_base_patch16_224''', type=str, help='''Name of the ViT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) snake_case__ : int = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
60
0
class __snake_case : def __init__( self ): '''simple docstring''' lowercase : Union[str, Any] = {} def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' print(self.vertex ) for i in self.vertex: print(snake_case ,""" -> """ ,""" -> """.join([str(snake_case ) for j in self.vertex[i]] ) ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ): '''simple docstring''' if from_vertex in self.vertex: self.vertex[from_vertex].append(snake_case ) else: # else make a new vertex lowercase : int = [to_vertex] def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowercase : List[Any] = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(snake_case ,snake_case ) def _SCREAMING_SNAKE_CASE ( self ,snake_case ,snake_case ): '''simple docstring''' lowercase : List[str] = True print(snake_case ,end=""" """ ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(snake_case ,snake_case ) if __name__ == "__main__": lowercase : Tuple = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print("""DFS:""") g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
362
import argparse import json import os from collections import OrderedDict import torch from transformers import LukeConfig, LukeForMaskedLM, MLukeTokenizer, XLMRobertaTokenizer from transformers.tokenization_utils_base import AddedToken @torch.no_grad() def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> int: # Load configuration defined in the metadata file with open(SCREAMING_SNAKE_CASE__ ) as metadata_file: lowercase : Union[str, Any] = json.load(SCREAMING_SNAKE_CASE__ ) lowercase : Tuple = LukeConfig(use_entity_aware_attention=SCREAMING_SNAKE_CASE__ , **metadata["""model_config"""] ) # Load in the weights from the checkpoint_path lowercase : Tuple = torch.load(SCREAMING_SNAKE_CASE__ , map_location="""cpu""" )["""module"""] # Load the entity vocab file lowercase : str = load_original_entity_vocab(SCREAMING_SNAKE_CASE__ ) # add an entry for [MASK2] lowercase : Any = max(entity_vocab.values() ) + 1 config.entity_vocab_size += 1 lowercase : Dict = XLMRobertaTokenizer.from_pretrained(metadata["""model_config"""]["""bert_model_name"""] ) # Add special tokens to the token vocabulary for downstream tasks lowercase : List[Any] = AddedToken("""<ent>""" , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) lowercase : int = AddedToken("""<ent2>""" , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) tokenizer.add_special_tokens({"""additional_special_tokens""": [entity_token_a, entity_token_a]} ) config.vocab_size += 2 print(f"Saving tokenizer to {pytorch_dump_folder_path}" ) tokenizer.save_pretrained(SCREAMING_SNAKE_CASE__ ) with open(os.path.join(SCREAMING_SNAKE_CASE__ , """tokenizer_config.json""" ) , """r""" ) as f: lowercase : List[str] = json.load(SCREAMING_SNAKE_CASE__ ) lowercase : Optional[Any] = """MLukeTokenizer""" with open(os.path.join(SCREAMING_SNAKE_CASE__ , """tokenizer_config.json""" ) , """w""" ) as f: json.dump(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) with open(os.path.join(SCREAMING_SNAKE_CASE__ , MLukeTokenizer.vocab_files_names["""entity_vocab_file"""] ) , """w""" ) as f: json.dump(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowercase : int = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ ) # Initialize the embeddings of the special tokens lowercase : Dict = tokenizer.convert_tokens_to_ids(["""@"""] )[0] lowercase : Dict = tokenizer.convert_tokens_to_ids(["""#"""] )[0] lowercase : int = state_dict["""embeddings.word_embeddings.weight"""] lowercase : Union[str, Any] = word_emb[ent_init_index].unsqueeze(0 ) lowercase : List[str] = word_emb[enta_init_index].unsqueeze(0 ) lowercase : str = torch.cat([word_emb, ent_emb, enta_emb] ) # add special tokens for 'entity_predictions.bias' for bias_name in ["lm_head.decoder.bias", "lm_head.bias"]: lowercase : List[Any] = state_dict[bias_name] lowercase : Any = decoder_bias[ent_init_index].unsqueeze(0 ) lowercase : Tuple = decoder_bias[enta_init_index].unsqueeze(0 ) lowercase : int = torch.cat([decoder_bias, ent_decoder_bias, enta_decoder_bias] ) # Initialize the query layers of the entity-aware self-attention mechanism for layer_index in range(config.num_hidden_layers ): for matrix_name in ["query.weight", "query.bias"]: lowercase : Union[str, Any] = f"encoder.layer.{layer_index}.attention.self." lowercase : List[str] = state_dict[prefix + matrix_name] lowercase : Any = state_dict[prefix + matrix_name] lowercase : str = state_dict[prefix + matrix_name] # Initialize the embedding of the [MASK2] entity using that of the [MASK] entity for downstream tasks lowercase : Any = state_dict["""entity_embeddings.entity_embeddings.weight"""] lowercase : Tuple = entity_emb[entity_vocab["""[MASK]"""]].unsqueeze(0 ) lowercase : Optional[Any] = torch.cat([entity_emb, entity_mask_emb] ) # add [MASK2] for 'entity_predictions.bias' lowercase : Optional[Any] = state_dict["""entity_predictions.bias"""] lowercase : str = entity_prediction_bias[entity_vocab["""[MASK]"""]].unsqueeze(0 ) lowercase : List[str] = torch.cat([entity_prediction_bias, entity_mask_bias] ) lowercase : List[str] = LukeForMaskedLM(config=SCREAMING_SNAKE_CASE__ ).eval() state_dict.pop("""entity_predictions.decoder.weight""" ) state_dict.pop("""lm_head.decoder.weight""" ) state_dict.pop("""lm_head.decoder.bias""" ) lowercase : List[str] = OrderedDict() for key, value in state_dict.items(): if not (key.startswith("""lm_head""" ) or key.startswith("""entity_predictions""" )): lowercase : List[Any] = state_dict[key] else: lowercase : Union[str, Any] = state_dict[key] lowercase , lowercase : int = model.load_state_dict(SCREAMING_SNAKE_CASE__ , strict=SCREAMING_SNAKE_CASE__ ) if set(SCREAMING_SNAKE_CASE__ ) != {"luke.embeddings.position_ids"}: raise ValueError(f"Unexpected unexpected_keys: {unexpected_keys}" ) if set(SCREAMING_SNAKE_CASE__ ) != { "lm_head.decoder.weight", "lm_head.decoder.bias", "entity_predictions.decoder.weight", }: raise ValueError(f"Unexpected missing_keys: {missing_keys}" ) model.tie_weights() assert (model.luke.embeddings.word_embeddings.weight == model.lm_head.decoder.weight).all() assert (model.luke.entity_embeddings.entity_embeddings.weight == model.entity_predictions.decoder.weight).all() # Check outputs lowercase : str = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ , task="""entity_classification""" ) lowercase : str = """ISO 639-3 uses the code fas for the dialects spoken across Iran and アフガニスタン (Afghanistan).""" lowercase : str = (0, 9) lowercase : Dict = tokenizer(SCREAMING_SNAKE_CASE__ , entity_spans=[span] , return_tensors="""pt""" ) lowercase : Any = model(**SCREAMING_SNAKE_CASE__ ) # Verify word hidden states if model_size == "large": raise NotImplementedError else: # base lowercase : List[Any] = torch.Size((1, 33, 768) ) lowercase : str = torch.tensor([[0.0892, 0.0596, -0.2819], [0.0134, 0.1199, 0.0573], [-0.0169, 0.0927, 0.0644]] ) if not (outputs.last_hidden_state.shape == expected_shape): raise ValueError( f"Outputs.last_hidden_state.shape is {outputs.last_hidden_state.shape}, Expected shape is {expected_shape}" ) if not torch.allclose(outputs.last_hidden_state[0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=1e-4 ): raise ValueError # Verify entity hidden states if model_size == "large": raise NotImplementedError else: # base lowercase : Optional[int] = torch.Size((1, 1, 768) ) lowercase : List[Any] = torch.tensor([[-0.1482, 0.0609, 0.0322]] ) if not (outputs.entity_last_hidden_state.shape == expected_shape): raise ValueError( f"Outputs.entity_last_hidden_state.shape is {outputs.entity_last_hidden_state.shape}, Expected shape is" f" {expected_shape}" ) if not torch.allclose(outputs.entity_last_hidden_state[0, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=1e-4 ): raise ValueError # Verify masked word/entity prediction lowercase : Any = MLukeTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ ) lowercase : List[Any] = """Tokyo is the capital of <mask>.""" lowercase : List[Any] = (24, 30) lowercase : int = tokenizer(SCREAMING_SNAKE_CASE__ , entity_spans=[span] , return_tensors="""pt""" ) lowercase : Dict = model(**SCREAMING_SNAKE_CASE__ ) lowercase : Union[str, Any] = encoding["""input_ids"""][0].tolist() lowercase : List[Any] = input_ids.index(tokenizer.convert_tokens_to_ids("""<mask>""" ) ) lowercase : Dict = outputs.logits[0][mask_position_id].argmax(dim=-1 ) assert "Japan" == tokenizer.decode(SCREAMING_SNAKE_CASE__ ) lowercase : Union[str, Any] = outputs.entity_logits[0][0].argmax().item() lowercase : int = [ entity for entity, entity_id in tokenizer.entity_vocab.items() if entity_id == predicted_entity_id ] assert [e for e in multilingual_predicted_entities if e.startswith("""en:""" )][0] == "en:Japan" # Finally, save our PyTorch model and tokenizer print("""Saving PyTorch model to {}""".format(SCREAMING_SNAKE_CASE__ ) ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) def _snake_case( SCREAMING_SNAKE_CASE__ ) -> Dict: lowercase : Optional[int] = ["""[MASK]""", """[PAD]""", """[UNK]"""] lowercase : List[str] = [json.loads(SCREAMING_SNAKE_CASE__ ) for line in open(SCREAMING_SNAKE_CASE__ )] lowercase : int = {} for entry in data: lowercase : Optional[Any] = entry["""id"""] for entity_name, language in entry["entities"]: if entity_name in SPECIAL_TOKENS: lowercase : Optional[Any] = entity_id break lowercase : List[Any] = f"{language}:{entity_name}" lowercase : Union[str, Any] = entity_id return new_mapping if __name__ == "__main__": lowercase : Any = argparse.ArgumentParser() # Required parameters parser.add_argument("""--checkpoint_path""", type=str, help="""Path to a pytorch_model.bin file.""") parser.add_argument( """--metadata_path""", default=None, type=str, help="""Path to a metadata.json file, defining the configuration.""" ) parser.add_argument( """--entity_vocab_path""", default=None, type=str, help="""Path to an entity_vocab.tsv file, containing the entity vocabulary.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to where to dump the output PyTorch model.""" ) parser.add_argument( """--model_size""", default="""base""", type=str, choices=["""base""", """large"""], help="""Size of the model to be converted.""" ) lowercase : str = parser.parse_args() convert_luke_checkpoint( args.checkpoint_path, args.metadata_path, args.entity_vocab_path, args.pytorch_dump_folder_path, args.model_size, )
285
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _UpperCamelCase : int = { "configuration_longt5": ["LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP", "LongT5Config", "LongT5OnnxConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Dict = [ "LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST", "LongT5EncoderModel", "LongT5ForConditionalGeneration", "LongT5Model", "LongT5PreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : Any = [ "FlaxLongT5ForConditionalGeneration", "FlaxLongT5Model", "FlaxLongT5PreTrainedModel", ] if TYPE_CHECKING: from .configuration_longta import LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP, LongTaConfig, LongTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longta import ( LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST, LongTaEncoderModel, LongTaForConditionalGeneration, LongTaModel, LongTaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_longta import ( FlaxLongTaForConditionalGeneration, FlaxLongTaModel, FlaxLongTaPreTrainedModel, ) else: import sys _UpperCamelCase : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
77
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def lowerCAmelCase_ ( A_ ,A_ ,A_): UpperCamelCase__: List[Any] = { "en": "Machine learning is great, isn't it?", "ru": "Машинное обучение - это здорово, не так ли?", "de": "Maschinelles Lernen ist großartig, oder?", } # BLUE scores as follows: # "pair": [fairseq, transformers] UpperCamelCase__: str = { "ru-en": ["[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)", "39.20"], "en-ru": ["[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)", "33.47"], "en-de": ["[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)", "42.83"], "de-en": ["[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)", "41.35"], } UpperCamelCase__: str = F"{src_lang}-{tgt_lang}" UpperCamelCase__: Optional[Any] = F"\n---\nlanguage: \n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt19\n- facebook\nlicense: apache-2.0\ndatasets:\n- wmt19\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}.\n\nFor more details, please see, [Facebook FAIR's WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616).\n\nThe abbreviation FSMT stands for FairSeqMachineTranslation\n\nAll four models are available:\n\n* [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru)\n* [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en)\n* [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de)\n* [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en)\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = \"facebook/wmt19-{src_lang}-{tgt_lang}\"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = \"{texts[src_lang]}\"\ninput_ids = tokenizer.encode(input, return_tensors=\"pt\")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n- The original (and this ported model) doesn't seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981)\n\n## Training data\n\nPretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616).\n\n## Eval results\n\npair | fairseq | transformers\n-------|---------|----------\n{pair} | {scores[pair][0]} | {scores[pair][1]}\n\nThe score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn't support:\n- model ensemble, therefore the best performing checkpoint was ported (``model4.pt``).\n- re-ranking\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=15\nmkdir -p $DATA_DIR\nsacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH=\"src:examples/seq2seq\" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\nnote: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`.\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt19/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)\n\n\n### BibTeX entry and citation info\n\n```bibtex\n@inproceedings{{...,\n year={{2020}},\n title={{Facebook FAIR's WMT19 News Translation Task Submission}},\n author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}},\n booktitle={{Proc. of WMT}},\n}}\n```\n\n\n## TODO\n\n- port model ensemble (fairseq uses 4 model checkpoints)\n\n" os.makedirs(A_ ,exist_ok=A_) UpperCamelCase__: Union[str, Any] = os.path.join(A_ ,"README.md") print(F"Generating {path}") with open(A_ ,"w" ,encoding="utf-8") as f: f.write(A_) # make sure we are under the root of the project A__: Optional[Any] = Path(__file__).resolve().parent.parent.parent A__: Optional[int] = repo_dir / '''model_cards''' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: A__ , A__ , A__: Optional[Any] = model_name.split('''-''') A__: List[str] = model_cards_dir / '''facebook''' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
149
0
'''simple docstring''' from __future__ import annotations def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' if len(snake_case__ ) == 0: return [] A, A : Union[str, Any] = min(snake_case__ ), max(snake_case__ ) A : Tuple = int(max_value - min_value ) + 1 A : list[list] = [[] for _ in range(snake_case__ )] for i in my_list: buckets[int(i - min_value )].append(snake_case__ ) return [v for bucket in buckets for v in sorted(snake_case__ )] if __name__ == "__main__": from doctest import testmod testmod() assert bucket_sort([4, 5, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bucket_sort([0, 1, -10, 15, 2, -2]) == [-10, -2, 0, 1, 2, 15]
311
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class A ( __snake_case ): def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Any: """simple docstring""" super().__init__() # make sure scheduler can always be converted to DDIM A : Dict = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=SCREAMING_SNAKE_CASE , scheduler=SCREAMING_SNAKE_CASE ) @torch.no_grad() def __call__( self , SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = 0.0 , SCREAMING_SNAKE_CASE = 50 , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = "pil" , SCREAMING_SNAKE_CASE = True , ) -> Union[ImagePipelineOutput, Tuple]: """simple docstring""" if isinstance(self.unet.config.sample_size , SCREAMING_SNAKE_CASE ): A : List[Any] = ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size, ) else: A : Optional[int] = (batch_size, self.unet.config.in_channels, *self.unet.config.sample_size) if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) and len(SCREAMING_SNAKE_CASE ) != batch_size: raise ValueError( F'You have passed a list of generators of length {len(SCREAMING_SNAKE_CASE )}, but requested an effective batch' F' size of {batch_size}. Make sure the batch size matches the length of the generators.' ) A : str = randn_tensor(SCREAMING_SNAKE_CASE , generator=SCREAMING_SNAKE_CASE , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(SCREAMING_SNAKE_CASE ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output A : Any = self.unet(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 A : int = self.scheduler.step( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , eta=SCREAMING_SNAKE_CASE , use_clipped_model_output=SCREAMING_SNAKE_CASE , generator=SCREAMING_SNAKE_CASE ).prev_sample A : Dict = (image / 2 + 0.5).clamp(0 , 1 ) A : Optional[int] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": A : int = self.numpy_to_pil(SCREAMING_SNAKE_CASE ) if not return_dict: return (image,) return ImagePipelineOutput(images=SCREAMING_SNAKE_CASE )
311
1
from argparse import ArgumentParser from . import BaseTransformersCLICommand def _UpperCAmelCase ( snake_case ): """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class __lowerCAmelCase ( lowerCamelCase__ ): @staticmethod def snake_case ( _snake_case ): """simple docstring""" _lowerCAmelCase = parser.add_parser("""download""" ) download_parser.add_argument( """--cache-dir""" , type=_snake_case , default=_snake_case , help="""Path to location to store the models""" ) download_parser.add_argument( """--force""" , action="""store_true""" , help="""Force the model to be download even if already in cache-dir""" ) download_parser.add_argument( """--trust-remote-code""" , action="""store_true""" , help="""Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you've reviewed the code as it will execute on your local machine""" , ) download_parser.add_argument("""model""" , type=_snake_case , help="""Name of the model to download""" ) download_parser.set_defaults(func=_snake_case ) def __init__( self , _snake_case , _snake_case , _snake_case , _snake_case ): """simple docstring""" _lowerCAmelCase = model _lowerCAmelCase = cache _lowerCAmelCase = force _lowerCAmelCase = trust_remote_code def snake_case ( self ): """simple docstring""" from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model , cache_dir=self._cache , force_download=self._force , trust_remote_code=self._trust_remote_code )
82
from collections.abc import Iterable from typing import Generic, TypeVar A__ = TypeVar("""_T""") class __lowerCAmelCase ( Generic[_T] ): def __init__( self , _snake_case = None ): """simple docstring""" _lowerCAmelCase = list(iterable or [] ) _lowerCAmelCase = [] def __len__( self ): """simple docstring""" return len(self._stacka ) + len(self._stacka ) def __repr__( self ): """simple docstring""" return F'Queue({tuple(self._stacka[::-1] + self._stacka )})' def snake_case ( self , _snake_case ): """simple docstring""" self._stacka.append(_snake_case ) def snake_case ( self ): """simple docstring""" _lowerCAmelCase = self._stacka.pop _lowerCAmelCase = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop() ) if not self._stacka: raise IndexError("""Queue is empty""" ) return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
82
1
'''simple docstring''' import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_MULTIPLE_CHOICE_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, ) from transformers.models.layoutlmva.modeling_layoutlmva import LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class UpperCAmelCase : def __init__( self : int , __snake_case : Any , __snake_case : List[Any]=2 , __snake_case : List[str]=3 , __snake_case : List[str]=4 , __snake_case : Union[str, Any]=2 , __snake_case : Union[str, Any]=7 , __snake_case : Union[str, Any]=True , __snake_case : Dict=True , __snake_case : Optional[int]=True , __snake_case : Optional[int]=True , __snake_case : int=99 , __snake_case : Optional[int]=36 , __snake_case : Dict=3 , __snake_case : Optional[Any]=4 , __snake_case : Union[str, Any]=37 , __snake_case : str="gelu" , __snake_case : int=0.1 , __snake_case : Optional[int]=0.1 , __snake_case : Dict=5_12 , __snake_case : int=16 , __snake_case : int=2 , __snake_case : List[str]=0.02 , __snake_case : Optional[int]=6 , __snake_case : Optional[int]=6 , __snake_case : Any=3 , __snake_case : Union[str, Any]=4 , __snake_case : Any=None , __snake_case : str=10_00 , ) -> Union[str, Any]: _lowerCAmelCase = parent _lowerCAmelCase = batch_size _lowerCAmelCase = num_channels _lowerCAmelCase = image_size _lowerCAmelCase = patch_size _lowerCAmelCase = text_seq_length _lowerCAmelCase = is_training _lowerCAmelCase = use_input_mask _lowerCAmelCase = use_token_type_ids _lowerCAmelCase = use_labels _lowerCAmelCase = vocab_size _lowerCAmelCase = hidden_size _lowerCAmelCase = num_hidden_layers _lowerCAmelCase = num_attention_heads _lowerCAmelCase = intermediate_size _lowerCAmelCase = hidden_act _lowerCAmelCase = hidden_dropout_prob _lowerCAmelCase = attention_probs_dropout_prob _lowerCAmelCase = max_position_embeddings _lowerCAmelCase = type_vocab_size _lowerCAmelCase = type_sequence_label_size _lowerCAmelCase = initializer_range _lowerCAmelCase = coordinate_size _lowerCAmelCase = shape_size _lowerCAmelCase = num_labels _lowerCAmelCase = num_choices _lowerCAmelCase = scope _lowerCAmelCase = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) _lowerCAmelCase = text_seq_length _lowerCAmelCase = (image_size // patch_size) ** 2 + 1 _lowerCAmelCase = self.text_seq_length + self.image_seq_length def lowercase__ ( self : List[str] ) -> Any: _lowerCAmelCase = ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) _lowerCAmelCase = ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: _lowerCAmelCase = bbox[i, j, 3] _lowerCAmelCase = bbox[i, j, 1] _lowerCAmelCase = t if bbox[i, j, 2] < bbox[i, j, 0]: _lowerCAmelCase = bbox[i, j, 2] _lowerCAmelCase = bbox[i, j, 0] _lowerCAmelCase = t _lowerCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCAmelCase = None if self.use_input_mask: _lowerCAmelCase = random_attention_mask([self.batch_size, self.text_seq_length] ) _lowerCAmelCase = None if self.use_token_type_ids: _lowerCAmelCase = ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) _lowerCAmelCase = None _lowerCAmelCase = None if self.use_labels: _lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCAmelCase = ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) _lowerCAmelCase = LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def lowercase__ ( self : Union[str, Any] , __snake_case : Dict , __snake_case : List[Any] , __snake_case : Optional[Any] , __snake_case : str , __snake_case : List[str] , __snake_case : str , __snake_case : Union[str, Any] , __snake_case : List[str] ) -> Any: _lowerCAmelCase = LayoutLMvaModel(config=__snake_case ) model.to(__snake_case ) model.eval() # text + image _lowerCAmelCase = model(__snake_case , pixel_values=__snake_case ) _lowerCAmelCase = model( __snake_case , bbox=__snake_case , pixel_values=__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case ) _lowerCAmelCase = model(__snake_case , bbox=__snake_case , pixel_values=__snake_case , token_type_ids=__snake_case ) _lowerCAmelCase = model(__snake_case , bbox=__snake_case , pixel_values=__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only _lowerCAmelCase = model(__snake_case ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only _lowerCAmelCase = model(pixel_values=__snake_case ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def lowercase__ ( self : int , __snake_case : Union[str, Any] , __snake_case : List[str] , __snake_case : Optional[int] , __snake_case : List[str] , __snake_case : List[str] , __snake_case : Dict , __snake_case : Union[str, Any] , __snake_case : Optional[Any] ) -> Optional[int]: _lowerCAmelCase = self.num_labels _lowerCAmelCase = LayoutLMvaForSequenceClassification(__snake_case ) model.to(__snake_case ) model.eval() _lowerCAmelCase = model( __snake_case , bbox=__snake_case , pixel_values=__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase__ ( self : Union[str, Any] , __snake_case : List[str] , __snake_case : Optional[Any] , __snake_case : Union[str, Any] , __snake_case : List[Any] , __snake_case : Any , __snake_case : Tuple , __snake_case : Optional[Any] , __snake_case : Dict ) -> Any: _lowerCAmelCase = self.num_labels _lowerCAmelCase = LayoutLMvaForTokenClassification(config=__snake_case ) model.to(__snake_case ) model.eval() _lowerCAmelCase = model( __snake_case , bbox=__snake_case , pixel_values=__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def lowercase__ ( self : int , __snake_case : Dict , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Optional[int] , __snake_case : str , __snake_case : Optional[Any] , __snake_case : Tuple , __snake_case : Union[str, Any] ) -> Dict: _lowerCAmelCase = LayoutLMvaForQuestionAnswering(config=__snake_case ) model.to(__snake_case ) model.eval() _lowerCAmelCase = model( __snake_case , bbox=__snake_case , pixel_values=__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , start_positions=__snake_case , end_positions=__snake_case , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowercase__ ( self : Union[str, Any] ) -> Optional[int]: _lowerCAmelCase = self.prepare_config_and_inputs() ( ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ( _lowerCAmelCase ) , ) = config_and_inputs _lowerCAmelCase = { """input_ids""": input_ids, """bbox""": bbox, """pixel_values""": pixel_values, """token_type_ids""": token_type_ids, """attention_mask""": input_mask, } return config, inputs_dict @require_torch class UpperCAmelCase ( snake_case_ , snake_case_ , unittest.TestCase ): _lowercase: Optional[int] = False _lowercase: Tuple = False _lowercase: Dict = False _lowercase: Dict = ( ( LayoutLMvaModel, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaForQuestionAnswering, ) if is_torch_available() else () ) _lowercase: Optional[int] = ( {'''document-question-answering''': LayoutLMvaForQuestionAnswering, '''feature-extraction''': LayoutLMvaModel} if is_torch_available() else {} ) def lowercase__ ( self : int , __snake_case : Tuple , __snake_case : str , __snake_case : Dict , __snake_case : str , __snake_case : str ) -> int: # `DocumentQuestionAnsweringPipeline` is expected to work with this model, but it combines the text and visual # embedding along the sequence dimension (dim 1), which causes an error during post-processing as `p_mask` has # the sequence dimension of the text embedding only. # (see the line `embedding_output = torch.cat([embedding_output, visual_embeddings], dim=1)`) return True def lowercase__ ( self : List[Any] ) -> str: _lowerCAmelCase = LayoutLMvaModelTester(self ) _lowerCAmelCase = ConfigTester(self , config_class=__snake_case , hidden_size=37 ) def lowercase__ ( self : str , __snake_case : str , __snake_case : Dict , __snake_case : Optional[int]=False ) -> Optional[int]: _lowerCAmelCase = copy.deepcopy(__snake_case ) if model_class in get_values(__snake_case ): _lowerCAmelCase = { k: v.unsqueeze(1 ).expand(-1 , self.model_tester.num_choices , -1 ).contiguous() if isinstance(__snake_case , torch.Tensor ) and v.ndim > 1 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(__snake_case ): _lowerCAmelCase = torch.ones(self.model_tester.batch_size , dtype=torch.long , device=__snake_case ) elif model_class in get_values(__snake_case ): _lowerCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__snake_case ) _lowerCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__snake_case ) elif model_class in [ *get_values(__snake_case ), ]: _lowerCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__snake_case ) elif model_class in [ *get_values(__snake_case ), ]: _lowerCAmelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=torch.long , device=__snake_case , ) return inputs_dict def lowercase__ ( self : List[str] ) -> List[Any]: self.config_tester.run_common_tests() def lowercase__ ( self : str ) -> Tuple: _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__snake_case ) def lowercase__ ( self : Optional[int] ) -> List[str]: _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: _lowerCAmelCase = type self.model_tester.create_and_check_model(*__snake_case ) def lowercase__ ( self : Tuple ) -> List[str]: _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__snake_case ) def lowercase__ ( self : List[Any] ) -> Optional[Any]: _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__snake_case ) def lowercase__ ( self : Dict ) -> Any: _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__snake_case ) @slow def lowercase__ ( self : int ) -> Any: for model_name in LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCAmelCase = LayoutLMvaModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) def UpperCamelCase__ ( ): """simple docstring""" _lowerCAmelCase = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch class UpperCAmelCase ( unittest.TestCase ): @cached_property def lowercase__ ( self : List[str] ) -> Optional[int]: return LayoutLMvaImageProcessor(apply_ocr=__snake_case ) if is_vision_available() else None @slow def lowercase__ ( self : int ) -> Optional[Any]: _lowerCAmelCase = LayoutLMvaModel.from_pretrained("""microsoft/layoutlmv3-base""" ).to(__snake_case ) _lowerCAmelCase = self.default_image_processor _lowerCAmelCase = prepare_img() _lowerCAmelCase = image_processor(images=__snake_case , return_tensors="""pt""" ).pixel_values.to(__snake_case ) _lowerCAmelCase = torch.tensor([[1, 2]] ) _lowerCAmelCase = torch.tensor([[1, 2, 3, 4], [5, 6, 7, 8]] ).unsqueeze(0 ) # forward pass _lowerCAmelCase = model( input_ids=input_ids.to(__snake_case ) , bbox=bbox.to(__snake_case ) , pixel_values=pixel_values.to(__snake_case ) , ) # verify the logits _lowerCAmelCase = torch.Size((1, 1_99, 7_68) ) self.assertEqual(outputs.last_hidden_state.shape , __snake_case ) _lowerCAmelCase = torch.tensor( [[-0.05_29, 0.36_18, 0.16_32], [-0.15_87, -0.16_67, -0.04_00], [-0.15_57, -0.16_71, -0.05_05]] ).to(__snake_case ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , __snake_case , atol=1E-4 ) )
220
'''simple docstring''' import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration A__ : str =5_00_00 A__ : Optional[int] =50_00 A__ , A__ : Optional[int] =os.path.split(__file__) A__ : Tuple =os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" for i in range(lowerCAmelCase ): _lowerCAmelCase = dataset[i] @get_duration def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" for i in range(0 , len(lowerCAmelCase ) , lowerCAmelCase ): _lowerCAmelCase = dataset[i : i + batch_size] @get_duration def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" with dataset.formatted_as(type=lowerCAmelCase ): for i in range(lowerCAmelCase ): _lowerCAmelCase = dataset[i] @get_duration def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" with dataset.formatted_as(type=lowerCAmelCase ): for i in range(0 , lowerCAmelCase , lowerCAmelCase ): _lowerCAmelCase = dataset[i : i + batch_size] def UpperCamelCase__ ( ): """simple docstring""" _lowerCAmelCase = {"""num examples""": SPEED_TEST_N_EXAMPLES} _lowerCAmelCase = [ (read, {"""length""": SMALL_TEST}), (read, {"""length""": SPEED_TEST_N_EXAMPLES}), (read_batch, {"""length""": SPEED_TEST_N_EXAMPLES, """batch_size""": 10}), (read_batch, {"""length""": SPEED_TEST_N_EXAMPLES, """batch_size""": 1_00}), (read_batch, {"""length""": SPEED_TEST_N_EXAMPLES, """batch_size""": 10_00}), (read_formatted, {"""type""": """numpy""", """length""": SMALL_TEST}), (read_formatted, {"""type""": """pandas""", """length""": SMALL_TEST}), (read_formatted, {"""type""": """torch""", """length""": SMALL_TEST}), (read_formatted, {"""type""": """tensorflow""", """length""": SMALL_TEST}), (read_formatted_batch, {"""type""": """numpy""", """length""": SMALL_TEST, """batch_size""": 10}), (read_formatted_batch, {"""type""": """numpy""", """length""": SMALL_TEST, """batch_size""": 10_00}), ] _lowerCAmelCase = [ (read, {"""length""": SMALL_TEST}), (read, {"""length""": SPEED_TEST_N_EXAMPLES}), (read_batch, {"""length""": SPEED_TEST_N_EXAMPLES, """batch_size""": 10}), (read_batch, {"""length""": SPEED_TEST_N_EXAMPLES, """batch_size""": 1_00}), (read_batch, {"""length""": SPEED_TEST_N_EXAMPLES, """batch_size""": 10_00}), (read_formatted, {"""type""": """numpy""", """length""": SMALL_TEST}), (read_formatted_batch, {"""type""": """numpy""", """length""": SMALL_TEST, """batch_size""": 10}), (read_formatted_batch, {"""type""": """numpy""", """length""": SMALL_TEST, """batch_size""": 10_00}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("""generating dataset""" ) _lowerCAmelCase = datasets.Features( {"""list""": datasets.Sequence(datasets.Value("""float32""" ) ), """numbers""": datasets.Value("""float32""" )} ) _lowerCAmelCase = generate_example_dataset( os.path.join(lowerCAmelCase , """dataset.arrow""" ) , lowerCAmelCase , num_examples=lowerCAmelCase , seq_shapes={"""list""": (1_00,)} , ) print("""first set of iterations""" ) for func, kwargs in functions: print(func.__name__ , str(lowerCAmelCase ) ) _lowerCAmelCase = func(lowerCAmelCase , **lowerCAmelCase ) print("""shuffling dataset""" ) _lowerCAmelCase = dataset.shuffle() print("""Second set of iterations (after shuffling""" ) for func, kwargs in functions_shuffled: print("""shuffled """ , func.__name__ , str(lowerCAmelCase ) ) _lowerCAmelCase = func( lowerCAmelCase , **lowerCAmelCase ) with open(lowerCAmelCase , """wb""" ) as f: f.write(json.dumps(lowerCAmelCase ).encode("""utf-8""" ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
220
1
'''simple docstring''' import inspect import math import tempfile import unittest import numpy as np from transformers import ViTMAEConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMAEForPreTraining, ViTMAEModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowercase_ : """simple docstring""" def __init__( self : str ,lowercase__ : Tuple ,lowercase__ : Dict=1_3 ,lowercase__ : List[str]=3_0 ,lowercase__ : Tuple=2 ,lowercase__ : Optional[int]=3 ,lowercase__ : List[str]=True ,lowercase__ : Tuple=True ,lowercase__ : int=3_2 ,lowercase__ : List[str]=5 ,lowercase__ : Tuple=4 ,lowercase__ : Any=3_7 ,lowercase__ : Any="gelu" ,lowercase__ : Union[str, Any]=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : str=1_0 ,lowercase__ : Optional[int]=0.0_2 ,lowercase__ : Union[str, Any]=3 ,lowercase__ : Optional[int]=0.6 ,lowercase__ : List[Any]=None ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = is_training __lowercase = use_labels __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = mask_ratio __lowercase = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) __lowercase = (image_size // patch_size) ** 2 __lowercase = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return ViTMAEConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=lowercase__ ,initializer_range=self.initializer_range ,mask_ratio=self.mask_ratio ,) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : str ,lowercase__ : Optional[int] ,lowercase__ : List[str] ): __lowercase = ViTMAEModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : List[Any] ,lowercase__ : int ,lowercase__ : Optional[Any] ): __lowercase = ViTMAEForPreTraining(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = (self.image_size // self.patch_size) ** 2 __lowercase = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape ,(self.batch_size, num_patches, expected_num_channels) ) # test greyscale images __lowercase = 1 __lowercase = ViTMAEForPreTraining(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowercase = model(lowercase__ ) __lowercase = self.patch_size**2 self.parent.assertEqual(result.logits.shape ,(self.batch_size, num_patches, expected_num_channels) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : str = (ViTMAEModel, ViTMAEForPreTraining) if is_torch_available() else () SCREAMING_SNAKE_CASE : Dict = {'feature-extraction': ViTMAEModel} if is_torch_available() else {} SCREAMING_SNAKE_CASE : Any = False SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : List[str] = False SCREAMING_SNAKE_CASE : Union[str, Any] = False def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = ViTMAEModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.config_tester.run_common_tests() @unittest.skip(reason='''ViTMAE does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase__ ,nn.Linear ) ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Optional[int] ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ): # make masks reproducible np.random.seed(2 ) __lowercase = int((pt_model.config.image_size // pt_model.config.patch_size) ** 2 ) __lowercase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) __lowercase = torch.from_numpy(lowercase__ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument __lowercase = pt_noise super().check_pt_tf_models(lowercase__ ,lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs[0].cpu().numpy() __lowercase = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowercase__ ) __lowercase = model_class.from_pretrained(lowercase__ ) model.to(lowercase__ ) # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) # Make sure we don't have nans __lowercase = after_outputs[0].cpu().numpy() __lowercase = 0 __lowercase = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowercase__ ,1e-5 ) @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.''' ) def SCREAMING_SNAKE_CASE ( self : Tuple ): pass @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.''' ) def SCREAMING_SNAKE_CASE ( self : int ): pass @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load''' ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ViTMAEModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : str ): return ViTImageProcessor.from_pretrained('''facebook/vit-mae-base''' ) if is_vision_available() else None @slow def SCREAMING_SNAKE_CASE ( self : str ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) __lowercase = ViTMAEForPreTraining.from_pretrained('''facebook/vit-mae-base''' ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) __lowercase = ViTMAEConfig() __lowercase = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) __lowercase = np.random.uniform(size=(1, num_patches) ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ,noise=torch.from_numpy(lowercase__ ).to(device=lowercase__ ) ) # verify the logits __lowercase = torch.Size((1, 1_9_6, 7_6_8) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor( [[-0.0_5_4_8, -1.7_0_2_3, -0.9_3_2_5], [0.3_7_2_1, -0.5_6_7_0, -0.2_2_3_3], [0.8_2_3_5, -1.3_8_7_8, -0.3_5_2_4]] ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] ,expected_slice.to(lowercase__ ) ,atol=1e-4 ) )
104
'''simple docstring''' from __future__ import annotations def _A ( A__ ): """simple docstring""" __lowercase = len(A__ ) # We need to create solution object to save path. __lowercase = [[0 for _ in range(A__ )] for _ in range(A__ )] __lowercase = run_maze(A__ , 0 , 0 , A__ ) if solved: print('''\n'''.join(str(A__ ) for row in solutions ) ) else: print('''No solution exists!''' ) return solved def _A ( A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = len(A__ ) # Final check point. if i == j == (size - 1): __lowercase = 1 return True __lowercase = (not i < 0) and (not j < 0) # Check lower bounds __lowercase = (i < size) and (j < size) # Check upper bounds if lower_flag and upper_flag: # check for already visited and block points. __lowercase = (not solutions[i][j]) and (not maze[i][j]) if block_flag: # check visited __lowercase = 1 # check for directions if ( run_maze(A__ , i + 1 , A__ , A__ ) or run_maze(A__ , A__ , j + 1 , A__ ) or run_maze(A__ , i - 1 , A__ , A__ ) or run_maze(A__ , A__ , j - 1 , A__ ) ): return True __lowercase = 0 return False return False if __name__ == "__main__": import doctest doctest.testmod()
104
1
from __future__ import annotations import queue class __lowerCAmelCase : def __init__( self :Union[str, Any] , __magic_name__ :Optional[Any] ): '''simple docstring''' a = data a = None a = None def __A ( ) -> TreeNode: print("""\n********Press N to stop entering at any point of time********\n""" ) a = input("""Enter the value of the root node: """ ).strip().lower() a = queue.Queue() a = TreeNode(int(__lowerCamelCase ) ) q.put(__lowerCamelCase ) while not q.empty(): a = q.get() a = f'Enter the left node of {node_found.data}: ' a = input(__lowerCamelCase ).strip().lower() or """n""" if check == "n": return tree_node a = TreeNode(int(__lowerCamelCase ) ) a = left_node q.put(__lowerCamelCase ) a = f'Enter the right node of {node_found.data}: ' a = input(__lowerCamelCase ).strip().lower() or """n""" if check == "n": return tree_node a = TreeNode(int(__lowerCamelCase ) ) a = right_node q.put(__lowerCamelCase ) raise def __A ( __lowerCamelCase ) -> None: if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return print(node.data , end=""",""" ) pre_order(node.left ) pre_order(node.right ) def __A ( __lowerCamelCase ) -> None: if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return in_order(node.left ) print(node.data , end=""",""" ) in_order(node.right ) def __A ( __lowerCamelCase ) -> None: if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return post_order(node.left ) post_order(node.right ) print(node.data , end=""",""" ) def __A ( __lowerCamelCase ) -> None: if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return a = queue.Queue() q.put(__lowerCamelCase ) while not q.empty(): a = q.get() print(node_dequeued.data , end=""",""" ) if node_dequeued.left: q.put(node_dequeued.left ) if node_dequeued.right: q.put(node_dequeued.right ) def __A ( __lowerCamelCase ) -> None: if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return a = queue.Queue() q.put(__lowerCamelCase ) while not q.empty(): a = [] while not q.empty(): a = q.get() print(node_dequeued.data , end=""",""" ) if node_dequeued.left: list_.append(node_dequeued.left ) if node_dequeued.right: list_.append(node_dequeued.right ) print() for node in list_: q.put(__lowerCamelCase ) def __A ( __lowerCamelCase ) -> None: if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return a = [] a = node while n or stack: while n: # start from root node, find its left child print(n.data , end=""",""" ) stack.append(__lowerCamelCase ) a = n.left # end of while means current node doesn't have left child a = stack.pop() # start to traverse its right child a = n.right def __A ( __lowerCamelCase ) -> None: if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return a = [] a = node while n or stack: while n: stack.append(__lowerCamelCase ) a = n.left a = stack.pop() print(n.data , end=""",""" ) a = n.right def __A ( __lowerCamelCase ) -> None: if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return a , a = [], [] a = node stacka.append(__lowerCamelCase ) while stacka: # to find the reversed order of post order, store it in stack2 a = stacka.pop() if n.left: stacka.append(n.left ) if n.right: stacka.append(n.right ) stacka.append(__lowerCamelCase ) while stacka: # pop up from stack2 will be the post order print(stacka.pop().data , end=""",""" ) def __A ( __lowerCamelCase = "" , __lowerCamelCase=50 , __lowerCamelCase="*" ) -> str: if not s: return "\n" + width * char a , a = divmod(width - len(__lowerCamelCase ) - 2 , 2 ) return f'{left * char} {s} {(left + extra) * char}' if __name__ == "__main__": import doctest doctest.testmod() print(prompt("Binary Tree Traversals")) __UpperCamelCase : TreeNode = build_tree() print(prompt("Pre Order Traversal")) pre_order(node) print(prompt() + "\n") print(prompt("In Order Traversal")) in_order(node) print(prompt() + "\n") print(prompt("Post Order Traversal")) post_order(node) print(prompt() + "\n") print(prompt("Level Order Traversal")) level_order(node) print(prompt() + "\n") print(prompt("Actual Level Order Traversal")) level_order_actual(node) print("*" * 50 + "\n") print(prompt("Pre Order Traversal - Iteration Version")) pre_order_iter(node) print(prompt() + "\n") print(prompt("In Order Traversal - Iteration Version")) in_order_iter(node) print(prompt() + "\n") print(prompt("Post Order Traversal - Iteration Version")) post_order_iter(node) print(prompt())
352
import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_MULTIPLE_CHOICE_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, ) from transformers.models.layoutlmva.modeling_layoutlmva import LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class __lowerCAmelCase : def __init__( self :Optional[int] , __magic_name__ :str , __magic_name__ :int=2 , __magic_name__ :List[str]=3 , __magic_name__ :Optional[int]=4 , __magic_name__ :str=2 , __magic_name__ :Any=7 , __magic_name__ :Optional[Any]=True , __magic_name__ :List[Any]=True , __magic_name__ :Optional[int]=True , __magic_name__ :Dict=True , __magic_name__ :List[Any]=99 , __magic_name__ :Dict=36 , __magic_name__ :Optional[Any]=3 , __magic_name__ :str=4 , __magic_name__ :Optional[Any]=37 , __magic_name__ :Dict="gelu" , __magic_name__ :Any=0.1 , __magic_name__ :Union[str, Any]=0.1 , __magic_name__ :Dict=512 , __magic_name__ :str=16 , __magic_name__ :List[Any]=2 , __magic_name__ :Tuple=0.02 , __magic_name__ :Any=6 , __magic_name__ :Optional[int]=6 , __magic_name__ :Tuple=3 , __magic_name__ :str=4 , __magic_name__ :List[str]=None , __magic_name__ :str=1000 , ): '''simple docstring''' a = parent a = batch_size a = num_channels a = image_size a = patch_size a = text_seq_length a = is_training a = use_input_mask a = use_token_type_ids a = use_labels a = vocab_size a = hidden_size a = num_hidden_layers a = num_attention_heads a = intermediate_size a = hidden_act a = hidden_dropout_prob a = attention_probs_dropout_prob a = max_position_embeddings a = type_vocab_size a = type_sequence_label_size a = initializer_range a = coordinate_size a = shape_size a = num_labels a = num_choices a = scope a = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) a = text_seq_length a = (image_size // patch_size) ** 2 + 1 a = self.text_seq_length + self.image_seq_length def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' a = ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) a = ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: a = bbox[i, j, 3] a = bbox[i, j, 1] a = t if bbox[i, j, 2] < bbox[i, j, 0]: a = bbox[i, j, 2] a = bbox[i, j, 0] a = t a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) a = None if self.use_input_mask: a = random_attention_mask([self.batch_size, self.text_seq_length] ) a = None if self.use_token_type_ids: a = ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) a = None a = None if self.use_labels: a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a = ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) a = LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def lowerCamelCase__ ( self :int , __magic_name__ :Union[str, Any] , __magic_name__ :List[str] , __magic_name__ :List[str] , __magic_name__ :int , __magic_name__ :Tuple , __magic_name__ :int , __magic_name__ :Tuple , __magic_name__ :int ): '''simple docstring''' a = LayoutLMvaModel(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() # text + image a = model(__magic_name__ , pixel_values=__magic_name__ ) a = model( __magic_name__ , bbox=__magic_name__ , pixel_values=__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ ) a = model(__magic_name__ , bbox=__magic_name__ , pixel_values=__magic_name__ , token_type_ids=__magic_name__ ) a = model(__magic_name__ , bbox=__magic_name__ , pixel_values=__magic_name__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only a = model(__magic_name__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only a = model(pixel_values=__magic_name__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def lowerCamelCase__ ( self :Optional[Any] , __magic_name__ :List[Any] , __magic_name__ :Optional[int] , __magic_name__ :Tuple , __magic_name__ :Union[str, Any] , __magic_name__ :int , __magic_name__ :Optional[int] , __magic_name__ :List[str] , __magic_name__ :List[str] ): '''simple docstring''' a = self.num_labels a = LayoutLMvaForSequenceClassification(__magic_name__ ) model.to(__magic_name__ ) model.eval() a = model( __magic_name__ , bbox=__magic_name__ , pixel_values=__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=__magic_name__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase__ ( self :Tuple , __magic_name__ :Dict , __magic_name__ :Optional[Any] , __magic_name__ :Optional[Any] , __magic_name__ :Dict , __magic_name__ :Dict , __magic_name__ :int , __magic_name__ :List[str] , __magic_name__ :Tuple ): '''simple docstring''' a = self.num_labels a = LayoutLMvaForTokenClassification(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() a = model( __magic_name__ , bbox=__magic_name__ , pixel_values=__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=__magic_name__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def lowerCamelCase__ ( self :Optional[int] , __magic_name__ :Tuple , __magic_name__ :str , __magic_name__ :int , __magic_name__ :Optional[int] , __magic_name__ :str , __magic_name__ :List[str] , __magic_name__ :Optional[int] , __magic_name__ :Optional[Any] ): '''simple docstring''' a = LayoutLMvaForQuestionAnswering(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() a = model( __magic_name__ , bbox=__magic_name__ , pixel_values=__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , start_positions=__magic_name__ , end_positions=__magic_name__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase__ ( self :Union[str, Any] ): '''simple docstring''' a = self.prepare_config_and_inputs() ( ( a ) , ( a ) , ( a ) , ( a ) , ( a ) , ( a ) , ( a ) , ( a ) , ) = config_and_inputs a = { """input_ids""": input_ids, """bbox""": bbox, """pixel_values""": pixel_values, """token_type_ids""": token_type_ids, """attention_mask""": input_mask, } return config, inputs_dict @require_torch class __lowerCAmelCase ( __magic_name__ , __magic_name__ , unittest.TestCase ): UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = False UpperCamelCase__ = ( ( LayoutLMvaModel, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaForQuestionAnswering, ) if is_torch_available() else () ) UpperCamelCase__ = ( {'''document-question-answering''': LayoutLMvaForQuestionAnswering, '''feature-extraction''': LayoutLMvaModel} if is_torch_available() else {} ) def lowerCamelCase__ ( self :List[str] , __magic_name__ :Union[str, Any] , __magic_name__ :List[str] , __magic_name__ :Optional[Any] , __magic_name__ :List[str] , __magic_name__ :List[Any] ): '''simple docstring''' return True def lowerCamelCase__ ( self :int ): '''simple docstring''' a = LayoutLMvaModelTester(self ) a = ConfigTester(self , config_class=__magic_name__ , hidden_size=37 ) def lowerCamelCase__ ( self :Optional[Any] , __magic_name__ :Dict , __magic_name__ :Dict , __magic_name__ :Any=False ): '''simple docstring''' a = copy.deepcopy(__magic_name__ ) if model_class in get_values(__magic_name__ ): a = { k: v.unsqueeze(1 ).expand(-1 , self.model_tester.num_choices , -1 ).contiguous() if isinstance(__magic_name__ , torch.Tensor ) and v.ndim > 1 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(__magic_name__ ): a = torch.ones(self.model_tester.batch_size , dtype=torch.long , device=__magic_name__ ) elif model_class in get_values(__magic_name__ ): a = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__magic_name__ ) a = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__magic_name__ ) elif model_class in [ *get_values(__magic_name__ ), ]: a = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__magic_name__ ) elif model_class in [ *get_values(__magic_name__ ), ]: a = torch.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=torch.long , device=__magic_name__ , ) return inputs_dict def lowerCamelCase__ ( self :str ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase__ ( self :Union[str, Any] ): '''simple docstring''' a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__magic_name__ ) def lowerCamelCase__ ( self :Tuple ): '''simple docstring''' a = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: a = type self.model_tester.create_and_check_model(*__magic_name__ ) def lowerCamelCase__ ( self :Tuple ): '''simple docstring''' a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__magic_name__ ) def lowerCamelCase__ ( self :Union[str, Any] ): '''simple docstring''' a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__magic_name__ ) def lowerCamelCase__ ( self :int ): '''simple docstring''' a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__magic_name__ ) @slow def lowerCamelCase__ ( self :int ): '''simple docstring''' for model_name in LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a = LayoutLMvaModel.from_pretrained(__magic_name__ ) self.assertIsNotNone(__magic_name__ ) def __A ( ) -> str: a = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch class __lowerCAmelCase ( unittest.TestCase ): @cached_property def lowerCamelCase__ ( self :Optional[Any] ): '''simple docstring''' return LayoutLMvaImageProcessor(apply_ocr=__magic_name__ ) if is_vision_available() else None @slow def lowerCamelCase__ ( self :Optional[Any] ): '''simple docstring''' a = LayoutLMvaModel.from_pretrained("""microsoft/layoutlmv3-base""" ).to(__magic_name__ ) a = self.default_image_processor a = prepare_img() a = image_processor(images=__magic_name__ , return_tensors="""pt""" ).pixel_values.to(__magic_name__ ) a = torch.tensor([[1, 2]] ) a = torch.tensor([[1, 2, 3, 4], [5, 6, 7, 8]] ).unsqueeze(0 ) # forward pass a = model( input_ids=input_ids.to(__magic_name__ ) , bbox=bbox.to(__magic_name__ ) , pixel_values=pixel_values.to(__magic_name__ ) , ) # verify the logits a = torch.Size((1, 199, 768) ) self.assertEqual(outputs.last_hidden_state.shape , __magic_name__ ) a = torch.tensor( [[-0.0529, 0.3618, 0.1632], [-0.1587, -0.1667, -0.0400], [-0.1557, -0.1671, -0.0505]] ).to(__magic_name__ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , __magic_name__ , atol=1E-4 ) )
347
0
from unittest.mock import patch import pyspark from datasets.packaged_modules.spark.spark import ( Spark, SparkExamplesIterable, _generate_iterable_examples, ) from ..utils import ( require_dill_gt_0_3_2, require_not_windows, ) def UpperCamelCase_( _snake_case : Optional[int] , _snake_case : Optional[Any] ): """simple docstring""" __a =[] for part_id in partition_order: __a =df.where(F'SPARK_PARTITION_ID() = {part_id}' ).collect() for row_idx, row in enumerate(_snake_case ): expected_row_ids_and_row_dicts.append((F'{part_id}_{row_idx}', row.asDict()) ) return expected_row_ids_and_row_dicts @require_not_windows @require_dill_gt_0_3_2 def UpperCamelCase_( ): """simple docstring""" __a =pyspark.sql.SparkSession.builder.master('local[*]' ).appName('pyspark' ).getOrCreate() __a =spark.range(100 ).repartition(1 ) __a =Spark(_snake_case ) # The id ints will be converted to Pyarrow int64s, so each row will be 8 bytes. Setting a max_shard_size of 16 means # that each partition can hold 2 rows. spark_builder._repartition_df_if_needed(max_shard_size=16 ) # Given that the dataframe has 100 rows and each partition has 2 rows, we expect 50 partitions. assert spark_builder.df.rdd.getNumPartitions() == 50 @require_not_windows @require_dill_gt_0_3_2 def UpperCamelCase_( ): """simple docstring""" __a =pyspark.sql.SparkSession.builder.master('local[*]' ).appName('pyspark' ).getOrCreate() __a =spark.range(10 ).repartition(2 ) __a =[1, 0] __a =_generate_iterable_examples(_snake_case , _snake_case ) # Reverse the partitions. __a =_get_expected_row_ids_and_row_dicts_for_partition_order(_snake_case , _snake_case ) for i, (row_id, row_dict) in enumerate(generate_fn() ): __a , __a =expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def UpperCamelCase_( ): """simple docstring""" __a =pyspark.sql.SparkSession.builder.master('local[*]' ).appName('pyspark' ).getOrCreate() __a =spark.range(10 ).repartition(1 ) __a =SparkExamplesIterable(_snake_case ) assert it.n_shards == 1 for i, (row_id, row_dict) in enumerate(_snake_case ): assert row_id == F'0_{i}' assert row_dict == {"id": i} @require_not_windows @require_dill_gt_0_3_2 def UpperCamelCase_( ): """simple docstring""" __a =pyspark.sql.SparkSession.builder.master('local[*]' ).appName('pyspark' ).getOrCreate() __a =spark.range(30 ).repartition(3 ) # Mock the generator so that shuffle reverses the partition indices. with patch('numpy.random.Generator' ) as generator_mock: __a =lambda _snake_case : x.reverse() __a =_get_expected_row_ids_and_row_dicts_for_partition_order(_snake_case , [2, 1, 0] ) __a =SparkExamplesIterable(_snake_case ).shuffle_data_sources(_snake_case ) assert shuffled_it.n_shards == 3 for i, (row_id, row_dict) in enumerate(_snake_case ): __a , __a =expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def UpperCamelCase_( ): """simple docstring""" __a =pyspark.sql.SparkSession.builder.master('local[*]' ).appName('pyspark' ).getOrCreate() __a =spark.range(20 ).repartition(4 ) # Partitions 0 and 2 __a =SparkExamplesIterable(_snake_case ).shard_data_sources(worker_id=0 , num_workers=2 ) assert shard_it_a.n_shards == 2 __a =_get_expected_row_ids_and_row_dicts_for_partition_order(_snake_case , [0, 2] ) for i, (row_id, row_dict) in enumerate(_snake_case ): __a , __a =expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict # Partitions 1 and 3 __a =SparkExamplesIterable(_snake_case ).shard_data_sources(worker_id=1 , num_workers=2 ) assert shard_it_a.n_shards == 2 __a =_get_expected_row_ids_and_row_dicts_for_partition_order(_snake_case , [1, 3] ) for i, (row_id, row_dict) in enumerate(_snake_case ): __a , __a =expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def UpperCamelCase_( ): """simple docstring""" __a =pyspark.sql.SparkSession.builder.master('local[*]' ).appName('pyspark' ).getOrCreate() __a =spark.range(100 ).repartition(1 ) __a =Spark(_snake_case ) # Choose a small max_shard_size for maximum partitioning. spark_builder._repartition_df_if_needed(max_shard_size=1 ) # The new number of partitions should not be greater than the number of rows. assert spark_builder.df.rdd.getNumPartitions() == 100
218
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _lowerCAmelCase : Tuple = { "configuration_llama": ["LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP", "LlamaConfig"], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Optional[int] = ["LlamaTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Optional[int] = ["LlamaTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : List[Any] = [ "LlamaForCausalLM", "LlamaModel", "LlamaPreTrainedModel", "LlamaForSequenceClassification", ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys _lowerCAmelCase : int = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
218
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Union[str, Any] = { """configuration_x_clip""": [ """XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XCLIPConfig""", """XCLIPTextConfig""", """XCLIPVisionConfig""", ], """processing_x_clip""": ["""XCLIPProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : List[str] = [ """XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """XCLIPModel""", """XCLIPPreTrainedModel""", """XCLIPTextModel""", """XCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys lowerCAmelCase : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
127
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import convert_to_rgb, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) class __lowercase ( UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : Any = ['''pixel_values'''] def __init__( self : List[Any] , lowerCAmelCase__ : bool = True , lowerCAmelCase__ : Dict[str, int] = None , lowerCAmelCase__ : PILImageResampling = PILImageResampling.BICUBIC , lowerCAmelCase__ : bool = True , lowerCAmelCase__ : Union[int, float] = 1 / 255 , lowerCAmelCase__ : bool = True , lowerCAmelCase__ : Optional[Union[float, List[float]]] = None , lowerCAmelCase__ : Optional[Union[float, List[float]]] = None , lowerCAmelCase__ : bool = True , **lowerCAmelCase__ : Tuple , ): super().__init__(**lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[str] = size if size is not None else {"height": 384, "width": 384} SCREAMING_SNAKE_CASE_: Union[str, Any] = get_size_dict(lowerCAmelCase__ , default_to_square=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: int = do_resize SCREAMING_SNAKE_CASE_: Dict = size SCREAMING_SNAKE_CASE_: int = resample SCREAMING_SNAKE_CASE_: str = do_rescale SCREAMING_SNAKE_CASE_: str = rescale_factor SCREAMING_SNAKE_CASE_: Optional[Any] = do_normalize SCREAMING_SNAKE_CASE_: Union[str, Any] = image_mean if image_mean is not None else OPENAI_CLIP_MEAN SCREAMING_SNAKE_CASE_: List[str] = image_std if image_std is not None else OPENAI_CLIP_STD SCREAMING_SNAKE_CASE_: Optional[int] = do_convert_rgb def _SCREAMING_SNAKE_CASE ( self : str , lowerCAmelCase__ : np.ndarray , lowerCAmelCase__ : Dict[str, int] , lowerCAmelCase__ : PILImageResampling = PILImageResampling.BICUBIC , lowerCAmelCase__ : Optional[Union[str, ChannelDimension]] = None , **lowerCAmelCase__ : Tuple , ): SCREAMING_SNAKE_CASE_: List[Any] = get_size_dict(lowerCAmelCase__ , default_to_square=lowerCAmelCase__) if "height" not in size or "width" not in size: raise ValueError(F"The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}") SCREAMING_SNAKE_CASE_: int = (size["height"], size["width"]) return resize(lowerCAmelCase__ , size=lowerCAmelCase__ , resample=lowerCAmelCase__ , data_format=lowerCAmelCase__ , **lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Tuple , lowerCAmelCase__ : np.ndarray , lowerCAmelCase__ : Union[int, float] , lowerCAmelCase__ : Optional[Union[str, ChannelDimension]] = None , **lowerCAmelCase__ : List[Any] , ): return rescale(lowerCAmelCase__ , scale=lowerCAmelCase__ , data_format=lowerCAmelCase__ , **lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : List[str] , lowerCAmelCase__ : np.ndarray , lowerCAmelCase__ : Union[float, List[float]] , lowerCAmelCase__ : Union[float, List[float]] , lowerCAmelCase__ : Optional[Union[str, ChannelDimension]] = None , **lowerCAmelCase__ : List[Any] , ): return normalize(lowerCAmelCase__ , mean=lowerCAmelCase__ , std=lowerCAmelCase__ , data_format=lowerCAmelCase__ , **lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase__ : ImageInput , lowerCAmelCase__ : Optional[bool] = None , lowerCAmelCase__ : Optional[Dict[str, int]] = None , lowerCAmelCase__ : PILImageResampling = None , lowerCAmelCase__ : Optional[bool] = None , lowerCAmelCase__ : Optional[float] = None , lowerCAmelCase__ : Optional[bool] = None , lowerCAmelCase__ : Optional[Union[float, List[float]]] = None , lowerCAmelCase__ : Optional[Union[float, List[float]]] = None , lowerCAmelCase__ : Optional[Union[str, TensorType]] = None , lowerCAmelCase__ : bool = None , lowerCAmelCase__ : ChannelDimension = ChannelDimension.FIRST , **lowerCAmelCase__ : Dict , ): SCREAMING_SNAKE_CASE_: Any = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE_: Optional[int] = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE_: Dict = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE_: Any = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE_: Optional[int] = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE_: Optional[Any] = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE_: Optional[Any] = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE_: Optional[int] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb SCREAMING_SNAKE_CASE_: str = size if size is not None else self.size SCREAMING_SNAKE_CASE_: List[Any] = get_size_dict(lowerCAmelCase__ , default_to_square=lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[Any] = make_list_of_images(lowerCAmelCase__) if not valid_images(lowerCAmelCase__): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray.") if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True.") if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True.") if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True.") # PIL RGBA images are converted to RGB if do_convert_rgb: SCREAMING_SNAKE_CASE_: List[Any] = [convert_to_rgb(lowerCAmelCase__) for image in images] # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE_: List[Any] = [to_numpy_array(lowerCAmelCase__) for image in images] if do_resize: SCREAMING_SNAKE_CASE_: int = [self.resize(image=lowerCAmelCase__ , size=lowerCAmelCase__ , resample=lowerCAmelCase__) for image in images] if do_rescale: SCREAMING_SNAKE_CASE_: Optional[Any] = [self.rescale(image=lowerCAmelCase__ , scale=lowerCAmelCase__) for image in images] if do_normalize: SCREAMING_SNAKE_CASE_: List[str] = [self.normalize(image=lowerCAmelCase__ , mean=lowerCAmelCase__ , std=lowerCAmelCase__) for image in images] SCREAMING_SNAKE_CASE_: str = [to_channel_dimension_format(lowerCAmelCase__ , lowerCAmelCase__) for image in images] SCREAMING_SNAKE_CASE_: List[str] = BatchFeature(data={"pixel_values": images} , tensor_type=lowerCAmelCase__) return encoded_outputs
127
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase : int = logging.get_logger(__name__) UpperCAmelCase : List[str] = { "uw-madison/mra-base-512-4": "https://huggingface.co/uw-madison/mra-base-512-4/resolve/main/config.json", } class SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase ): lowercase__ = "mra" def __init__( self : str , lowerCAmelCase_ : int=5_0_2_6_5 , lowerCAmelCase_ : Tuple=7_6_8 , lowerCAmelCase_ : Tuple=1_2 , lowerCAmelCase_ : Optional[Any]=1_2 , lowerCAmelCase_ : List[Any]=3_0_7_2 , lowerCAmelCase_ : List[Any]="gelu" , lowerCAmelCase_ : List[str]=0.1 , lowerCAmelCase_ : List[Any]=0.1 , lowerCAmelCase_ : int=5_1_2 , lowerCAmelCase_ : Optional[int]=1 , lowerCAmelCase_ : Optional[Any]=0.02 , lowerCAmelCase_ : List[Any]=1E-5 , lowerCAmelCase_ : Union[str, Any]="absolute" , lowerCAmelCase_ : str=4 , lowerCAmelCase_ : Optional[int]="full" , lowerCAmelCase_ : int=0 , lowerCAmelCase_ : List[str]=0 , lowerCAmelCase_ : Optional[int]=1 , lowerCAmelCase_ : Optional[Any]=0 , lowerCAmelCase_ : Optional[int]=2 , **lowerCAmelCase_ : List[str] , ): """simple docstring""" super().__init__(pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_) lowercase_ = vocab_size lowercase_ = max_position_embeddings lowercase_ = hidden_size lowercase_ = num_hidden_layers lowercase_ = num_attention_heads lowercase_ = intermediate_size lowercase_ = hidden_act lowercase_ = hidden_dropout_prob lowercase_ = attention_probs_dropout_prob lowercase_ = initializer_range lowercase_ = type_vocab_size lowercase_ = layer_norm_eps lowercase_ = position_embedding_type lowercase_ = block_per_row lowercase_ = approx_mode lowercase_ = initial_prior_first_n_blocks lowercase_ = initial_prior_diagonal_n_blocks
136
"""simple docstring""" import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder UpperCAmelCase : Union[str, Any] = "__DUMMY_TRANSFORMERS_USER__" UpperCAmelCase : Dict = "Dummy User" UpperCAmelCase : Optional[int] = "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" UpperCAmelCase : Tuple = "https://hub-ci.huggingface.co" UpperCAmelCase : Optional[Any] = CI_HUB_ENDPOINT + "/datasets/{repo_id}/resolve/{revision}/{path}" UpperCAmelCase : Tuple = CI_HUB_ENDPOINT + "/{repo_id}/resolve/{revision}/{filename}" UpperCAmelCase : int = Path("~/.huggingface/hub_ci_token").expanduser() @pytest.fixture def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> Dict: '''simple docstring''' monkeypatch.setattr( """huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE""" , __lowerCAmelCase ) @pytest.fixture def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> Union[str, Any]: '''simple docstring''' monkeypatch.setattr("""datasets.config.HF_ENDPOINT""" , __lowerCAmelCase ) monkeypatch.setattr("""datasets.config.HUB_DATASETS_URL""" , __lowerCAmelCase ) @pytest.fixture def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> int: '''simple docstring''' monkeypatch.setattr("""huggingface_hub.hf_api.HfFolder.path_token""" , __lowerCAmelCase ) @pytest.fixture def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> Optional[Any]: '''simple docstring''' HfFolder.save_token(__lowerCAmelCase ) yield HfFolder.delete_token() @pytest.fixture(scope="""session""" ) def _SCREAMING_SNAKE_CASE () -> Dict: '''simple docstring''' return HfApi(endpoint=__lowerCAmelCase ) @pytest.fixture(scope="""session""" ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> List[str]: '''simple docstring''' lowercase_ = HfFolder.get_token() HfFolder.save_token(__lowerCAmelCase ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(__lowerCAmelCase ) @pytest.fixture def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> Optional[Any]: '''simple docstring''' def _cleanup_repo(__lowerCAmelCase ): hf_api.delete_repo(__lowerCAmelCase , token=__lowerCAmelCase , repo_type="""dataset""" ) return _cleanup_repo @pytest.fixture def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> Optional[Any]: '''simple docstring''' @contextmanager def _temporary_repo(__lowerCAmelCase ): try: yield repo_id finally: cleanup_repo(__lowerCAmelCase ) return _temporary_repo @pytest.fixture(scope="""session""" ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> List[Any]: '''simple docstring''' lowercase_ = F'''repo_txt_data-{int(time.time() * 10E3 )}''' lowercase_ = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(__lowerCAmelCase , token=__lowerCAmelCase , repo_type="""dataset""" , private=__lowerCAmelCase ) hf_api.upload_file( token=__lowerCAmelCase , path_or_fileobj=str(__lowerCAmelCase ) , path_in_repo="""data/text_data.txt""" , repo_id=__lowerCAmelCase , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(__lowerCAmelCase , token=__lowerCAmelCase , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Any: '''simple docstring''' return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope="""session""" ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Union[str, Any]: '''simple docstring''' lowercase_ = F'''repo_zipped_txt_data-{int(time.time() * 10E3 )}''' lowercase_ = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(__lowerCAmelCase , token=__lowerCAmelCase , repo_type="""dataset""" , private=__lowerCAmelCase ) hf_api.upload_file( token=__lowerCAmelCase , path_or_fileobj=str(__lowerCAmelCase ) , path_in_repo="""data.zip""" , repo_id=__lowerCAmelCase , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(__lowerCAmelCase , token=__lowerCAmelCase , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Optional[Any]: '''simple docstring''' return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope="""session""" ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> str: '''simple docstring''' lowercase_ = F'''repo_zipped_img_data-{int(time.time() * 10E3 )}''' lowercase_ = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(__lowerCAmelCase , token=__lowerCAmelCase , repo_type="""dataset""" , private=__lowerCAmelCase ) hf_api.upload_file( token=__lowerCAmelCase , path_or_fileobj=str(__lowerCAmelCase ) , path_in_repo="""data.zip""" , repo_id=__lowerCAmelCase , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(__lowerCAmelCase , token=__lowerCAmelCase , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Dict: '''simple docstring''' return hf_private_dataset_repo_zipped_img_data_
136
1
'''simple docstring''' import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import SegformerImageProcessor, SwinConfig, UperNetConfig, UperNetForSemanticSegmentation def _UpperCamelCase ( UpperCamelCase__ ): UpperCAmelCase__ : Any = 3_8_4 UpperCAmelCase__ : int = 7 if "tiny" in model_name: UpperCAmelCase__ : Union[str, Any] = 9_6 UpperCAmelCase__ : Any = (2, 2, 6, 2) UpperCAmelCase__ : List[str] = (3, 6, 1_2, 2_4) elif "small" in model_name: UpperCAmelCase__ : Tuple = 9_6 UpperCAmelCase__ : Union[str, Any] = (2, 2, 1_8, 2) UpperCAmelCase__ : Dict = (3, 6, 1_2, 2_4) elif "base" in model_name: UpperCAmelCase__ : List[str] = 1_2_8 UpperCAmelCase__ : Optional[int] = (2, 2, 1_8, 2) UpperCAmelCase__ : Union[str, Any] = (4, 8, 1_6, 3_2) UpperCAmelCase__ : Optional[int] = 1_2 UpperCAmelCase__ : List[str] = 5_1_2 elif "large" in model_name: UpperCAmelCase__ : List[str] = 1_9_2 UpperCAmelCase__ : List[Any] = (2, 2, 1_8, 2) UpperCAmelCase__ : List[Any] = (6, 1_2, 2_4, 4_8) UpperCAmelCase__ : Dict = 1_2 UpperCAmelCase__ : str = 7_6_8 # set label information UpperCAmelCase__ : Optional[int] = 1_5_0 UpperCAmelCase__ : Optional[int] = """huggingface/label-files""" UpperCAmelCase__ : Dict = """ade20k-id2label.json""" UpperCAmelCase__ : Tuple = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type="""dataset""" ) , """r""" ) ) UpperCAmelCase__ : List[str] = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} UpperCAmelCase__ : Optional[Any] = {v: k for k, v in idalabel.items()} UpperCAmelCase__ : List[Any] = SwinConfig( embed_dim=UpperCamelCase__ , depths=UpperCamelCase__ , num_heads=UpperCamelCase__ , window_size=UpperCamelCase__ , out_features=["""stage1""", """stage2""", """stage3""", """stage4"""] , ) UpperCAmelCase__ : Union[str, Any] = UperNetConfig( backbone_config=UpperCamelCase__ , auxiliary_in_channels=UpperCamelCase__ , num_labels=UpperCamelCase__ , idalabel=UpperCamelCase__ , labelaid=UpperCamelCase__ , ) return config def _UpperCamelCase ( UpperCamelCase__ ): UpperCAmelCase__ : List[Any] = [] # fmt: off # stem rename_keys.append(("""backbone.patch_embed.projection.weight""", """backbone.embeddings.patch_embeddings.projection.weight""") ) rename_keys.append(("""backbone.patch_embed.projection.bias""", """backbone.embeddings.patch_embeddings.projection.bias""") ) rename_keys.append(("""backbone.patch_embed.norm.weight""", """backbone.embeddings.norm.weight""") ) rename_keys.append(("""backbone.patch_embed.norm.bias""", """backbone.embeddings.norm.bias""") ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.norm1.weight''', f'''backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.weight''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.norm1.bias''', f'''backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.bias''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_bias_table''', f'''backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_index''', f'''backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.weight''', f'''backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.bias''', f'''backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.norm2.weight''', f'''backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.weight''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.norm2.bias''', f'''backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.bias''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.weight''', f'''backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.bias''', f'''backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.ffn.layers.1.weight''', f'''backbone.encoder.layers.{i}.blocks.{j}.output.dense.weight''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.ffn.layers.1.bias''', f'''backbone.encoder.layers.{i}.blocks.{j}.output.dense.bias''') ) if i < 3: rename_keys.append((f'''backbone.stages.{i}.downsample.reduction.weight''', f'''backbone.encoder.layers.{i}.downsample.reduction.weight''') ) rename_keys.append((f'''backbone.stages.{i}.downsample.norm.weight''', f'''backbone.encoder.layers.{i}.downsample.norm.weight''') ) rename_keys.append((f'''backbone.stages.{i}.downsample.norm.bias''', f'''backbone.encoder.layers.{i}.downsample.norm.bias''') ) rename_keys.append((f'''backbone.norm{i}.weight''', f'''backbone.hidden_states_norms.stage{i+1}.weight''') ) rename_keys.append((f'''backbone.norm{i}.bias''', f'''backbone.hidden_states_norms.stage{i+1}.bias''') ) # decode head rename_keys.extend( [ ("""decode_head.conv_seg.weight""", """decode_head.classifier.weight"""), ("""decode_head.conv_seg.bias""", """decode_head.classifier.bias"""), ("""auxiliary_head.conv_seg.weight""", """auxiliary_head.classifier.weight"""), ("""auxiliary_head.conv_seg.bias""", """auxiliary_head.classifier.bias"""), ] ) # fmt: on return rename_keys def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): UpperCAmelCase__ : Dict = dct.pop(UpperCamelCase__ ) UpperCAmelCase__ : Optional[int] = val def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): UpperCAmelCase__ : int = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): UpperCAmelCase__ : Tuple = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) UpperCAmelCase__ : str = state_dict.pop(f'''backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.weight''' ) UpperCAmelCase__ : Union[str, Any] = state_dict.pop(f'''backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase__ : str = in_proj_weight[:dim, :] UpperCAmelCase__ : int = in_proj_bias[: dim] UpperCAmelCase__ : Dict = in_proj_weight[ dim : dim * 2, : ] UpperCAmelCase__ : List[str] = in_proj_bias[ dim : dim * 2 ] UpperCAmelCase__ : Union[str, Any] = in_proj_weight[ -dim :, : ] UpperCAmelCase__ : Optional[Any] = in_proj_bias[-dim :] # fmt: on def _UpperCamelCase ( UpperCamelCase__ ): UpperCAmelCase__ , UpperCAmelCase__ : List[str] = x.shape UpperCAmelCase__ : Union[str, Any] = x.reshape(UpperCamelCase__ , 4 , in_channel // 4 ) UpperCAmelCase__ : int = x[:, [0, 2, 1, 3], :].transpose(1 , 2 ).reshape(UpperCamelCase__ , UpperCamelCase__ ) return x def _UpperCamelCase ( UpperCamelCase__ ): UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = x.shape UpperCAmelCase__ : Optional[Any] = x.reshape(UpperCamelCase__ , in_channel // 4 , 4 ) UpperCAmelCase__ : Optional[int] = x[:, :, [0, 2, 1, 3]].transpose(1 , 2 ).reshape(UpperCamelCase__ , UpperCamelCase__ ) return x def _UpperCamelCase ( UpperCamelCase__ ): UpperCAmelCase__ : int = x.shape[0] UpperCAmelCase__ : Union[str, Any] = x.reshape(4 , in_channel // 4 ) UpperCAmelCase__ : str = x[[0, 2, 1, 3], :].transpose(0 , 1 ).reshape(UpperCamelCase__ ) return x def _UpperCamelCase ( UpperCamelCase__ ): UpperCAmelCase__ : Any = x.shape[0] UpperCAmelCase__ : int = x.reshape(in_channel // 4 , 4 ) UpperCAmelCase__ : List[str] = x[:, [0, 2, 1, 3]].transpose(0 , 1 ).reshape(UpperCamelCase__ ) return x def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): UpperCAmelCase__ : List[Any] = { """upernet-swin-tiny""": """https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210531_112542-e380ad3e.pth""", """upernet-swin-small""": """https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210526_192015-ee2fff1c.pth""", """upernet-swin-base""": """https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K_20210531_125459-429057bf.pth""", """upernet-swin-large""": """https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k_20220318_091743-9ba68901.pth""", } UpperCAmelCase__ : Tuple = model_name_to_url[model_name] UpperCAmelCase__ : Union[str, Any] = torch.hub.load_state_dict_from_url(UpperCamelCase__ , map_location="""cpu""" , file_name=UpperCamelCase__ )[ """state_dict""" ] for name, param in state_dict.items(): print(UpperCamelCase__ , param.shape ) UpperCAmelCase__ : Optional[int] = get_upernet_config(UpperCamelCase__ ) UpperCAmelCase__ : Tuple = UperNetForSemanticSegmentation(UpperCamelCase__ ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): UpperCAmelCase__ : List[Any] = state_dict.pop(UpperCamelCase__ ) if "bn" in key: UpperCAmelCase__ : int = key.replace("""bn""" , """batch_norm""" ) UpperCAmelCase__ : str = val # rename keys UpperCAmelCase__ : List[Any] = create_rename_keys(UpperCamelCase__ ) for src, dest in rename_keys: rename_key(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) read_in_q_k_v(UpperCamelCase__ , config.backbone_config ) # fix downsample parameters for key, value in state_dict.items(): if "downsample" in key: if "reduction" in key: UpperCAmelCase__ : Optional[Any] = reverse_correct_unfold_reduction_order(UpperCamelCase__ ) if "norm" in key: UpperCAmelCase__ : Optional[int] = reverse_correct_unfold_norm_order(UpperCamelCase__ ) model.load_state_dict(UpperCamelCase__ ) # verify on image UpperCAmelCase__ : List[Any] = """https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg""" UpperCAmelCase__ : Any = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ).convert("""RGB""" ) UpperCAmelCase__ : Dict = SegformerImageProcessor() UpperCAmelCase__ : List[Any] = processor(UpperCamelCase__ , return_tensors="""pt""" ).pixel_values with torch.no_grad(): UpperCAmelCase__ : Union[str, Any] = model(UpperCamelCase__ ) UpperCAmelCase__ : str = outputs.logits print(logits.shape ) print("""First values of logits:""" , logits[0, 0, :3, :3] ) # assert values if model_name == "upernet-swin-tiny": UpperCAmelCase__ : str = torch.tensor( [[-7.59_58, -7.59_58, -7.43_02], [-7.59_58, -7.59_58, -7.43_02], [-7.47_97, -7.47_97, -7.30_68]] ) elif model_name == "upernet-swin-small": UpperCAmelCase__ : Optional[Any] = torch.tensor( [[-7.19_21, -7.19_21, -6.95_32], [-7.19_21, -7.19_21, -6.95_32], [-7.09_08, -7.09_08, -6.85_34]] ) elif model_name == "upernet-swin-base": UpperCAmelCase__ : Union[str, Any] = torch.tensor( [[-6.58_51, -6.58_51, -6.43_30], [-6.58_51, -6.58_51, -6.43_30], [-6.47_63, -6.47_63, -6.32_54]] ) elif model_name == "upernet-swin-large": UpperCAmelCase__ : Union[str, Any] = torch.tensor( [[-7.52_97, -7.52_97, -7.38_02], [-7.52_97, -7.52_97, -7.38_02], [-7.40_44, -7.40_44, -7.25_86]] ) print("""Logits:""" , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCamelCase__ , atol=1e-4 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: print(f'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(UpperCamelCase__ ) print(f'''Saving processor to {pytorch_dump_folder_path}''' ) processor.save_pretrained(UpperCamelCase__ ) if push_to_hub: print(f'''Pushing model and processor for {model_name} to hub''' ) model.push_to_hub(f'''openmmlab/{model_name}''' ) processor.push_to_hub(f'''openmmlab/{model_name}''' ) if __name__ == "__main__": __A =argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='upernet-swin-tiny', type=str, choices=[f"""upernet-swin-{size}""" for size in ['tiny', 'small', 'base', 'large']], help='Name of the Swin + UperNet model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) __A =parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
283
'''simple docstring''' import functools def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): # Validation if not isinstance(UpperCamelCase__ , UpperCamelCase__ ) or not all(isinstance(UpperCamelCase__ , UpperCamelCase__ ) for day in days ): raise ValueError("""The parameter days should be a list of integers""" ) if len(UpperCamelCase__ ) != 3 or not all(isinstance(UpperCamelCase__ , UpperCamelCase__ ) for cost in costs ): raise ValueError("""The parameter costs should be a list of three integers""" ) if len(UpperCamelCase__ ) == 0: return 0 if min(UpperCamelCase__ ) <= 0: raise ValueError("""All days elements should be greater than 0""" ) if max(UpperCamelCase__ ) >= 3_6_6: raise ValueError("""All days elements should be less than 366""" ) UpperCAmelCase__ : Union[str, Any] = set(UpperCamelCase__ ) @functools.cache def dynamic_programming(UpperCamelCase__ ) -> int: if index > 3_6_5: return 0 if index not in days_set: return dynamic_programming(index + 1 ) return min( costs[0] + dynamic_programming(index + 1 ) , costs[1] + dynamic_programming(index + 7 ) , costs[2] + dynamic_programming(index + 3_0 ) , ) return dynamic_programming(1 ) if __name__ == "__main__": import doctest doctest.testmod()
283
1
'''simple docstring''' __lowercase : Tuple = { '''A''': '''.-''', '''B''': '''-...''', '''C''': '''-.-.''', '''D''': '''-..''', '''E''': '''.''', '''F''': '''..-.''', '''G''': '''--.''', '''H''': '''....''', '''I''': '''..''', '''J''': '''.---''', '''K''': '''-.-''', '''L''': '''.-..''', '''M''': '''--''', '''N''': '''-.''', '''O''': '''---''', '''P''': '''.--.''', '''Q''': '''--.-''', '''R''': '''.-.''', '''S''': '''...''', '''T''': '''-''', '''U''': '''..-''', '''V''': '''...-''', '''W''': '''.--''', '''X''': '''-..-''', '''Y''': '''-.--''', '''Z''': '''--..''', '''1''': '''.----''', '''2''': '''..---''', '''3''': '''...--''', '''4''': '''....-''', '''5''': '''.....''', '''6''': '''-....''', '''7''': '''--...''', '''8''': '''---..''', '''9''': '''----.''', '''0''': '''-----''', '''&''': '''.-...''', '''@''': '''.--.-.''', ''':''': '''---...''', ''',''': '''--..--''', '''.''': '''.-.-.-''', '''\'''': '''.----.''', '''"''': '''.-..-.''', '''?''': '''..--..''', '''/''': '''-..-.''', '''=''': '''-...-''', '''+''': '''.-.-.''', '''-''': '''-....-''', '''(''': '''-.--.''', ''')''': '''-.--.-''', '''!''': '''-.-.--''', ''' ''': '''/''' } # Exclamation mark is not in ITU-R recommendation # fmt: on __lowercase : Any = {value: key for key, value in MORSE_CODE_DICT.items()} def lowercase_ ( _lowercase ) -> str: '''simple docstring''' return " ".join(MORSE_CODE_DICT[char] for char in message.upper() ) def lowercase_ ( _lowercase ) -> str: '''simple docstring''' return "".join(REVERSE_DICT[char] for char in message.split() ) def lowercase_ ( ) -> None: '''simple docstring''' lowerCamelCase_ : Optional[int] = '''Morse code here!''' print(_lowercase ) lowerCamelCase_ : List[str] = encrypt(_lowercase ) print(_lowercase ) lowerCamelCase_ : Tuple = decrypt(_lowercase ) print(_lowercase ) if __name__ == "__main__": main()
318
'''simple docstring''' from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time __lowercase : str = Lock() def lowercase_ ( _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ) -> List[Any]: '''simple docstring''' global process_lock # we perform n swaps since after n swaps we know we are sorted # we *could* stop early if we are sorted already, but it takes as long to # find out we are sorted as it does to sort the list with this algorithm for i in range(0 , 10 ): if (i + position) % 2 == 0 and r_send is not None: # send your value to your right neighbor process_lock.acquire() r_send[1].send(_lowercase ) process_lock.release() # receive your right neighbor's value process_lock.acquire() lowerCamelCase_ : Dict = rr_cv[0].recv() process_lock.release() # take the lower value since you are on the left lowerCamelCase_ : Union[str, Any] = min(_lowercase , _lowercase ) elif (i + position) % 2 != 0 and l_send is not None: # send your value to your left neighbor process_lock.acquire() l_send[1].send(_lowercase ) process_lock.release() # receive your left neighbor's value process_lock.acquire() lowerCamelCase_ : str = lr_cv[0].recv() process_lock.release() # take the higher value since you are on the right lowerCamelCase_ : Any = max(_lowercase , _lowercase ) # after all swaps are performed, send the values back to main result_pipe[1].send(_lowercase ) def lowercase_ ( _lowercase ) -> int: '''simple docstring''' lowerCamelCase_ : int = [] lowerCamelCase_ : Tuple = [] # initialize the list of pipes where the values will be retrieved for _ in arr: result_pipe.append(Pipe() ) # creates the processes # the first and last process only have one neighbor so they are made outside # of the loop lowerCamelCase_ : str = Pipe() lowerCamelCase_ : List[Any] = Pipe() process_array_.append( Process( target=_lowercase , args=(0, arr[0], None, temp_rs, None, temp_rr, result_pipe[0]) , ) ) lowerCamelCase_ : Optional[Any] = temp_rs lowerCamelCase_ : List[str] = temp_rr for i in range(1 , len(_lowercase ) - 1 ): lowerCamelCase_ : str = Pipe() lowerCamelCase_ : Any = Pipe() process_array_.append( Process( target=_lowercase , args=(i, arr[i], temp_ls, temp_rs, temp_lr, temp_rr, result_pipe[i]) , ) ) lowerCamelCase_ : Dict = temp_rs lowerCamelCase_ : Tuple = temp_rr process_array_.append( Process( target=_lowercase , args=( len(_lowercase ) - 1, arr[len(_lowercase ) - 1], temp_ls, None, temp_lr, None, result_pipe[len(_lowercase ) - 1], ) , ) ) # start the processes for p in process_array_: p.start() # wait for the processes to end and write their values to the list for p in range(0 , len(_lowercase ) ): lowerCamelCase_ : Optional[Any] = result_pipe[p][0].recv() process_array_[p].join() return arr def lowercase_ ( ) -> Any: '''simple docstring''' lowerCamelCase_ : Union[str, Any] = list(range(10 , 0 , -1 ) ) print('''Initial List''' ) print(*_lowercase ) lowerCamelCase_ : Optional[int] = odd_even_transposition(_lowercase ) print('''Sorted List\n''' ) print(*_lowercase ) if __name__ == "__main__": main()
318
1
import warnings from typing import Dict import numpy as np from ..utils import ExplicitEnum, add_end_docstrings, is_tf_available, is_torch_available from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING def _lowercase ( UpperCamelCase_ ) -> Optional[int]: '''simple docstring''' return 1.0 / (1.0 + np.exp(-_outputs )) def _lowercase ( UpperCamelCase_ ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE__ = np.max(_outputs , axis=-1 , keepdims=UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = np.exp(_outputs - maxes ) return shifted_exp / shifted_exp.sum(axis=-1 , keepdims=UpperCamelCase_ ) class lowercase__ ( _UpperCAmelCase ): A__ : Optional[int] ="""sigmoid""" A__ : List[Any] ="""softmax""" A__ : str ="""none""" @add_end_docstrings( _UpperCAmelCase , R""" return_all_scores (`bool`, *optional*, defaults to `False`): Whether to return all prediction scores or just the one of the predicted class. function_to_apply (`str`, *optional*, defaults to `\"default\"`): The function to apply to the model outputs in order to retrieve the scores. Accepts four different values: - `\"default\"`: if the model has a single label, will apply the sigmoid function on the output. If the model has several labels, will apply the softmax function on the output. - `\"sigmoid\"`: Applies the sigmoid function on the output. - `\"softmax\"`: Applies the softmax function on the output. - `\"none\"`: Does not apply any function on the output. """ , ) class lowercase__ ( _UpperCAmelCase ): A__ : int =False A__ : Dict =ClassificationFunction.NONE def __init__( self : str , **UpperCAmelCase_ : int ): super().__init__(**UpperCAmelCase_ ) self.check_model_type( TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING ) def A_ ( self : Any , UpperCAmelCase_ : List[str]=None , UpperCAmelCase_ : int=None , UpperCAmelCase_ : Tuple="" , **UpperCAmelCase_ : Dict ): # Using "" as default argument because we're going to use `top_k=None` in user code to declare # "No top_k" SCREAMING_SNAKE_CASE__ = tokenizer_kwargs SCREAMING_SNAKE_CASE__ = {} if hasattr(self.model.config , 'return_all_scores' ) and return_all_scores is None: SCREAMING_SNAKE_CASE__ = self.model.config.return_all_scores if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) or top_k is None: SCREAMING_SNAKE_CASE__ = top_k SCREAMING_SNAKE_CASE__ = False elif return_all_scores is not None: warnings.warn( '`return_all_scores` is now deprecated, if want a similar functionality use `top_k=None` instead of' ' `return_all_scores=True` or `top_k=1` instead of `return_all_scores=False`.' , UpperCAmelCase_ , ) if return_all_scores: SCREAMING_SNAKE_CASE__ = None else: SCREAMING_SNAKE_CASE__ = 1 if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): SCREAMING_SNAKE_CASE__ = ClassificationFunction[function_to_apply.upper()] if function_to_apply is not None: SCREAMING_SNAKE_CASE__ = function_to_apply return preprocess_params, {}, postprocess_params def __call__( self : int , *UpperCAmelCase_ : Optional[int] , **UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE__ = super().__call__(*UpperCAmelCase_ , **UpperCAmelCase_ ) # TODO try and retrieve it in a nicer way from _sanitize_parameters. SCREAMING_SNAKE_CASE__ = 'top_k' not in kwargs if isinstance(args[0] , UpperCAmelCase_ ) and _legacy: # This pipeline is odd, and return a list when single item is run return [result] else: return result def A_ ( self : Dict , UpperCAmelCase_ : Dict , **UpperCAmelCase_ : Optional[Any] ): SCREAMING_SNAKE_CASE__ = self.framework if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): return self.tokenizer(**UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ ) elif isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) and len(UpperCAmelCase_ ) == 1 and isinstance(inputs[0] , UpperCAmelCase_ ) and len(inputs[0] ) == 2: # It used to be valid to use a list of list of list for text pairs, keeping this path for BC return self.tokenizer( text=inputs[0][0] , text_pair=inputs[0][1] , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ ) elif isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): # This is likely an invalid usage of the pipeline attempting to pass text pairs. raise ValueError( 'The pipeline received invalid inputs, if you are trying to send text pairs, you can try to send a' ' dictionary `{"text": "My text", "text_pair": "My pair"}` in order to send a text pair.' ) return self.tokenizer(UpperCAmelCase_ , return_tensors=UpperCAmelCase_ , **UpperCAmelCase_ ) def A_ ( self : Optional[Any] , UpperCAmelCase_ : Optional[Any] ): return self.model(**UpperCAmelCase_ ) def A_ ( self : Dict , UpperCAmelCase_ : str , UpperCAmelCase_ : Tuple=None , UpperCAmelCase_ : List[str]=1 , UpperCAmelCase_ : int=True ): # `_legacy` is used to determine if we're running the naked pipeline and in backward # compatibility mode, or if running the pipeline with `pipeline(..., top_k=1)` we're running # the more natural result containing the list. # Default value before `set_parameters` if function_to_apply is None: if self.model.config.problem_type == "multi_label_classification" or self.model.config.num_labels == 1: SCREAMING_SNAKE_CASE__ = ClassificationFunction.SIGMOID elif self.model.config.problem_type == "single_label_classification" or self.model.config.num_labels > 1: SCREAMING_SNAKE_CASE__ = ClassificationFunction.SOFTMAX elif hasattr(self.model.config , 'function_to_apply' ) and function_to_apply is None: SCREAMING_SNAKE_CASE__ = self.model.config.function_to_apply else: SCREAMING_SNAKE_CASE__ = ClassificationFunction.NONE SCREAMING_SNAKE_CASE__ = model_outputs['logits'][0] SCREAMING_SNAKE_CASE__ = outputs.numpy() if function_to_apply == ClassificationFunction.SIGMOID: SCREAMING_SNAKE_CASE__ = sigmoid(UpperCAmelCase_ ) elif function_to_apply == ClassificationFunction.SOFTMAX: SCREAMING_SNAKE_CASE__ = softmax(UpperCAmelCase_ ) elif function_to_apply == ClassificationFunction.NONE: SCREAMING_SNAKE_CASE__ = outputs else: raise ValueError(F'Unrecognized `function_to_apply` argument: {function_to_apply}' ) if top_k == 1 and _legacy: return {"label": self.model.config.idalabel[scores.argmax().item()], "score": scores.max().item()} SCREAMING_SNAKE_CASE__ = [ {'label': self.model.config.idalabel[i], 'score': score.item()} for i, score in enumerate(UpperCAmelCase_ ) ] if not _legacy: dict_scores.sort(key=lambda UpperCAmelCase_ : x["score"] , reverse=UpperCAmelCase_ ) if top_k is not None: SCREAMING_SNAKE_CASE__ = dict_scores[:top_k] return dict_scores
169
import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=None ) -> Optional[Any]: '''simple docstring''' assert torch_layer.weight.shape == weight.shape, F'{torch_layer} layer.weight does not match' SCREAMING_SNAKE_CASE__ = nn.Parameter(UpperCamelCase_ ) if bias is not None: assert torch_layer.bias.shape == bias.shape, F'{torch_layer} layer.bias does not match' SCREAMING_SNAKE_CASE__ = nn.Parameter(UpperCamelCase_ ) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ = np.asarray(weights[0] ) SCREAMING_SNAKE_CASE__ = np.asarray(weights[1] ) SCREAMING_SNAKE_CASE__ = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(UpperCamelCase_ ).transpose(1 , 2 ).contiguous().view(-1 , UpperCamelCase_ ) , ) set_param( torch_layer.self_attention.value , torch.tensor(UpperCamelCase_ ).transpose(1 , 2 ).contiguous().view(-1 , UpperCamelCase_ ) , ) set_param( torch_layer.output.dense , torch.tensor(UpperCamelCase_ ).view(-1 , UpperCamelCase_ ).contiguous().transpose(0 , 1 ) , ) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE__ = np.asarray(weights[0] ) SCREAMING_SNAKE_CASE__ = np.asarray(weights[1] ) SCREAMING_SNAKE_CASE__ = np.asarray(weights[2] ) SCREAMING_SNAKE_CASE__ = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(UpperCamelCase_ ).transpose(1 , 2 ).contiguous().view(-1 , UpperCamelCase_ ) , ) set_param( torch_layer.self_attention.key , torch.tensor(UpperCamelCase_ ).transpose(1 , 2 ).contiguous().view(-1 , UpperCamelCase_ ) , ) set_param( torch_layer.self_attention.value , torch.tensor(UpperCamelCase_ ).transpose(1 , 2 ).contiguous().view(-1 , UpperCamelCase_ ) , ) set_param( torch_layer.output.dense , torch.tensor(UpperCamelCase_ ).view(-1 , UpperCamelCase_ ).contiguous().transpose(0 , 1 ) , ) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> Tuple: '''simple docstring''' SCREAMING_SNAKE_CASE__ = weights[0][0][0] SCREAMING_SNAKE_CASE__ = np.asarray(layer_norm_a[0] ) SCREAMING_SNAKE_CASE__ = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(UpperCamelCase_ ) , torch.tensor(UpperCamelCase_ ) , ) # lsh weights + output SCREAMING_SNAKE_CASE__ = weights[0][1] if len(UpperCamelCase_ ) < 4: set_layer_weights_in_torch_lsh(UpperCamelCase_ , torch_block.attention , UpperCamelCase_ ) else: set_layer_weights_in_torch_local(UpperCamelCase_ , torch_block.attention , UpperCamelCase_ ) # intermediate weighs SCREAMING_SNAKE_CASE__ = weights[2][0][1][2] # Chunked Feed Forward if len(UpperCamelCase_ ) == 4: SCREAMING_SNAKE_CASE__ = intermediate_weights[2] # layernorm 2 SCREAMING_SNAKE_CASE__ = np.asarray(intermediate_weights[0][0] ) SCREAMING_SNAKE_CASE__ = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(UpperCamelCase_ ) , torch.tensor(UpperCamelCase_ ) , ) # intermediate dense SCREAMING_SNAKE_CASE__ = np.asarray(intermediate_weights[1][0] ) SCREAMING_SNAKE_CASE__ = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(UpperCamelCase_ ).transpose(0 , 1 ).contiguous() , torch.tensor(UpperCamelCase_ ) , ) # intermediate out SCREAMING_SNAKE_CASE__ = np.asarray(intermediate_weights[4][0] ) SCREAMING_SNAKE_CASE__ = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(UpperCamelCase_ ).transpose(0 , 1 ).contiguous() , torch.tensor(UpperCamelCase_ ) , ) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE__ = torch_model.reformer # word embeds SCREAMING_SNAKE_CASE__ = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(UpperCamelCase_ ) , ) if isinstance(weights[3] , UpperCamelCase_ ): SCREAMING_SNAKE_CASE__ = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): SCREAMING_SNAKE_CASE__ = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), F'{position_embeddings[emb_idx]} emb does not match' SCREAMING_SNAKE_CASE__ = nn.Parameter(torch.tensor(UpperCamelCase_ ) ) SCREAMING_SNAKE_CASE__ = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( UpperCamelCase_ ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): SCREAMING_SNAKE_CASE__ = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # output layer norm SCREAMING_SNAKE_CASE__ = np.asarray(weights[7][0] ) SCREAMING_SNAKE_CASE__ = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(UpperCamelCase_ ) , torch.tensor(UpperCamelCase_ ) , ) # output embeddings SCREAMING_SNAKE_CASE__ = np.asarray(weights[9][0] ) SCREAMING_SNAKE_CASE__ = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(UpperCamelCase_ ).transpose(0 , 1 ).contiguous() , torch.tensor(UpperCamelCase_ ) , ) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ = ReformerConfig.from_json_file(UpperCamelCase_ ) print(F'Building PyTorch model from configuration: {config}' ) SCREAMING_SNAKE_CASE__ = ReformerModelWithLMHead(UpperCamelCase_ ) with open(UpperCamelCase_ , 'rb' ) as f: SCREAMING_SNAKE_CASE__ = pickle.load(UpperCamelCase_ )['weights'] set_model_weights_in_torch(UpperCamelCase_ , UpperCamelCase_ , config.hidden_size ) # Save pytorch-model print(F'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict() , UpperCamelCase_ ) if __name__ == "__main__": __snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( """--trax_model_pkl_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained Reformer model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __snake_case = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
169
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_rembert import RemBertTokenizer else: snake_case_ = None snake_case_ = logging.get_logger(__name__) snake_case_ = {'vocab_file': 'sentencepiece.model', 'tokenizer_file': 'tokenizer.json'} snake_case_ = { 'vocab_file': { 'google/rembert': 'https://huggingface.co/google/rembert/resolve/main/sentencepiece.model', }, 'tokenizer_file': { 'google/rembert': 'https://huggingface.co/google/rembert/resolve/main/tokenizer.json', }, } snake_case_ = { 'google/rembert': 256, } snake_case_ = '▁' class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ): A_ : Optional[int] = VOCAB_FILES_NAMES A_ : Optional[int] = PRETRAINED_VOCAB_FILES_MAP A_ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ : Tuple = RemBertTokenizer def __init__(self : Union[str, Any] , a__ : List[str]=None , a__ : Dict=None , a__ : Union[str, Any]=True , a__ : Tuple=True , a__ : Dict=False , a__ : Tuple="[CLS]" , a__ : Tuple="[SEP]" , a__ : List[str]="<unk>" , a__ : str="[SEP]" , a__ : Any="<pad>" , a__ : List[Any]="[CLS]" , a__ : str="[MASK]" , **a__ : Tuple , ): """simple docstring""" __snake_case = AddedToken(a__ , lstrip=a__ , rstrip=a__ ) if isinstance(a__ , a__ ) else mask_token super().__init__( a__ , tokenizer_file=a__ , do_lower_case=a__ , remove_space=a__ , keep_accents=a__ , bos_token=a__ , eos_token=a__ , unk_token=a__ , sep_token=a__ , pad_token=a__ , cls_token=a__ , mask_token=a__ , **a__ , ) __snake_case = do_lower_case __snake_case = remove_space __snake_case = keep_accents __snake_case = vocab_file __snake_case = False if not self.vocab_file else True def a (self : List[Any] , a__ : List[int] , a__ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def a (self : Tuple , a__ : List[int] , a__ : Optional[List[int]] = None , a__ : bool = False ): """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a__ )) + [1] + ([0] * len(a__ )) + [1] return [1] + ([0] * len(a__ )) + [1] def a (self : str , a__ : List[int] , a__ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def a (self : Any , a__ : str , a__ : Optional[str] = None ): """simple docstring""" if not os.path.isdir(a__ ): logger.error('''Vocabulary path ({}) should be a directory'''.format(a__ ) ) return __snake_case = os.path.join( a__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a__ ): copyfile(self.vocab_file , a__ ) return (out_vocab_file,)
24
from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING lowercase__ : Dict = logging.get_logger(__name__) @add_end_docstrings(UpperCAmelCase__ ) class UpperCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' def __init__( self : Optional[Any] , **__lowercase : Union[str, Any] ): """simple docstring""" super().__init__(**__lowercase ) requires_backends(self , "vision" ) requires_backends(self , "torch" ) if self.framework != "pt": raise ValueError(f"The {self.__class__} is only available in PyTorch." ) self.check_model_type(__lowercase ) def snake_case__ ( self : Optional[int] , **__lowercase : Optional[Any] ): """simple docstring""" snake_case_ = {} snake_case_ = {} snake_case_ = {} # preprocess args if "points_per_batch" in kwargs: snake_case_ = kwargs["points_per_batch"] if "points_per_crop" in kwargs: snake_case_ = kwargs["points_per_crop"] if "crops_n_layers" in kwargs: snake_case_ = kwargs["crops_n_layers"] if "crop_overlap_ratio" in kwargs: snake_case_ = kwargs["crop_overlap_ratio"] if "crop_n_points_downscale_factor" in kwargs: snake_case_ = kwargs["crop_n_points_downscale_factor"] # postprocess args if "pred_iou_thresh" in kwargs: snake_case_ = kwargs["pred_iou_thresh"] if "stability_score_offset" in kwargs: snake_case_ = kwargs["stability_score_offset"] if "mask_threshold" in kwargs: snake_case_ = kwargs["mask_threshold"] if "stability_score_thresh" in kwargs: snake_case_ = kwargs["stability_score_thresh"] if "crops_nms_thresh" in kwargs: snake_case_ = kwargs["crops_nms_thresh"] if "output_rle_mask" in kwargs: snake_case_ = kwargs["output_rle_mask"] if "output_bboxes_mask" in kwargs: snake_case_ = kwargs["output_bboxes_mask"] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__( self : Optional[int] , __lowercase : List[str] , *__lowercase : Optional[Any] , __lowercase : Dict=None , __lowercase : List[str]=None , **__lowercase : Optional[Any] ): """simple docstring""" return super().__call__(__lowercase , *__lowercase , num_workers=__lowercase , batch_size=__lowercase , **__lowercase ) def snake_case__ ( self : str , __lowercase : int , __lowercase : List[str]=64 , __lowercase : int = 0 , __lowercase : float = 5_12 / 15_00 , __lowercase : Optional[int] = 32 , __lowercase : Optional[int] = 1 , ): """simple docstring""" snake_case_ = load_image(__lowercase ) snake_case_ = self.image_processor.size["longest_edge"] snake_case_ , snake_case_ , snake_case_ , snake_case_ = self.image_processor.generate_crop_boxes( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) snake_case_ = self.image_processor(images=__lowercase , return_tensors="pt" ) with self.device_placement(): if self.framework == "pt": snake_case_ = self.get_inference_context() with inference_context(): snake_case_ = self._ensure_tensor_on_device(__lowercase , device=self.device ) snake_case_ = self.model.get_image_embeddings(model_inputs.pop("pixel_values" ) ) snake_case_ = image_embeddings snake_case_ = grid_points.shape[1] snake_case_ = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( "Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. " "To return all points at once, set points_per_batch to None" ) for i in range(0 , __lowercase , __lowercase ): snake_case_ = grid_points[:, i : i + points_per_batch, :, :] snake_case_ = input_labels[:, i : i + points_per_batch] snake_case_ = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def snake_case__ ( self : Tuple , __lowercase : Union[str, Any] , __lowercase : Union[str, Any]=0.88 , __lowercase : Union[str, Any]=0.95 , __lowercase : int=0 , __lowercase : int=1 , ): """simple docstring""" snake_case_ = model_inputs.pop("input_boxes" ) snake_case_ = model_inputs.pop("is_last" ) snake_case_ = model_inputs.pop("original_sizes" ).tolist() snake_case_ = model_inputs.pop("reshaped_input_sizes" ).tolist() snake_case_ = self.model(**__lowercase ) # post processing happens here in order to avoid CPU GPU copies of ALL the masks snake_case_ = model_outputs["pred_masks"] snake_case_ = self.image_processor.post_process_masks( __lowercase , __lowercase , __lowercase , __lowercase , binarize=__lowercase ) snake_case_ = model_outputs["iou_scores"] snake_case_ , snake_case_ , snake_case_ = self.image_processor.filter_masks( masks[0] , iou_scores[0] , original_sizes[0] , input_boxes[0] , __lowercase , __lowercase , __lowercase , __lowercase , ) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def snake_case__ ( self : str , __lowercase : Any , __lowercase : Optional[int]=False , __lowercase : int=False , __lowercase : List[str]=0.7 , ): """simple docstring""" snake_case_ = [] snake_case_ = [] snake_case_ = [] for model_output in model_outputs: all_scores.append(model_output.pop("iou_scores" ) ) all_masks.extend(model_output.pop("masks" ) ) all_boxes.append(model_output.pop("boxes" ) ) snake_case_ = torch.cat(__lowercase ) snake_case_ = torch.cat(__lowercase ) snake_case_ , snake_case_ , snake_case_ , snake_case_ = self.image_processor.post_process_for_mask_generation( __lowercase , __lowercase , __lowercase , __lowercase ) snake_case_ = defaultdict(__lowercase ) for output in model_outputs: for k, v in output.items(): extra[k].append(__lowercase ) snake_case_ = {} if output_rle_mask: snake_case_ = rle_mask if output_bboxes_mask: snake_case_ = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
187
0
"""simple docstring""" import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow a = logging.getLogger() @unittest.skip('''Temporarily disable the doc tests.''' ) @require_torch @require_tf @slow class lowercase_ ( unittest.TestCase ): '''simple docstring''' def lowerCAmelCase_ ( self : Union[str, Any] , _UpperCAmelCase : Path , _UpperCAmelCase : Union[str, None] = None , _UpperCAmelCase : Union[List[str], None] = None , _UpperCAmelCase : Union[str, List[str], None] = None , _UpperCAmelCase : bool = True , ): _A = [file for file in os.listdir(_UpperCAmelCase ) if os.path.isfile(os.path.join(_UpperCAmelCase , _UpperCAmelCase ) )] if identifier is not None: _A = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(_UpperCAmelCase , _UpperCAmelCase ): for n_ in n_identifier: _A = [file for file in files if n_ not in file] else: _A = [file for file in files if n_identifier not in file] _A = ignore_files or [] ignore_files.append('__init__.py' ) _A = [file for file in files if file not in ignore_files] for file in files: # Open all files print('Testing' , _UpperCAmelCase ) if only_modules: _A = file.split('.' )[0] try: _A = getattr(_UpperCAmelCase , _UpperCAmelCase ) _A = doctest.DocTestSuite(_UpperCAmelCase ) _A = unittest.TextTestRunner().run(_UpperCAmelCase ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(F'''{module_identifier} is not a module.''' ) else: _A = doctest.testfile(str('..' / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def lowerCAmelCase_ ( self : Any ): _A = Path('src/transformers' ) _A = 'modeling' _A = [ 'modeling_ctrl.py', 'modeling_tf_ctrl.py', ] self.analyze_directory(_UpperCAmelCase , identifier=_UpperCAmelCase , ignore_files=_UpperCAmelCase ) def lowerCAmelCase_ ( self : List[Any] ): _A = Path('src/transformers' ) _A = 'tokenization' self.analyze_directory(_UpperCAmelCase , identifier=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Tuple ): _A = Path('src/transformers' ) _A = 'configuration' self.analyze_directory(_UpperCAmelCase , identifier=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Union[str, Any] ): _A = Path('src/transformers' ) _A = ['configuration', 'modeling', 'tokenization'] self.analyze_directory(_UpperCAmelCase , n_identifier=_UpperCAmelCase ) def lowerCAmelCase_ ( self : List[Any] ): _A = Path('docs/source' ) _A = ['favicon.ico'] self.analyze_directory(_UpperCAmelCase , ignore_files=_UpperCAmelCase , only_modules=_UpperCAmelCase )
271
"""simple docstring""" import tempfile import torch from diffusers import IPNDMScheduler from .test_schedulers import SchedulerCommonTest class lowercase_ ( __lowerCAmelCase ): '''simple docstring''' UpperCAmelCase : Union[str, Any] = (IPNDMScheduler,) UpperCAmelCase : Optional[Any] = (('''num_inference_steps''', 50),) def lowerCAmelCase_ ( self : Union[str, Any] , **_UpperCAmelCase : List[Any] ): _A = {'num_train_timesteps': 1_000} config.update(**_UpperCAmelCase ) return config def lowerCAmelCase_ ( self : Tuple , _UpperCAmelCase : Optional[int]=0 , **_UpperCAmelCase : Union[str, Any] ): _A = dict(self.forward_default_kwargs ) _A = kwargs.pop('num_inference_steps' , _UpperCAmelCase ) _A = self.dummy_sample _A = 0.1 * sample _A = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: _A = self.get_scheduler_config(**_UpperCAmelCase ) _A = scheduler_class(**_UpperCAmelCase ) scheduler.set_timesteps(_UpperCAmelCase ) # copy over dummy past residuals _A = dummy_past_residuals[:] if time_step is None: _A = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_UpperCAmelCase ) _A = scheduler_class.from_pretrained(_UpperCAmelCase ) new_scheduler.set_timesteps(_UpperCAmelCase ) # copy over dummy past residuals _A = dummy_past_residuals[:] _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = new_scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = new_scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def lowerCAmelCase_ ( self : str ): pass def lowerCAmelCase_ ( self : Optional[Any] , _UpperCAmelCase : Any=0 , **_UpperCAmelCase : Any ): _A = dict(self.forward_default_kwargs ) _A = kwargs.pop('num_inference_steps' , _UpperCAmelCase ) _A = self.dummy_sample _A = 0.1 * sample _A = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: _A = self.get_scheduler_config() _A = scheduler_class(**_UpperCAmelCase ) scheduler.set_timesteps(_UpperCAmelCase ) # copy over dummy past residuals (must be after setting timesteps) _A = dummy_past_residuals[:] if time_step is None: _A = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_UpperCAmelCase ) _A = scheduler_class.from_pretrained(_UpperCAmelCase ) # copy over dummy past residuals new_scheduler.set_timesteps(_UpperCAmelCase ) # copy over dummy past residual (must be after setting timesteps) _A = dummy_past_residuals[:] _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = new_scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = new_scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def lowerCAmelCase_ ( self : List[str] , **_UpperCAmelCase : Optional[int] ): _A = self.scheduler_classes[0] _A = self.get_scheduler_config(**_UpperCAmelCase ) _A = scheduler_class(**_UpperCAmelCase ) _A = 10 _A = self.dummy_model() _A = self.dummy_sample_deter scheduler.set_timesteps(_UpperCAmelCase ) for i, t in enumerate(scheduler.timesteps ): _A = model(_UpperCAmelCase , _UpperCAmelCase ) _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ).prev_sample for i, t in enumerate(scheduler.timesteps ): _A = model(_UpperCAmelCase , _UpperCAmelCase ) _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ).prev_sample return sample def lowerCAmelCase_ ( self : Union[str, Any] ): _A = dict(self.forward_default_kwargs ) _A = kwargs.pop('num_inference_steps' , _UpperCAmelCase ) for scheduler_class in self.scheduler_classes: _A = self.get_scheduler_config() _A = scheduler_class(**_UpperCAmelCase ) _A = self.dummy_sample _A = 0.1 * sample if num_inference_steps is not None and hasattr(_UpperCAmelCase , 'set_timesteps' ): scheduler.set_timesteps(_UpperCAmelCase ) elif num_inference_steps is not None and not hasattr(_UpperCAmelCase , 'set_timesteps' ): _A = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) _A = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] _A = dummy_past_residuals[:] _A = scheduler.timesteps[5] _A = scheduler.timesteps[6] _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample _A = scheduler.step(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def lowerCAmelCase_ ( self : Tuple ): for timesteps in [100, 1_000]: self.check_over_configs(num_train_timesteps=_UpperCAmelCase , time_step=_UpperCAmelCase ) def lowerCAmelCase_ ( self : List[Any] ): for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=_UpperCAmelCase , time_step=_UpperCAmelCase ) def lowerCAmelCase_ ( self : Optional[int] ): _A = self.full_loop() _A = torch.mean(torch.abs(_UpperCAmelCase ) ) assert abs(result_mean.item() - 2_540_529 ) < 10
271
1
"""simple docstring""" _lowercase = '''Tobias Carryer''' from time import time class lowerCAmelCase_ : '''simple docstring''' def __init__( self : Optional[Any] ,A_ : Optional[int] ,A_ : Any ,A_ : Union[str, Any] ,A_ : Dict=int(time() ) ) -> str: # noqa: B008 A = multiplier A = increment A = modulo A = seed def _SCREAMING_SNAKE_CASE ( self : str ) -> Optional[Any]: A = (self.multiplier * self.seed + self.increment) % self.modulo return self.seed if __name__ == "__main__": # Show the LCG in action. _lowercase = LinearCongruentialGenerator(1_66_45_25, 10_13_90_42_23, 2 << 31) while True: print(lcg.next_number())
74
import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels __lowerCamelCase : List[str] = object() # For specifying empty leaf dict `{}` __lowerCamelCase : Optional[int] = object() def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : List[Any] , __UpperCamelCase : List[str] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ = tuple((re.compile(x + """$""" ) for x in qs) ) for i in range(len(__UpperCamelCase ) - len(__UpperCamelCase ) + 1 ): SCREAMING_SNAKE_CASE__ = [x.match(__UpperCamelCase ) for x, y in zip(__UpperCamelCase , ks[i:] )] if matches and all(__UpperCamelCase ): return True return False def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Optional[int] ) -> List[str]: """simple docstring""" def replace(__UpperCamelCase : Tuple , __UpperCamelCase : Any ): for rule, replacement in rules: if _match(__UpperCamelCase , __UpperCamelCase ): return replacement return val return replace def __SCREAMING_SNAKE_CASE ( ) -> List[Any]: """simple docstring""" return [ # embeddings (("transformer", "wpe", "embedding"), P("""mp""" , __UpperCamelCase )), (("transformer", "wte", "embedding"), P("""mp""" , __UpperCamelCase )), # atention (("attention", "(q_proj|k_proj|v_proj)", "kernel"), P(__UpperCamelCase , """mp""" )), (("attention", "out_proj", "kernel"), P("""mp""" , __UpperCamelCase )), (("attention", "out_proj", "bias"), None), # mlp (("mlp", "c_fc", "kernel"), P(__UpperCamelCase , """mp""" )), (("mlp", "c_fc", "bias"), P("""mp""" )), (("mlp", "c_proj", "kernel"), P("""mp""" , __UpperCamelCase )), (("mlp", "c_proj", "bias"), None), # layer norms ((r"ln_\d+", "bias"), None), ((r"\d+", r"ln_\d+", "scale"), None), (("ln_f", "bias"), None), (("ln_f", "scale"), None), ] def __SCREAMING_SNAKE_CASE ( __UpperCamelCase : Union[str, Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ = _get_partition_rules() SCREAMING_SNAKE_CASE__ = _replacement_rules(__UpperCamelCase ) SCREAMING_SNAKE_CASE__ = {k: _unmatched for k in flatten_dict(__UpperCamelCase )} SCREAMING_SNAKE_CASE__ = {k: replace(__UpperCamelCase , __UpperCamelCase ) for k, v in initd.items()} assert _unmatched not in result.values(), "Incomplete partition spec." return freeze(unflatten_dict(__UpperCamelCase ) )
219
0
def a( A : int = 100 ) -> int: """simple docstring""" a = (n * (n + 1) // 2) ** 2 a = n * (n + 1) * (2 * n + 1) // 6 return sum_cubes - sum_squares if __name__ == "__main__": print(F"""{solution() = }""")
71
import itertools import os from collections import Counter, defaultdict from concurrent.futures import ThreadPoolExecutor, as_completed import numpy as np import datasets from .execute import check_correctness _lowercase: Optional[Any] = "\\n@misc{chen2021evaluating,\n title={Evaluating Large Language Models Trained on Code},\n author={Mark Chen and Jerry Tworek and Heewoo Jun and Qiming Yuan \\nand Henrique Ponde de Oliveira Pinto and Jared Kaplan and Harri Edwards \\nand Yuri Burda and Nicholas Joseph and Greg Brockman and Alex Ray \\nand Raul Puri and Gretchen Krueger and Michael Petrov and Heidy Khlaaf \\nand Girish Sastry and Pamela Mishkin and Brooke Chan and Scott Gray \\nand Nick Ryder and Mikhail Pavlov and Alethea Power and Lukasz Kaiser \\nand Mohammad Bavarian and Clemens Winter and Philippe Tillet \\nand Felipe Petroski Such and Dave Cummings and Matthias Plappert \\nand Fotios Chantzis and Elizabeth Barnes and Ariel Herbert-Voss \\nand William Hebgen Guss and Alex Nichol and Alex Paino and Nikolas Tezak \\nand Jie Tang and Igor Babuschkin and Suchir Balaji and Shantanu Jain \\nand William Saunders and Christopher Hesse and Andrew N. Carr \\nand Jan Leike and Josh Achiam and Vedant Misra and Evan Morikawa \\nand Alec Radford and Matthew Knight and Miles Brundage and Mira Murati \\nand Katie Mayer and Peter Welinder and Bob McGrew and Dario Amodei \\nand Sam McCandlish and Ilya Sutskever and Wojciech Zaremba},\n year={2021},\n eprint={2107.03374},\n archivePrefix={arXiv},\n primaryClass={cs.LG}\n}\n" _lowercase: Union[str, Any] = "\\nThis metric implements the evaluation harness for the HumanEval problem solving dataset\ndescribed in the paper \"Evaluating Large Language Models Trained on Code\"\n(https://arxiv.org/abs/2107.03374).\n" _lowercase: Union[str, Any] = "\nCalculates how good are predictions given some references, using certain scores\nArgs:\n predictions: list of candidates to evaluate. Each candidates should be a list\n of strings with several code candidates to solve the problem.\n references: a list with a test for each prediction. Each test should evaluate the\n correctness of a code candidate.\n k: number of code candidates to consider in the evaluation (Default: [1, 10, 100])\n num_workers: number of workers used to evaluate the canidate programs (Default: 4).\n timeout:\nReturns:\n pass_at_k: dict with pass rates for each k\n results: dict with granular results of each unittest\nExamples:\n >>> code_eval = datasets.load_metric(\"code_eval\")\n >>> test_cases = [\"assert add(2,3)==5\"]\n >>> candidates = [[\"def add(a,b): return a*b\", \"def add(a, b): return a+b\"]]\n >>> pass_at_k, results = code_eval.compute(references=test_cases, predictions=candidates, k=[1, 2])\n >>> print(pass_at_k)\n {'pass@1': 0.5, 'pass@2': 1.0}\n" _lowercase: List[Any] = "\n################################################################################\n !!!WARNING!!!\n################################################################################\nThe \"code_eval\" metric executes untrusted model-generated code in Python.\nAlthough it is highly unlikely that model-generated code will do something\novertly malicious in response to this test suite, model-generated code may act\ndestructively due to a lack of model capability or alignment.\nUsers are strongly encouraged to sandbox this evaluation suite so that it\ndoes not perform destructive actions on their host or network. For more\ninformation on how OpenAI sandboxes its code, see the paper \"Evaluating Large\nLanguage Models Trained on Code\" (https://arxiv.org/abs/2107.03374).\n\nOnce you have read this disclaimer and taken appropriate precautions,\nset the environment variable HF_ALLOW_CODE_EVAL=\"1\". Within Python you can to this\nwith:\n\n>>> import os\n>>> os.environ[\"HF_ALLOW_CODE_EVAL\"] = \"1\"\n\n################################################################################\\n" _lowercase: List[Any] = "The MIT License\n\nCopyright (c) OpenAI (https://openai.com)\n\nPermission is hereby granted, free of charge, to any person obtaining a copy\nof this software and associated documentation files (the \"Software\"), to deal\nin the Software without restriction, including without limitation the rights\nto use, copy, modify, merge, publish, distribute, sublicense, and/or sell\ncopies of the Software, and to permit persons to whom the Software is\nfurnished to do so, subject to the following conditions:\n\nThe above copyright notice and this permission notice shall be included in\nall copies or substantial portions of the Software.\n\nTHE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\nIMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\nFITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\nAUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\nLIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\nOUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\nTHE SOFTWARE." @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class _lowercase ( datasets.Metric ): """simple docstring""" def UpperCamelCase_ (self ): """simple docstring""" return datasets.MetricInfo( # This is the description that will appear on the metrics page. description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Sequence(datasets.Value("string" ) ), "references": datasets.Value("string" ), } ) , homepage="https://github.com/openai/human-eval" , codebase_urls=["https://github.com/openai/human-eval"] , reference_urls=["https://github.com/openai/human-eval"] , license=_LICENSE , ) def UpperCamelCase_ (self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=[1, 10, 100] , lowerCamelCase_=4 , lowerCamelCase_=3.0 ): """simple docstring""" if os.getenv("HF_ALLOW_CODE_EVAL" , 0 ) != "1": raise ValueError(_WARNING ) if os.name == "nt": raise NotImplementedError("This metric is currently not supported on Windows." ) with ThreadPoolExecutor(max_workers=lowerCamelCase_ ) as executor: a = [] a = Counter() a = 0 a = defaultdict(lowerCamelCase_ ) for task_id, (candidates, test_case) in enumerate(zip(lowerCamelCase_ , lowerCamelCase_ ) ): for candidate in candidates: a = candidate + "\n" + test_case a = (test_program, timeout, task_id, completion_id[task_id]) a = executor.submit(lowerCamelCase_ , *lowerCamelCase_ ) futures.append(lowerCamelCase_ ) completion_id[task_id] += 1 n_samples += 1 for future in as_completed(lowerCamelCase_ ): a = future.result() results[result["task_id"]].append((result["completion_id"], result) ) a , a = [], [] for result in results.values(): result.sort() a = [r[1]["passed"] for r in result] total.append(len(lowerCamelCase_ ) ) correct.append(sum(lowerCamelCase_ ) ) a = np.array(lowerCamelCase_ ) a = np.array(lowerCamelCase_ ) a = k a = {F'''pass@{k}''': estimate_pass_at_k(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ).mean() for k in ks if (total >= k).all()} return pass_at_k, results def a( A : Optional[Any] , A : str , A : Dict ) -> Optional[int]: """simple docstring""" def estimator(A : int , A : int , A : int ) -> float: if n - c < k: return 1.0 return 1.0 - np.prod(1.0 - k / np.arange(n - c + 1 , n + 1 ) ) if isinstance(A , A ): a = itertools.repeat(A , len(A ) ) else: assert len(A ) == len(A ) a = iter(A ) return np.array([estimator(int(A ) , int(A ) , A ) for n, c in zip(A , A )] )
71
1
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import platform import numpy as np import psutil import torch from accelerate import __version__ as version from accelerate.commands.config import default_config_file, load_config_from_file from ..utils import is_npu_available, is_xpu_available def SCREAMING_SNAKE_CASE ( _lowerCamelCase : str=None ) -> Tuple: if subparsers is not None: _lowerCAmelCase : Optional[int] = subparsers.add_parser("""env""" ) else: _lowerCAmelCase : List[Any] = argparse.ArgumentParser("""Accelerate env command""" ) parser.add_argument( """--config_file""" ,default=__lowerCAmelCase ,help="""The config file to use for the default values in the launching script.""" ) if subparsers is not None: parser.set_defaults(func=__lowerCAmelCase ) return parser def SCREAMING_SNAKE_CASE ( _lowerCamelCase : Optional[int] ) -> Any: _lowerCAmelCase : List[str] = torch.__version__ _lowerCAmelCase : Tuple = torch.cuda.is_available() _lowerCAmelCase : int = is_xpu_available() _lowerCAmelCase : Union[str, Any] = is_npu_available() _lowerCAmelCase : Union[str, Any] = '''Not found''' # Get the default from the config file. if args.config_file is not None or os.path.isfile(__lowerCAmelCase ): _lowerCAmelCase : int = load_config_from_file(args.config_file ).to_dict() _lowerCAmelCase : List[str] = { '''`Accelerate` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''Numpy version''': np.__version__, '''PyTorch version (GPU?)''': f"{pt_version} ({pt_cuda_available})", '''PyTorch XPU available''': str(__lowerCAmelCase ), '''PyTorch NPU available''': str(__lowerCAmelCase ), '''System RAM''': f"{psutil.virtual_memory().total / 1024 ** 3:.2f} GB", } if pt_cuda_available: _lowerCAmelCase : List[str] = torch.cuda.get_device_name() print("""\nCopy-and-paste the text below in your GitHub issue\n""" ) print("""\n""".join([f"- {prop}: {val}" for prop, val in info.items()] ) ) print("""- `Accelerate` default config:""" if args.config_file is None else """- `Accelerate` config passed:""" ) _lowerCAmelCase : List[str] = ( '''\n'''.join([f"\t- {prop}: {val}" for prop, val in accelerate_config.items()] ) if isinstance(__lowerCAmelCase ,__lowerCAmelCase ) else f"\t{accelerate_config}" ) print(__lowerCAmelCase ) _lowerCAmelCase : List[str] = accelerate_config return info def SCREAMING_SNAKE_CASE ( ) -> int: _lowerCAmelCase : Optional[Any] = env_command_parser() _lowerCAmelCase : List[str] = parser.parse_args() env_command(__lowerCAmelCase ) return 0 if __name__ == "__main__": raise SystemExit(main())
44
A__ = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] A__ = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] A__ = { 0: '''Sunday''', 1: '''Monday''', 2: '''Tuesday''', 3: '''Wednesday''', 4: '''Thursday''', 5: '''Friday''', 6: '''Saturday''', } def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> str: """simple docstring""" assert len(str(__lowerCAmelCase ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: snake_case__ : Optional[int] = year // 100 snake_case__ : List[str] = (5 * (century % 4) + 2) % 7 snake_case__ : Dict = year % 100 snake_case__ : Union[str, Any] = centurian % 12 snake_case__ : List[Any] = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 snake_case__ : List[str] = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 400) == 0) else DOOMSDAY_LEAP[month - 1] ) snake_case__ : List[str] = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
230
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, is_vision_available, ) _A = {"""configuration_vit""": ["""VIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ViTConfig""", """ViTOnnxConfig"""]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = ["""ViTFeatureExtractor"""] _A = ["""ViTImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ """VIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """ViTForImageClassification""", """ViTForMaskedImageModeling""", """ViTModel""", """ViTPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ """TFViTForImageClassification""", """TFViTModel""", """TFViTPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ """FlaxViTForImageClassification""", """FlaxViTModel""", """FlaxViTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_vit import VIT_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTConfig, ViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_vit import ViTFeatureExtractor from .image_processing_vit import ViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit import ( VIT_PRETRAINED_MODEL_ARCHIVE_LIST, ViTForImageClassification, ViTForMaskedImageModeling, ViTModel, ViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit import TFViTForImageClassification, TFViTModel, TFViTPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel, FlaxViTPreTrainedModel else: import sys _A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
212
"""simple docstring""" import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask _A = logging.getLogger(__name__) class _lowerCamelCase ( a_ ): def __init__( self : List[Any] , UpperCamelCase : Dict=-1 ) -> List[Any]: """simple docstring""" # in NER datasets, the last column is usually reserved for NER label lowerCAmelCase__ : Optional[int] = label_idx def _lowerCAmelCase ( self : Optional[int] , UpperCamelCase : Optional[int] , UpperCamelCase : Union[Split, str] ) -> List[InputExample]: """simple docstring""" if isinstance(UpperCamelCase , UpperCamelCase ): lowerCAmelCase__ : int = mode.value lowerCAmelCase__ : Dict = os.path.join(UpperCamelCase , f"""{mode}.txt""" ) lowerCAmelCase__ : str = 1 lowerCAmelCase__ : List[Any] = [] with open(UpperCamelCase , encoding="""utf-8""" ) as f: lowerCAmelCase__ : str = [] lowerCAmelCase__ : Any = [] for line in f: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCamelCase , labels=UpperCamelCase ) ) guid_index += 1 lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Tuple = [] else: lowerCAmelCase__ : Optional[int] = line.split(""" """ ) words.append(splits[0] ) if len(UpperCamelCase ) > 1: labels.append(splits[self.label_idx].replace("""\n""" , """""" ) ) else: # Examples could have no label for mode = "test" labels.append("""O""" ) if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCamelCase , labels=UpperCamelCase ) ) return examples def _lowerCAmelCase ( self : Any , UpperCamelCase : TextIO , UpperCamelCase : TextIO , UpperCamelCase : List ) -> List[Any]: """simple docstring""" lowerCAmelCase__ : Dict = 0 for line in test_input_reader: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": writer.write(UpperCamelCase ) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: lowerCAmelCase__ : Union[str, Any] = line.split()[0] + """ """ + preds_list[example_id].pop(0 ) + """\n""" writer.write(UpperCamelCase ) else: logger.warning("""Maximum sequence length exceeded: No prediction for '%s'.""" , line.split()[0] ) def _lowerCAmelCase ( self : str , UpperCamelCase : str ) -> List[str]: """simple docstring""" if path: with open(UpperCamelCase , """r""" ) as f: lowerCAmelCase__ : Any = f.read().splitlines() if "O" not in labels: lowerCAmelCase__ : List[str] = ["""O"""] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class _lowerCamelCase ( a_ ): def __init__( self : Union[str, Any] ) -> Any: """simple docstring""" # in CONLL2003 dataset chunk column is second-to-last super().__init__(label_idx=-2 ) def _lowerCAmelCase ( self : Optional[Any] , UpperCamelCase : str ) -> List[str]: """simple docstring""" if path: with open(UpperCamelCase , """r""" ) as f: lowerCAmelCase__ : Any = f.read().splitlines() if "O" not in labels: lowerCAmelCase__ : str = ["""O"""] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class _lowerCamelCase ( a_ ): def _lowerCAmelCase ( self : Union[str, Any] , UpperCamelCase : Any , UpperCamelCase : Union[Split, str] ) -> List[InputExample]: """simple docstring""" if isinstance(UpperCamelCase , UpperCamelCase ): lowerCAmelCase__ : Optional[Any] = mode.value lowerCAmelCase__ : int = os.path.join(UpperCamelCase , f"""{mode}.txt""" ) lowerCAmelCase__ : Optional[int] = 1 lowerCAmelCase__ : List[Any] = [] with open(UpperCamelCase , encoding="""utf-8""" ) as f: for sentence in parse_incr(UpperCamelCase ): lowerCAmelCase__ : Union[str, Any] = [] lowerCAmelCase__ : List[Any] = [] for token in sentence: words.append(token["""form"""] ) labels.append(token["""upos"""] ) assert len(UpperCamelCase ) == len(UpperCamelCase ) if words: examples.append(InputExample(guid=f"""{mode}-{guid_index}""" , words=UpperCamelCase , labels=UpperCamelCase ) ) guid_index += 1 return examples def _lowerCAmelCase ( self : List[str] , UpperCamelCase : TextIO , UpperCamelCase : TextIO , UpperCamelCase : List ) -> List[Any]: """simple docstring""" lowerCAmelCase__ : Any = 0 for sentence in parse_incr(UpperCamelCase ): lowerCAmelCase__ : Union[str, Any] = preds_list[example_id] lowerCAmelCase__ : List[Any] = """""" for token in sentence: out += f"""{token['form']} ({token['upos']}|{s_p.pop(0 )}) """ out += "\n" writer.write(UpperCamelCase ) example_id += 1 def _lowerCAmelCase ( self : Dict , UpperCamelCase : str ) -> List[str]: """simple docstring""" if path: with open(UpperCamelCase , """r""" ) as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
212
1
'''simple docstring''' from typing import Any class _UpperCamelCase : '''simple docstring''' def __init__( self : int , lowerCAmelCase__ : Any ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = data __SCREAMING_SNAKE_CASE : int = None def __repr__( self : Any ): """simple docstring""" return F"Node({self.data})" class _UpperCamelCase : '''simple docstring''' def __init__( self : Dict ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = None def __iter__( self : Optional[int] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Union[str, Any] = self.head while node: yield node.data __SCREAMING_SNAKE_CASE : Union[str, Any] = node.next def __len__( self : Tuple ): """simple docstring""" return sum(1 for _ in self ) def __repr__( self : Any ): """simple docstring""" return "->".join([str(lowerCAmelCase__ ) for item in self] ) def __getitem__( self : List[str] , lowerCAmelCase__ : int ): """simple docstring""" if not 0 <= index < len(self ): raise ValueError("""list index out of range.""" ) for i, node in enumerate(self ): if i == index: return node return None def __setitem__( self : str , lowerCAmelCase__ : int , lowerCAmelCase__ : Any ): """simple docstring""" if not 0 <= index < len(self ): raise ValueError("""list index out of range.""" ) __SCREAMING_SNAKE_CASE : Tuple = self.head for _ in range(lowerCAmelCase__ ): __SCREAMING_SNAKE_CASE : Union[str, Any] = current.next __SCREAMING_SNAKE_CASE : Tuple = data def UpperCamelCase__ ( self : Union[str, Any] , lowerCAmelCase__ : Any ): """simple docstring""" self.insert_nth(len(self ) , lowerCAmelCase__ ) def UpperCamelCase__ ( self : Any , lowerCAmelCase__ : Any ): """simple docstring""" self.insert_nth(0 , lowerCAmelCase__ ) def UpperCamelCase__ ( self : Dict , lowerCAmelCase__ : int , lowerCAmelCase__ : Any ): """simple docstring""" if not 0 <= index <= len(self ): raise IndexError("""list index out of range""" ) __SCREAMING_SNAKE_CASE : List[str] = Node(lowerCAmelCase__ ) if self.head is None: __SCREAMING_SNAKE_CASE : Optional[Any] = new_node elif index == 0: __SCREAMING_SNAKE_CASE : Tuple = self.head # link new_node to head __SCREAMING_SNAKE_CASE : str = new_node else: __SCREAMING_SNAKE_CASE : str = self.head for _ in range(index - 1 ): __SCREAMING_SNAKE_CASE : Union[str, Any] = temp.next __SCREAMING_SNAKE_CASE : List[Any] = temp.next __SCREAMING_SNAKE_CASE : Dict = new_node def UpperCamelCase__ ( self : int ): # print every node data """simple docstring""" print(self ) def UpperCamelCase__ ( self : Optional[int] ): """simple docstring""" return self.delete_nth(0 ) def UpperCamelCase__ ( self : Any ): # delete from tail """simple docstring""" return self.delete_nth(len(self ) - 1 ) def UpperCamelCase__ ( self : str , lowerCAmelCase__ : int = 0 ): """simple docstring""" if not 0 <= index <= len(self ) - 1: # test if index is valid raise IndexError("""List index out of range.""" ) __SCREAMING_SNAKE_CASE : Dict = self.head # default first node if index == 0: __SCREAMING_SNAKE_CASE : Union[str, Any] = self.head.next else: __SCREAMING_SNAKE_CASE : int = self.head for _ in range(index - 1 ): __SCREAMING_SNAKE_CASE : List[str] = temp.next __SCREAMING_SNAKE_CASE : int = temp.next __SCREAMING_SNAKE_CASE : Optional[Any] = temp.next.next return delete_node.data def UpperCamelCase__ ( self : Tuple ): """simple docstring""" return self.head is None def UpperCamelCase__ ( self : str ): """simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = None __SCREAMING_SNAKE_CASE : Tuple = self.head while current: # Store the current node's next node. __SCREAMING_SNAKE_CASE : Optional[int] = current.next # Make the current node's next point backwards __SCREAMING_SNAKE_CASE : Optional[Any] = prev # Make the previous node be the current node __SCREAMING_SNAKE_CASE : List[Any] = current # Make the current node the next node (to progress iteration) __SCREAMING_SNAKE_CASE : Tuple = next_node # Return prev in order to put the head at the end __SCREAMING_SNAKE_CASE : Any = prev def lowerCAmelCase_ ( ): __SCREAMING_SNAKE_CASE : Union[str, Any] = LinkedList() assert linked_list.is_empty() is True assert str(_lowerCamelCase ) == "" try: linked_list.delete_head() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. try: linked_list.delete_tail() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. for i in range(10 ): assert len(_lowerCamelCase ) == i linked_list.insert_nth(_lowerCamelCase , i + 1 ) assert str(_lowerCamelCase ) == "->".join(str(_lowerCamelCase ) for i in range(1 , 11 ) ) linked_list.insert_head(0 ) linked_list.insert_tail(11 ) assert str(_lowerCamelCase ) == "->".join(str(_lowerCamelCase ) for i in range(0 , 12 ) ) assert linked_list.delete_head() == 0 assert linked_list.delete_nth(9 ) == 10 assert linked_list.delete_tail() == 11 assert len(_lowerCamelCase ) == 9 assert str(_lowerCamelCase ) == "->".join(str(_lowerCamelCase ) for i in range(1 , 10 ) ) assert all(linked_list[i] == i + 1 for i in range(0 , 9 ) ) is True for i in range(0 , 9 ): __SCREAMING_SNAKE_CASE : List[Any] = -i assert all(linked_list[i] == -i for i in range(0 , 9 ) ) is True linked_list.reverse() assert str(_lowerCamelCase ) == "->".join(str(_lowerCamelCase ) for i in range(-8 , 1 ) ) def lowerCAmelCase_ ( ): __SCREAMING_SNAKE_CASE : str = [ -9, 1_00, Node(77_34_51_12 ), """dlrow olleH""", 7, 55_55, 0, -1_92.5_55_55, """Hello, world!""", 77.9, Node(10 ), None, None, 12.20, ] __SCREAMING_SNAKE_CASE : Any = LinkedList() for i in test_input: linked_list.insert_tail(_lowerCamelCase ) # Check if it's empty or not assert linked_list.is_empty() is False assert ( str(_lowerCamelCase ) == "-9->100->Node(77345112)->dlrow olleH->7->5555->0->" "-192.55555->Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the head __SCREAMING_SNAKE_CASE : Optional[Any] = linked_list.delete_head() assert result == -9 assert ( str(_lowerCamelCase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the tail __SCREAMING_SNAKE_CASE : Tuple = linked_list.delete_tail() assert result == 12.2 assert ( str(_lowerCamelCase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None" ) # Delete a node in specific location in linked list __SCREAMING_SNAKE_CASE : Any = linked_list.delete_nth(10 ) assert result is None assert ( str(_lowerCamelCase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None" ) # Add a Node instance to its head linked_list.insert_head(Node("""Hello again, world!""" ) ) assert ( str(_lowerCamelCase ) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None" ) # Add None to its tail linked_list.insert_tail(_lowerCamelCase ) assert ( str(_lowerCamelCase ) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None->None" ) # Reverse the linked list linked_list.reverse() assert ( str(_lowerCamelCase ) == "None->None->Node(10)->77.9->Hello, world!->-192.55555->0->5555->" "7->dlrow olleH->Node(77345112)->100->Node(Hello again, world!)" ) def lowerCAmelCase_ ( ): from doctest import testmod testmod() __SCREAMING_SNAKE_CASE : List[Any] = LinkedList() linked_list.insert_head(input("""Inserting 1st at head """ ).strip() ) linked_list.insert_head(input("""Inserting 2nd at head """ ).strip() ) print("""\nPrint list:""" ) linked_list.print_list() linked_list.insert_tail(input("""\nInserting 1st at tail """ ).strip() ) linked_list.insert_tail(input("""Inserting 2nd at tail """ ).strip() ) print("""\nPrint list:""" ) linked_list.print_list() print("""\nDelete head""" ) linked_list.delete_head() print("""Delete tail""" ) linked_list.delete_tail() print("""\nPrint list:""" ) linked_list.print_list() print("""\nReverse linked list""" ) linked_list.reverse() print("""\nPrint list:""" ) linked_list.print_list() print("""\nString representation of linked list:""" ) print(_lowerCamelCase ) print("""\nReading/changing Node data using indexing:""" ) print(F"Element at Position 1: {linked_list[1]}" ) __SCREAMING_SNAKE_CASE : Dict = input("""Enter New Value: """ ).strip() print("""New list:""" ) print(_lowerCamelCase ) print(F"length of linked_list is : {len(_lowerCamelCase )}" ) if __name__ == "__main__": main()
112
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase__ : Any = { '''configuration_instructblip''': [ '''INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InstructBlipConfig''', '''InstructBlipQFormerConfig''', '''InstructBlipVisionConfig''', ], '''processing_instructblip''': ['''InstructBlipProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : Optional[Any] = [ '''INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''InstructBlipQFormerModel''', '''InstructBlipPreTrainedModel''', '''InstructBlipForConditionalGeneration''', '''InstructBlipVisionModel''', ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys UpperCamelCase__ : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
112
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available lowerCAmelCase = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase = ['BartphoTokenizer'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
356
import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger lowerCAmelCase = '<<<<<<< This should probably be modified because it mentions: ' lowerCAmelCase = '=======\n>>>>>>>\n' lowerCAmelCase = [ 'TextEncoderConfig', 'ByteTextEncoder', 'SubwordTextEncoder', 'encoder_config', 'maybe_build_from_corpus', 'manual_dir', ] lowerCAmelCase = [ # (pattern, replacement) # Order is important here for some replacements (R'tfds\.core', R'datasets'), (R'tf\.io\.gfile\.GFile', R'open'), (R'tf\.([\w\d]+)', R'datasets.Value(\'\1\')'), (R'tfds\.features\.Text\(\)', R'datasets.Value(\'string\')'), (R'tfds\.features\.Text\(', R'datasets.Value(\'string\'),'), (R'features\s*=\s*tfds.features.FeaturesDict\(', R'features=datasets.Features('), (R'tfds\.features\.FeaturesDict\(', R'dict('), (R'The TensorFlow Datasets Authors', R'The TensorFlow Datasets Authors and the HuggingFace Datasets Authors'), (R'tfds\.', R'datasets.'), (R'dl_manager\.manual_dir', R'self.config.data_dir'), (R'self\.builder_config', R'self.config'), ] def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" return ConvertCommand(args.tfds_path , args.datasets_directory ) class _a ( UpperCamelCase__ ): @staticmethod def lowerCamelCase_ ( UpperCamelCase_: ArgumentParser ) -> int: """simple docstring""" lowercase__ = parser.add_parser( '''convert''' , help='''Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.''' , ) train_parser.add_argument( '''--tfds_path''' , type=UpperCamelCase_ , required=UpperCamelCase_ , help='''Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.''' , ) train_parser.add_argument( '''--datasets_directory''' , type=UpperCamelCase_ , required=UpperCamelCase_ , help='''Path to the HuggingFace Datasets folder.''' ) train_parser.set_defaults(func=UpperCamelCase_ ) def __init__( self: Dict , UpperCamelCase_: str , UpperCamelCase_: str , *UpperCamelCase_: Union[str, Any] ) -> List[str]: """simple docstring""" lowercase__ = get_logger('''datasets-cli/converting''' ) lowercase__ = tfds_path lowercase__ = datasets_directory def lowerCamelCase_ ( self: str ) -> Dict: """simple docstring""" if os.path.isdir(self._tfds_path ): lowercase__ = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): lowercase__ = os.path.dirname(self._tfds_path ) else: raise ValueError('''--tfds_path is neither a directory nor a file. Please check path.''' ) lowercase__ = os.path.abspath(self._datasets_directory ) self._logger.info(f'Converting datasets from {abs_tfds_path} to {abs_datasets_path}' ) lowercase__ = [] lowercase__ = [] lowercase__ = {} if os.path.isdir(self._tfds_path ): lowercase__ = os.listdir(UpperCamelCase_ ) else: lowercase__ = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(f'Looking at file {f_name}' ) lowercase__ = os.path.join(UpperCamelCase_ , UpperCamelCase_ ) lowercase__ = os.path.join(UpperCamelCase_ , UpperCamelCase_ ) if not os.path.isfile(UpperCamelCase_ ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info('''Skipping file''' ) continue with open(UpperCamelCase_ , encoding='''utf-8''' ) as f: lowercase__ = f.readlines() lowercase__ = [] lowercase__ = False lowercase__ = False lowercase__ = [] for line in lines: lowercase__ = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: lowercase__ = '''import datasets\n''' elif "import tensorflow" in out_line: # order is important here lowercase__ = '''''' continue elif "from absl import logging" in out_line: lowercase__ = '''from datasets import logging\n''' elif "getLogger" in out_line: lowercase__ = out_line.replace('''getLogger''' , '''get_logger''' ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): lowercase__ = True lowercase__ = list(filter(lambda UpperCamelCase_ : e in out_line , UpperCamelCase_ ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(UpperCamelCase_ ) + '''\n''' ) out_lines.append(UpperCamelCase_ ) out_lines.append(UpperCamelCase_ ) continue else: for pattern, replacement in TO_CONVERT: lowercase__ = re.sub(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: lowercase__ = re.match(r'''from\stensorflow_datasets.*import\s([^\.\r\n]+)''' , UpperCamelCase_ ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split(''',''' ) ) lowercase__ = '''from . import ''' + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(f'Error converting {out_line.strip()}' ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: lowercase__ = True out_lines.append(UpperCamelCase_ ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset lowercase__ = f_name.replace('''.py''' , '''''' ) lowercase__ = os.path.join(UpperCamelCase_ , UpperCamelCase_ ) lowercase__ = os.path.join(UpperCamelCase_ , UpperCamelCase_ ) os.makedirs(UpperCamelCase_ , exist_ok=UpperCamelCase_ ) self._logger.info(f'Adding directory {output_dir}' ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(UpperCamelCase_ ) if needs_manual_update: with_manual_update.append(UpperCamelCase_ ) with open(UpperCamelCase_ , '''w''' , encoding='''utf-8''' ) as f: f.writelines(UpperCamelCase_ ) self._logger.info(f'Converted in {output_file}' ) for utils_file in utils_files: try: lowercase__ = os.path.basename(UpperCamelCase_ ) lowercase__ = imports_to_builder_map[f_name.replace('''.py''' , '''''' )] self._logger.info(f'Moving {dest_folder} to {utils_file}' ) shutil.copy(UpperCamelCase_ , UpperCamelCase_ ) except KeyError: self._logger.error(f'Cannot find destination folder for {utils_file}. Please copy manually.' ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( f'You need to manually update file {file_path} to remove configurations using \'TextEncoderConfig\'.' )
93
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import BeitConfig, BeitForImageClassification, BeitForMaskedImageModeling, BeitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() lowercase : Dict = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : Optional[Any] , _lowerCamelCase : Union[str, Any]=False , _lowerCamelCase : Optional[int]=False) -> Any: '''simple docstring''' __UpperCamelCase : Optional[int] = "backbone." if is_semantic else "" __UpperCamelCase : Optional[Any] = [] for i in range(config.num_hidden_layers): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'{prefix}blocks.{i}.norm1.weight', F'beit.encoder.layer.{i}.layernorm_before.weight')) rename_keys.append((F'{prefix}blocks.{i}.norm1.bias', F'beit.encoder.layer.{i}.layernorm_before.bias')) rename_keys.append( (F'{prefix}blocks.{i}.attn.proj.weight', F'beit.encoder.layer.{i}.attention.output.dense.weight')) rename_keys.append( (F'{prefix}blocks.{i}.attn.proj.bias', F'beit.encoder.layer.{i}.attention.output.dense.bias')) rename_keys.append((F'{prefix}blocks.{i}.norm2.weight', F'beit.encoder.layer.{i}.layernorm_after.weight')) rename_keys.append((F'{prefix}blocks.{i}.norm2.bias', F'beit.encoder.layer.{i}.layernorm_after.bias')) rename_keys.append((F'{prefix}blocks.{i}.mlp.fc1.weight', F'beit.encoder.layer.{i}.intermediate.dense.weight')) rename_keys.append((F'{prefix}blocks.{i}.mlp.fc1.bias', F'beit.encoder.layer.{i}.intermediate.dense.bias')) rename_keys.append((F'{prefix}blocks.{i}.mlp.fc2.weight', F'beit.encoder.layer.{i}.output.dense.weight')) rename_keys.append((F'{prefix}blocks.{i}.mlp.fc2.bias', F'beit.encoder.layer.{i}.output.dense.bias')) # projection layer + position embeddings rename_keys.extend( [ (F'{prefix}cls_token', "beit.embeddings.cls_token"), (F'{prefix}patch_embed.proj.weight', "beit.embeddings.patch_embeddings.projection.weight"), (F'{prefix}patch_embed.proj.bias', "beit.embeddings.patch_embeddings.projection.bias"), (F'{prefix}pos_embed', "beit.embeddings.position_embeddings"), ]) if has_lm_head: # mask token + layernorm rename_keys.extend( [ ("mask_token", "beit.embeddings.mask_token"), ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ]) else: # layernorm + classification head rename_keys.extend( [ ("fc_norm.weight", "beit.pooler.layernorm.weight"), ("fc_norm.bias", "beit.pooler.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ]) return rename_keys def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : Tuple , _lowerCamelCase : List[str] , _lowerCamelCase : Union[str, Any]=False , _lowerCamelCase : Optional[int]=False) -> List[str]: '''simple docstring''' for i in range(config.num_hidden_layers): __UpperCamelCase : List[Any] = "backbone." if is_semantic else "" # queries, keys and values __UpperCamelCase : Tuple = state_dict.pop(F'{prefix}blocks.{i}.attn.qkv.weight') __UpperCamelCase : Optional[Any] = state_dict.pop(F'{prefix}blocks.{i}.attn.q_bias') __UpperCamelCase : int = state_dict.pop(F'{prefix}blocks.{i}.attn.v_bias') __UpperCamelCase : Optional[int] = in_proj_weight[ : config.hidden_size, : ] __UpperCamelCase : Any = q_bias __UpperCamelCase : Dict = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] __UpperCamelCase : List[Any] = in_proj_weight[ -config.hidden_size :, : ] __UpperCamelCase : Tuple = v_bias # gamma_1 and gamma_2 # we call them lambda because otherwise they are renamed when using .from_pretrained __UpperCamelCase : Union[str, Any] = state_dict.pop(F'{prefix}blocks.{i}.gamma_1') __UpperCamelCase : Optional[int] = state_dict.pop(F'{prefix}blocks.{i}.gamma_2') __UpperCamelCase : Any = gamma_a __UpperCamelCase : List[Any] = gamma_a def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : str , _lowerCamelCase : List[str] , _lowerCamelCase : Optional[Any]) -> Any: '''simple docstring''' __UpperCamelCase : str = dct.pop(_lowerCamelCase) __UpperCamelCase : str = val def _SCREAMING_SNAKE_CASE ( ) -> Optional[int]: '''simple docstring''' __UpperCamelCase : Any = "http://images.cocodataset.org/val2017/000000039769.jpg" __UpperCamelCase : Dict = Image.open(requests.get(_lowerCamelCase , stream=_lowerCamelCase).raw) return im @torch.no_grad() def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : Optional[int] , _lowerCamelCase : List[str] , _lowerCamelCase : Dict=False) -> str: '''simple docstring''' __UpperCamelCase : str = False if "rvlcdip" in checkpoint_url else True __UpperCamelCase : int = BeitConfig(use_absolute_position_embeddings=_lowerCamelCase , use_mask_token=_lowerCamelCase) # size of the architecture if "large" in checkpoint_url or "dit-l" in checkpoint_url: __UpperCamelCase : Optional[int] = 1_024 __UpperCamelCase : Union[str, Any] = 4_096 __UpperCamelCase : Tuple = 24 __UpperCamelCase : int = 16 # labels if "rvlcdip" in checkpoint_url: __UpperCamelCase : List[str] = 16 __UpperCamelCase : str = "huggingface/label-files" __UpperCamelCase : str = "rvlcdip-id2label.json" __UpperCamelCase : Tuple = json.load(open(hf_hub_download(_lowerCamelCase , _lowerCamelCase , repo_type="dataset") , "r")) __UpperCamelCase : Dict = {int(_lowerCamelCase): v for k, v in idalabel.items()} __UpperCamelCase : Optional[Any] = idalabel __UpperCamelCase : List[Any] = {v: k for k, v in idalabel.items()} # load state_dict of original model, remove and rename some keys __UpperCamelCase : str = torch.hub.load_state_dict_from_url(_lowerCamelCase , map_location="cpu")["model"] __UpperCamelCase : Optional[Any] = create_rename_keys(_lowerCamelCase , has_lm_head=_lowerCamelCase) for src, dest in rename_keys: rename_key(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase) read_in_q_k_v(_lowerCamelCase , _lowerCamelCase , has_lm_head=_lowerCamelCase) # load HuggingFace model __UpperCamelCase : List[Any] = BeitForMaskedImageModeling(_lowerCamelCase) if has_lm_head else BeitForImageClassification(_lowerCamelCase) model.eval() model.load_state_dict(_lowerCamelCase) # Check outputs on an image __UpperCamelCase : Dict = BeitImageProcessor( size=config.image_size , resample=PILImageResampling.BILINEAR , do_center_crop=_lowerCamelCase) __UpperCamelCase : Union[str, Any] = prepare_img() __UpperCamelCase : Union[str, Any] = image_processor(images=_lowerCamelCase , return_tensors="pt") __UpperCamelCase : str = encoding["pixel_values"] __UpperCamelCase : str = model(_lowerCamelCase) __UpperCamelCase : Union[str, Any] = outputs.logits # verify logits __UpperCamelCase : Tuple = [1, 16] if "rvlcdip" in checkpoint_url else [1, 196, 8_192] assert logits.shape == torch.Size(_lowerCamelCase), "Shape of logits not as expected" Path(_lowerCamelCase).mkdir(exist_ok=_lowerCamelCase) print(F'Saving model to {pytorch_dump_folder_path}') model.save_pretrained(_lowerCamelCase) print(F'Saving image processor to {pytorch_dump_folder_path}') image_processor.save_pretrained(_lowerCamelCase) if push_to_hub: if has_lm_head: __UpperCamelCase : List[str] = "dit-base" if "base" in checkpoint_url else "dit-large" else: __UpperCamelCase : Optional[Any] = "dit-base-finetuned-rvlcdip" if "dit-b" in checkpoint_url else "dit-large-finetuned-rvlcdip" image_processor.push_to_hub( repo_path_or_name=Path(_lowerCamelCase , _lowerCamelCase) , organization="nielsr" , commit_message="Add image processor" , use_temp_dir=_lowerCamelCase , ) model.push_to_hub( repo_path_or_name=Path(_lowerCamelCase , _lowerCamelCase) , organization="nielsr" , commit_message="Add model" , use_temp_dir=_lowerCamelCase , ) if __name__ == "__main__": lowercase : Optional[Any] = argparse.ArgumentParser() parser.add_argument( '--checkpoint_url', default='https://layoutlm.blob.core.windows.net/dit/dit-pts/dit-base-224-p16-500k-62d53a.pth', type=str, help='URL to the original PyTorch checkpoint (.pth file).', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument( '--push_to_hub', action='store_true', ) lowercase : Tuple = parser.parse_args() convert_dit_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
232
# XXX: we want transformers master here - in the absense of conftest manipulating sys.path: # hack it in for now: import sys from pathlib import Path lowercase : Optional[int] = Path(__file__).resolve().parents[3] / 'src' sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa import io # noqa import itertools # noqa import json # noqa import os # noqa import unittest # noqa from copy import deepcopy # noqa from parameterized import parameterized # noqa from transformers import TrainingArguments, is_torch_available # noqa from transformers.deepspeed import is_deepspeed_available # noqa from transformers.file_utils import WEIGHTS_NAME # noqa from transformers.testing_utils import ( # noqa CaptureLogger, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, mockenv_context, require_deepspeed, require_torch_gpu, require_torch_multi_gpu, slow, ) from transformers.trainer_utils import set_seed # noqa set_seed(42) lowercase : Optional[Any] = {'base': 'patrickvonplaten/wav2vec2_tiny_random', 'robust': 'patrickvonplaten/wav2vec2_tiny_random_robust'} lowercase : str = 'zero2' lowercase : Optional[int] = 'zero3' lowercase : Optional[Any] = [ZEROa, ZEROa] def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : str , _lowerCamelCase : Union[str, Any] , _lowerCamelCase : List[str]) -> int: '''simple docstring''' __UpperCamelCase : Union[str, Any] = parameterized.to_safe_name("_".join(str(_lowerCamelCase) for x in param.args)) return F'{func.__name__}_{param_based_name}' # Cartesian-product of zero stages with models to test lowercase : List[str] = list(itertools.product(stages, models.keys())) @slow @require_deepspeed @require_torch_gpu class lowerCamelCase__ ( __lowercase): '''simple docstring''' @parameterized.expand(a , name_func=a ) def _lowerCamelCase ( self :Dict , a :Optional[Any] , a :str ) -> Optional[int]: self.run_and_check( stage=a , model=a , distributed=a , fpaa=a , ) @require_torch_multi_gpu @parameterized.expand(a , name_func=a ) def _lowerCamelCase ( self :List[str] , a :str , a :str ) -> List[Any]: self.run_and_check( stage=a , model=a , distributed=a , fpaa=a , ) @parameterized.expand(a , name_func=a ) def _lowerCamelCase ( self :List[Any] , a :List[str] , a :int ) -> Optional[int]: self.run_and_check( stage=a , model=a , distributed=a , fpaa=a , ) @require_torch_multi_gpu @parameterized.expand(a , name_func=a ) def _lowerCamelCase ( self :List[str] , a :List[Any] , a :Dict ) -> int: self.run_and_check( stage=a , model=a , distributed=a , fpaa=a , ) def _lowerCamelCase ( self :Any , a :List[str] ) -> Optional[Any]: # XXX: run_asr is premature and doesn't save any results # so all we check for now is that the process didn't fail pass def _lowerCamelCase ( self :Optional[Any] , a :str , a :str , a :int = 1_0 , a :bool = True , a :bool = True , a :bool = True , ) -> Any: __UpperCamelCase : Optional[Any] = models[model] __UpperCamelCase : List[Any] = self.run_trainer( stage=a , model_name=a , eval_steps=a , num_train_epochs=1 , distributed=a , fpaa=a , ) self.do_checks(a ) return output_dir def _lowerCamelCase ( self :List[str] , a :str , a :str , a :int = 1_0 , a :int = 1 , a :bool = True , a :bool = True , ) -> Dict: __UpperCamelCase : int = self.get_auto_remove_tmp_dir("./xxx" , after=a ) __UpperCamelCase : int = f'\n --model_name_or_path {model_name}\n --dataset_name hf-internal-testing/librispeech_asr_dummy\n --dataset_config_name clean\n --train_split_name validation\n --validation_split_name validation\n --output_dir {output_dir}\n --num_train_epochs {str(a )}\n --per_device_train_batch_size 2\n --per_device_eval_batch_size 2\n --evaluation_strategy steps\n --learning_rate 5e-4\n --warmup_steps 8\n --orthography timit\n --preprocessing_num_workers 1\n --group_by_length\n --freeze_feature_extractor\n --report_to none\n --save_steps 0\n --eval_steps {eval_steps}\n --report_to none\n '.split() if fpaa: args.extend(["--fp16"] ) # currently ds_config_wav2vec2_zero.json requires "zero_optimization.find_unused_parameters": true, # hence the separate config files __UpperCamelCase : Dict = f'--deepspeed {self.test_file_dir_str}/ds_config_wav2vec2_{stage}.json'.split() __UpperCamelCase : int = [f'{self.examples_dir_str}/research_projects/wav2vec2/run_asr.py'] __UpperCamelCase : Optional[Any] = self.get_launcher(a ) __UpperCamelCase : Optional[int] = launcher + script + args + ds_args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(a , env=self.get_env() ) return output_dir def _lowerCamelCase ( self :Any , a :List[Any]=False ) -> List[Any]: # 1. explicitly set --num_nodes=1 just in case these tests end up run on a multi-node setup # - it won't be able to handle that # 2. for now testing with just 2 gpus max (since some quality tests may give different # results with mode gpus because we use very little data) __UpperCamelCase : List[Any] = min(2 , get_gpu_count() ) if distributed else 1 return f'deepspeed --num_nodes 1 --num_gpus {num_gpus}'.split()
232
1
"""simple docstring""" from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class SCREAMING_SNAKE_CASE__ ( __a ): lowercase__ = ["image_processor", "tokenizer"] lowercase__ = "BridgeTowerImageProcessor" lowercase__ = ("RobertaTokenizer", "RobertaTokenizerFast") def __init__( self : str , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : Dict): """simple docstring""" super().__init__(UpperCamelCase__ , UpperCamelCase__) def __call__( self : List[str] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Union[bool, str, PaddingStrategy] = False , lowerCAmelCase_ : Union[bool, str, TruncationStrategy] = None , lowerCAmelCase_ : Optional[int] = None , lowerCAmelCase_ : int = 0 , lowerCAmelCase_ : Optional[int] = None , lowerCAmelCase_ : Optional[bool] = None , lowerCAmelCase_ : Optional[bool] = None , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Optional[Union[str, TensorType]] = None , **lowerCAmelCase_ : List[Any] , ): """simple docstring""" lowercase_ = self.tokenizer( text=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , padding=UpperCamelCase__ , truncation=UpperCamelCase__ , max_length=UpperCamelCase__ , stride=UpperCamelCase__ , pad_to_multiple_of=UpperCamelCase__ , return_token_type_ids=UpperCamelCase__ , return_attention_mask=UpperCamelCase__ , return_overflowing_tokens=UpperCamelCase__ , return_special_tokens_mask=UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , return_length=UpperCamelCase__ , verbose=UpperCamelCase__ , return_tensors=UpperCamelCase__ , **UpperCamelCase__ , ) # add pixel_values + pixel_mask lowercase_ = self.image_processor( UpperCamelCase__ , return_tensors=UpperCamelCase__ , do_normalize=UpperCamelCase__ , do_center_crop=UpperCamelCase__ , **UpperCamelCase__) encoding.update(UpperCamelCase__) return encoding def _UpperCAmelCase ( self : List[str] , *lowerCAmelCase_ : Union[str, Any] , **lowerCAmelCase_ : List[str]): """simple docstring""" return self.tokenizer.batch_decode(*UpperCamelCase__ , **UpperCamelCase__) def _UpperCAmelCase ( self : int , *lowerCAmelCase_ : int , **lowerCAmelCase_ : Tuple): """simple docstring""" return self.tokenizer.decode(*UpperCamelCase__ , **UpperCamelCase__) @property def _UpperCAmelCase ( self : Any): """simple docstring""" lowercase_ = self.tokenizer.model_input_names lowercase_ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names))
369
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase=False ) -> Union[str, Any]: '''simple docstring''' lowercase_ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''module.blocks.{i}.norm1.weight''', F'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''module.blocks.{i}.norm1.bias''', F'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (F'''module.blocks.{i}.attn.proj.weight''', F'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''module.blocks.{i}.attn.proj.bias''', F'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''module.blocks.{i}.norm2.weight''', F'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''module.blocks.{i}.norm2.bias''', F'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''module.blocks.{i}.mlp.fc1.weight''', F'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''module.blocks.{i}.mlp.fc1.bias''', F'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''module.blocks.{i}.mlp.fc2.weight''', F'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''module.blocks.{i}.mlp.fc2.bias''', F'''vit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ("""module.cls_token""", """vit.embeddings.cls_token"""), ("""module.patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight"""), ("""module.patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias"""), ("""module.pos_embed""", """vit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""module.norm.weight""", """layernorm.weight"""), ("""module.norm.bias""", """layernorm.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" lowercase_ = [(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False ) -> Any: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: lowercase_ = """""" else: lowercase_ = """vit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowercase_ = state_dict.pop(F'''module.blocks.{i}.attn.qkv.weight''' ) lowercase_ = state_dict.pop(F'''module.blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict lowercase_ = in_proj_weight[ : config.hidden_size, : ] lowercase_ = in_proj_bias[: config.hidden_size] lowercase_ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowercase_ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowercase_ = in_proj_weight[ -config.hidden_size :, : ] lowercase_ = in_proj_bias[-config.hidden_size :] def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> List[str]: '''simple docstring''' lowercase_ = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> str: '''simple docstring''' lowercase_ = [ """module.fc.fc1.weight""", """module.fc.fc1.bias""", """module.fc.bn1.weight""", """module.fc.bn1.bias""", """module.fc.bn1.running_mean""", """module.fc.bn1.running_var""", """module.fc.bn1.num_batches_tracked""", """module.fc.fc2.weight""", """module.fc.fc2.bias""", """module.fc.bn2.weight""", """module.fc.bn2.bias""", """module.fc.bn2.running_mean""", """module.fc.bn2.running_var""", """module.fc.bn2.num_batches_tracked""", """module.fc.fc3.weight""", """module.fc.fc3.bias""", ] for k in ignore_keys: state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Tuple: '''simple docstring''' lowercase_ = dct.pop(__lowerCAmelCase ) lowercase_ = val def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> Tuple: '''simple docstring''' lowercase_ = ViTMSNConfig() lowercase_ = 10_00 lowercase_ = """datasets/huggingface/label-files""" lowercase_ = """imagenet-1k-id2label.json""" lowercase_ = json.load(open(hf_hub_download(__lowerCAmelCase , __lowerCAmelCase ) , """r""" ) ) lowercase_ = {int(__lowerCAmelCase ): v for k, v in idalabel.items()} lowercase_ = idalabel lowercase_ = {v: k for k, v in idalabel.items()} if "s16" in checkpoint_url: lowercase_ = 3_84 lowercase_ = 15_36 lowercase_ = 6 elif "l16" in checkpoint_url: lowercase_ = 10_24 lowercase_ = 40_96 lowercase_ = 24 lowercase_ = 16 lowercase_ = 0.1 elif "b4" in checkpoint_url: lowercase_ = 4 elif "l7" in checkpoint_url: lowercase_ = 7 lowercase_ = 10_24 lowercase_ = 40_96 lowercase_ = 24 lowercase_ = 16 lowercase_ = 0.1 lowercase_ = ViTMSNModel(__lowerCAmelCase ) lowercase_ = torch.hub.load_state_dict_from_url(__lowerCAmelCase , map_location="""cpu""" )["""target_encoder"""] lowercase_ = ViTImageProcessor(size=config.image_size ) remove_projection_head(__lowerCAmelCase ) lowercase_ = create_rename_keys(__lowerCAmelCase , base_model=__lowerCAmelCase ) for src, dest in rename_keys: rename_key(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) read_in_q_k_v(__lowerCAmelCase , __lowerCAmelCase , base_model=__lowerCAmelCase ) model.load_state_dict(__lowerCAmelCase ) model.eval() lowercase_ = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowercase_ = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ) lowercase_ = ViTImageProcessor( size=config.image_size , image_mean=__lowerCAmelCase , image_std=__lowerCAmelCase ) lowercase_ = image_processor(images=__lowerCAmelCase , return_tensors="""pt""" ) # forward pass torch.manual_seed(2 ) lowercase_ = model(**__lowerCAmelCase ) lowercase_ = outputs.last_hidden_state # The following Colab Notebook was used to generate these outputs: # https://colab.research.google.com/gist/sayakpaul/3672419a04f5997827503fd84079bdd1/scratchpad.ipynb if "s16" in checkpoint_url: lowercase_ = torch.tensor([[-1.0_915, -1.4_876, -1.1_809]] ) elif "b16" in checkpoint_url: lowercase_ = torch.tensor([[14.2_889, -18.9_045, 11.7_281]] ) elif "l16" in checkpoint_url: lowercase_ = torch.tensor([[41.5_028, -22.8_681, 45.6_475]] ) elif "b4" in checkpoint_url: lowercase_ = torch.tensor([[-4.3_868, 5.2_932, -0.4_137]] ) else: lowercase_ = torch.tensor([[-0.1_792, -0.6_465, 2.4_263]] ) # verify logits assert torch.allclose(last_hidden_state[:, 0, :3] , __lowerCAmelCase , atol=1E-4 ) print(F'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(__lowerCAmelCase ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": UpperCAmelCase : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint_url", default="https://dl.fbaipublicfiles.com/msn/vits16_800ep.pth.tar", type=str, help="URL of the checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) UpperCAmelCase : Tuple = parser.parse_args() convert_vit_msn_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
313
0
"""simple docstring""" from typing import Any def __UpperCAmelCase ( __UpperCamelCase ): if not input_list: return [] __lowercase : Dict = [input_list.count(__UpperCamelCase ) for value in input_list] __lowercase : Tuple = max(__UpperCamelCase ) # Gets the maximum count in the input list. # Gets values of modes return sorted({input_list[i] for i, value in enumerate(__UpperCamelCase ) if value == y} ) if __name__ == "__main__": import doctest doctest.testmod()
249
"""simple docstring""" import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast a_ = datasets.utils.logging.get_logger(__name__) @dataclass class UpperCAmelCase_ ( datasets.BuilderConfig ): UpperCamelCase =1_00_00 UpperCamelCase =None UpperCamelCase =None class UpperCAmelCase_ ( datasets.ArrowBasedBuilder ): UpperCamelCase =ParquetConfig def _lowerCamelCase ( self ) -> List[str]: return datasets.DatasetInfo(features=self.config.features ) def _lowerCamelCase ( self , UpperCamelCase_ ) -> Tuple: if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) __lowercase : Optional[Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(UpperCamelCase_ , (str, list, tuple) ): __lowercase : str = data_files if isinstance(UpperCamelCase_ , UpperCamelCase_ ): __lowercase : Union[str, Any] = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive __lowercase : int = [dl_manager.iter_files(UpperCamelCase_ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''files''': files} )] __lowercase : int = [] for split_name, files in data_files.items(): if isinstance(UpperCamelCase_ , UpperCamelCase_ ): __lowercase : List[str] = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive __lowercase : Any = [dl_manager.iter_files(UpperCamelCase_ ) for file in files] # Infer features is they are stoed in the arrow schema if self.info.features is None: for file in itertools.chain.from_iterable(UpperCamelCase_ ): with open(UpperCamelCase_ , '''rb''' ) as f: __lowercase : Any = datasets.Features.from_arrow_schema(pq.read_schema(UpperCamelCase_ ) ) break splits.append(datasets.SplitGenerator(name=UpperCamelCase_ , gen_kwargs={'''files''': files} ) ) return splits def _lowerCamelCase ( self , UpperCamelCase_ ) -> pa.Table: if self.info.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example __lowercase : Tuple = table_cast(UpperCamelCase_ , self.info.features.arrow_schema ) return pa_table def _lowerCamelCase ( self , UpperCamelCase_ ) -> Tuple: __lowercase : Union[str, Any] = self.info.features.arrow_schema if self.info.features is not None else None if self.info.features is not None and self.config.columns is not None: if sorted(field.name for field in schema ) != sorted(self.config.columns ): raise ValueError( F"""Tried to load parquet data with columns '{self.config.columns}' with mismatching features '{self.info.features}'""" ) for file_idx, file in enumerate(itertools.chain.from_iterable(UpperCamelCase_ ) ): with open(UpperCamelCase_ , '''rb''' ) as f: __lowercase : Union[str, Any] = pq.ParquetFile(UpperCamelCase_ ) try: for batch_idx, record_batch in enumerate( parquet_file.iter_batches(batch_size=self.config.batch_size , columns=self.config.columns ) ): __lowercase : Dict = pa.Table.from_batches([record_batch] ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield F"""{file_idx}_{batch_idx}""", self._cast_table(UpperCamelCase_ ) except ValueError as e: logger.error(F"""Failed to read file '{file}' with error {type(UpperCamelCase_ )}: {e}""" ) raise
249
1
from typing import List, Optional, Tuple, Union import torch from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class _lowerCamelCase( _a ): def __init__( self, lowerCamelCase, lowerCamelCase) -> Dict: """simple docstring""" super().__init__() # make sure scheduler can always be converted to DDIM _lowercase : List[str] = DDIMScheduler.from_config(scheduler.config) self.register_modules(unet=lowerCamelCase, scheduler=lowerCamelCase) @torch.no_grad() def __call__( self, lowerCamelCase = 1, lowerCamelCase = None, lowerCamelCase = 0.0, lowerCamelCase = 50, lowerCamelCase = None, lowerCamelCase = "pil", lowerCamelCase = True, ) -> Union[ImagePipelineOutput, Tuple]: """simple docstring""" if isinstance(self.unet.config.sample_size, lowerCamelCase): _lowercase : Optional[int] = ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size, ) else: _lowercase : Union[str, Any] = (batch_size, self.unet.config.in_channels, *self.unet.config.sample_size) if isinstance(lowerCamelCase, lowerCamelCase) and len(lowerCamelCase) != batch_size: raise ValueError( F'''You have passed a list of generators of length {len(lowerCamelCase)}, but requested an effective batch''' F''' size of {batch_size}. Make sure the batch size matches the length of the generators.''') _lowercase : str = randn_tensor(lowerCamelCase, generator=lowerCamelCase, device=self.device, dtype=self.unet.dtype) # set step values self.scheduler.set_timesteps(lowerCamelCase) for t in self.progress_bar(self.scheduler.timesteps): # 1. predict noise model_output _lowercase : Union[str, Any] = self.unet(lowerCamelCase, lowerCamelCase).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 _lowercase : Optional[Any] = self.scheduler.step( lowerCamelCase, lowerCamelCase, lowerCamelCase, eta=lowerCamelCase, use_clipped_model_output=lowerCamelCase, generator=lowerCamelCase).prev_sample _lowercase : Any = (image / 2 + 0.5).clamp(0, 1) _lowercase : str = image.cpu().permute(0, 2, 3, 1).numpy() if output_type == "pil": _lowercase : Optional[int] = self.numpy_to_pil(lowerCamelCase) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase)
84
import argparse import os import transformers from .convert_slow_tokenizer import SLOW_TO_FAST_CONVERTERS from .utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Dict = {name: getattr(transformers, name + "Fast") for name in SLOW_TO_FAST_CONVERTERS} def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: if tokenizer_name is not None and tokenizer_name not in TOKENIZER_CLASSES: raise ValueError(F'''Unrecognized tokenizer name, should be one of {list(TOKENIZER_CLASSES.keys() )}.''' ) if tokenizer_name is None: _lowercase : Any = TOKENIZER_CLASSES else: _lowercase : Tuple = {tokenizer_name: getattr(lowerCamelCase_ , tokenizer_name + 'Fast' )} logger.info(F'''Loading tokenizer classes: {tokenizer_names}''' ) for tokenizer_name in tokenizer_names: _lowercase : Union[str, Any] = TOKENIZER_CLASSES[tokenizer_name] _lowercase : Any = True if checkpoint_name is None: _lowercase : int = list(tokenizer_class.max_model_input_sizes.keys() ) else: _lowercase : List[Any] = [checkpoint_name] logger.info(F'''For tokenizer {tokenizer_class.__class__.__name__} loading checkpoints: {checkpoint_names}''' ) for checkpoint in checkpoint_names: logger.info(F'''Loading {tokenizer_class.__class__.__name__} {checkpoint}''' ) # Load tokenizer _lowercase : Union[str, Any] = tokenizer_class.from_pretrained(lowerCamelCase_ , force_download=lowerCamelCase_ ) # Save fast tokenizer logger.info(F'''Save fast tokenizer to {dump_path} with prefix {checkpoint} add_prefix {add_prefix}''' ) # For organization names we create sub-directories if "/" in checkpoint: _lowercase , _lowercase : str = checkpoint.split('/' ) _lowercase : Any = os.path.join(lowerCamelCase_ , lowerCamelCase_ ) elif add_prefix: _lowercase : Union[str, Any] = checkpoint _lowercase : List[str] = dump_path else: _lowercase : str = None _lowercase : Any = dump_path logger.info(F'''=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}''' ) if checkpoint in list(tokenizer.pretrained_vocab_files_map.values() )[0]: _lowercase : Tuple = list(tokenizer.pretrained_vocab_files_map.values() )[0][checkpoint] _lowercase : List[Any] = file_path.split(lowerCamelCase_ )[-1][0] if next_char == "/": _lowercase : Any = os.path.join(lowerCamelCase_ , lowerCamelCase_ ) _lowercase : List[Any] = None logger.info(F'''=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}''' ) _lowercase : Optional[Any] = tokenizer.save_pretrained( lowerCamelCase_ , legacy_format=lowerCamelCase_ , filename_prefix=lowerCamelCase_ ) logger.info(F'''=> File names {file_names}''' ) for file_name in file_names: if not file_name.endswith('tokenizer.json' ): os.remove(lowerCamelCase_ ) logger.info(F'''=> removing {file_name}''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--dump_path", default=None, type=str, required=True, help="Path to output generated fast tokenizer files." ) parser.add_argument( "--tokenizer_name", default=None, type=str, help=( F"Optional tokenizer type selected in the list of {list(TOKENIZER_CLASSES.keys())}. If not given, will " "download and convert all the checkpoints from AWS." ), ) parser.add_argument( "--checkpoint_name", default=None, type=str, help="Optional checkpoint name. If not given, will download and convert the canonical checkpoints from AWS.", ) parser.add_argument( "--force_download", action="store_true", help="Re-download checkpoints.", ) SCREAMING_SNAKE_CASE : str = parser.parse_args() convert_slow_checkpoint_to_fast(args.tokenizer_name, args.checkpoint_name, args.dump_path, args.force_download)
84
1
import os import tempfile import unittest from transformers import FlaubertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FlaubertForMultipleChoice, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertModel, FlaubertWithLMHeadModel, ) from transformers.models.flaubert.modeling_flaubert import FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST class _snake_case ( _lowercase ): def __init__( self: Optional[Any] , __lowerCamelCase: str , __lowerCamelCase: str=13 , __lowerCamelCase: Optional[int]=7 , __lowerCamelCase: Union[str, Any]=True , __lowerCamelCase: Optional[Any]=True , __lowerCamelCase: Optional[Any]=True , __lowerCamelCase: Any=True , __lowerCamelCase: Optional[Any]=True , __lowerCamelCase: List[str]=False , __lowerCamelCase: Optional[Any]=False , __lowerCamelCase: Tuple=False , __lowerCamelCase: Optional[int]=2 , __lowerCamelCase: Dict=99 , __lowerCamelCase: Dict=0 , __lowerCamelCase: Tuple=32 , __lowerCamelCase: List[Any]=5 , __lowerCamelCase: Any=4 , __lowerCamelCase: Dict=0.1 , __lowerCamelCase: List[Any]=0.1 , __lowerCamelCase: Optional[Any]=5_12 , __lowerCamelCase: Union[str, Any]=12 , __lowerCamelCase: Tuple=2 , __lowerCamelCase: Tuple=0.02 , __lowerCamelCase: Union[str, Any]=3 , __lowerCamelCase: Tuple=4 , __lowerCamelCase: Any="last" , __lowerCamelCase: Tuple=None , __lowerCamelCase: str=None , ) -> str: __UpperCAmelCase : Union[str, Any] = parent __UpperCAmelCase : Optional[int] = batch_size __UpperCAmelCase : Dict = seq_length __UpperCAmelCase : Union[str, Any] = is_training __UpperCAmelCase : Union[str, Any] = use_input_lengths __UpperCAmelCase : Optional[int] = use_token_type_ids __UpperCAmelCase : str = use_labels __UpperCAmelCase : Optional[Any] = gelu_activation __UpperCAmelCase : str = sinusoidal_embeddings __UpperCAmelCase : Union[str, Any] = causal __UpperCAmelCase : List[str] = asm __UpperCAmelCase : int = n_langs __UpperCAmelCase : Dict = vocab_size __UpperCAmelCase : Dict = n_special __UpperCAmelCase : Optional[int] = hidden_size __UpperCAmelCase : Optional[Any] = num_hidden_layers __UpperCAmelCase : Dict = num_attention_heads __UpperCAmelCase : Optional[Any] = hidden_dropout_prob __UpperCAmelCase : List[Any] = attention_probs_dropout_prob __UpperCAmelCase : Any = max_position_embeddings __UpperCAmelCase : Optional[Any] = type_vocab_size __UpperCAmelCase : List[Any] = type_sequence_label_size __UpperCAmelCase : Union[str, Any] = initializer_range __UpperCAmelCase : List[str] = num_labels __UpperCAmelCase : int = num_choices __UpperCAmelCase : Dict = summary_type __UpperCAmelCase : List[Any] = use_proj __UpperCAmelCase : Union[str, Any] = scope def _lowerCamelCase ( self: str ) -> List[str]: __UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCAmelCase : str = random_attention_mask([self.batch_size, self.seq_length] ) __UpperCAmelCase : Dict = None if self.use_input_lengths: __UpperCAmelCase : Optional[Any] = ( ids_tensor([self.batch_size] , vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length __UpperCAmelCase : str = None if self.use_token_type_ids: __UpperCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , self.n_langs ) __UpperCAmelCase : Union[str, Any] = None __UpperCAmelCase : Any = None __UpperCAmelCase : str = None if self.use_labels: __UpperCAmelCase : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __UpperCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __UpperCAmelCase : str = ids_tensor([self.batch_size] , 2 ).float() __UpperCAmelCase : Any = ids_tensor([self.batch_size] , self.num_choices ) __UpperCAmelCase : List[str] = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def _lowerCamelCase ( self: Tuple ) -> str: return FlaubertConfig( vocab_size=self.vocab_size , n_special=self.n_special , emb_dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , gelu_activation=self.gelu_activation , sinusoidal_embeddings=self.sinusoidal_embeddings , asm=self.asm , causal=self.causal , n_langs=self.n_langs , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , summary_type=self.summary_type , use_proj=self.use_proj , ) def _lowerCamelCase ( self: Tuple , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: Any , __lowerCamelCase: List[str] , __lowerCamelCase: Tuple , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: Optional[int] , __lowerCamelCase: int , __lowerCamelCase: Optional[Any] , __lowerCamelCase: List[str] , ) -> List[Any]: __UpperCAmelCase : Tuple = FlaubertModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : Optional[Any] = model(__lowerCamelCase , lengths=__lowerCamelCase , langs=__lowerCamelCase ) __UpperCAmelCase : Optional[int] = model(__lowerCamelCase , langs=__lowerCamelCase ) __UpperCAmelCase : str = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self: int , __lowerCamelCase: Tuple , __lowerCamelCase: Any , __lowerCamelCase: Optional[Any] , __lowerCamelCase: List[str] , __lowerCamelCase: List[Any] , __lowerCamelCase: List[str] , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: str , __lowerCamelCase: Any , ) -> List[str]: __UpperCAmelCase : Any = FlaubertWithLMHeadModel(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : str = model(__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowerCamelCase ( self: Any , __lowerCamelCase: List[Any] , __lowerCamelCase: List[Any] , __lowerCamelCase: List[Any] , __lowerCamelCase: Optional[int] , __lowerCamelCase: int , __lowerCamelCase: Optional[Any] , __lowerCamelCase: str , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: Union[str, Any] , ) -> Union[str, Any]: __UpperCAmelCase : List[Any] = FlaubertForQuestionAnsweringSimple(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : int = model(__lowerCamelCase ) __UpperCAmelCase : int = model(__lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowerCamelCase ( self: Union[str, Any] , __lowerCamelCase: str , __lowerCamelCase: Dict , __lowerCamelCase: int , __lowerCamelCase: Any , __lowerCamelCase: List[Any] , __lowerCamelCase: Dict , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: Tuple , __lowerCamelCase: Any , ) -> Dict: __UpperCAmelCase : int = FlaubertForQuestionAnswering(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : Any = model(__lowerCamelCase ) __UpperCAmelCase : List[Any] = model( __lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase , cls_index=__lowerCamelCase , is_impossible=__lowerCamelCase , p_mask=__lowerCamelCase , ) __UpperCAmelCase : Optional[Any] = model( __lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase , cls_index=__lowerCamelCase , is_impossible=__lowerCamelCase , ) ((__UpperCAmelCase) , ) : Optional[Any] = result_with_labels.to_tuple() __UpperCAmelCase : Tuple = model(__lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase ) ((__UpperCAmelCase) , ) : List[str] = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape , () ) self.parent.assertEqual(result.start_top_log_probs.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual(result.start_top_index.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual( result.end_top_log_probs.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual( result.end_top_index.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual(result.cls_logits.shape , (self.batch_size,) ) def _lowerCamelCase ( self: List[str] , __lowerCamelCase: List[Any] , __lowerCamelCase: Dict , __lowerCamelCase: List[Any] , __lowerCamelCase: List[Any] , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: Optional[Any] , __lowerCamelCase: Any , __lowerCamelCase: str , __lowerCamelCase: Optional[int] , ) -> Any: __UpperCAmelCase : int = FlaubertForSequenceClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : Any = model(__lowerCamelCase ) __UpperCAmelCase : Tuple = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCamelCase ( self: List[Any] , __lowerCamelCase: int , __lowerCamelCase: List[str] , __lowerCamelCase: List[Any] , __lowerCamelCase: Optional[Any] , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: int , __lowerCamelCase: List[str] , __lowerCamelCase: Tuple , __lowerCamelCase: Any , ) -> List[str]: __UpperCAmelCase : Optional[int] = self.num_labels __UpperCAmelCase : List[str] = FlaubertForTokenClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : List[Any] = model(__lowerCamelCase , attention_mask=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowerCamelCase ( self: str , __lowerCamelCase: Dict , __lowerCamelCase: Any , __lowerCamelCase: List[str] , __lowerCamelCase: Dict , __lowerCamelCase: Any , __lowerCamelCase: str , __lowerCamelCase: Optional[Any] , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: Optional[Any] , ) -> Any: __UpperCAmelCase : Any = self.num_choices __UpperCAmelCase : Union[str, Any] = FlaubertForMultipleChoice(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : Tuple = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __UpperCAmelCase : Union[str, Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __UpperCAmelCase : str = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __UpperCAmelCase : Tuple = model( __lowerCamelCase , attention_mask=__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowerCamelCase ( self: Tuple ) -> Dict: __UpperCAmelCase : Dict = self.prepare_config_and_inputs() ( ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ) : Optional[Any] = config_and_inputs __UpperCAmelCase : List[str] = { "input_ids": input_ids, "token_type_ids": token_type_ids, "lengths": input_lengths, "attention_mask": input_mask, } return config, inputs_dict @require_torch class _snake_case ( _lowercase , _lowercase , unittest.TestCase ): lowerCamelCase__: Dict = ( ( FlaubertModel, FlaubertWithLMHeadModel, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertForMultipleChoice, ) if is_torch_available() else () ) lowerCamelCase__: Optional[Any] = ( { "feature-extraction": FlaubertModel, "fill-mask": FlaubertWithLMHeadModel, "question-answering": FlaubertForQuestionAnsweringSimple, "text-classification": FlaubertForSequenceClassification, "token-classification": FlaubertForTokenClassification, "zero-shot": FlaubertForSequenceClassification, } if is_torch_available() else {} ) def _lowerCamelCase ( self: List[Any] , __lowerCamelCase: Dict , __lowerCamelCase: List[str] , __lowerCamelCase: str , __lowerCamelCase: str , __lowerCamelCase: Any ) -> Tuple: if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith("Fast" ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def _lowerCamelCase ( self: Tuple , __lowerCamelCase: List[str] , __lowerCamelCase: Optional[int] , __lowerCamelCase: Dict=False ) -> Dict: __UpperCAmelCase : List[str] = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if model_class.__name__ == "FlaubertForQuestionAnswering": __UpperCAmelCase : Dict = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__lowerCamelCase ) __UpperCAmelCase : List[str] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__lowerCamelCase ) return inputs_dict def _lowerCamelCase ( self: List[str] ) -> int: __UpperCAmelCase : int = FlaubertModelTester(self ) __UpperCAmelCase : Optional[Any] = ConfigTester(self , config_class=__lowerCamelCase , emb_dim=37 ) def _lowerCamelCase ( self: List[Any] ) -> str: self.config_tester.run_common_tests() def _lowerCamelCase ( self: int ) -> Optional[Any]: __UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_model(*__lowerCamelCase ) def _lowerCamelCase ( self: Union[str, Any] ) -> Dict: __UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_lm_head(*__lowerCamelCase ) def _lowerCamelCase ( self: Any ) -> List[Any]: __UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_simple_qa(*__lowerCamelCase ) def _lowerCamelCase ( self: Any ) -> Optional[Any]: __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_qa(*__lowerCamelCase ) def _lowerCamelCase ( self: Tuple ) -> Any: __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_sequence_classif(*__lowerCamelCase ) def _lowerCamelCase ( self: Union[str, Any] ) -> Optional[Any]: __UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_token_classif(*__lowerCamelCase ) def _lowerCamelCase ( self: str ) -> Optional[int]: __UpperCAmelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_multiple_choice(*__lowerCamelCase ) @slow def _lowerCamelCase ( self: str ) -> Optional[int]: for model_name in FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCAmelCase : int = FlaubertModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) @slow @require_torch_gpu def _lowerCamelCase ( self: Optional[int] ) -> Union[str, Any]: __UpperCAmelCase , __UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # FlauBertForMultipleChoice behaves incorrectly in JIT environments. if model_class == FlaubertForMultipleChoice: return __UpperCAmelCase : int = True __UpperCAmelCase : Dict = model_class(config=__lowerCamelCase ) __UpperCAmelCase : str = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) __UpperCAmelCase : Optional[Any] = torch.jit.trace( __lowerCamelCase , (inputs_dict["input_ids"].to("cpu" ), inputs_dict["attention_mask"].to("cpu" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(__lowerCamelCase , os.path.join(__lowerCamelCase , "traced_model.pt" ) ) __UpperCAmelCase : Any = torch.jit.load(os.path.join(__lowerCamelCase , "traced_model.pt" ) , map_location=__lowerCamelCase ) loaded(inputs_dict["input_ids"].to(__lowerCamelCase ) , inputs_dict["attention_mask"].to(__lowerCamelCase ) ) @require_torch class _snake_case ( unittest.TestCase ): @slow def _lowerCamelCase ( self: List[Any] ) -> str: __UpperCAmelCase : Optional[int] = FlaubertModel.from_pretrained("flaubert/flaubert_base_cased" ) __UpperCAmelCase : List[Any] = torch.tensor([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]] ) with torch.no_grad(): __UpperCAmelCase : Tuple = model(__lowerCamelCase )[0] __UpperCAmelCase : Optional[Any] = torch.Size((1, 11, 7_68) ) self.assertEqual(output.shape , __lowerCamelCase ) __UpperCAmelCase : str = torch.tensor( [[[-2.62_51, -1.42_98, -0.02_27], [-2.85_10, -1.63_87, 0.22_58], [-2.81_14, -1.18_32, -0.30_66]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowerCamelCase , atol=1e-4 ) )
157
def _UpperCamelCase ( snake_case__ ) -> bool: if not isinstance(snake_case__, snake_case__ ): raise ValueError("check_bouncy() accepts only integer arguments" ) __UpperCAmelCase : Optional[int] = str(snake_case__ ) __UpperCAmelCase : Any = "".join(sorted(snake_case__ ) ) return sorted_str_n != str_n and sorted_str_n[::-1] != str_n def _UpperCamelCase ( snake_case__ = 99 ) -> int: if not 0 < percent < 100: raise ValueError("solution() only accepts values from 0 to 100" ) __UpperCAmelCase : List[str] = 0 __UpperCAmelCase : int = 1 while True: if check_bouncy(snake_case__ ): bouncy_num += 1 if (bouncy_num / num) * 100 >= percent: return num num += 1 if __name__ == "__main__": from doctest import testmod testmod() print(F'{solution(99)}')
157
1
def a( A : float ) -> float: """simple docstring""" return 10 - x * x def a( A : float , A : float ) -> float: """simple docstring""" if equation(A ) * equation(A ) >= 0: raise ValueError("Wrong space!" ) a = a while (b - a) >= 0.01: # Find middle point a = (a + b) / 2 # Check if middle point is root if equation(A ) == 0.0: break # Decide the side to repeat the steps if equation(A ) * equation(A ) < 0: a = c else: a = c return c if __name__ == "__main__": import doctest doctest.testmod() print(bisection(-2, 5)) print(bisection(0, 6))
71
import argparse import hashlib import os import urllib import warnings import torch from torch import nn from tqdm import tqdm from transformers import WhisperConfig, WhisperForConditionalGeneration _lowercase: List[Any] = { "tiny.en": "https://openaipublic.azureedge.net/main/whisper/models/d3dd57d32accea0b295c96e26691aa14d8822fac7d9d27d5dc00b4ca2826dd03/tiny.en.pt", "tiny": "https://openaipublic.azureedge.net/main/whisper/models/65147644a518d12f04e32d6f3b26facc3f8dd46e5390956a9424a650c0ce22b9/tiny.pt", "base.en": "https://openaipublic.azureedge.net/main/whisper/models/25a8566e1d0c1e2231d1c762132cd20e0f96a85d16145c3a00adf5d1ac670ead/base.en.pt", "base": "https://openaipublic.azureedge.net/main/whisper/models/ed3a0b6b1c0edf879ad9b11b1af5a0e6ab5db9205f891f668f8b0e6c6326e34e/base.pt", "small.en": "https://openaipublic.azureedge.net/main/whisper/models/f953ad0fd29cacd07d5a9eda5624af0f6bcf2258be67c92b79389873d91e0872/small.en.pt", "small": "https://openaipublic.azureedge.net/main/whisper/models/9ecf779972d90ba49c06d968637d720dd632c55bbf19d441fb42bf17a411e794/small.pt", "medium.en": "https://openaipublic.azureedge.net/main/whisper/models/d7440d1dc186f76616474e0ff0b3b6b879abc9d1a4926b7adfa41db2d497ab4f/medium.en.pt", "medium": "https://openaipublic.azureedge.net/main/whisper/models/345ae4da62f9b3d59415adc60127b97c714f32e89e936602e85993674d08dcb1/medium.pt", "large": "https://openaipublic.azureedge.net/main/whisper/models/e4b87e7e0bf463eb8e6956e646f1e277e901512310def2c24bf0e11bd3c28e9a/large.pt", "large-v2": "https://openaipublic.azureedge.net/main/whisper/models/81f7c96c852ee8fc832187b0132e569d6c3065a3252ed18e56effd0b6a73e524/large-v2.pt", } def a( A : List[str] ) -> Union[str, Any]: """simple docstring""" a = ["layers", "blocks"] for k in ignore_keys: state_dict.pop(A , A ) _lowercase: Optional[int] = { "blocks": "layers", "mlp.0": "fc1", "mlp.2": "fc2", "mlp_ln": "final_layer_norm", ".attn.query": ".self_attn.q_proj", ".attn.key": ".self_attn.k_proj", ".attn.value": ".self_attn.v_proj", ".attn_ln": ".self_attn_layer_norm", ".attn.out": ".self_attn.out_proj", ".cross_attn.query": ".encoder_attn.q_proj", ".cross_attn.key": ".encoder_attn.k_proj", ".cross_attn.value": ".encoder_attn.v_proj", ".cross_attn_ln": ".encoder_attn_layer_norm", ".cross_attn.out": ".encoder_attn.out_proj", "decoder.ln.": "decoder.layer_norm.", "encoder.ln.": "encoder.layer_norm.", "token_embedding": "embed_tokens", "encoder.positional_embedding": "encoder.embed_positions.weight", "decoder.positional_embedding": "decoder.embed_positions.weight", "ln_post": "layer_norm", } def a( A : str ) -> Union[str, Any]: """simple docstring""" a = list(s_dict.keys() ) for key in keys: a = key for k, v in WHISPER_MAPPING.items(): if k in key: a = new_key.replace(A , A ) print(f'''{key} -> {new_key}''' ) a = s_dict.pop(A ) return s_dict def a( A : Union[str, Any] ) -> Tuple: """simple docstring""" a , a = emb.weight.shape a = nn.Linear(A , A , bias=A ) a = emb.weight.data return lin_layer def a( A : str , A : str ) -> bytes: """simple docstring""" os.makedirs(A , exist_ok=A ) a = os.path.basename(A ) a = url.split("/" )[-2] a = os.path.join(A , A ) if os.path.exists(A ) and not os.path.isfile(A ): raise RuntimeError(f'''{download_target} exists and is not a regular file''' ) if os.path.isfile(A ): a = open(A , "rb" ).read() if hashlib.shaaaa(A ).hexdigest() == expected_shaaaa: return model_bytes else: warnings.warn(f'''{download_target} exists, but the SHA256 checksum does not match; re-downloading the file''' ) with urllib.request.urlopen(A ) as source, open(A , "wb" ) as output: with tqdm( total=int(source.info().get("Content-Length" ) ) , ncols=80 , unit="iB" , unit_scale=A , unit_divisor=1024 ) as loop: while True: a = source.read(8192 ) if not buffer: break output.write(A ) loop.update(len(A ) ) a = open(A , "rb" ).read() if hashlib.shaaaa(A ).hexdigest() != expected_shaaaa: raise RuntimeError( "Model has been downloaded but the SHA256 checksum does not not match. Please retry loading the model." ) return model_bytes def a( A : List[str] , A : Union[str, Any] ) -> str: """simple docstring""" if ".pt" not in checkpoint_path: a = _download(_MODELS[checkpoint_path] ) else: a = torch.load(A , map_location="cpu" ) a = original_checkpoint["dims"] a = original_checkpoint["model_state_dict"] a = state_dict["decoder.token_embedding.weight"] remove_ignore_keys_(A ) rename_keys(A ) a = True a = state_dict["decoder.layers.0.fc1.weight"].shape[0] a = WhisperConfig( vocab_size=dimensions["n_vocab"] , encoder_ffn_dim=A , decoder_ffn_dim=A , num_mel_bins=dimensions["n_mels"] , d_model=dimensions["n_audio_state"] , max_target_positions=dimensions["n_text_ctx"] , encoder_layers=dimensions["n_audio_layer"] , encoder_attention_heads=dimensions["n_audio_head"] , decoder_layers=dimensions["n_text_layer"] , decoder_attention_heads=dimensions["n_text_state"] , max_source_positions=dimensions["n_audio_ctx"] , ) a = WhisperForConditionalGeneration(A ) a , a = model.model.load_state_dict(A , strict=A ) if len(A ) > 0 and not set(A ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( "Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing," f''' but all the following weights are missing {missing}''' ) if tie_embeds: a = make_linear_from_emb(model.model.decoder.embed_tokens ) else: a = proj_out_weights model.save_pretrained(A ) if __name__ == "__main__": _lowercase: Dict = argparse.ArgumentParser() # # Required parameters parser.add_argument("--checkpoint_path", type=str, help="Patht to the downloaded checkpoints") parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") _lowercase: List[str] = parser.parse_args() convert_openai_whisper_to_tfms(args.checkpoint_path, args.pytorch_dump_folder_path)
71
1
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
174
'''simple docstring''' import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def __magic_name__( lowerCamelCase=None, lowerCamelCase=None): return field(default_factory=lambda: default, metadata=lowerCamelCase) @dataclass class a__ : """simple docstring""" __UpperCamelCase : str = field( metadata={'help': 'The csv file to plot.'} , ) __UpperCamelCase : bool = field( default=__A , metadata={'help': 'Whether to plot along batch size or sequence length. Defaults to sequence length.'} , ) __UpperCamelCase : bool = field( default=__A , metadata={'help': 'Whether the csv file has time results or memory results. Defaults to memory results.'} , ) __UpperCamelCase : bool = field( default=__A , metadata={'help': 'Disable logarithmic scale when plotting'} , ) __UpperCamelCase : bool = field( default=__A , metadata={ 'help': 'Whether the csv file has training results or inference results. Defaults to inference results.' } , ) __UpperCamelCase : Optional[str] = field( default=__A , metadata={'help': 'Filename under which the plot will be saved. If unused no plot is saved.'} , ) __UpperCamelCase : Optional[List[str]] = list_field( default=__A , metadata={'help': 'List of model names that are used instead of the ones in the csv file.'} ) def __magic_name__( lowerCamelCase): try: int(lowerCamelCase) return True except ValueError: return False def __magic_name__( lowerCamelCase): try: float(lowerCamelCase) return True except ValueError: return False class a__ : """simple docstring""" def __init__(self , __lowercase ): __lowerCAmelCase = args __lowerCAmelCase = defaultdict(lambda: {"bsz": [], "seq_len": [], "result": {}} ) with open(self.args.csv_file , newline='''''' ) as csv_file: __lowerCAmelCase = csv.DictReader(__lowercase ) for row in reader: __lowerCAmelCase = row['''model'''] self.result_dict[model_name]["bsz"].append(int(row['''batch_size'''] ) ) self.result_dict[model_name]["seq_len"].append(int(row['''sequence_length'''] ) ) if can_convert_to_int(row['''result'''] ): # value is not None __lowerCAmelCase = int(row['''result'''] ) elif can_convert_to_float(row['''result'''] ): # value is not None __lowerCAmelCase = float(row['''result'''] ) def _snake_case (self ): __lowerCAmelCase , __lowerCAmelCase = plt.subplots() __lowerCAmelCase = '''Time usage''' if self.args.is_time else '''Memory usage''' __lowerCAmelCase = title_str + ''' for training''' if self.args.is_train else title_str + ''' for inference''' if not self.args.no_log_scale: # set logarithm scales ax.set_xscale('''log''' ) ax.set_yscale('''log''' ) for axis in [ax.xaxis, ax.yaxis]: axis.set_major_formatter(ScalarFormatter() ) for model_name_idx, model_name in enumerate(self.result_dict.keys() ): __lowerCAmelCase = sorted(set(self.result_dict[model_name]['''bsz'''] ) ) __lowerCAmelCase = sorted(set(self.result_dict[model_name]['''seq_len'''] ) ) __lowerCAmelCase = self.result_dict[model_name]['''result'''] ((__lowerCAmelCase) , (__lowerCAmelCase)) = ( (batch_sizes, sequence_lengths) if self.args.plot_along_batch else (sequence_lengths, batch_sizes) ) __lowerCAmelCase = ( model_name if self.args.short_model_names is None else self.args.short_model_names[model_name_idx] ) for inner_loop_value in inner_loop_array: if self.args.plot_along_batch: __lowerCAmelCase = np.asarray( [results[(x, inner_loop_value)] for x in x_axis_array if (x, inner_loop_value) in results] , dtype=__lowercase , ) else: __lowerCAmelCase = np.asarray( [results[(inner_loop_value, x)] for x in x_axis_array if (inner_loop_value, x) in results] , dtype=np.floataa , ) ((__lowerCAmelCase) , (__lowerCAmelCase)) = ( ('''batch_size''', '''len''') if self.args.plot_along_batch else ('''in #tokens''', '''bsz''') ) __lowerCAmelCase = np.asarray(__lowercase , __lowercase )[: len(__lowercase )] plt.scatter( __lowercase , __lowercase , label=F"""{label_model_name} - {inner_loop_label}: {inner_loop_value}""" ) plt.plot(__lowercase , __lowercase , '''--''' ) title_str += F""" {label_model_name} vs.""" __lowerCAmelCase = title_str[:-4] __lowerCAmelCase = '''Time in s''' if self.args.is_time else '''Memory in MB''' # plot plt.title(__lowercase ) plt.xlabel(__lowercase ) plt.ylabel(__lowercase ) plt.legend() if self.args.figure_png_file is not None: plt.savefig(self.args.figure_png_file ) else: plt.show() def __magic_name__( ): __lowerCAmelCase = HfArgumentParser(lowerCamelCase) __lowerCAmelCase = parser.parse_args_into_dataclasses()[0] __lowerCAmelCase = Plot(args=lowerCamelCase) plot.plot() if __name__ == "__main__": main()
174
1
"""simple docstring""" from __future__ import annotations from collections import namedtuple def UpperCAmelCase__ (snake_case__ : float , snake_case__ : float , snake_case__ : float ): """simple docstring""" _snake_case : int = namedtuple("""result""" , """name value""" ) if (voltage, current, power).count(0 ) != 1: raise ValueError("""Only one argument must be 0""" ) elif power < 0: raise ValueError( """Power cannot be negative in any electrical/electronics system""" ) elif voltage == 0: return result("""voltage""" , power / current ) elif current == 0: return result("""current""" , power / voltage ) elif power == 0: return result("""power""" , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError("""Exactly one argument must be 0""" ) if __name__ == "__main__": import doctest doctest.testmod()
132
"""simple docstring""" import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger A_ = '''<<<<<<< This should probably be modified because it mentions: ''' A_ = '''======= >>>>>>> ''' A_ = [ '''TextEncoderConfig''', '''ByteTextEncoder''', '''SubwordTextEncoder''', '''encoder_config''', '''maybe_build_from_corpus''', '''manual_dir''', ] A_ = [ # (pattern, replacement) # Order is important here for some replacements (r'''tfds\.core''', r'''datasets'''), (r'''tf\.io\.gfile\.GFile''', r'''open'''), (r'''tf\.([\w\d]+)''', r'''datasets.Value(\'\1\')'''), (r'''tfds\.features\.Text\(\)''', r'''datasets.Value(\'string\')'''), (r'''tfds\.features\.Text\(''', r'''datasets.Value(\'string\'),'''), (r'''features\s*=\s*tfds.features.FeaturesDict\(''', r'''features=datasets.Features('''), (r'''tfds\.features\.FeaturesDict\(''', r'''dict('''), (r'''The TensorFlow Datasets Authors''', r'''The TensorFlow Datasets Authors and the HuggingFace Datasets Authors'''), (r'''tfds\.''', r'''datasets.'''), (r'''dl_manager\.manual_dir''', r'''self.config.data_dir'''), (r'''self\.builder_config''', r'''self.config'''), ] def UpperCAmelCase__ (snake_case__ : Namespace ): """simple docstring""" return ConvertCommand(args.tfds_path , args.datasets_directory ) class lowercase( __a ): '''simple docstring''' @staticmethod def UpperCamelCase_ ( a_: ArgumentParser ): '''simple docstring''' _snake_case : Tuple = parser.add_parser( """convert""", help="""Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.""", ) train_parser.add_argument( """--tfds_path""", type=a_, required=a_, help="""Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.""", ) train_parser.add_argument( """--datasets_directory""", type=a_, required=a_, help="""Path to the HuggingFace Datasets folder.""" ) train_parser.set_defaults(func=a_ ) def __init__( self: List[str], a_: str, a_: str, *a_: str ): '''simple docstring''' _snake_case : Optional[Any] = get_logger("""datasets-cli/converting""" ) _snake_case : Any = tfds_path _snake_case : Optional[Any] = datasets_directory def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' if os.path.isdir(self._tfds_path ): _snake_case : int = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): _snake_case : Any = os.path.dirname(self._tfds_path ) else: raise ValueError("""--tfds_path is neither a directory nor a file. Please check path.""" ) _snake_case : Union[str, Any] = os.path.abspath(self._datasets_directory ) self._logger.info(f"Converting datasets from {abs_tfds_path} to {abs_datasets_path}" ) _snake_case : Tuple = [] _snake_case : Dict = [] _snake_case : Optional[Any] = {} if os.path.isdir(self._tfds_path ): _snake_case : List[str] = os.listdir(a_ ) else: _snake_case : int = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(f"Looking at file {f_name}" ) _snake_case : Dict = os.path.join(a_, a_ ) _snake_case : Union[str, Any] = os.path.join(a_, a_ ) if not os.path.isfile(a_ ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info("""Skipping file""" ) continue with open(a_, encoding="""utf-8""" ) as f: _snake_case : str = f.readlines() _snake_case : List[str] = [] _snake_case : Any = False _snake_case : Union[str, Any] = False _snake_case : Optional[Any] = [] for line in lines: _snake_case : Optional[int] = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: _snake_case : Optional[Any] = """import datasets\n""" elif "import tensorflow" in out_line: # order is important here _snake_case : Optional[int] = """""" continue elif "from absl import logging" in out_line: _snake_case : int = """from datasets import logging\n""" elif "getLogger" in out_line: _snake_case : Any = out_line.replace("""getLogger""", """get_logger""" ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): _snake_case : Union[str, Any] = True _snake_case : Optional[Any] = list(filter(lambda a_ : e in out_line, a_ ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(a_ ) + """\n""" ) out_lines.append(a_ ) out_lines.append(a_ ) continue else: for pattern, replacement in TO_CONVERT: _snake_case : List[str] = re.sub(a_, a_, a_ ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: _snake_case : Dict = re.match(r"""from\stensorflow_datasets.*import\s([^\.\r\n]+)""", a_ ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split(""",""" ) ) _snake_case : Optional[Any] = """from . import """ + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(f"Error converting {out_line.strip()}" ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: _snake_case : Tuple = True out_lines.append(a_ ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset _snake_case : List[str] = f_name.replace(""".py""", """""" ) _snake_case : str = os.path.join(a_, a_ ) _snake_case : str = os.path.join(a_, a_ ) os.makedirs(a_, exist_ok=a_ ) self._logger.info(f"Adding directory {output_dir}" ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(a_ ) if needs_manual_update: with_manual_update.append(a_ ) with open(a_, """w""", encoding="""utf-8""" ) as f: f.writelines(a_ ) self._logger.info(f"Converted in {output_file}" ) for utils_file in utils_files: try: _snake_case : Optional[int] = os.path.basename(a_ ) _snake_case : Optional[Any] = imports_to_builder_map[f_name.replace(""".py""", """""" )] self._logger.info(f"Moving {dest_folder} to {utils_file}" ) shutil.copy(a_, a_ ) except KeyError: self._logger.error(f"Cannot find destination folder for {utils_file}. Please copy manually." ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( f"You need to manually update file {file_path} to remove configurations using 'TextEncoderConfig'." )
132
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { "configuration_time_series_transformer": [ "TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "TimeSeriesTransformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ "TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TimeSeriesTransformerForPrediction", "TimeSeriesTransformerModel", "TimeSeriesTransformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimeSeriesTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimeSeriesTransformerForPrediction, TimeSeriesTransformerModel, TimeSeriesTransformerPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
7
from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =CustomTokenizer pass
76
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowercase : str = logging.get_logger(__name__) class lowerCamelCase__ ( __SCREAMING_SNAKE_CASE): '''simple docstring''' _A = ["pixel_values"] def __init__( self :Union[str, Any] , a :List[Any] = True , a :Any = None , a :Any = 0.9 , a :Dict = PILImageResampling.BICUBIC , a :Tuple = True , a :List[str] = None , a :Optional[Any] = 1 / 2_5_5 , a :List[str] = True , a :Dict = True , a :List[Any] = None , a :str = None , **a :Optional[int] , ) -> None: super().__init__(**__UpperCAmelCase ) __UpperCamelCase : List[str] = size if size is not None else {"""shortest_edge""": 2_2_4} __UpperCamelCase : Any = get_size_dict(__UpperCAmelCase , default_to_square=__UpperCAmelCase ) __UpperCamelCase : Optional[int] = crop_size if crop_size is not None else {"""height""": 2_2_4, """width""": 2_2_4} __UpperCamelCase : Union[str, Any] = get_size_dict(__UpperCAmelCase , param_name="crop_size" ) __UpperCamelCase : Dict = do_resize __UpperCamelCase : int = size __UpperCamelCase : Optional[int] = crop_pct __UpperCamelCase : int = resample __UpperCamelCase : int = do_center_crop __UpperCamelCase : int = crop_size __UpperCamelCase : List[Any] = do_rescale __UpperCamelCase : int = rescale_factor __UpperCamelCase : Dict = do_normalize __UpperCamelCase : Any = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN __UpperCamelCase : Optional[Any] = image_std if image_std is not None else IMAGENET_DEFAULT_STD def _lowerCamelCase ( self :Optional[Any] , a :List[Any] , a :Optional[Any] , a :str = None , a :Union[str, Any] = PILImageResampling.BICUBIC , a :str = None , **a :int , ) -> np.ndarray: __UpperCamelCase : List[str] = get_size_dict(__UpperCAmelCase , default_to_square=__UpperCAmelCase ) if "shortest_edge" not in size and ("height" not in size or "width" not in size): raise ValueError(f'size must contain \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}' ) if crop_pct is not None: if "shortest_edge" in size: __UpperCamelCase : Tuple = int(size["shortest_edge"] / crop_pct ) elif "height" in size and "width" in size: if size["height"] == size["width"]: __UpperCamelCase : List[Any] = int(size["height"] / crop_pct ) else: __UpperCamelCase : str = (int(size["height"] / crop_pct ), int(size["width"] / crop_pct )) else: raise ValueError("Invalid size for resize: {}".format(__UpperCAmelCase ) ) __UpperCamelCase : Union[str, Any] = get_resize_output_image_size(__UpperCAmelCase , size=__UpperCAmelCase , default_to_square=__UpperCAmelCase ) else: if "shortest_edge" in size: __UpperCamelCase : Tuple = get_resize_output_image_size(__UpperCAmelCase , size=size["shortest_edge"] , default_to_square=__UpperCAmelCase ) elif "height" in size and "width" in size: __UpperCamelCase : int = (size["""height"""], size["""width"""]) else: raise ValueError("Invalid size for resize: {}".format(__UpperCAmelCase ) ) return resize(__UpperCAmelCase , size=__UpperCAmelCase , resample=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def _lowerCamelCase ( self :List[Any] , a :Tuple , a :Any , a :List[str] = None , **a :int , ) -> np.ndarray: __UpperCamelCase : Union[str, Any] = get_size_dict(__UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f'size must contain \'height\' and \'width\' as keys. Got {size.keys()}' ) return center_crop(__UpperCAmelCase , size=(size["height"], size["width"]) , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def _lowerCamelCase ( self :List[Any] , a :str , a :Optional[int] , a :List[str] = None , **a :Union[str, Any] , ) -> Union[str, Any]: return rescale(__UpperCAmelCase , scale=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def _lowerCamelCase ( self :List[Any] , a :Any , a :Dict , a :List[Any] , a :str = None , **a :Tuple , ) -> np.ndarray: return normalize(__UpperCAmelCase , mean=__UpperCAmelCase , std=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def _lowerCamelCase ( self :Union[str, Any] , a :List[Any] , a :Any = None , a :List[Any] = None , a :List[str] = None , a :List[Any] = None , a :Optional[int] = None , a :int = None , a :Any = None , a :Union[str, Any] = None , a :Optional[Any] = None , a :Optional[Any] = None , a :Union[str, Any] = None , a :str = None , a :Tuple = ChannelDimension.FIRST , **a :Optional[int] , ) -> PIL.Image.Image: __UpperCamelCase : Union[str, Any] = do_resize if do_resize is not None else self.do_resize __UpperCamelCase : int = crop_pct if crop_pct is not None else self.crop_pct __UpperCamelCase : Tuple = resample if resample is not None else self.resample __UpperCamelCase : int = do_center_crop if do_center_crop is not None else self.do_center_crop __UpperCamelCase : Optional[Any] = do_rescale if do_rescale is not None else self.do_rescale __UpperCamelCase : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __UpperCamelCase : Optional[Any] = do_normalize if do_normalize is not None else self.do_normalize __UpperCamelCase : List[Any] = image_mean if image_mean is not None else self.image_mean __UpperCamelCase : Dict = image_std if image_std is not None else self.image_std __UpperCamelCase : Any = size if size is not None else self.size __UpperCamelCase : Any = get_size_dict(__UpperCAmelCase , default_to_square=__UpperCAmelCase ) __UpperCamelCase : int = crop_size if crop_size is not None else self.crop_size __UpperCamelCase : Optional[Any] = get_size_dict(__UpperCAmelCase , param_name="crop_size" ) __UpperCamelCase : List[Any] = make_list_of_images(__UpperCAmelCase ) if not valid_images(__UpperCAmelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_pct is None: raise ValueError("Crop_pct must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. __UpperCamelCase : Union[str, Any] = [to_numpy_array(__UpperCAmelCase ) for image in images] if do_resize: __UpperCamelCase : int = [self.resize(image=__UpperCAmelCase , size=__UpperCAmelCase , crop_pct=__UpperCAmelCase , resample=__UpperCAmelCase ) for image in images] if do_center_crop: __UpperCamelCase : Union[str, Any] = [self.center_crop(image=__UpperCAmelCase , size=__UpperCAmelCase ) for image in images] if do_rescale: __UpperCamelCase : List[str] = [self.rescale(image=__UpperCAmelCase , scale=__UpperCAmelCase ) for image in images] if do_normalize: __UpperCamelCase : Optional[int] = [self.normalize(image=__UpperCAmelCase , mean=__UpperCAmelCase , std=__UpperCAmelCase ) for image in images] __UpperCamelCase : Union[str, Any] = [to_channel_dimension_format(__UpperCAmelCase , __UpperCAmelCase ) for image in images] __UpperCamelCase : Dict = {"""pixel_values""": images} return BatchFeature(data=__UpperCAmelCase , tensor_type=__UpperCAmelCase )
366
from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax lowercase : str = logging.get_logger(__name__) @add_end_docstrings(__lowercase) class lowerCamelCase__ ( __lowercase): '''simple docstring''' def __init__( self :Any , **a :Union[str, Any] ) -> Union[str, Any]: super().__init__(**a ) requires_backends(self , "vision" ) self.check_model_type( TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if self.framework == "tf" else MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING ) def __call__( self :Any , a :Union[str, List[str], "Image", List["Image"]] , **a :Tuple ) -> List[str]: return super().__call__(a , **a ) def _lowerCamelCase ( self :List[Any] , **a :List[str] ) -> List[Any]: __UpperCamelCase : List[Any] = {} if "candidate_labels" in kwargs: __UpperCamelCase : Optional[int] = kwargs["candidate_labels"] if "hypothesis_template" in kwargs: __UpperCamelCase : List[str] = kwargs["hypothesis_template"] return preprocess_params, {}, {} def _lowerCamelCase ( self :List[str] , a :Optional[int] , a :List[str]=None , a :Dict="This is a photo of {}." ) -> Any: __UpperCamelCase : Dict = load_image(a ) __UpperCamelCase : Any = self.image_processor(images=[image] , return_tensors=self.framework ) __UpperCamelCase : str = candidate_labels __UpperCamelCase : List[Any] = [hypothesis_template.format(a ) for x in candidate_labels] __UpperCamelCase : List[Any] = self.tokenizer(a , return_tensors=self.framework , padding=a ) __UpperCamelCase : Any = [text_inputs] return inputs def _lowerCamelCase ( self :Union[str, Any] , a :Optional[Any] ) -> List[Any]: __UpperCamelCase : List[str] = model_inputs.pop("candidate_labels" ) __UpperCamelCase : Dict = model_inputs.pop("text_inputs" ) if isinstance(text_inputs[0] , a ): __UpperCamelCase : Optional[Any] = text_inputs[0] else: # Batching case. __UpperCamelCase : int = text_inputs[0][0] __UpperCamelCase : str = self.model(**a , **a ) __UpperCamelCase : List[Any] = { "candidate_labels": candidate_labels, "logits": outputs.logits_per_image, } return model_outputs def _lowerCamelCase ( self :List[Any] , a :List[Any] ) -> Tuple: __UpperCamelCase : Any = model_outputs.pop("candidate_labels" ) __UpperCamelCase : Optional[Any] = model_outputs["logits"][0] if self.framework == "pt": __UpperCamelCase : int = logits.softmax(dim=-1 ).squeeze(-1 ) __UpperCamelCase : List[str] = probs.tolist() if not isinstance(a , a ): __UpperCamelCase : List[Any] = [scores] elif self.framework == "tf": __UpperCamelCase : Optional[int] = stable_softmax(a , axis=-1 ) __UpperCamelCase : Dict = probs.numpy().tolist() else: raise ValueError(f'Unsupported framework: {self.framework}' ) __UpperCamelCase : Tuple = [ {"score": score, "label": candidate_label} for score, candidate_label in sorted(zip(a , a ) , key=lambda a : -x[0] ) ] return result
151
0
"""simple docstring""" def _snake_case ( _snake_case : List[str] , _snake_case : Any ) -> float: '''simple docstring''' if discount_rate < 0: raise ValueError('Discount rate cannot be negative' ) if not cash_flows: raise ValueError('Cash flows list cannot be empty' ) _A = sum( cash_flow / ((1 + discount_rate) ** i) for i, cash_flow in enumerate(__lowerCAmelCase ) ) return round(__lowerCAmelCase , ndigits=2 ) if __name__ == "__main__": import doctest doctest.testmod()
315
import copy import os from collections import OrderedDict from typing import TYPE_CHECKING, Any, Dict, Mapping, Optional, Union if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase : Optional[Any] =logging.get_logger(__name__) lowerCamelCase : Optional[int] ={ '''google/owlvit-base-patch32''': '''https://huggingface.co/google/owlvit-base-patch32/resolve/main/config.json''', '''google/owlvit-base-patch16''': '''https://huggingface.co/google/owlvit-base-patch16/resolve/main/config.json''', '''google/owlvit-large-patch14''': '''https://huggingface.co/google/owlvit-large-patch14/resolve/main/config.json''', } class __a ( A__ ): _lowerCAmelCase : Optional[int] = '''owlvit_text_model''' def __init__( self : Any , SCREAMING_SNAKE_CASE : Union[str, Any]=4_94_08 , SCREAMING_SNAKE_CASE : List[str]=5_12 , SCREAMING_SNAKE_CASE : List[Any]=20_48 , SCREAMING_SNAKE_CASE : Any=12 , SCREAMING_SNAKE_CASE : Any=8 , SCREAMING_SNAKE_CASE : Dict=16 , SCREAMING_SNAKE_CASE : Union[str, Any]="quick_gelu" , SCREAMING_SNAKE_CASE : List[str]=1e-5 , SCREAMING_SNAKE_CASE : Union[str, Any]=0.0 , SCREAMING_SNAKE_CASE : Any=0.0_2 , SCREAMING_SNAKE_CASE : int=1.0 , SCREAMING_SNAKE_CASE : Any=0 , SCREAMING_SNAKE_CASE : int=4_94_06 , SCREAMING_SNAKE_CASE : List[str]=4_94_07 , **SCREAMING_SNAKE_CASE : Optional[Any] , ): '''simple docstring''' super().__init__(pad_token_id=SCREAMING_SNAKE_CASE , bos_token_id=SCREAMING_SNAKE_CASE , eos_token_id=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Tuple = vocab_size UpperCamelCase__ : int = hidden_size UpperCamelCase__ : List[str] = intermediate_size UpperCamelCase__ : Tuple = num_hidden_layers UpperCamelCase__ : str = num_attention_heads UpperCamelCase__ : Any = max_position_embeddings UpperCamelCase__ : List[Any] = hidden_act UpperCamelCase__ : str = layer_norm_eps UpperCamelCase__ : List[Any] = attention_dropout UpperCamelCase__ : Tuple = initializer_range UpperCamelCase__ : Optional[Any] = initializer_factor @classmethod def __lowercase ( cls : Optional[int] , SCREAMING_SNAKE_CASE : Union[str, os.PathLike] , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' cls._set_token_in_kwargs(SCREAMING_SNAKE_CASE ) UpperCamelCase__ , UpperCamelCase__ : Any = cls.get_config_dict(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) # get the text config dict if we are loading from OwlViTConfig if config_dict.get("model_type" ) == "owlvit": UpperCamelCase__ : Dict = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' F'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) class __a ( A__ ): _lowerCAmelCase : str = '''owlvit_vision_model''' def __init__( self : Optional[int] , SCREAMING_SNAKE_CASE : str=7_68 , SCREAMING_SNAKE_CASE : Dict=30_72 , SCREAMING_SNAKE_CASE : int=12 , SCREAMING_SNAKE_CASE : Union[str, Any]=12 , SCREAMING_SNAKE_CASE : Dict=3 , SCREAMING_SNAKE_CASE : Union[str, Any]=7_68 , SCREAMING_SNAKE_CASE : Optional[int]=32 , SCREAMING_SNAKE_CASE : Dict="quick_gelu" , SCREAMING_SNAKE_CASE : Optional[Any]=1e-5 , SCREAMING_SNAKE_CASE : List[str]=0.0 , SCREAMING_SNAKE_CASE : Dict=0.0_2 , SCREAMING_SNAKE_CASE : Optional[int]=1.0 , **SCREAMING_SNAKE_CASE : Tuple , ): '''simple docstring''' super().__init__(**SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Any = hidden_size UpperCamelCase__ : str = intermediate_size UpperCamelCase__ : Any = num_hidden_layers UpperCamelCase__ : str = num_attention_heads UpperCamelCase__ : int = num_channels UpperCamelCase__ : Union[str, Any] = image_size UpperCamelCase__ : List[Any] = patch_size UpperCamelCase__ : Tuple = hidden_act UpperCamelCase__ : Optional[int] = layer_norm_eps UpperCamelCase__ : Optional[Any] = attention_dropout UpperCamelCase__ : Dict = initializer_range UpperCamelCase__ : int = initializer_factor @classmethod def __lowercase ( cls : Dict , SCREAMING_SNAKE_CASE : Union[str, os.PathLike] , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' cls._set_token_in_kwargs(SCREAMING_SNAKE_CASE ) UpperCamelCase__ , UpperCamelCase__ : List[Any] = cls.get_config_dict(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) # get the vision config dict if we are loading from OwlViTConfig if config_dict.get("model_type" ) == "owlvit": UpperCamelCase__ : List[str] = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' F'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) class __a ( A__ ): _lowerCAmelCase : str = '''owlvit''' _lowerCAmelCase : Tuple = True def __init__( self : Any , SCREAMING_SNAKE_CASE : Optional[Any]=None , SCREAMING_SNAKE_CASE : int=None , SCREAMING_SNAKE_CASE : str=5_12 , SCREAMING_SNAKE_CASE : Any=2.6_5_9_2 , SCREAMING_SNAKE_CASE : Union[str, Any]=True , **SCREAMING_SNAKE_CASE : List[Any] , ): '''simple docstring''' super().__init__(**SCREAMING_SNAKE_CASE ) if text_config is None: UpperCamelCase__ : str = {} logger.info("text_config is None. Initializing the OwlViTTextConfig with default values." ) if vision_config is None: UpperCamelCase__ : List[str] = {} logger.info("vision_config is None. initializing the OwlViTVisionConfig with default values." ) UpperCamelCase__ : Dict = OwlViTTextConfig(**SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Dict = OwlViTVisionConfig(**SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Dict = projection_dim UpperCamelCase__ : Union[str, Any] = logit_scale_init_value UpperCamelCase__ : int = return_dict UpperCamelCase__ : Tuple = 1.0 @classmethod def __lowercase ( cls : Any , SCREAMING_SNAKE_CASE : Union[str, os.PathLike] , **SCREAMING_SNAKE_CASE : Dict ): '''simple docstring''' cls._set_token_in_kwargs(SCREAMING_SNAKE_CASE ) UpperCamelCase__ , UpperCamelCase__ : Dict = cls.get_config_dict(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' F'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) @classmethod def __lowercase ( cls : int , SCREAMING_SNAKE_CASE : Dict , SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' UpperCamelCase__ : List[Any] = {} UpperCamelCase__ : Union[str, Any] = text_config UpperCamelCase__ : Optional[int] = vision_config return cls.from_dict(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) def __lowercase ( self : Tuple ): '''simple docstring''' UpperCamelCase__ : Optional[int] = copy.deepcopy(self.__dict__ ) UpperCamelCase__ : Union[str, Any] = self.text_config.to_dict() UpperCamelCase__ : List[str] = self.vision_config.to_dict() UpperCamelCase__ : Optional[int] = self.__class__.model_type return output class __a ( A__ ): @property def __lowercase ( self : Any ): '''simple docstring''' return OrderedDict( [ ("input_ids", {0: "batch", 1: "sequence"}), ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("attention_mask", {0: "batch", 1: "sequence"}), ] ) @property def __lowercase ( self : Optional[int] ): '''simple docstring''' return OrderedDict( [ ("logits_per_image", {0: "batch"}), ("logits_per_text", {0: "batch"}), ("text_embeds", {0: "batch"}), ("image_embeds", {0: "batch"}), ] ) @property def __lowercase ( self : Dict ): '''simple docstring''' return 1e-4 def __lowercase ( self : Optional[int] , SCREAMING_SNAKE_CASE : "ProcessorMixin" , SCREAMING_SNAKE_CASE : int = -1 , SCREAMING_SNAKE_CASE : int = -1 , SCREAMING_SNAKE_CASE : Optional["TensorType"] = None , ): '''simple docstring''' UpperCamelCase__ : Optional[int] = super().generate_dummy_inputs( processor.tokenizer , batch_size=SCREAMING_SNAKE_CASE , seq_length=SCREAMING_SNAKE_CASE , framework=SCREAMING_SNAKE_CASE ) UpperCamelCase__ : List[str] = super().generate_dummy_inputs( processor.image_processor , batch_size=SCREAMING_SNAKE_CASE , framework=SCREAMING_SNAKE_CASE ) return {**text_input_dict, **image_input_dict} @property def __lowercase ( self : Tuple ): '''simple docstring''' return 14
189
0
import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class lowercase ( _UpperCamelCase ): '''simple docstring''' @require_torch def UpperCamelCase__ (self ) -> Dict: """simple docstring""" UpperCAmelCase__ = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' UpperCAmelCase__ = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' UpperCAmelCase__ = '\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache UpperCAmelCase__ = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(__a ) BertModel.from_pretrained(__a ) BertTokenizer.from_pretrained(__a ) pipeline(task='fill-mask' , model=__a ) # baseline - just load from_pretrained with normal network UpperCAmelCase__ = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed UpperCAmelCase__ = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files UpperCAmelCase__ = '1' UpperCAmelCase__ = subprocess.run(__a , env=__a , check=__a , capture_output=__a ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def UpperCamelCase__ (self ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' UpperCAmelCase__ = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' UpperCAmelCase__ = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache UpperCAmelCase__ = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(__a ) BertModel.from_pretrained(__a ) BertTokenizer.from_pretrained(__a ) pipeline(task='fill-mask' , model=__a ) # baseline - just load from_pretrained with normal network UpperCAmelCase__ = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed UpperCAmelCase__ = self.get_env() UpperCAmelCase__ = subprocess.run(__a , env=__a , check=__a , capture_output=__a ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def UpperCamelCase__ (self ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = '\nfrom transformers import BertConfig, BertModel, BertTokenizer\n ' UpperCAmelCase__ = '\nmname = "hf-internal-testing/tiny-random-bert-sharded"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint("success")\n ' UpperCAmelCase__ = '\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled")\nsocket.socket = offline_socket\n ' # baseline - just load from_pretrained with normal network UpperCAmelCase__ = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed UpperCAmelCase__ = self.get_env() UpperCAmelCase__ = subprocess.run(__a , env=__a , check=__a , capture_output=__a ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) # next emulate no network UpperCAmelCase__ = [sys.executable, '-c', '\n'.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files UpperCAmelCase__ = '1' UpperCAmelCase__ = subprocess.run(__a , env=__a , check=__a , capture_output=__a ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def UpperCamelCase__ (self ) -> List[str]: """simple docstring""" UpperCAmelCase__ = '\nfrom transformers import pipeline\n ' UpperCAmelCase__ = '\nmname = "hf-internal-testing/tiny-random-bert"\npipe = pipeline(model=mname)\n ' UpperCAmelCase__ = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled")\nsocket.socket = offline_socket\n ' UpperCAmelCase__ = self.get_env() UpperCAmelCase__ = '1' UpperCAmelCase__ = [sys.executable, '-c', '\n'.join([load, mock, run] )] UpperCAmelCase__ = subprocess.run(__a , env=__a , check=__a , capture_output=__a ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( 'You cannot infer task automatically within `pipeline` when using offline mode' , result.stderr.decode().replace('\n' , '' ) , ) @require_torch def UpperCamelCase__ (self ) -> int: """simple docstring""" UpperCAmelCase__ = '\nfrom transformers import AutoModel\n ' UpperCAmelCase__ = '\nmname = "hf-internal-testing/test_dynamic_model"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint("success")\n ' # baseline - just load from_pretrained with normal network UpperCAmelCase__ = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed UpperCAmelCase__ = self.get_env() UpperCAmelCase__ = subprocess.run(__a , env=__a , check=__a , capture_output=__a ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files UpperCAmelCase__ = '1' UpperCAmelCase__ = subprocess.run(__a , env=__a , check=__a , capture_output=__a ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() )
352
import copy import os import cva import numpy as np from matplotlib import pyplot as plt class lowercase : '''simple docstring''' def __init__(self ) -> str: """simple docstring""" UpperCAmelCase__ = '' UpperCAmelCase__ = '' UpperCAmelCase__ = [] UpperCAmelCase__ = 0 UpperCAmelCase__ = 256 UpperCAmelCase__ = 0 UpperCAmelCase__ = 0 UpperCAmelCase__ = 0 UpperCAmelCase__ = 0 def UpperCamelCase__ (self , __a ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = cva.imread(__a , 0 ) UpperCAmelCase__ = copy.deepcopy(self.img ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = plt.hist(self.img.ravel() , 256 , [0, 256] , label='x' ) UpperCAmelCase__ = np.sum(__a ) for i in range(len(__a ) ): UpperCAmelCase__ = x[i] / self.k self.sk += prk UpperCAmelCase__ = (self.L - 1) * self.sk if self.rem != 0: UpperCAmelCase__ = int(last % last ) UpperCAmelCase__ = int(last + 1 if self.rem >= 0.5 else last ) self.last_list.append(__a ) UpperCAmelCase__ = int(np.ma.count(self.img ) / self.img[1].size ) UpperCAmelCase__ = self.img[1].size for i in range(self.number_of_cols ): for j in range(self.number_of_rows ): UpperCAmelCase__ = self.img[j][i] if num != self.last_list[num]: UpperCAmelCase__ = self.last_list[num] cva.imwrite('output_data/output.jpg' , self.img ) def UpperCamelCase__ (self ) -> Optional[int]: """simple docstring""" plt.hist(self.img.ravel() , 256 , [0, 256] ) def UpperCamelCase__ (self ) -> Tuple: """simple docstring""" cva.imshow('Output-Image' , self.img ) cva.imshow('Input-Image' , self.original_image ) cva.waitKey(5000 ) cva.destroyAllWindows() if __name__ == "__main__": _UpperCamelCase = os.path.join(os.path.basename(__file__), '''image_data/input.jpg''') _UpperCamelCase = ConstantStretch() stretcher.stretch(file_path) stretcher.plot_histogram() stretcher.show_image()
335
0
import argparse import json import os import pickle import shutil import numpy as np import torch from distiller import Distiller from lm_seqs_dataset import LmSeqsDataset from transformers import ( BertConfig, BertForMaskedLM, BertTokenizer, DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer, GPTaConfig, GPTaLMHeadModel, GPTaTokenizer, RobertaConfig, RobertaForMaskedLM, RobertaTokenizer, ) from utils import git_log, init_gpu_params, logger, set_seed UpperCamelCase = { '''distilbert''': (DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer), '''roberta''': (RobertaConfig, RobertaForMaskedLM, RobertaTokenizer), '''bert''': (BertConfig, BertForMaskedLM, BertTokenizer), '''gpt2''': (GPTaConfig, GPTaLMHeadModel, GPTaTokenizer), } def lowercase_ ( _lowerCamelCase : Optional[Any]): assert (args.mlm and args.alpha_mlm > 0.0) or (not args.mlm and args.alpha_mlm == 0.0) assert (args.alpha_mlm > 0.0 and args.alpha_clm == 0.0) or (args.alpha_mlm == 0.0 and args.alpha_clm > 0.0) if args.mlm: assert os.path.isfile(args.token_counts) assert (args.student_type in ["roberta", "distilbert"]) and (args.teacher_type in ["roberta", "bert"]) else: assert (args.student_type in ["gpt2"]) and (args.teacher_type in ["gpt2"]) assert args.teacher_type == args.student_type or ( args.student_type == "distilbert" and args.teacher_type == "bert" ) assert os.path.isfile(args.student_config) if args.student_pretrained_weights is not None: assert os.path.isfile(args.student_pretrained_weights) if args.freeze_token_type_embds: assert args.student_type in ["roberta"] assert args.alpha_ce >= 0.0 assert args.alpha_mlm >= 0.0 assert args.alpha_clm >= 0.0 assert args.alpha_mse >= 0.0 assert args.alpha_cos >= 0.0 assert args.alpha_ce + args.alpha_mlm + args.alpha_clm + args.alpha_mse + args.alpha_cos > 0.0 def lowercase_ ( _lowerCamelCase : Dict , _lowerCamelCase : Union[str, Any]): if args.student_type == "roberta": lowercase__ : List[Any] = False elif args.student_type == "gpt2": lowercase__ : str = False def lowercase_ ( _lowerCamelCase : Optional[Any] , _lowerCamelCase : int): if args.student_type == "roberta": lowercase__ : List[Any] = False def lowercase_ ( ): lowercase__ : Any = argparse.ArgumentParser(description="Training") parser.add_argument("--force" , action="store_true" , help="Overwrite dump_path if it already exists.") parser.add_argument( "--dump_path" , type=_UpperCAmelCase , required=_UpperCAmelCase , help="The output directory (log, checkpoints, parameters, etc.)") parser.add_argument( "--data_file" , type=_UpperCAmelCase , required=_UpperCAmelCase , help="The binarized file (tokenized + tokens_to_ids) and grouped by sequence." , ) parser.add_argument( "--student_type" , type=_UpperCAmelCase , choices=["distilbert", "roberta", "gpt2"] , required=_UpperCAmelCase , help="The student type (DistilBERT, RoBERTa)." , ) parser.add_argument("--student_config" , type=_UpperCAmelCase , required=_UpperCAmelCase , help="Path to the student configuration.") parser.add_argument( "--student_pretrained_weights" , default=_UpperCAmelCase , type=_UpperCAmelCase , help="Load student initialization checkpoint.") parser.add_argument( "--teacher_type" , choices=["bert", "roberta", "gpt2"] , required=_UpperCAmelCase , help="Teacher type (BERT, RoBERTa).") parser.add_argument("--teacher_name" , type=_UpperCAmelCase , required=_UpperCAmelCase , help="The teacher model.") parser.add_argument("--temperature" , default=2.0 , type=_UpperCAmelCase , help="Temperature for the softmax temperature.") parser.add_argument( "--alpha_ce" , default=0.5 , type=_UpperCAmelCase , help="Linear weight for the distillation loss. Must be >=0.") parser.add_argument( "--alpha_mlm" , default=0.0 , type=_UpperCAmelCase , help="Linear weight for the MLM loss. Must be >=0. Should be used in conjunction with `mlm` flag." , ) parser.add_argument("--alpha_clm" , default=0.5 , type=_UpperCAmelCase , help="Linear weight for the CLM loss. Must be >=0.") parser.add_argument("--alpha_mse" , default=0.0 , type=_UpperCAmelCase , help="Linear weight of the MSE loss. Must be >=0.") parser.add_argument( "--alpha_cos" , default=0.0 , type=_UpperCAmelCase , help="Linear weight of the cosine embedding loss. Must be >=0.") parser.add_argument( "--mlm" , action="store_true" , help="The LM step: MLM or CLM. If `mlm` is True, the MLM is used over CLM.") parser.add_argument( "--mlm_mask_prop" , default=0.15 , type=_UpperCAmelCase , help="Proportion of tokens for which we need to make a prediction." , ) parser.add_argument("--word_mask" , default=0.8 , type=_UpperCAmelCase , help="Proportion of tokens to mask out.") parser.add_argument("--word_keep" , default=0.1 , type=_UpperCAmelCase , help="Proportion of tokens to keep.") parser.add_argument("--word_rand" , default=0.1 , type=_UpperCAmelCase , help="Proportion of tokens to randomly replace.") parser.add_argument( "--mlm_smoothing" , default=0.7 , type=_UpperCAmelCase , help="Smoothing parameter to emphasize more rare tokens (see XLM, similar to word2vec)." , ) parser.add_argument("--token_counts" , type=_UpperCAmelCase , help="The token counts in the data_file for MLM.") parser.add_argument( "--restrict_ce_to_mask" , action="store_true" , help="If true, compute the distillation loss only the [MLM] prediction distribution." , ) parser.add_argument( "--freeze_pos_embs" , action="store_true" , help="Freeze positional embeddings during distillation. For student_type in [\'roberta\', \'gpt2\'] only." , ) parser.add_argument( "--freeze_token_type_embds" , action="store_true" , help="Freeze token type embeddings during distillation if existent. For student_type in [\'roberta\'] only." , ) parser.add_argument("--n_epoch" , type=_UpperCAmelCase , default=3 , help="Number of pass on the whole dataset.") parser.add_argument("--batch_size" , type=_UpperCAmelCase , default=5 , help="Batch size (for each process).") parser.add_argument( "--group_by_size" , action="store_false" , help="If true, group sequences that have similar length into the same batch. Default is true." , ) parser.add_argument( "--gradient_accumulation_steps" , type=_UpperCAmelCase , default=50 , help="Gradient accumulation for larger training batches." , ) parser.add_argument("--warmup_prop" , default=0.05 , type=_UpperCAmelCase , help="Linear warmup proportion.") parser.add_argument("--weight_decay" , default=0.0 , type=_UpperCAmelCase , help="Weight decay if we apply some.") parser.add_argument("--learning_rate" , default=5E-4 , type=_UpperCAmelCase , help="The initial learning rate for Adam.") parser.add_argument("--adam_epsilon" , default=1E-6 , type=_UpperCAmelCase , help="Epsilon for Adam optimizer.") parser.add_argument("--max_grad_norm" , default=5.0 , type=_UpperCAmelCase , help="Max gradient norm.") parser.add_argument("--initializer_range" , default=0.02 , type=_UpperCAmelCase , help="Random initialization range.") parser.add_argument( "--fp16" , action="store_true" , help="Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit" , ) parser.add_argument( "--fp16_opt_level" , type=_UpperCAmelCase , default="O1" , help=( "For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\']." "See details at https://nvidia.github.io/apex/amp.html" ) , ) parser.add_argument("--n_gpu" , type=_UpperCAmelCase , default=1 , help="Number of GPUs in the node.") parser.add_argument("--local_rank" , type=_UpperCAmelCase , default=-1 , help="Distributed training - Local rank") parser.add_argument("--seed" , type=_UpperCAmelCase , default=56 , help="Random seed") parser.add_argument("--log_interval" , type=_UpperCAmelCase , default=500 , help="Tensorboard logging interval.") parser.add_argument("--checkpoint_interval" , type=_UpperCAmelCase , default=4000 , help="Checkpoint interval.") lowercase__ : List[Any] = parser.parse_args() sanity_checks(_UpperCAmelCase) # ARGS # init_gpu_params(_UpperCAmelCase) set_seed(_UpperCAmelCase) if args.is_master: if os.path.exists(args.dump_path): if not args.force: raise ValueError( f'''Serialization dir {args.dump_path} already exists, but you have not precised wheter to overwrite''' " itUse `--force` if you want to overwrite it") else: shutil.rmtree(args.dump_path) if not os.path.exists(args.dump_path): os.makedirs(args.dump_path) logger.info(f'''Experiment will be dumped and logged in {args.dump_path}''') # SAVE PARAMS # logger.info(f'''Param: {args}''') with open(os.path.join(args.dump_path , "parameters.json") , "w") as f: json.dump(vars(_UpperCAmelCase) , _UpperCAmelCase , indent=4) git_log(args.dump_path) lowercase__ , lowercase__ , lowercase__ : int = MODEL_CLASSES[args.student_type] lowercase__ , lowercase__ , lowercase__ : str = MODEL_CLASSES[args.teacher_type] # TOKENIZER # lowercase__ : Union[str, Any] = teacher_tokenizer_class.from_pretrained(args.teacher_name) lowercase__ : List[Any] = {} for tok_name, tok_symbol in tokenizer.special_tokens_map.items(): lowercase__ : Optional[int] = tokenizer.all_special_tokens.index(_UpperCAmelCase) lowercase__ : Any = tokenizer.all_special_ids[idx] logger.info(f'''Special tokens {special_tok_ids}''') lowercase__ : int = special_tok_ids lowercase__ : int = tokenizer.max_model_input_sizes[args.teacher_name] # DATA LOADER # logger.info(f'''Loading data from {args.data_file}''') with open(args.data_file , "rb") as fp: lowercase__ : Union[str, Any] = pickle.load(_UpperCAmelCase) if args.mlm: logger.info(f'''Loading token counts from {args.token_counts} (already pre-computed)''') with open(args.token_counts , "rb") as fp: lowercase__ : Dict = pickle.load(_UpperCAmelCase) lowercase__ : str = np.maximum(_UpperCAmelCase , 1) ** -args.mlm_smoothing for idx in special_tok_ids.values(): lowercase__ : Tuple = 0.0 # do not predict special tokens lowercase__ : Optional[int] = torch.from_numpy(_UpperCAmelCase) else: lowercase__ : List[str] = None lowercase__ : Optional[int] = LmSeqsDataset(params=_UpperCAmelCase , data=_UpperCAmelCase) logger.info("Data loader created.") # STUDENT # logger.info(f'''Loading student config from {args.student_config}''') lowercase__ : Optional[int] = student_config_class.from_pretrained(args.student_config) lowercase__ : Optional[Any] = True if args.student_pretrained_weights is not None: logger.info(f'''Loading pretrained weights from {args.student_pretrained_weights}''') lowercase__ : Any = student_model_class.from_pretrained(args.student_pretrained_weights , config=_UpperCAmelCase) else: lowercase__ : Tuple = student_model_class(_UpperCAmelCase) if args.n_gpu > 0: student.to(f'''cuda:{args.local_rank}''') logger.info("Student loaded.") # TEACHER # lowercase__ : Tuple = teacher_model_class.from_pretrained(args.teacher_name , output_hidden_states=_UpperCAmelCase) if args.n_gpu > 0: teacher.to(f'''cuda:{args.local_rank}''') logger.info(f'''Teacher loaded from {args.teacher_name}.''') # FREEZING # if args.freeze_pos_embs: freeze_pos_embeddings(_UpperCAmelCase , _UpperCAmelCase) if args.freeze_token_type_embds: freeze_token_type_embeddings(_UpperCAmelCase , _UpperCAmelCase) # SANITY CHECKS # assert student.config.vocab_size == teacher.config.vocab_size assert student.config.hidden_size == teacher.config.hidden_size assert student.config.max_position_embeddings == teacher.config.max_position_embeddings if args.mlm: assert token_probs.size(0) == stu_architecture_config.vocab_size # DISTILLER # torch.cuda.empty_cache() lowercase__ : Any = Distiller( params=_UpperCAmelCase , dataset=_UpperCAmelCase , token_probs=_UpperCAmelCase , student=_UpperCAmelCase , teacher=_UpperCAmelCase) distiller.train() logger.info("Let\'s go get some drinks.") if __name__ == "__main__": main()
87
def A ( _UpperCAmelCase : str ) -> bool: '''simple docstring''' return credit_card_number.startswith(('34', '35', '37', '4', '5', '6') ) def A ( _UpperCAmelCase : str ) -> bool: '''simple docstring''' _UpperCAmelCase = credit_card_number _UpperCAmelCase = 0 _UpperCAmelCase = len(_UpperCAmelCase ) - 2 for i in range(_UpperCAmelCase , -1 , -2 ): # double the value of every second digit _UpperCAmelCase = int(cc_number[i] ) digit *= 2 # If doubling of a number results in a two digit number # i.e greater than 9(e.g., 6 × 2 = 12), # then add the digits of the product (e.g., 12: 1 + 2 = 3, 15: 1 + 5 = 6), # to get a single digit number. if digit > 9: digit %= 10 digit += 1 _UpperCAmelCase = cc_number[:i] + str(_UpperCAmelCase ) + cc_number[i + 1 :] total += digit # Sum up the remaining digits for i in range(len(_UpperCAmelCase ) - 1 , -1 , -2 ): total += int(cc_number[i] ) return total % 10 == 0 def A ( _UpperCAmelCase : str ) -> bool: '''simple docstring''' _UpperCAmelCase = F"{credit_card_number} is an invalid credit card number because" if not credit_card_number.isdigit(): print(F"{error_message} it has nonnumerical characters." ) return False if not 13 <= len(_UpperCAmelCase ) <= 16: print(F"{error_message} of its length." ) return False if not validate_initial_digits(_UpperCAmelCase ): print(F"{error_message} of its first two digits." ) return False if not luhn_validation(_UpperCAmelCase ): print(F"{error_message} it fails the Luhn check." ) return False print(F"{credit_card_number} is a valid credit card number." ) return True if __name__ == "__main__": import doctest doctest.testmod() validate_credit_card_number("4111111111111111") validate_credit_card_number("32323")
339
0
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device __UpperCAmelCase =False class a__ ( unittest.TestCase ): pass @slow @require_torch_gpu class a__ ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self : Dict ): """simple docstring""" __lowerCamelCase = VersatileDiffusionImageVariationPipeline.from_pretrained('''shi-labs/versatile-diffusion''' ) pipe.to(a ) pipe.set_progress_bar_config(disable=a ) __lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) __lowerCamelCase = torch.manual_seed(0 ) __lowerCamelCase = pipe( image=a , generator=a , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' , ).images __lowerCamelCase = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) __lowerCamelCase = np.array([0.04_41, 0.04_69, 0.05_07, 0.05_75, 0.06_32, 0.06_50, 0.08_65, 0.09_09, 0.09_45] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
360
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool __UpperCAmelCase ={ "Acehnese Arabic": "ace_Arab", "Acehnese Latin": "ace_Latn", "Mesopotamian Arabic": "acm_Arab", "Ta'izzi-Adeni Arabic": "acq_Arab", "Tunisian Arabic": "aeb_Arab", "Afrikaans": "afr_Latn", "South Levantine Arabic": "ajp_Arab", "Akan": "aka_Latn", "Amharic": "amh_Ethi", "North Levantine Arabic": "apc_Arab", "Modern Standard Arabic": "arb_Arab", "Modern Standard Arabic Romanized": "arb_Latn", "Najdi Arabic": "ars_Arab", "Moroccan Arabic": "ary_Arab", "Egyptian Arabic": "arz_Arab", "Assamese": "asm_Beng", "Asturian": "ast_Latn", "Awadhi": "awa_Deva", "Central Aymara": "ayr_Latn", "South Azerbaijani": "azb_Arab", "North Azerbaijani": "azj_Latn", "Bashkir": "bak_Cyrl", "Bambara": "bam_Latn", "Balinese": "ban_Latn", "Belarusian": "bel_Cyrl", "Bemba": "bem_Latn", "Bengali": "ben_Beng", "Bhojpuri": "bho_Deva", "Banjar Arabic": "bjn_Arab", "Banjar Latin": "bjn_Latn", "Standard Tibetan": "bod_Tibt", "Bosnian": "bos_Latn", "Buginese": "bug_Latn", "Bulgarian": "bul_Cyrl", "Catalan": "cat_Latn", "Cebuano": "ceb_Latn", "Czech": "ces_Latn", "Chokwe": "cjk_Latn", "Central Kurdish": "ckb_Arab", "Crimean Tatar": "crh_Latn", "Welsh": "cym_Latn", "Danish": "dan_Latn", "German": "deu_Latn", "Southwestern Dinka": "dik_Latn", "Dyula": "dyu_Latn", "Dzongkha": "dzo_Tibt", "Greek": "ell_Grek", "English": "eng_Latn", "Esperanto": "epo_Latn", "Estonian": "est_Latn", "Basque": "eus_Latn", "Ewe": "ewe_Latn", "Faroese": "fao_Latn", "Fijian": "fij_Latn", "Finnish": "fin_Latn", "Fon": "fon_Latn", "French": "fra_Latn", "Friulian": "fur_Latn", "Nigerian Fulfulde": "fuv_Latn", "Scottish Gaelic": "gla_Latn", "Irish": "gle_Latn", "Galician": "glg_Latn", "Guarani": "grn_Latn", "Gujarati": "guj_Gujr", "Haitian Creole": "hat_Latn", "Hausa": "hau_Latn", "Hebrew": "heb_Hebr", "Hindi": "hin_Deva", "Chhattisgarhi": "hne_Deva", "Croatian": "hrv_Latn", "Hungarian": "hun_Latn", "Armenian": "hye_Armn", "Igbo": "ibo_Latn", "Ilocano": "ilo_Latn", "Indonesian": "ind_Latn", "Icelandic": "isl_Latn", "Italian": "ita_Latn", "Javanese": "jav_Latn", "Japanese": "jpn_Jpan", "Kabyle": "kab_Latn", "Jingpho": "kac_Latn", "Kamba": "kam_Latn", "Kannada": "kan_Knda", "Kashmiri Arabic": "kas_Arab", "Kashmiri Devanagari": "kas_Deva", "Georgian": "kat_Geor", "Central Kanuri Arabic": "knc_Arab", "Central Kanuri Latin": "knc_Latn", "Kazakh": "kaz_Cyrl", "Kabiyè": "kbp_Latn", "Kabuverdianu": "kea_Latn", "Khmer": "khm_Khmr", "Kikuyu": "kik_Latn", "Kinyarwanda": "kin_Latn", "Kyrgyz": "kir_Cyrl", "Kimbundu": "kmb_Latn", "Northern Kurdish": "kmr_Latn", "Kikongo": "kon_Latn", "Korean": "kor_Hang", "Lao": "lao_Laoo", "Ligurian": "lij_Latn", "Limburgish": "lim_Latn", "Lingala": "lin_Latn", "Lithuanian": "lit_Latn", "Lombard": "lmo_Latn", "Latgalian": "ltg_Latn", "Luxembourgish": "ltz_Latn", "Luba-Kasai": "lua_Latn", "Ganda": "lug_Latn", "Luo": "luo_Latn", "Mizo": "lus_Latn", "Standard Latvian": "lvs_Latn", "Magahi": "mag_Deva", "Maithili": "mai_Deva", "Malayalam": "mal_Mlym", "Marathi": "mar_Deva", "Minangkabau Arabic ": "min_Arab", "Minangkabau Latin": "min_Latn", "Macedonian": "mkd_Cyrl", "Plateau Malagasy": "plt_Latn", "Maltese": "mlt_Latn", "Meitei Bengali": "mni_Beng", "Halh Mongolian": "khk_Cyrl", "Mossi": "mos_Latn", "Maori": "mri_Latn", "Burmese": "mya_Mymr", "Dutch": "nld_Latn", "Norwegian Nynorsk": "nno_Latn", "Norwegian Bokmål": "nob_Latn", "Nepali": "npi_Deva", "Northern Sotho": "nso_Latn", "Nuer": "nus_Latn", "Nyanja": "nya_Latn", "Occitan": "oci_Latn", "West Central Oromo": "gaz_Latn", "Odia": "ory_Orya", "Pangasinan": "pag_Latn", "Eastern Panjabi": "pan_Guru", "Papiamento": "pap_Latn", "Western Persian": "pes_Arab", "Polish": "pol_Latn", "Portuguese": "por_Latn", "Dari": "prs_Arab", "Southern Pashto": "pbt_Arab", "Ayacucho Quechua": "quy_Latn", "Romanian": "ron_Latn", "Rundi": "run_Latn", "Russian": "rus_Cyrl", "Sango": "sag_Latn", "Sanskrit": "san_Deva", "Santali": "sat_Olck", "Sicilian": "scn_Latn", "Shan": "shn_Mymr", "Sinhala": "sin_Sinh", "Slovak": "slk_Latn", "Slovenian": "slv_Latn", "Samoan": "smo_Latn", "Shona": "sna_Latn", "Sindhi": "snd_Arab", "Somali": "som_Latn", "Southern Sotho": "sot_Latn", "Spanish": "spa_Latn", "Tosk Albanian": "als_Latn", "Sardinian": "srd_Latn", "Serbian": "srp_Cyrl", "Swati": "ssw_Latn", "Sundanese": "sun_Latn", "Swedish": "swe_Latn", "Swahili": "swh_Latn", "Silesian": "szl_Latn", "Tamil": "tam_Taml", "Tatar": "tat_Cyrl", "Telugu": "tel_Telu", "Tajik": "tgk_Cyrl", "Tagalog": "tgl_Latn", "Thai": "tha_Thai", "Tigrinya": "tir_Ethi", "Tamasheq Latin": "taq_Latn", "Tamasheq Tifinagh": "taq_Tfng", "Tok Pisin": "tpi_Latn", "Tswana": "tsn_Latn", "Tsonga": "tso_Latn", "Turkmen": "tuk_Latn", "Tumbuka": "tum_Latn", "Turkish": "tur_Latn", "Twi": "twi_Latn", "Central Atlas Tamazight": "tzm_Tfng", "Uyghur": "uig_Arab", "Ukrainian": "ukr_Cyrl", "Umbundu": "umb_Latn", "Urdu": "urd_Arab", "Northern Uzbek": "uzn_Latn", "Venetian": "vec_Latn", "Vietnamese": "vie_Latn", "Waray": "war_Latn", "Wolof": "wol_Latn", "Xhosa": "xho_Latn", "Eastern Yiddish": "ydd_Hebr", "Yoruba": "yor_Latn", "Yue Chinese": "yue_Hant", "Chinese Simplified": "zho_Hans", "Chinese Traditional": "zho_Hant", "Standard Malay": "zsm_Latn", "Zulu": "zul_Latn", } class a__ ( UpperCAmelCase__ ): lowerCamelCase : Any ="facebook/nllb-200-distilled-600M" lowerCamelCase : Optional[Any] =( "This is a tool that translates text from a language to another. It takes three inputs: `text`, which should " "be the text to translate, `src_lang`, which should be the language of the text to translate and `tgt_lang`, " "which should be the language for the desired ouput language. Both `src_lang` and `tgt_lang` are written in " "plain English, such as 'Romanian', or 'Albanian'. It returns the text translated in `tgt_lang`." ) lowerCamelCase : Tuple ="translator" lowerCamelCase : Any =AutoTokenizer lowerCamelCase : Dict =AutoModelForSeqaSeqLM lowerCamelCase : Union[str, Any] =LANGUAGE_CODES lowerCamelCase : Any =["text", "text", "text"] lowerCamelCase : Union[str, Any] =["text"] def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , a : Any , a : List[str] , a : Any ): """simple docstring""" if src_lang not in self.lang_to_code: raise ValueError(f"""{src_lang} is not a supported language.""" ) if tgt_lang not in self.lang_to_code: raise ValueError(f"""{tgt_lang} is not a supported language.""" ) __lowerCamelCase = self.lang_to_code[src_lang] __lowerCamelCase = self.lang_to_code[tgt_lang] return self.pre_processor._build_translation_inputs( a , return_tensors='''pt''' , src_lang=a , tgt_lang=a ) def SCREAMING_SNAKE_CASE__ ( self : str , a : Optional[Any] ): """simple docstring""" return self.model.generate(**a ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , a : str ): """simple docstring""" return self.post_processor.decode(outputs[0].tolist() , skip_special_tokens=a )
237
0
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase : List[str] ={"""configuration_focalnet""": ["""FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """FocalNetConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : List[Any] =[ """FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """FocalNetForImageClassification""", """FocalNetForMaskedImageModeling""", """FocalNetBackbone""", """FocalNetModel""", """FocalNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_focalnet import FOCALNET_PRETRAINED_CONFIG_ARCHIVE_MAP, FocalNetConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_focalnet import ( FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST, FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, FocalNetPreTrainedModel, ) else: import sys UpperCAmelCase : List[Any] =_LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
128
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar UpperCAmelCase : Dict =TypeVar("""T""") class _lowercase (Generic[T] ): '''simple docstring''' def __init__( self , snake_case__ ): '''simple docstring''' UpperCamelCase_ = data UpperCamelCase_ = None def __str__( self ): '''simple docstring''' return F"""{self.data}""" class _lowercase (Generic[T] ): '''simple docstring''' def __init__( self ): '''simple docstring''' UpperCamelCase_ = None def __iter__( self ): '''simple docstring''' UpperCamelCase_ = self.top while node: yield node.data UpperCamelCase_ = node.next def __str__( self ): '''simple docstring''' return "->".join([str(snake_case__ ) for item in self] ) def __len__( self ): '''simple docstring''' return len(tuple(iter(self ) ) ) def _lowerCamelCase ( self ): '''simple docstring''' return self.top is None def _lowerCamelCase ( self , snake_case__ ): '''simple docstring''' UpperCamelCase_ = Node(snake_case__ ) if not self.is_empty(): UpperCamelCase_ = self.top UpperCamelCase_ = node def _lowerCamelCase ( self ): '''simple docstring''' if self.is_empty(): raise IndexError("pop from empty stack" ) assert isinstance(self.top , snake_case__ ) UpperCamelCase_ = self.top UpperCamelCase_ = self.top.next return pop_node.data def _lowerCamelCase ( self ): '''simple docstring''' if self.is_empty(): raise IndexError("peek from empty stack" ) assert self.top is not None return self.top.data def _lowerCamelCase ( self ): '''simple docstring''' UpperCamelCase_ = None if __name__ == "__main__": from doctest import testmod testmod()
128
1
def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :int , SCREAMING_SNAKE_CASE :bool = False ) -> bool: if n == 2: return True if not n % 2 or n < 2: return False if n > 5 and n % 10 not in (1, 3, 7, 9): # can quickly check last digit return False if n > 3_317_044_064_679_887_385_961_981 and not allow_probable: raise ValueError( """Warning: upper bound of deterministic test is exceeded. """ """Pass allow_probable=True to allow probabilistic test. """ """A return value of True indicates a probable prime.""" ) # array bounds provided by analysis __lowerCAmelCase : str = [ 2_047, 1_373_653, 25_326_001, 3_215_031_751, 2_152_302_898_747, 3_474_749_660_383, 341_550_071_728_321, 1, 3_825_123_056_546_413_051, 1, 1, 318_665_857_834_031_151_167_461, 3_317_044_064_679_887_385_961_981, ] __lowerCAmelCase : Tuple = [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41] for idx, _p in enumerate(SCREAMING_SNAKE_CASE , 1 ): if n < _p: # then we have our last prime to check __lowerCAmelCase : Any = primes[:idx] break __lowerCAmelCase , __lowerCAmelCase : Any = n - 1, 0 # break up n -1 into a power of 2 (s) and # remaining odd component # essentially, solve for d * 2 ** s == n - 1 while d % 2 == 0: d //= 2 s += 1 for prime in plist: __lowerCAmelCase : List[Any] = False for r in range(SCREAMING_SNAKE_CASE ): __lowerCAmelCase : List[Any] = pow(SCREAMING_SNAKE_CASE , d * 2**r , SCREAMING_SNAKE_CASE ) # see article for analysis explanation for m if (r == 0 and m == 1) or ((m + 1) % n == 0): __lowerCAmelCase : int = True # this loop will not determine compositeness break if pr: continue # if pr is False, then the above loop never evaluated to true, # and the n MUST be composite return False return True def _SCREAMING_SNAKE_CASE ( ) -> None: assert not miller_rabin(561 ) assert miller_rabin(563 ) # 2047 assert not miller_rabin(838_201 ) assert miller_rabin(838_207 ) # 1_373_653 assert not miller_rabin(17_316_001 ) assert miller_rabin(17_316_017 ) # 25_326_001 assert not miller_rabin(3_078_386_641 ) assert miller_rabin(3_078_386_653 ) # 3_215_031_751 assert not miller_rabin(1_713_045_574_801 ) assert miller_rabin(1_713_045_574_819 ) # 2_152_302_898_747 assert not miller_rabin(2_779_799_728_307 ) assert miller_rabin(2_779_799_728_327 ) # 3_474_749_660_383 assert not miller_rabin(113_850_023_909_441 ) assert miller_rabin(113_850_023_909_527 ) # 341_550_071_728_321 assert not miller_rabin(1_275_041_018_848_804_351 ) assert miller_rabin(1_275_041_018_848_804_391 ) # 3_825_123_056_546_413_051 assert not miller_rabin(79_666_464_458_507_787_791_867 ) assert miller_rabin(79_666_464_458_507_787_791_951 ) # 318_665_857_834_031_151_167_461 assert not miller_rabin(552_840_677_446_647_897_660_333 ) assert miller_rabin(552_840_677_446_647_897_660_359 ) # 3_317_044_064_679_887_385_961_981 # upper limit for probabilistic test if __name__ == "__main__": test_miller_rabin()
232
import sacrebleu as scb from packaging import version from sacrebleu import TER import datasets _UpperCAmelCase = '\\n@inproceedings{snover-etal-2006-study,\n title = "A Study of Translation Edit Rate with Targeted Human Annotation",\n author = "Snover, Matthew and\n Dorr, Bonnie and\n Schwartz, Rich and\n Micciulla, Linnea and\n Makhoul, John",\n booktitle = "Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers",\n month = aug # " 8-12",\n year = "2006",\n address = "Cambridge, Massachusetts, USA",\n publisher = "Association for Machine Translation in the Americas",\n url = "https://aclanthology.org/2006.amta-papers.25",\n pages = "223--231",\n}\n@inproceedings{post-2018-call,\n title = "A Call for Clarity in Reporting {BLEU} Scores",\n author = "Post, Matt",\n booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers",\n month = oct,\n year = "2018",\n address = "Belgium, Brussels",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/W18-6319",\n pages = "186--191",\n}\n' _UpperCAmelCase = '\\nTER (Translation Edit Rate, also called Translation Error Rate) is a metric to quantify the edit operations that a\nhypothesis requires to match a reference translation. We use the implementation that is already present in sacrebleu\n(https://github.com/mjpost/sacreBLEU#ter), which in turn is inspired by the TERCOM implementation, which can be found\nhere: https://github.com/jhclark/tercom.\n\nThe implementation here is slightly different from sacrebleu in terms of the required input format. The length of\nthe references and hypotheses lists need to be the same, so you may need to transpose your references compared to\nsacrebleu\'s required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534\n\nSee the README.md file at https://github.com/mjpost/sacreBLEU#ter for more information.\n' _UpperCAmelCase = '\nProduces TER scores alongside the number of edits and reference length.\n\nArgs:\n predictions (list of str): The system stream (a sequence of segments).\n references (list of list of str): A list of one or more reference streams (each a sequence of segments).\n normalized (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`.\n ignore_punct (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`.\n support_zh_ja_chars (boolean): If `True`, tokenization/normalization supports processing of Chinese characters,\n as well as Japanese Kanji, Hiragana, Katakana, and Phonetic Extensions of Katakana.\n Only applies if `normalized = True`. Defaults to `False`.\n case_sensitive (boolean): If `False`, makes all predictions and references lowercase to ignore differences in case. Defaults to `False`.\n\nReturns:\n \'score\' (float): TER score (num_edits / sum_ref_lengths * 100)\n \'num_edits\' (int): The cumulative number of edits\n \'ref_length\' (float): The cumulative average reference length\n\nExamples:\n Example 1:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?",\n ... "What did the TER metric user say to the developer?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"],\n ... ["Your jokes are...", "...TERrible"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... case_sensitive=True)\n >>> print(results)\n {\'score\': 150.0, \'num_edits\': 15, \'ref_length\': 10.0}\n\n Example 2:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... case_sensitive=True)\n >>> print(results)\n {\'score\': 62.5, \'num_edits\': 5, \'ref_length\': 8.0}\n\n Example 3:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... normalized=True,\n ... case_sensitive=True)\n >>> print(results)\n {\'score\': 57.14285714285714, \'num_edits\': 6, \'ref_length\': 10.5}\n\n Example 4:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... ignore_punct=True,\n ... case_sensitive=False)\n >>> print(results)\n {\'score\': 0.0, \'num_edits\': 0, \'ref_length\': 8.0}\n\n Example 5:\n >>> predictions = ["does this sentence match??",\n ... "what about this sentence?",\n ... "What did the TER metric user say to the developer?"]\n >>> references = [["does this sentence match", "does this sentence match!?!"],\n ... ["wHaT aBoUt ThIs SeNtEnCe?", "wHaT aBoUt ThIs SeNtEnCe?"],\n ... ["Your jokes are...", "...TERrible"]]\n >>> ter = datasets.load_metric("ter")\n >>> results = ter.compute(predictions=predictions,\n ... references=references,\n ... ignore_punct=True,\n ... case_sensitive=False)\n >>> print(results)\n {\'score\': 100.0, \'num_edits\': 10, \'ref_length\': 10.0}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class snake_case_ ( datasets.Metric ): def UpperCAmelCase__ ( self : Tuple )->str: '''simple docstring''' if version.parse(scb.__version__ ) < version.parse("""1.4.12""" ): raise ImportWarning( """To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n""" """You can install it with `pip install \"sacrebleu>=1.4.12\"`.""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="""http://www.cs.umd.edu/~snover/tercom/""" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Sequence(datasets.Value("""string""" , id="""sequence""" ) , id="""references""" ), } ) , codebase_urls=["""https://github.com/mjpost/sacreBLEU#ter"""] , reference_urls=[ """https://github.com/jhclark/tercom""", ] , ) def UpperCAmelCase__ ( self : Dict , _snake_case : List[Any] , _snake_case : int , _snake_case : bool = False , _snake_case : bool = False , _snake_case : bool = False , _snake_case : bool = False , )->str: '''simple docstring''' __lowerCAmelCase : List[str] = len(references[0] ) if any(len(_snake_case ) != references_per_prediction for refs in references ): raise ValueError("""Sacrebleu requires the same number of references for each prediction""" ) __lowerCAmelCase : str = [[refs[i] for refs in references] for i in range(_snake_case )] __lowerCAmelCase : Tuple = TER( normalized=_snake_case , no_punct=_snake_case , asian_support=_snake_case , case_sensitive=_snake_case , ) __lowerCAmelCase : List[Any] = sb_ter.corpus_score(_snake_case , _snake_case ) return {"score": output.score, "num_edits": output.num_edits, "ref_length": output.ref_length}
232
1
import argparse import os import re import packaging.version UpperCamelCase__ = """examples/""" UpperCamelCase__ = { """examples""": (re.compile(R"""^check_min_version\(\"[^\"]+\"\)\s*$""", re.MULTILINE), """check_min_version(\"VERSION\")\n"""), """init""": (re.compile(R"""^__version__\s+=\s+\"([^\"]+)\"\s*$""", re.MULTILINE), """__version__ = \"VERSION\"\n"""), """setup""": (re.compile(R"""^(\s*)version\s*=\s*\"[^\"]+\",""", re.MULTILINE), R"""\1version=\"VERSION\","""), """doc""": (re.compile(R"""^(\s*)release\s*=\s*\"[^\"]+\"$""", re.MULTILINE), """release = \"VERSION\"\n"""), } UpperCamelCase__ = { """init""": """src/transformers/__init__.py""", """setup""": """setup.py""", } UpperCamelCase__ = """README.md""" def _a ( SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Any , SCREAMING_SNAKE_CASE_ : List[str] ): with open(SCREAMING_SNAKE_CASE_ , "r" , encoding="utf-8" , newline="\n" ) as f: __lowerCAmelCase = f.read() __lowerCAmelCase , __lowerCAmelCase = REPLACE_PATTERNS[pattern] __lowerCAmelCase = replace.replace("VERSION" , SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase = re_pattern.sub(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) with open(SCREAMING_SNAKE_CASE_ , "w" , encoding="utf-8" , newline="\n" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) def _a ( SCREAMING_SNAKE_CASE_ : List[Any] ): for folder, directories, fnames in os.walk(SCREAMING_SNAKE_CASE_ ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove("research_projects" ) if "legacy" in directories: directories.remove("legacy" ) for fname in fnames: if fname.endswith(".py" ): update_version_in_file(os.path.join(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ , pattern="examples" ) def _a ( SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : Optional[int]=False ): for pattern, fname in REPLACE_FILES.items(): update_version_in_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if not patch: update_version_in_examples(SCREAMING_SNAKE_CASE_ ) def _a ( ): __lowerCAmelCase = "🤗 Transformers currently provides the following architectures" __lowerCAmelCase = "1. Want to contribute a new model?" with open(SCREAMING_SNAKE_CASE_ , "r" , encoding="utf-8" , newline="\n" ) as f: __lowerCAmelCase = f.readlines() # Find the start of the list. __lowerCAmelCase = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 __lowerCAmelCase = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith("1." ): __lowerCAmelCase = lines[index].replace( "https://huggingface.co/docs/transformers/main/model_doc" , "https://huggingface.co/docs/transformers/model_doc" , ) index += 1 with open(SCREAMING_SNAKE_CASE_ , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(SCREAMING_SNAKE_CASE_ ) def _a ( ): with open(REPLACE_FILES["init"] , "r" ) as f: __lowerCAmelCase = f.read() __lowerCAmelCase = REPLACE_PATTERNS["init"][0].search(SCREAMING_SNAKE_CASE_ ).groups()[0] return packaging.version.parse(SCREAMING_SNAKE_CASE_ ) def _a ( SCREAMING_SNAKE_CASE_ : List[Any]=False ): __lowerCAmelCase = get_version() if patch and default_version.is_devrelease: raise ValueError("Can't create a patch version from the dev branch, checkout a released version!" ) if default_version.is_devrelease: __lowerCAmelCase = default_version.base_version elif patch: __lowerCAmelCase = F"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: __lowerCAmelCase = F"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. __lowerCAmelCase = input(F"""Which version are you releasing? [{default_version}]""" ) if len(SCREAMING_SNAKE_CASE_ ) == 0: __lowerCAmelCase = default_version print(F"""Updating version to {version}.""" ) global_version_update(SCREAMING_SNAKE_CASE_ , patch=SCREAMING_SNAKE_CASE_ ) if not patch: print("Cleaning main README, don't forget to run `make fix-copies`." ) clean_main_ref_in_model_list() def _a ( ): __lowerCAmelCase = get_version() __lowerCAmelCase = F"""{current_version.major}.{current_version.minor + 1}.0.dev0""" __lowerCAmelCase = current_version.base_version # Check with the user we got that right. __lowerCAmelCase = input(F"""Which version are we developing now? [{dev_version}]""" ) if len(SCREAMING_SNAKE_CASE_ ) == 0: __lowerCAmelCase = dev_version print(F"""Updating version to {version}.""" ) global_version_update(SCREAMING_SNAKE_CASE_ ) print("Cleaning main README, don't forget to run `make fix-copies`." ) clean_main_ref_in_model_list() if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() parser.add_argument("""--post_release""", action="""store_true""", help="""Whether this is pre or post release.""") parser.add_argument("""--patch""", action="""store_true""", help="""Whether or not this is a patch release.""") UpperCamelCase__ = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("""Nothing to do after a patch :-)""") else: post_release_work()
92
"""simple docstring""" import re def _lowercase ( __snake_case ) -> str: if len(re.findall("[ATCG]" ,__snake_case ) ) != len(__snake_case ): raise ValueError("Invalid Strand" ) return dna.translate(dna.maketrans("ATCG" ,"TAGC" ) ) if __name__ == "__main__": import doctest doctest.testmod()
269
0
"""simple docstring""" import math def UpperCamelCase_ ( ) -> None: """simple docstring""" lowerCAmelCase_ : List[Any] = input('Enter message: ' ) lowerCAmelCase_ : Any = int(input(f"Enter key [2-{len(lowerCAmelCase__ ) - 1}]: " ) ) lowerCAmelCase_ : Union[str, Any] = input('Encryption/Decryption [e/d]: ' ) if mode.lower().startswith('e' ): lowerCAmelCase_ : List[str] = encrypt_message(lowerCAmelCase__ , lowerCAmelCase__ ) elif mode.lower().startswith('d' ): lowerCAmelCase_ : List[Any] = decrypt_message(lowerCAmelCase__ , lowerCAmelCase__ ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(f"Output:\n{text + '|'}" ) def UpperCamelCase_ ( lowerCAmelCase__ : int , lowerCAmelCase__ : str ) -> str: """simple docstring""" lowerCAmelCase_ : List[Any] = [''] * key for col in range(lowerCAmelCase__ ): lowerCAmelCase_ : Tuple = col while pointer < len(lowerCAmelCase__ ): cipher_text[col] += message[pointer] pointer += key return "".join(lowerCAmelCase__ ) def UpperCamelCase_ ( lowerCAmelCase__ : int , lowerCAmelCase__ : str ) -> str: """simple docstring""" lowerCAmelCase_ : Tuple = math.ceil(len(lowerCAmelCase__ ) / key ) lowerCAmelCase_ : Dict = key lowerCAmelCase_ : int = (num_cols * num_rows) - len(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = [''] * num_cols lowerCAmelCase_ : int = 0 lowerCAmelCase_ : Dict = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): lowerCAmelCase_ : List[str] = 0 row += 1 return "".join(lowerCAmelCase__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
359
"""simple docstring""" import os from bleurt import score # From: git+https://github.com/google-research/bleurt.git import datasets lowercase__ : Tuple = datasets.logging.get_logger(__name__) lowercase__ : List[Any] = """\ @inproceedings{bleurt, title={BLEURT: Learning Robust Metrics for Text Generation}, author={Thibault Sellam and Dipanjan Das and Ankur P. Parikh}, booktitle={ACL}, year={2020}, url={https://arxiv.org/abs/2004.04696} } """ lowercase__ : Tuple = """\ BLEURT a learnt evaluation metric for Natural Language Generation. It is built using multiple phases of transfer learning starting from a pretrained BERT model (Devlin et al. 2018) and then employing another pre-training phrase using synthetic data. Finally it is trained on WMT human annotations. You may run BLEURT out-of-the-box or fine-tune it for your specific application (the latter is expected to perform better). See the project's README at https://github.com/google-research/bleurt#readme for more information. """ lowercase__ : List[Any] = """ BLEURT score. Args: `predictions` (list of str): prediction/candidate sentences `references` (list of str): reference sentences `checkpoint` BLEURT checkpoint. Will default to BLEURT-tiny if None. Returns: 'scores': List of scores. Examples: >>> predictions = [\"hello there\", \"general kenobi\"] >>> references = [\"hello there\", \"general kenobi\"] >>> bleurt = datasets.load_metric(\"bleurt\") >>> results = bleurt.compute(predictions=predictions, references=references) >>> print([round(v, 2) for v in results[\"scores\"]]) [1.03, 1.04] """ lowercase__ : List[Any] = { """bleurt-tiny-128""": """https://storage.googleapis.com/bleurt-oss/bleurt-tiny-128.zip""", """bleurt-tiny-512""": """https://storage.googleapis.com/bleurt-oss/bleurt-tiny-512.zip""", """bleurt-base-128""": """https://storage.googleapis.com/bleurt-oss/bleurt-base-128.zip""", """bleurt-base-512""": """https://storage.googleapis.com/bleurt-oss/bleurt-base-512.zip""", """bleurt-large-128""": """https://storage.googleapis.com/bleurt-oss/bleurt-large-128.zip""", """bleurt-large-512""": """https://storage.googleapis.com/bleurt-oss/bleurt-large-512.zip""", """BLEURT-20-D3""": """https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D3.zip""", """BLEURT-20-D6""": """https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D6.zip""", """BLEURT-20-D12""": """https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D12.zip""", """BLEURT-20""": """https://storage.googleapis.com/bleurt-oss-21/BLEURT-20.zip""", } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class UpperCamelCase__ ( datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE__ ( self : List[str] ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='https://github.com/google-research/bleurt' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/google-research/bleurt'] , reference_urls=['https://github.com/google-research/bleurt', 'https://arxiv.org/abs/2004.04696'] , ) def SCREAMING_SNAKE_CASE__ ( self : str , SCREAMING_SNAKE_CASE_ : str ): # check that config name specifies a valid BLEURT model if self.config_name == "default": logger.warning( 'Using default BLEURT-Base checkpoint for sequence maximum length 128. ' 'You can use a bigger model for better results with e.g.: datasets.load_metric(\'bleurt\', \'bleurt-large-512\').' ) lowerCAmelCase_ : List[Any] = 'bleurt-base-128' if self.config_name.lower() in CHECKPOINT_URLS: lowerCAmelCase_ : List[Any] = self.config_name.lower() elif self.config_name.upper() in CHECKPOINT_URLS: lowerCAmelCase_ : Tuple = self.config_name.upper() else: raise KeyError( F"{self.config_name} model not found. You should supply the name of a model checkpoint for bleurt in {CHECKPOINT_URLS.keys()}" ) # download the model checkpoint specified by self.config_name and set up the scorer lowerCAmelCase_ : List[Any] = dl_manager.download_and_extract(CHECKPOINT_URLS[checkpoint_name] ) lowerCAmelCase_ : List[str] = score.BleurtScorer(os.path.join(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) def SCREAMING_SNAKE_CASE__ ( self : str , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : Any ): lowerCAmelCase_ : Tuple = self.scorer.score(references=SCREAMING_SNAKE_CASE_ , candidates=SCREAMING_SNAKE_CASE_ ) return {"scores": scores}
289
0
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
40
"""simple docstring""" import unittest from knapsack import knapsack as k class _A ( unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[Any]): a : str = 0 a : Optional[int] = [0] a : Union[str, Any] = [0] a : Any = len(__UpperCAmelCase) self.assertEqual(k.knapsack(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) , 0) a : List[str] = [60] a : str = [10] a : Optional[int] = len(__UpperCAmelCase) self.assertEqual(k.knapsack(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) , 0) def __snake_case ( self : Optional[int]): a : Any = 3 a : str = [1, 2, 3] a : Tuple = [3, 2, 1] a : Any = len(__UpperCAmelCase) self.assertEqual(k.knapsack(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) , 5) def __snake_case ( self : Tuple): a : int = 50 a : List[Any] = [60, 100, 120] a : Optional[int] = [10, 20, 30] a : str = len(__UpperCAmelCase) self.assertEqual(k.knapsack(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) , 220) if __name__ == "__main__": unittest.main()
40
1
import math def _a ( a :int = 100 ) -> int: a = sum(i * i for i in range(1 , n + 1 ) ) a = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(f"""{solution() = }""")
26
def _a ( a :int = 600_851_475_143 ) -> int: try: a = int(a ) except (TypeError, ValueError): raise TypeError('''Parameter n must be int or castable to int.''' ) if n <= 0: raise ValueError('''Parameter n must be greater than or equal to one.''' ) a = 2 a = 0 if n == 2: return 2 while n > 2: while n % i != 0: i += 1 a = i while n % i == 0: a = n // i i += 1 return int(a ) if __name__ == "__main__": print(f"""{solution() = }""")
26
1
"""simple docstring""" import re import time from typing import Optional import IPython.display as disp from ..trainer_callback import TrainerCallback from ..trainer_utils import IntervalStrategy, has_length def __UpperCAmelCase ( __UpperCamelCase ): __lowercase : Optional[int] = int(_a ) __lowercase : Optional[Any] = t // 36_00, (t // 60) % 60, t % 60 return f"""{h}:{m:02d}:{s:02d}""" if h != 0 else f"""{m:02d}:{s:02d}""" def __UpperCAmelCase ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=3_00 ): return f"""\n <div>\n {prefix}\n <progress value=\'{value}\' max=\'{total}\' style=\'width:{width}px; height:20px; vertical-align: middle;\'></progress>\n {label}\n </div>\n """ def __UpperCAmelCase ( __UpperCamelCase ): __lowercase : str = '''<table border="1" class="dataframe">\n''' html_code += """ <thead>\n <tr style="text-align: left;">\n""" for i in items[0]: html_code += f""" <th>{i}</th>\n""" html_code += " </tr>\n </thead>\n <tbody>\n" for line in items[1:]: html_code += " <tr>\n" for elt in line: __lowercase : Optional[Any] = f"""{elt:.6f}""" if isinstance(_a , _a ) else str(_a ) html_code += f""" <td>{elt}</td>\n""" html_code += " </tr>\n" html_code += " </tbody>\n</table><p>" return html_code class UpperCAmelCase_ : UpperCamelCase =5 UpperCamelCase =0.2 def __init__( self , UpperCamelCase_ , UpperCamelCase_ = None , UpperCamelCase_ = True , UpperCamelCase_ = None , UpperCamelCase_ = 3_00 , ) -> List[str]: __lowercase : Optional[int] = total __lowercase : Tuple = '''''' if prefix is None else prefix __lowercase : int = leave __lowercase : Any = parent __lowercase : List[Any] = width __lowercase : int = None __lowercase : str = None __lowercase : Optional[Any] = None def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ = False , UpperCamelCase_ = None ) -> List[str]: __lowercase : Tuple = value if comment is not None: __lowercase : List[Any] = comment if self.last_value is None: __lowercase : Tuple = time.time() __lowercase : str = value __lowercase : Union[str, Any] = None __lowercase : Union[str, Any] = self.warmup __lowercase : List[str] = 1 self.update_bar(_SCREAMING_SNAKE_CASE ) elif value <= self.last_value and not force_update: return elif force_update or self.first_calls > 0 or value >= min(self.last_value + self.wait_for , self.total ): if self.first_calls > 0: self.first_calls -= 1 __lowercase : Union[str, Any] = time.time() __lowercase : int = current_time - self.start_time # We could have value = self.start_value if the update is called twixe with the same start value. if value > self.start_value: __lowercase : List[str] = self.elapsed_time / (value - self.start_value) else: __lowercase : int = None if value >= self.total: __lowercase : Union[str, Any] = self.total __lowercase : List[Any] = None if not self.leave: self.close() elif self.average_time_per_item is not None: __lowercase : Optional[Any] = self.average_time_per_item * (self.total - value) self.update_bar(_SCREAMING_SNAKE_CASE ) __lowercase : Optional[Any] = value __lowercase : Optional[int] = current_time if self.average_time_per_item is None: __lowercase : str = 1 else: __lowercase : Optional[Any] = max(int(self.update_every / self.average_time_per_item ) , 1 ) def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_=None ) -> str: __lowercase : List[str] = ''' ''' * (len(str(self.total ) ) - len(str(_SCREAMING_SNAKE_CASE ) )) + str(_SCREAMING_SNAKE_CASE ) if self.elapsed_time is None: __lowercase : Union[str, Any] = F"""[{spaced_value}/{self.total} : < :""" elif self.predicted_remaining is None: __lowercase : List[str] = F"""[{spaced_value}/{self.total} {format_time(self.elapsed_time )}""" else: __lowercase : Dict = ( F"""[{spaced_value}/{self.total} {format_time(self.elapsed_time )} <""" F""" {format_time(self.predicted_remaining )}""" ) self.label += F""", {1/self.average_time_per_item:.2f} it/s""" self.label += "]" if self.comment is None or len(self.comment ) == 0 else F""", {self.comment}]""" self.display() def _lowerCamelCase ( self ) -> Optional[int]: __lowercase : Optional[int] = html_progress_bar(self.value , self.total , self.prefix , self.label , self.width ) if self.parent is not None: # If this is a child bar, the parent will take care of the display. self.parent.display() return if self.output is None: __lowercase : List[str] = disp.display(disp.HTML(self.html_code ) , display_id=_SCREAMING_SNAKE_CASE ) else: self.output.update(disp.HTML(self.html_code ) ) def _lowerCamelCase ( self ) -> List[str]: if self.parent is None and self.output is not None: self.output.update(disp.HTML('''''' ) ) class UpperCAmelCase_ ( _lowercase ): def __init__( self , UpperCamelCase_ , UpperCamelCase_=None ) -> Optional[Any]: super().__init__(_SCREAMING_SNAKE_CASE ) __lowercase : Tuple = None if column_names is None else [column_names] __lowercase : Optional[int] = None def _lowerCamelCase ( self ) -> Union[str, Any]: __lowercase : Optional[int] = html_progress_bar(self.value , self.total , self.prefix , self.label , self.width ) if self.inner_table is not None: self.html_code += text_to_html_table(self.inner_table ) if self.child_bar is not None: self.html_code += self.child_bar.html_code if self.output is None: __lowercase : Optional[int] = disp.display(disp.HTML(self.html_code ) , display_id=_SCREAMING_SNAKE_CASE ) else: self.output.update(disp.HTML(self.html_code ) ) def _lowerCamelCase ( self , UpperCamelCase_ ) -> Union[str, Any]: if self.inner_table is None: __lowercase : str = [list(values.keys() ), list(values.values() )] else: __lowercase : Tuple = self.inner_table[0] if len(self.inner_table ) == 1: # We give a chance to update the column names at the first iteration for key in values.keys(): if key not in columns: columns.append(_SCREAMING_SNAKE_CASE ) __lowercase : List[Any] = columns self.inner_table.append([values[c] for c in columns] ) def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_=None , UpperCamelCase_=3_00 ) -> Optional[Any]: __lowercase : Optional[int] = NotebookProgressBar(_SCREAMING_SNAKE_CASE , prefix=_SCREAMING_SNAKE_CASE , parent=self , width=_SCREAMING_SNAKE_CASE ) return self.child_bar def _lowerCamelCase ( self ) -> Any: __lowercase : List[Any] = None self.display() class UpperCAmelCase_ ( _lowercase ): def __init__( self ) -> int: __lowercase : Any = None __lowercase : Any = None __lowercase : Any = False def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , **UpperCamelCase_ ) -> List[str]: __lowercase : str = '''Epoch''' if args.evaluation_strategy == IntervalStrategy.EPOCH else '''Step''' __lowercase : Union[str, Any] = 0 __lowercase : List[Any] = 0 __lowercase : Dict = [self.first_column] + ['''Training Loss'''] if args.evaluation_strategy != IntervalStrategy.NO: column_names.append('''Validation Loss''' ) __lowercase : List[Any] = NotebookTrainingTracker(state.max_steps , _SCREAMING_SNAKE_CASE ) def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , **UpperCamelCase_ ) -> Union[str, Any]: __lowercase : Dict = int(state.epoch ) if int(state.epoch ) == state.epoch else F"""{state.epoch:.2f}""" self.training_tracker.update( state.global_step + 1 , comment=F"""Epoch {epoch}/{state.num_train_epochs}""" , force_update=self._force_next_update , ) __lowercase : Optional[Any] = False def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=None , **UpperCamelCase_ ) -> List[str]: if not has_length(_SCREAMING_SNAKE_CASE ): return if self.prediction_bar is None: if self.training_tracker is not None: __lowercase : List[Any] = self.training_tracker.add_child(len(_SCREAMING_SNAKE_CASE ) ) else: __lowercase : List[str] = NotebookProgressBar(len(_SCREAMING_SNAKE_CASE ) ) self.prediction_bar.update(1 ) else: self.prediction_bar.update(self.prediction_bar.value + 1 ) def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , **UpperCamelCase_ ) -> Optional[int]: if self.prediction_bar is not None: self.prediction_bar.close() __lowercase : List[Any] = None def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=None , **UpperCamelCase_ ) -> Union[str, Any]: # Only for when there is no evaluation if args.evaluation_strategy == IntervalStrategy.NO and "loss" in logs: __lowercase : Optional[Any] = {'''Training Loss''': logs['''loss''']} # First column is necessarily Step sine we're not in epoch eval strategy __lowercase : Optional[Any] = state.global_step self.training_tracker.write_line(_SCREAMING_SNAKE_CASE ) def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=None , **UpperCamelCase_ ) -> Any: if self.training_tracker is not None: __lowercase : List[str] = {'''Training Loss''': '''No log''', '''Validation Loss''': '''No log'''} for log in reversed(state.log_history ): if "loss" in log: __lowercase : int = log['''loss'''] break if self.first_column == "Epoch": __lowercase : Optional[int] = int(state.epoch ) else: __lowercase : Optional[int] = state.global_step __lowercase : Optional[Any] = '''eval''' for k in metrics: if k.endswith('''_loss''' ): __lowercase : Optional[Any] = re.sub(R'''\_loss$''' , '''''' , _SCREAMING_SNAKE_CASE ) __lowercase : Union[str, Any] = metrics.pop('''total_flos''' , _SCREAMING_SNAKE_CASE ) __lowercase : Union[str, Any] = metrics.pop('''epoch''' , _SCREAMING_SNAKE_CASE ) __lowercase : Any = metrics.pop(F"""{metric_key_prefix}_runtime""" , _SCREAMING_SNAKE_CASE ) __lowercase : int = metrics.pop(F"""{metric_key_prefix}_samples_per_second""" , _SCREAMING_SNAKE_CASE ) __lowercase : List[str] = metrics.pop(F"""{metric_key_prefix}_steps_per_second""" , _SCREAMING_SNAKE_CASE ) __lowercase : List[Any] = metrics.pop(F"""{metric_key_prefix}_jit_compilation_time""" , _SCREAMING_SNAKE_CASE ) for k, v in metrics.items(): if k == F"""{metric_key_prefix}_loss""": __lowercase : Union[str, Any] = v else: __lowercase : List[str] = k.split('''_''' ) __lowercase : Union[str, Any] = ''' '''.join([part.capitalize() for part in splits[1:]] ) __lowercase : Optional[int] = v self.training_tracker.write_line(_SCREAMING_SNAKE_CASE ) self.training_tracker.remove_child() __lowercase : List[Any] = None # Evaluation takes a long time so we should force the next update. __lowercase : str = True def _lowerCamelCase ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , **UpperCamelCase_ ) -> Optional[Any]: self.training_tracker.update( state.global_step , comment=F"""Epoch {int(state.epoch )}/{state.num_train_epochs}""" , force_update=_SCREAMING_SNAKE_CASE ) __lowercase : Dict = None
249
import argparse import glob import logging import os import sys import time from collections import defaultdict from pathlib import Path from typing import Dict, List, Tuple import numpy as np import pytorch_lightning as pl import torch from callbacks import SeqaSeqLoggingCallback, get_checkpoint_callback, get_early_stopping_callback from torch import nn from torch.utils.data import DataLoader from transformers import MBartTokenizer, TaForConditionalGeneration from transformers.models.bart.modeling_bart import shift_tokens_right from utils import ( ROUGE_KEYS, LegacySeqaSeqDataset, SeqaSeqDataset, assert_all_frozen, calculate_bleu, calculate_rouge, check_output_dir, flatten_list, freeze_embeds, freeze_params, get_git_info, label_smoothed_nll_loss, lmap, pickle_save, save_git_info, save_json, use_task_specific_params, ) # need the parent dir module sys.path.insert(2, str(Path(__file__).resolve().parents[1])) from lightning_base import BaseTransformer, add_generic_args, generic_train # noqa lowerCamelCase = logging.getLogger(__name__) class _a ( _lowercase): _a : Dict = '''summarization''' _a : int = ['''loss'''] _a : Optional[Any] = ROUGE_KEYS _a : Any = '''rouge2''' def __init__( self : Optional[Any] , _SCREAMING_SNAKE_CASE : List[str] , **_SCREAMING_SNAKE_CASE : Dict )-> int: if hparams.sortish_sampler and hparams.gpus > 1: lowerCAmelCase__ : int = False elif hparams.max_tokens_per_batch is not None: if hparams.gpus > 1: raise NotImplementedError('''Dynamic Batch size does not work for multi-gpu training''' ) if hparams.sortish_sampler: raise ValueError('''--sortish_sampler and --max_tokens_per_batch may not be used simultaneously''' ) super().__init__(_SCREAMING_SNAKE_CASE , num_labels=_SCREAMING_SNAKE_CASE , mode=self.mode , **_SCREAMING_SNAKE_CASE ) use_task_specific_params(self.model , '''summarization''' ) save_git_info(self.hparams.output_dir ) lowerCAmelCase__ : List[Any] = Path(self.output_dir ) / '''metrics.json''' lowerCAmelCase__ : int = Path(self.output_dir ) / '''hparams.pkl''' pickle_save(self.hparams , self.hparams_save_path ) lowerCAmelCase__ : Tuple = 0 lowerCAmelCase__ : List[Any] = defaultdict(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Any = self.config.model_type lowerCAmelCase__ : Any = self.config.tgt_vocab_size if self.model_type == '''fsmt''' else self.config.vocab_size lowerCAmelCase__ : dict = { "data_dir": self.hparams.data_dir, "max_source_length": self.hparams.max_source_length, "prefix": self.model.config.prefix or "", } lowerCAmelCase__ : List[Any] = { '''train''': self.hparams.n_train, '''val''': self.hparams.n_val, '''test''': self.hparams.n_test, } lowerCAmelCase__ : Union[str, Any] = {k: v if v >= 0 else None for k, v in n_observations_per_split.items()} lowerCAmelCase__ : Tuple = { '''train''': self.hparams.max_target_length, '''val''': self.hparams.val_max_target_length, '''test''': self.hparams.test_max_target_length, } assert self.target_lens["train"] <= self.target_lens["val"], F'target_lens: {self.target_lens}' assert self.target_lens["train"] <= self.target_lens["test"], F'target_lens: {self.target_lens}' if self.hparams.freeze_embeds: freeze_embeds(self.model ) if self.hparams.freeze_encoder: freeze_params(self.model.get_encoder() ) assert_all_frozen(self.model.get_encoder() ) lowerCAmelCase__ : List[str] = get_git_info()['''repo_sha'''] lowerCAmelCase__ : int = hparams.num_workers lowerCAmelCase__ : Optional[int] = None # default to config if self.model.config.decoder_start_token_id is None and isinstance(self.tokenizer , _SCREAMING_SNAKE_CASE ): lowerCAmelCase__ : Optional[int] = self.tokenizer.lang_code_to_id[hparams.tgt_lang] lowerCAmelCase__ : List[Any] = self.decoder_start_token_id lowerCAmelCase__ : int = ( SeqaSeqDataset if hasattr(self.tokenizer , '''prepare_seq2seq_batch''' ) else LegacySeqaSeqDataset ) lowerCAmelCase__ : Union[str, Any] = False lowerCAmelCase__ : List[str] = self.model.config.num_beams if self.hparams.eval_beams is None else self.hparams.eval_beams if self.hparams.eval_max_gen_length is not None: lowerCAmelCase__ : int = self.hparams.eval_max_gen_length else: lowerCAmelCase__ : Tuple = self.model.config.max_length lowerCAmelCase__ : List[str] = self.default_val_metric if self.hparams.val_metric is None else self.hparams.val_metric def UpperCAmelCase__( self : Dict , _SCREAMING_SNAKE_CASE : Dict[str, torch.Tensor] )-> Dict[str, List[str]]: lowerCAmelCase__ : Any = { k: self.tokenizer.batch_decode(v.tolist() ) if '''mask''' not in k else v.shape for k, v in batch.items() } save_json(_SCREAMING_SNAKE_CASE , Path(self.output_dir ) / '''text_batch.json''' ) save_json({k: v.tolist() for k, v in batch.items()} , Path(self.output_dir ) / '''tok_batch.json''' ) lowerCAmelCase__ : Union[str, Any] = True return readable_batch def UpperCAmelCase__( self : Optional[Any] , _SCREAMING_SNAKE_CASE : Tuple , **_SCREAMING_SNAKE_CASE : Tuple )-> Tuple: return self.model(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Optional[int] , _SCREAMING_SNAKE_CASE : List[int] )-> Tuple: lowerCAmelCase__ : int = self.tokenizer.batch_decode( _SCREAMING_SNAKE_CASE , skip_special_tokens=_SCREAMING_SNAKE_CASE , clean_up_tokenization_spaces=_SCREAMING_SNAKE_CASE ) return lmap(str.strip , _SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Optional[Any] , _SCREAMING_SNAKE_CASE : dict )-> Tuple: lowerCAmelCase__ : Any = self.tokenizer.pad_token_id lowerCAmelCase__ , lowerCAmelCase__ : int = batch['''input_ids'''], batch['''attention_mask'''] lowerCAmelCase__ : Optional[Any] = batch['''labels'''] if isinstance(self.model , _SCREAMING_SNAKE_CASE ): lowerCAmelCase__ : List[str] = self.model._shift_right(_SCREAMING_SNAKE_CASE ) else: lowerCAmelCase__ : Any = shift_tokens_right(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if not self.already_saved_batch: # This would be slightly better if it only happened on rank zero lowerCAmelCase__ : Any = decoder_input_ids self.save_readable_batch(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Tuple = self(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE , decoder_input_ids=_SCREAMING_SNAKE_CASE , use_cache=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[str] = outputs['''logits'''] if self.hparams.label_smoothing == 0: # Same behavior as modeling_bart.py, besides ignoring pad_token_id lowerCAmelCase__ : Optional[int] = nn.CrossEntropyLoss(ignore_index=_SCREAMING_SNAKE_CASE ) assert lm_logits.shape[-1] == self.vocab_size lowerCAmelCase__ : Optional[Any] = ce_loss_fct(lm_logits.view(-1 , lm_logits.shape[-1] ) , tgt_ids.view(-1 ) ) else: lowerCAmelCase__ : Union[str, Any] = nn.functional.log_softmax(_SCREAMING_SNAKE_CASE , dim=-1 ) lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = label_smoothed_nll_loss( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , self.hparams.label_smoothing , ignore_index=_SCREAMING_SNAKE_CASE ) return (loss,) @property def UpperCAmelCase__( self : Optional[Any] )-> int: return self.tokenizer.pad_token_id def UpperCAmelCase__( self : Dict , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : List[str] )-> Dict: lowerCAmelCase__ : List[Any] = self._step(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : int = dict(zip(self.loss_names , _SCREAMING_SNAKE_CASE ) ) # tokens per batch lowerCAmelCase__ : Dict = batch['''input_ids'''].ne(self.pad ).sum() + batch['''labels'''].ne(self.pad ).sum() lowerCAmelCase__ : Optional[int] = batch['''input_ids'''].shape[0] lowerCAmelCase__ : List[Any] = batch['''input_ids'''].eq(self.pad ).sum() lowerCAmelCase__ : List[str] = batch['''input_ids'''].eq(self.pad ).float().mean() # TODO(SS): make a wandb summary metric for this return {"loss": loss_tensors[0], "log": logs} def UpperCAmelCase__( self : Tuple , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : Optional[int] )-> Dict: return self._generative_step(_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : List[Any] , _SCREAMING_SNAKE_CASE : Optional[Any] , _SCREAMING_SNAKE_CASE : Tuple="val" )-> Dict: self.step_count += 1 lowerCAmelCase__ : Any = {k: torch.stack([x[k] for x in outputs] ).mean() for k in self.loss_names} lowerCAmelCase__ : Union[str, Any] = losses['''loss'''] lowerCAmelCase__ : Union[str, Any] = { k: np.array([x[k] for x in outputs] ).mean() for k in self.metric_names + ['''gen_time''', '''gen_len'''] } lowerCAmelCase__ : Optional[Any] = ( generative_metrics[self.val_metric] if self.val_metric in generative_metrics else losses[self.val_metric] ) lowerCAmelCase__ : torch.FloatTensor = torch.tensor(_SCREAMING_SNAKE_CASE ).type_as(_SCREAMING_SNAKE_CASE ) generative_metrics.update({k: v.item() for k, v in losses.items()} ) losses.update(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Dict = {F'{prefix}_avg_{k}': x for k, x in losses.items()} lowerCAmelCase__ : Optional[Any] = self.step_count self.metrics[prefix].append(_SCREAMING_SNAKE_CASE ) # callback writes this to self.metrics_save_path lowerCAmelCase__ : Optional[int] = flatten_list([x['''preds'''] for x in outputs] ) return { "log": all_metrics, "preds": preds, F'{prefix}_loss': loss, F'{prefix}_{self.val_metric}': metric_tensor, } def UpperCAmelCase__( self : Union[str, Any] , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : List[Any] )-> Dict: return calculate_rouge(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Tuple , _SCREAMING_SNAKE_CASE : dict )-> dict: lowerCAmelCase__ : Any = time.time() # parser.add_argument('--eval_max_gen_length', type=int, default=None, help='never generate more than n tokens') lowerCAmelCase__ : str = self.model.generate( batch['''input_ids'''] , attention_mask=batch['''attention_mask'''] , use_cache=_SCREAMING_SNAKE_CASE , decoder_start_token_id=self.decoder_start_token_id , num_beams=self.eval_beams , max_length=self.eval_max_length , ) lowerCAmelCase__ : List[str] = (time.time() - ta) / batch['''input_ids'''].shape[0] lowerCAmelCase__ : List[str] = self.ids_to_clean_text(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[str] = self.ids_to_clean_text(batch['''labels'''] ) lowerCAmelCase__ : Any = self._step(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : int = dict(zip(self.loss_names , _SCREAMING_SNAKE_CASE ) ) lowerCAmelCase__ : Dict = self.calc_generative_metrics(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : str = np.mean(lmap(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) base_metrics.update(gen_time=_SCREAMING_SNAKE_CASE , gen_len=_SCREAMING_SNAKE_CASE , preds=_SCREAMING_SNAKE_CASE , target=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) return base_metrics def UpperCAmelCase__( self : Any , _SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : Optional[Any] )-> Optional[int]: return self._generative_step(_SCREAMING_SNAKE_CASE ) def UpperCAmelCase__( self : Tuple , _SCREAMING_SNAKE_CASE : int )-> str: return self.validation_epoch_end(_SCREAMING_SNAKE_CASE , prefix='''test''' ) def UpperCAmelCase__( self : Dict , _SCREAMING_SNAKE_CASE : int )-> SeqaSeqDataset: lowerCAmelCase__ : Dict = self.n_obs[type_path] lowerCAmelCase__ : Dict = self.target_lens[type_path] lowerCAmelCase__ : int = self.dataset_class( self.tokenizer , type_path=_SCREAMING_SNAKE_CASE , n_obs=_SCREAMING_SNAKE_CASE , max_target_length=_SCREAMING_SNAKE_CASE , **self.dataset_kwargs , ) return dataset def UpperCAmelCase__( self : List[str] , _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : bool = False )-> DataLoader: lowerCAmelCase__ : str = self.get_dataset(_SCREAMING_SNAKE_CASE ) if self.hparams.sortish_sampler and type_path != "test" and type_path != "val": lowerCAmelCase__ : Any = dataset.make_sortish_sampler(_SCREAMING_SNAKE_CASE , distributed=self.hparams.gpus > 1 ) return DataLoader( _SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE , collate_fn=dataset.collate_fn , shuffle=_SCREAMING_SNAKE_CASE , num_workers=self.num_workers , sampler=_SCREAMING_SNAKE_CASE , ) elif self.hparams.max_tokens_per_batch is not None and type_path != "test" and type_path != "val": lowerCAmelCase__ : Optional[int] = dataset.make_dynamic_sampler( self.hparams.max_tokens_per_batch , distributed=self.hparams.gpus > 1 ) return DataLoader( _SCREAMING_SNAKE_CASE , batch_sampler=_SCREAMING_SNAKE_CASE , collate_fn=dataset.collate_fn , num_workers=self.num_workers , ) else: return DataLoader( _SCREAMING_SNAKE_CASE , batch_size=_SCREAMING_SNAKE_CASE , collate_fn=dataset.collate_fn , shuffle=_SCREAMING_SNAKE_CASE , num_workers=self.num_workers , sampler=_SCREAMING_SNAKE_CASE , ) def UpperCAmelCase__( self : List[str] )-> DataLoader: lowerCAmelCase__ : Dict = self.get_dataloader('''train''' , batch_size=self.hparams.train_batch_size , shuffle=_SCREAMING_SNAKE_CASE ) return dataloader def UpperCAmelCase__( self : List[str] )-> DataLoader: return self.get_dataloader('''val''' , batch_size=self.hparams.eval_batch_size ) def UpperCAmelCase__( self : List[Any] )-> DataLoader: return self.get_dataloader('''test''' , batch_size=self.hparams.eval_batch_size ) @staticmethod def UpperCAmelCase__( _SCREAMING_SNAKE_CASE : List[str] , _SCREAMING_SNAKE_CASE : str )-> List[str]: BaseTransformer.add_model_specific_args(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) add_generic_args(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) parser.add_argument( '''--max_source_length''' , default=1024 , type=_SCREAMING_SNAKE_CASE , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--max_target_length''' , default=56 , type=_SCREAMING_SNAKE_CASE , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--val_max_target_length''' , default=142 , type=_SCREAMING_SNAKE_CASE , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--test_max_target_length''' , default=142 , type=_SCREAMING_SNAKE_CASE , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument('''--freeze_encoder''' , action='''store_true''' ) parser.add_argument('''--freeze_embeds''' , action='''store_true''' ) parser.add_argument('''--sortish_sampler''' , action='''store_true''' , default=_SCREAMING_SNAKE_CASE ) parser.add_argument('''--overwrite_output_dir''' , action='''store_true''' , default=_SCREAMING_SNAKE_CASE ) parser.add_argument('''--max_tokens_per_batch''' , type=_SCREAMING_SNAKE_CASE , default=_SCREAMING_SNAKE_CASE ) parser.add_argument('''--logger_name''' , type=_SCREAMING_SNAKE_CASE , choices=['''default''', '''wandb''', '''wandb_shared'''] , default='''default''' ) parser.add_argument('''--n_train''' , type=_SCREAMING_SNAKE_CASE , default=-1 , required=_SCREAMING_SNAKE_CASE , help='''# examples. -1 means use all.''' ) parser.add_argument('''--n_val''' , type=_SCREAMING_SNAKE_CASE , default=500 , required=_SCREAMING_SNAKE_CASE , help='''# examples. -1 means use all.''' ) parser.add_argument('''--n_test''' , type=_SCREAMING_SNAKE_CASE , default=-1 , required=_SCREAMING_SNAKE_CASE , help='''# examples. -1 means use all.''' ) parser.add_argument( '''--task''' , type=_SCREAMING_SNAKE_CASE , default='''summarization''' , required=_SCREAMING_SNAKE_CASE , help='''# examples. -1 means use all.''' ) parser.add_argument('''--label_smoothing''' , type=_SCREAMING_SNAKE_CASE , default=0.0 , required=_SCREAMING_SNAKE_CASE ) parser.add_argument('''--src_lang''' , type=_SCREAMING_SNAKE_CASE , default='''''' , required=_SCREAMING_SNAKE_CASE ) parser.add_argument('''--tgt_lang''' , type=_SCREAMING_SNAKE_CASE , default='''''' , required=_SCREAMING_SNAKE_CASE ) parser.add_argument('''--eval_beams''' , type=_SCREAMING_SNAKE_CASE , default=_SCREAMING_SNAKE_CASE , required=_SCREAMING_SNAKE_CASE ) parser.add_argument( '''--val_metric''' , type=_SCREAMING_SNAKE_CASE , default=_SCREAMING_SNAKE_CASE , required=_SCREAMING_SNAKE_CASE , choices=['''bleu''', '''rouge2''', '''loss''', None] ) parser.add_argument('''--eval_max_gen_length''' , type=_SCREAMING_SNAKE_CASE , default=_SCREAMING_SNAKE_CASE , help='''never generate more than n tokens''' ) parser.add_argument('''--save_top_k''' , type=_SCREAMING_SNAKE_CASE , default=1 , required=_SCREAMING_SNAKE_CASE , help='''How many checkpoints to save''' ) parser.add_argument( '''--early_stopping_patience''' , type=_SCREAMING_SNAKE_CASE , default=-1 , required=_SCREAMING_SNAKE_CASE , help=( '''-1 means never early stop. early_stopping_patience is measured in validation checks, not epochs. So''' ''' val_check_interval will effect it.''' ) , ) return parser class _a ( _lowercase): _a : Tuple = '''translation''' _a : Union[str, Any] = ['''loss'''] _a : int = ['''bleu'''] _a : List[str] = '''bleu''' def __init__( self : Optional[int] , _SCREAMING_SNAKE_CASE : int , **_SCREAMING_SNAKE_CASE : Any )-> Optional[int]: super().__init__(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Dict = hparams.src_lang lowerCAmelCase__ : str = hparams.tgt_lang def UpperCAmelCase__( self : int , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : Tuple )-> dict: return calculate_bleu(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def lowerCamelCase_ ( _a , _a=None ): """simple docstring""" Path(args.output_dir ).mkdir(exist_ok=_a ) check_output_dir(_a , expected_items=3 ) if model is None: if "summarization" in args.task: lowerCAmelCase__ : SummarizationModule = SummarizationModule(_a ) else: lowerCAmelCase__ : SummarizationModule = TranslationModule(_a ) lowerCAmelCase__ : Optional[int] = Path(args.data_dir ).name if ( args.logger_name == "default" or args.fast_dev_run or str(args.output_dir ).startswith('''/tmp''' ) or str(args.output_dir ).startswith('''/var''' ) ): lowerCAmelCase__ : Any = True # don't pollute wandb logs unnecessarily elif args.logger_name == "wandb": from pytorch_lightning.loggers import WandbLogger lowerCAmelCase__ : Optional[int] = os.environ.get('''WANDB_PROJECT''' , _a ) lowerCAmelCase__ : Union[str, Any] = WandbLogger(name=model.output_dir.name , project=_a ) elif args.logger_name == "wandb_shared": from pytorch_lightning.loggers import WandbLogger lowerCAmelCase__ : List[Any] = WandbLogger(name=model.output_dir.name , project=f'hf_{dataset}' ) if args.early_stopping_patience >= 0: lowerCAmelCase__ : Optional[Any] = get_early_stopping_callback(model.val_metric , args.early_stopping_patience ) else: lowerCAmelCase__ : int = False lowerCAmelCase__ : str = args.val_metric == '''loss''' lowerCAmelCase__ : pl.Trainer = generic_train( _a , _a , logging_callback=SeqaSeqLoggingCallback() , checkpoint_callback=get_checkpoint_callback( args.output_dir , model.val_metric , args.save_top_k , _a ) , early_stopping_callback=_a , logger=_a , ) pickle_save(model.hparams , model.output_dir / '''hparams.pkl''' ) if not args.do_predict: return model lowerCAmelCase__ : Any = '''''' lowerCAmelCase__ : Tuple = sorted(glob.glob(os.path.join(args.output_dir , '''*.ckpt''' ) , recursive=_a ) ) if checkpoints: lowerCAmelCase__ : List[str] = checkpoints[-1] lowerCAmelCase__ : Any = checkpoints[-1] trainer.logger.log_hyperparams(model.hparams ) # test() without a model tests using the best checkpoint automatically trainer.test() return model if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() lowerCamelCase = pl.Trainer.add_argparse_args(parser) lowerCamelCase = SummarizationModule.add_model_specific_args(parser, os.getcwd()) lowerCamelCase = parser.parse_args() main(args)
131
0
def A__ ( lowerCamelCase , lowerCamelCase ) -> int: return int(input_a == input_a == 0 ) def A__ ( ) -> None: print("""Truth Table of NOR Gate:""" ) print("""| Input 1 | Input 2 | Output |""" ) print(F'''| 0 | 0 | {nor_gate(0 , 0 )} |''' ) print(F'''| 0 | 1 | {nor_gate(0 , 1 )} |''' ) print(F'''| 1 | 0 | {nor_gate(1 , 0 )} |''' ) print(F'''| 1 | 1 | {nor_gate(1 , 1 )} |''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
223
import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline lowerCamelCase_ : Optional[Any] = datasets.utils.logging.get_logger(__name__) @dataclass class _UpperCamelCase ( datasets.BuilderConfig ): '''simple docstring''' __UpperCamelCase : Optional[datasets.Features] = None __UpperCamelCase : str = "utf-8" __UpperCamelCase : Optional[str] = None __UpperCamelCase : Optional[str] = None __UpperCamelCase : bool = True # deprecated __UpperCamelCase : Optional[int] = None # deprecated __UpperCamelCase : int = 10 << 20 # 10MB __UpperCamelCase : Optional[bool] = None class _UpperCamelCase ( datasets.ArrowBasedBuilder ): '''simple docstring''' __UpperCamelCase : Tuple = JsonConfig def lowerCAmelCase__ ( self : int ): if self.config.block_size is not None: logger.warning("""The JSON loader parameter `block_size` is deprecated. Please use `chunksize` instead""" ) UpperCamelCase_: List[str] = self.config.block_size if self.config.use_threads is not True: logger.warning( """The JSON loader parameter `use_threads` is deprecated and doesn't have any effect anymore.""" ) if self.config.newlines_in_values is not None: raise ValueError("""The JSON loader parameter `newlines_in_values` is no longer supported""" ) return datasets.DatasetInfo(features=self.config.features ) def lowerCAmelCase__ ( self : Dict , snake_case_ : str ): if not self.config.data_files: raise ValueError(f'''At least one data file must be specified, but got data_files={self.config.data_files}''' ) UpperCamelCase_: Dict = dl_manager.download_and_extract(self.config.data_files ) if isinstance(snake_case_ , (str, list, tuple) ): UpperCamelCase_: List[Any] = data_files if isinstance(snake_case_ , snake_case_ ): UpperCamelCase_: str = [files] UpperCamelCase_: Any = [dl_manager.iter_files(snake_case_ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""files""": files} )] UpperCamelCase_: Dict = [] for split_name, files in data_files.items(): if isinstance(snake_case_ , snake_case_ ): UpperCamelCase_: Tuple = [files] UpperCamelCase_: Optional[int] = [dl_manager.iter_files(snake_case_ ) for file in files] splits.append(datasets.SplitGenerator(name=snake_case_ , gen_kwargs={"""files""": files} ) ) return splits def lowerCAmelCase__ ( self : str , snake_case_ : pa.Table ): if self.config.features is not None: # adding missing columns for column_name in set(self.config.features ) - set(pa_table.column_names ): UpperCamelCase_: Union[str, Any] = self.config.features.arrow_schema.field(snake_case_ ).type UpperCamelCase_: Tuple = pa_table.append_column(snake_case_ , pa.array([None] * len(snake_case_ ) , type=snake_case_ ) ) # more expensive cast to support nested structures with keys in a different order # allows str <-> int/float or str to Audio for example UpperCamelCase_: int = table_cast(snake_case_ , self.config.features.arrow_schema ) return pa_table def lowerCAmelCase__ ( self : Dict , snake_case_ : Optional[Any] ): for file_idx, file in enumerate(itertools.chain.from_iterable(snake_case_ ) ): # If the file is one json object and if we need to look at the list of items in one specific field if self.config.field is not None: with open(snake_case_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: UpperCamelCase_: Dict = json.load(snake_case_ ) # We keep only the field we are interested in UpperCamelCase_: Optional[int] = dataset[self.config.field] # We accept two format: a list of dicts or a dict of lists if isinstance(snake_case_ , (list, tuple) ): UpperCamelCase_: Optional[int] = set().union(*[row.keys() for row in dataset] ) UpperCamelCase_: int = {col: [row.get(snake_case_ ) for row in dataset] for col in keys} else: UpperCamelCase_: Optional[int] = dataset UpperCamelCase_: List[str] = pa.Table.from_pydict(snake_case_ ) yield file_idx, self._cast_table(snake_case_ ) # If the file has one json object per line else: with open(snake_case_ , """rb""" ) as f: UpperCamelCase_: Optional[int] = 0 # Use block_size equal to the chunk size divided by 32 to leverage multithreading # Set a default minimum value of 16kB if the chunk size is really small UpperCamelCase_: Optional[int] = max(self.config.chunksize // 32 , 16 << 10 ) UpperCamelCase_: Tuple = ( self.config.encoding_errors if self.config.encoding_errors is not None else """strict""" ) while True: UpperCamelCase_: int = f.read(self.config.chunksize ) if not batch: break # Finish current line try: batch += f.readline() except (AttributeError, io.UnsupportedOperation): batch += readline(snake_case_ ) # PyArrow only accepts utf-8 encoded bytes if self.config.encoding != "utf-8": UpperCamelCase_: Tuple = batch.decode(self.config.encoding , errors=snake_case_ ).encode("""utf-8""" ) try: while True: try: UpperCamelCase_: Tuple = paj.read_json( io.BytesIO(snake_case_ ) , read_options=paj.ReadOptions(block_size=snake_case_ ) ) break except (pa.ArrowInvalid, pa.ArrowNotImplementedError) as e: if ( isinstance(snake_case_ , pa.ArrowInvalid ) and "straddling" not in str(snake_case_ ) or block_size > len(snake_case_ ) ): raise else: # Increase the block size in case it was too small. # The block size will be reset for the next file. logger.debug( f'''Batch of {len(snake_case_ )} bytes couldn\'t be parsed with block_size={block_size}. Retrying with block_size={block_size * 2}.''' ) block_size *= 2 except pa.ArrowInvalid as e: try: with open( snake_case_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: UpperCamelCase_: Optional[Any] = json.load(snake_case_ ) except json.JSONDecodeError: logger.error(f'''Failed to read file \'{file}\' with error {type(snake_case_ )}: {e}''' ) raise e # If possible, parse the file as a list of json objects and exit the loop if isinstance(snake_case_ , snake_case_ ): # list is the only sequence type supported in JSON try: UpperCamelCase_: Any = set().union(*[row.keys() for row in dataset] ) UpperCamelCase_: List[str] = {col: [row.get(snake_case_ ) for row in dataset] for col in keys} UpperCamelCase_: int = pa.Table.from_pydict(snake_case_ ) except (pa.ArrowInvalid, AttributeError) as e: logger.error(f'''Failed to read file \'{file}\' with error {type(snake_case_ )}: {e}''' ) raise ValueError(f'''Not able to read records in the JSON file at {file}.''' ) from None yield file_idx, self._cast_table(snake_case_ ) break else: logger.error(f'''Failed to read file \'{file}\' with error {type(snake_case_ )}: {e}''' ) raise ValueError( f'''Not able to read records in the JSON file at {file}. ''' f'''You should probably indicate the field of the JSON file containing your records. ''' f'''This JSON file contain the following fields: {str(list(dataset.keys() ) )}. ''' f'''Select the correct one and provide it as `field=\'XXX\'` to the dataset loading method. ''' ) from None # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(snake_case_ ) batch_idx += 1
223
1
"""simple docstring""" import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _a = logging.get_logger(__name__) _a = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt'} # See all LED models at https://huggingface.co/models?filter=LED _a = { 'vocab_file': { 'allenai/led-base-16384': 'https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json', }, 'merges_file': { 'allenai/led-base-16384': 'https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt', }, 'tokenizer_file': { 'allenai/led-base-16384': 'https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json', }, } _a = { 'allenai/led-base-16384': 1_63_84, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def _A ( ) -> Union[str, Any]: '''simple docstring''' __lowercase = ( list(range(ord("!"), ord("~") + 1)) + list(range(ord("¡"), ord("¬") + 1)) + list(range(ord("®"), ord("ÿ") + 1)) ) __lowercase = bs[:] __lowercase = 0 for b in range(2**8): if b not in bs: bs.append(UpperCamelCase_) cs.append(2**8 + n) n += 1 __lowercase = [chr(UpperCamelCase_) for n in cs] return dict(zip(UpperCamelCase_, UpperCamelCase_)) def _A ( UpperCamelCase_ : List[Any]) -> Any: '''simple docstring''' __lowercase = set() __lowercase = word[0] for char in word[1:]: pairs.add((prev_char, char)) __lowercase = char return pairs class _lowerCAmelCase ( lowercase ): """simple docstring""" __UpperCAmelCase : Any = VOCAB_FILES_NAMES __UpperCAmelCase : Tuple = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : List[Any] = ["input_ids", "attention_mask"] def __init__( self : Optional[Any], UpperCAmelCase__ : List[str], UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Union[str, Any]="replace", UpperCAmelCase__ : Tuple="<s>", UpperCAmelCase__ : Optional[Any]="</s>", UpperCAmelCase__ : Any="</s>", UpperCAmelCase__ : Optional[int]="<s>", UpperCAmelCase__ : Tuple="<unk>", UpperCAmelCase__ : Dict="<pad>", UpperCAmelCase__ : int="<mask>", UpperCAmelCase__ : Any=False, **UpperCAmelCase__ : List[Any], ): __lowercase = AddedToken(UpperCAmelCase__, lstrip=UpperCAmelCase__, rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__, UpperCAmelCase__ ) else bos_token __lowercase = AddedToken(UpperCAmelCase__, lstrip=UpperCAmelCase__, rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__, UpperCAmelCase__ ) else eos_token __lowercase = AddedToken(UpperCAmelCase__, lstrip=UpperCAmelCase__, rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__, UpperCAmelCase__ ) else sep_token __lowercase = AddedToken(UpperCAmelCase__, lstrip=UpperCAmelCase__, rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__, UpperCAmelCase__ ) else cls_token __lowercase = AddedToken(UpperCAmelCase__, lstrip=UpperCAmelCase__, rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__, UpperCAmelCase__ ) else unk_token __lowercase = AddedToken(UpperCAmelCase__, lstrip=UpperCAmelCase__, rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__, UpperCAmelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it __lowercase = AddedToken(UpperCAmelCase__, lstrip=UpperCAmelCase__, rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__, UpperCAmelCase__ ) else mask_token super().__init__( errors=UpperCAmelCase__, bos_token=UpperCAmelCase__, eos_token=UpperCAmelCase__, unk_token=UpperCAmelCase__, sep_token=UpperCAmelCase__, cls_token=UpperCAmelCase__, pad_token=UpperCAmelCase__, mask_token=UpperCAmelCase__, add_prefix_space=UpperCAmelCase__, **UpperCAmelCase__, ) with open(UpperCAmelCase__, encoding="utf-8" ) as vocab_handle: __lowercase = json.load(UpperCAmelCase__ ) __lowercase = {v: k for k, v in self.encoder.items()} __lowercase = errors # how to handle errors in decoding __lowercase = bytes_to_unicode() __lowercase = {v: k for k, v in self.byte_encoder.items()} with open(UpperCAmelCase__, encoding="utf-8" ) as merges_handle: __lowercase = merges_handle.read().split("\n" )[1:-1] __lowercase = [tuple(merge.split() ) for merge in bpe_merges] __lowercase = dict(zip(UpperCAmelCase__, range(len(UpperCAmelCase__ ) ) ) ) __lowercase = {} __lowercase = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions __lowercase = re.compile(r"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def _lowercase ( self : List[str] ): return len(self.encoder ) def _lowercase ( self : int ): return dict(self.encoder, **self.added_tokens_encoder ) def _lowercase ( self : str, UpperCAmelCase__ : str ): if token in self.cache: return self.cache[token] __lowercase = tuple(UpperCAmelCase__ ) __lowercase = get_pairs(UpperCAmelCase__ ) if not pairs: return token while True: __lowercase = min(UpperCAmelCase__, key=lambda UpperCAmelCase__ : self.bpe_ranks.get(UpperCAmelCase__, float("inf" ) ) ) if bigram not in self.bpe_ranks: break __lowercase ,__lowercase = bigram __lowercase = [] __lowercase = 0 while i < len(UpperCAmelCase__ ): try: __lowercase = word.index(UpperCAmelCase__, UpperCAmelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) __lowercase = j if word[i] == first and i < len(UpperCAmelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __lowercase = tuple(UpperCAmelCase__ ) __lowercase = new_word if len(UpperCAmelCase__ ) == 1: break else: __lowercase = get_pairs(UpperCAmelCase__ ) __lowercase = " ".join(UpperCAmelCase__ ) __lowercase = word return word def _lowercase ( self : Tuple, UpperCAmelCase__ : Tuple ): __lowercase = [] for token in re.findall(self.pat, UpperCAmelCase__ ): __lowercase = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(UpperCAmelCase__ ).split(" " ) ) return bpe_tokens def _lowercase ( self : Optional[int], UpperCAmelCase__ : int ): return self.encoder.get(UpperCAmelCase__, self.encoder.get(self.unk_token ) ) def _lowercase ( self : Optional[Any], UpperCAmelCase__ : str ): return self.decoder.get(UpperCAmelCase__ ) def _lowercase ( self : List[Any], UpperCAmelCase__ : Any ): __lowercase = "".join(UpperCAmelCase__ ) __lowercase = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8", errors=self.errors ) return text def _lowercase ( self : str, UpperCAmelCase__ : str, UpperCAmelCase__ : Optional[str] = None ): if not os.path.isdir(UpperCAmelCase__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowercase = os.path.join( UpperCAmelCase__, (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) __lowercase = os.path.join( UpperCAmelCase__, (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(UpperCAmelCase__, "w", encoding="utf-8" ) as f: f.write(json.dumps(self.encoder, indent=2, sort_keys=UpperCAmelCase__, ensure_ascii=UpperCAmelCase__ ) + "\n" ) __lowercase = 0 with open(UpperCAmelCase__, "w", encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items(), key=lambda UpperCAmelCase__ : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" " Please check that the tokenizer is not corrupted!" ) __lowercase = token_index writer.write(" ".join(UpperCAmelCase__ ) + "\n" ) index += 1 return vocab_file, merge_file def _lowercase ( self : Union[str, Any], UpperCAmelCase__ : List[int], UpperCAmelCase__ : Optional[List[int]] = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __lowercase = [self.cls_token_id] __lowercase = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _lowercase ( self : Any, UpperCAmelCase__ : List[int], UpperCAmelCase__ : Optional[List[int]] = None, UpperCAmelCase__ : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase__, token_ids_a=UpperCAmelCase__, already_has_special_tokens=UpperCAmelCase__ ) if token_ids_a is None: return [1] + ([0] * len(UpperCAmelCase__ )) + [1] return [1] + ([0] * len(UpperCAmelCase__ )) + [1, 1] + ([0] * len(UpperCAmelCase__ )) + [1] def _lowercase ( self : List[Any], UpperCAmelCase__ : List[int], UpperCAmelCase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _lowercase ( self : Dict, UpperCAmelCase__ : List[Any], UpperCAmelCase__ : str=False, **UpperCAmelCase__ : List[Any] ): __lowercase = kwargs.pop("add_prefix_space", self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(UpperCAmelCase__ ) > 0 and not text[0].isspace()): __lowercase = " " + text return (text, kwargs) def _lowercase ( self : Dict, UpperCAmelCase__ : Union[Dict[str, EncodedInput], BatchEncoding], UpperCAmelCase__ : Optional[int] = None, UpperCAmelCase__ : PaddingStrategy = PaddingStrategy.DO_NOT_PAD, UpperCAmelCase__ : Optional[int] = None, UpperCAmelCase__ : Optional[bool] = None, ): __lowercase = super()._pad( encoded_inputs=UpperCAmelCase__, max_length=UpperCAmelCase__, padding_strategy=UpperCAmelCase__, pad_to_multiple_of=UpperCAmelCase__, return_attention_mask=UpperCAmelCase__, ) # Load from model defaults if return_attention_mask is None: __lowercase = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: __lowercase = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. __lowercase = len(encoded_inputs["global_attention_mask"] ) != len(UpperCAmelCase__ ) if needs_to_be_padded: __lowercase = len(UpperCAmelCase__ ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` __lowercase = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": __lowercase = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
17
# DISCLAIMER: This code is strongly influenced by https://github.com/pesser/pytorch_diffusion # and https://github.com/hojonathanho/diffusion import math from dataclasses import dataclass from typing import List, Optional, Tuple, Union import numpy as np import torch from diffusers.configuration_utils import ConfigMixin, register_to_config from diffusers.schedulers.scheduling_utils import SchedulerMixin from diffusers.utils import BaseOutput, deprecate @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->DDIM class _lowercase ( lowerCAmelCase ): """simple docstring""" __A = 42 __A = None def a( A : Optional[Any] , A : Any=0.999 , A : Dict="cosine" , ) -> Optional[int]: """simple docstring""" if alpha_transform_type == "cosine": def alpha_bar_fn(A : Optional[int] ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(A : Any ): return math.exp(t * -12.0 ) else: raise ValueError(f'''Unsupported alpha_tranform_type: {alpha_transform_type}''' ) a = [] for i in range(A ): a = i / num_diffusion_timesteps a = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(A ) / alpha_bar_fn(A ) , A ) ) return torch.tensor(A , dtype=torch.floataa ) class _lowercase ( lowerCAmelCase, lowerCAmelCase ): """simple docstring""" __A = 1 @register_to_config def __init__(self , lowerCamelCase_ = 1000 , lowerCamelCase_ = 0.0001 , lowerCamelCase_ = 0.02 , lowerCamelCase_ = "linear" , lowerCamelCase_ = None , lowerCamelCase_ = True , lowerCamelCase_ = True , lowerCamelCase_ = 0 , lowerCamelCase_ = "epsilon" , lowerCamelCase_ = 1.0 , **lowerCamelCase_ , ): """simple docstring""" if kwargs.get("set_alpha_to_one" , lowerCamelCase_ ) is not None: a = ( "The `set_alpha_to_one` argument is deprecated. Please use `set_alpha_to_zero` instead." ) deprecate("set_alpha_to_one" , "1.0.0" , lowerCamelCase_ , standard_warn=lowerCamelCase_ ) a = kwargs["set_alpha_to_one"] if trained_betas is not None: a = torch.tensor(lowerCamelCase_ , dtype=torch.floataa ) elif beta_schedule == "linear": a = torch.linspace(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. a = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , lowerCamelCase_ , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule a = betas_for_alpha_bar(lowerCamelCase_ ) else: raise NotImplementedError(F'''{beta_schedule} does is not implemented for {self.__class__}''' ) a = 1.0 - self.betas a = torch.cumprod(self.alphas , dim=0 ) # At every step in inverted ddim, we are looking into the next alphas_cumprod # For the final step, there is no next alphas_cumprod, and the index is out of bounds # `set_alpha_to_zero` decides whether we set this parameter simply to zero # in this case, self.step() just output the predicted noise # or whether we use the final alpha of the "non-previous" one. a = torch.tensor(0.0 ) if set_alpha_to_zero else self.alphas_cumprod[-1] # standard deviation of the initial noise distribution a = 1.0 # setable values a = None a = torch.from_numpy(np.arange(0 , lowerCamelCase_ ).copy().astype(np.intaa ) ) def UpperCamelCase_ (self , lowerCamelCase_ , lowerCamelCase_ = None ): """simple docstring""" return sample def UpperCamelCase_ (self , lowerCamelCase_ , lowerCamelCase_ = None ): """simple docstring""" if num_inference_steps > self.config.num_train_timesteps: raise ValueError( F'''`num_inference_steps`: {num_inference_steps} cannot be larger than `self.config.train_timesteps`:''' F''' {self.config.num_train_timesteps} as the unet model trained with this scheduler can only handle''' F''' maximal {self.config.num_train_timesteps} timesteps.''' ) a = num_inference_steps a = self.config.num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 a = (np.arange(0 , lowerCamelCase_ ) * step_ratio).round().copy().astype(np.intaa ) a = torch.from_numpy(lowerCamelCase_ ).to(lowerCamelCase_ ) self.timesteps += self.config.steps_offset def UpperCamelCase_ (self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = 0.0 , lowerCamelCase_ = False , lowerCamelCase_ = None , lowerCamelCase_ = True , ): """simple docstring""" a = timestep + self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas # change original implementation to exactly match noise levels for analogous forward process a = self.alphas_cumprod[timestep] a = ( self.alphas_cumprod[prev_timestep] if prev_timestep < self.config.num_train_timesteps else self.final_alpha_cumprod ) a = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf if self.config.prediction_type == "epsilon": a = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 a = model_output elif self.config.prediction_type == "sample": a = model_output a = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 elif self.config.prediction_type == "v_prediction": a = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output a = (alpha_prod_t**0.5) * model_output + (beta_prod_t**0.5) * sample else: raise ValueError( F'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample`, or''' " `v_prediction`" ) # 4. Clip or threshold "predicted x_0" if self.config.clip_sample: a = pred_original_sample.clamp( -self.config.clip_sample_range , self.config.clip_sample_range ) # 5. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf a = (1 - alpha_prod_t_prev) ** 0.5 * pred_epsilon # 6. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf a = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if not return_dict: return (prev_sample, pred_original_sample) return DDIMSchedulerOutput(prev_sample=lowerCamelCase_ , pred_original_sample=lowerCamelCase_ ) def __len__(self ): """simple docstring""" return self.config.num_train_timesteps
227
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase_ : Union[str, Any] = { '''configuration_convbert''': ['''CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ConvBertConfig''', '''ConvBertOnnxConfig'''], '''tokenization_convbert''': ['''ConvBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Tuple = ['''ConvBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : str = [ '''CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ConvBertForMaskedLM''', '''ConvBertForMultipleChoice''', '''ConvBertForQuestionAnswering''', '''ConvBertForSequenceClassification''', '''ConvBertForTokenClassification''', '''ConvBertLayer''', '''ConvBertModel''', '''ConvBertPreTrainedModel''', '''load_tf_weights_in_convbert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : str = [ '''TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFConvBertForMaskedLM''', '''TFConvBertForMultipleChoice''', '''TFConvBertForQuestionAnswering''', '''TFConvBertForSequenceClassification''', '''TFConvBertForTokenClassification''', '''TFConvBertLayer''', '''TFConvBertModel''', '''TFConvBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_convbert import CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvBertConfig, ConvBertOnnxConfig from .tokenization_convbert import ConvBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_convbert_fast import ConvBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convbert import ( CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvBertForMaskedLM, ConvBertForMultipleChoice, ConvBertForQuestionAnswering, ConvBertForSequenceClassification, ConvBertForTokenClassification, ConvBertLayer, ConvBertModel, ConvBertPreTrainedModel, load_tf_weights_in_convbert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convbert import ( TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertLayer, TFConvBertModel, TFConvBertPreTrainedModel, ) else: import sys UpperCAmelCase_ : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
62
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DeformableDetrImageProcessor class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __init__( self : Union[str, Any] , __lowerCamelCase : int , __lowerCamelCase : Any=7 , __lowerCamelCase : Tuple=3 , __lowerCamelCase : Optional[Any]=30 , __lowerCamelCase : Union[str, Any]=400 , __lowerCamelCase : Optional[int]=True , __lowerCamelCase : Tuple=None , __lowerCamelCase : int=True , __lowerCamelCase : Dict=[0.5, 0.5, 0.5] , __lowerCamelCase : int=[0.5, 0.5, 0.5] , __lowerCamelCase : Optional[int]=True , __lowerCamelCase : str=1 / 255 , __lowerCamelCase : str=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p UpperCamelCase :List[Any] = size if size is not None else {"""shortest_edge""": 18, """longest_edge""": 1_333} UpperCamelCase :Tuple = parent UpperCamelCase :int = batch_size UpperCamelCase :str = num_channels UpperCamelCase :Dict = min_resolution UpperCamelCase :Any = max_resolution UpperCamelCase :int = do_resize UpperCamelCase :str = size UpperCamelCase :Dict = do_normalize UpperCamelCase :Tuple = image_mean UpperCamelCase :Optional[int] = image_std UpperCamelCase :Tuple = do_rescale UpperCamelCase :Optional[Any] = rescale_factor UpperCamelCase :List[Any] = do_pad def _A ( self : List[Any] ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def _A ( self : Dict , __lowerCamelCase : Any , __lowerCamelCase : Optional[int]=False ): if not batched: UpperCamelCase :Optional[Any] = image_inputs[0] if isinstance(__lowerCamelCase , Image.Image ): UpperCamelCase , UpperCamelCase :Union[str, Any] = image.size else: UpperCamelCase , UpperCamelCase :Optional[int] = image.shape[1], image.shape[2] if w < h: UpperCamelCase :int = int(self.size["""shortest_edge"""] * h / w ) UpperCamelCase :Tuple = self.size["""shortest_edge"""] elif w > h: UpperCamelCase :List[Any] = self.size["""shortest_edge"""] UpperCamelCase :str = int(self.size["""shortest_edge"""] * w / h ) else: UpperCamelCase :List[Any] = self.size["""shortest_edge"""] UpperCamelCase :str = self.size["""shortest_edge"""] else: UpperCamelCase :List[Any] = [] for image in image_inputs: UpperCamelCase , UpperCamelCase :int = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) UpperCamelCase :int = max(__lowerCamelCase , key=lambda __lowerCamelCase : item[0] )[0] UpperCamelCase :Tuple = max(__lowerCamelCase , key=lambda __lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _SCREAMING_SNAKE_CASE ( _a , unittest.TestCase ): snake_case__ : Optional[int] = DeformableDetrImageProcessor if is_vision_available() else None def _A ( self : Optional[Any] ): UpperCamelCase :str = DeformableDetrImageProcessingTester(self ) @property def _A ( self : Optional[Any] ): return self.image_processor_tester.prepare_image_processor_dict() def _A ( self : Dict ): UpperCamelCase :int = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowerCamelCase , """image_mean""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """image_std""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_normalize""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_resize""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_rescale""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """do_pad""" ) ) self.assertTrue(hasattr(__lowerCamelCase , """size""" ) ) def _A ( self : str ): UpperCamelCase :Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 18, """longest_edge""": 1_333} ) self.assertEqual(image_processor.do_pad , __lowerCamelCase ) UpperCamelCase :int = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=__lowerCamelCase ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42, """longest_edge""": 84} ) self.assertEqual(image_processor.do_pad , __lowerCamelCase ) def _A ( self : List[Any] ): pass def _A ( self : Dict ): # Initialize image_processing UpperCamelCase :List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCamelCase :List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowerCamelCase ) for image in image_inputs: self.assertIsInstance(__lowerCamelCase , Image.Image ) # Test not batched input UpperCamelCase :Dict = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCamelCase , UpperCamelCase :Optional[int] = self.image_processor_tester.get_expected_values(__lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase , UpperCamelCase :str = self.image_processor_tester.get_expected_values(__lowerCamelCase , batched=__lowerCamelCase ) UpperCamelCase :int = image_processing(__lowerCamelCase , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _A ( self : Tuple ): # Initialize image_processing UpperCamelCase :Any = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase :Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowerCamelCase , numpify=__lowerCamelCase ) for image in image_inputs: self.assertIsInstance(__lowerCamelCase , np.ndarray ) # Test not batched input UpperCamelCase :Union[str, Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCamelCase , UpperCamelCase :Any = self.image_processor_tester.get_expected_values(__lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase :Dict = image_processing(__lowerCamelCase , return_tensors="""pt""" ).pixel_values UpperCamelCase , UpperCamelCase :Optional[Any] = self.image_processor_tester.get_expected_values(__lowerCamelCase , batched=__lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _A ( self : Any ): # Initialize image_processing UpperCamelCase :Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase :List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowerCamelCase , torchify=__lowerCamelCase ) for image in image_inputs: self.assertIsInstance(__lowerCamelCase , torch.Tensor ) # Test not batched input UpperCamelCase :Tuple = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values UpperCamelCase , UpperCamelCase :List[str] = self.image_processor_tester.get_expected_values(__lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase :Union[str, Any] = image_processing(__lowerCamelCase , return_tensors="""pt""" ).pixel_values UpperCamelCase , UpperCamelCase :List[str] = self.image_processor_tester.get_expected_values(__lowerCamelCase , batched=__lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def _A ( self : Optional[Any] ): # prepare image and target UpperCamelCase :int = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_annotations.txt""" , """r""" ) as f: UpperCamelCase :str = json.loads(f.read() ) UpperCamelCase :List[Any] = {"""image_id""": 39_769, """annotations""": target} # encode them UpperCamelCase :Optional[int] = DeformableDetrImageProcessor() UpperCamelCase :Dict = image_processing(images=__lowerCamelCase , annotations=__lowerCamelCase , return_tensors="""pt""" ) # verify pixel values UpperCamelCase :Union[str, Any] = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding["""pixel_values"""].shape , __lowerCamelCase ) UpperCamelCase :Optional[Any] = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , __lowerCamelCase , atol=1E-4 ) ) # verify area UpperCamelCase :str = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , __lowerCamelCase ) ) # verify boxes UpperCamelCase :List[Any] = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , __lowerCamelCase ) UpperCamelCase :List[str] = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , __lowerCamelCase , atol=1E-3 ) ) # verify image_id UpperCamelCase :Tuple = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , __lowerCamelCase ) ) # verify is_crowd UpperCamelCase :List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , __lowerCamelCase ) ) # verify class_labels UpperCamelCase :Union[str, Any] = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , __lowerCamelCase ) ) # verify orig_size UpperCamelCase :Dict = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , __lowerCamelCase ) ) # verify size UpperCamelCase :int = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , __lowerCamelCase ) ) @slow def _A ( self : str ): # prepare image, target and masks_path UpperCamelCase :Any = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) with open("""./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt""" , """r""" ) as f: UpperCamelCase :Any = json.loads(f.read() ) UpperCamelCase :int = {"""file_name""": """000000039769.png""", """image_id""": 39_769, """segments_info""": target} UpperCamelCase :Any = pathlib.Path("""./tests/fixtures/tests_samples/COCO/coco_panoptic""" ) # encode them UpperCamelCase :Tuple = DeformableDetrImageProcessor(format="""coco_panoptic""" ) UpperCamelCase :Dict = image_processing(images=__lowerCamelCase , annotations=__lowerCamelCase , masks_path=__lowerCamelCase , return_tensors="""pt""" ) # verify pixel values UpperCamelCase :Optional[int] = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding["""pixel_values"""].shape , __lowerCamelCase ) UpperCamelCase :Optional[int] = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding["""pixel_values"""][0, 0, 0, :3] , __lowerCamelCase , atol=1E-4 ) ) # verify area UpperCamelCase :List[str] = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""area"""] , __lowerCamelCase ) ) # verify boxes UpperCamelCase :List[str] = torch.Size([6, 4] ) self.assertEqual(encoding["""labels"""][0]["""boxes"""].shape , __lowerCamelCase ) UpperCamelCase :List[Any] = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""boxes"""][0] , __lowerCamelCase , atol=1E-3 ) ) # verify image_id UpperCamelCase :str = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""image_id"""] , __lowerCamelCase ) ) # verify is_crowd UpperCamelCase :Tuple = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""iscrowd"""] , __lowerCamelCase ) ) # verify class_labels UpperCamelCase :List[Any] = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""class_labels"""] , __lowerCamelCase ) ) # verify masks UpperCamelCase :Union[str, Any] = 822_873 self.assertEqual(encoding["""labels"""][0]["""masks"""].sum().item() , __lowerCamelCase ) # verify orig_size UpperCamelCase :Tuple = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""orig_size"""] , __lowerCamelCase ) ) # verify size UpperCamelCase :str = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding["""labels"""][0]["""size"""] , __lowerCamelCase ) )
62
1
"""simple docstring""" import argparse import re from typing import Dict import torch from datasets import Audio, Dataset, load_dataset, load_metric from transformers import AutoFeatureExtractor, pipeline def __lowerCamelCase ( a_ : Tuple , a_ : Optional[Any] ) -> str: __SCREAMING_SNAKE_CASE :Union[str, Any] = args.log_outputs __SCREAMING_SNAKE_CASE :Optional[int] = '''_'''.join(args.dataset.split('''/''' ) + [args.config, args.split] ) # load metric __SCREAMING_SNAKE_CASE :List[Any] = load_metric('''wer''' ) __SCREAMING_SNAKE_CASE :Any = load_metric('''cer''' ) # compute metrics __SCREAMING_SNAKE_CASE :Union[str, Any] = wer.compute(references=result['''target'''] , predictions=result['''prediction'''] ) __SCREAMING_SNAKE_CASE :List[str] = cer.compute(references=result['''target'''] , predictions=result['''prediction'''] ) # print & log results __SCREAMING_SNAKE_CASE :Union[str, Any] = f'''WER: {wer_result}\nCER: {cer_result}''' print(_A ) with open(f'''{dataset_id}_eval_results.txt''' , '''w''' ) as f: f.write(_A ) # log all results in text file. Possibly interesting for analysis if log_outputs is not None: __SCREAMING_SNAKE_CASE :Union[str, Any] = f'''log_{dataset_id}_predictions.txt''' __SCREAMING_SNAKE_CASE :Optional[Any] = f'''log_{dataset_id}_targets.txt''' with open(_A , '''w''' ) as p, open(_A , '''w''' ) as t: # mapping function to write output def write_to_file(a_ : Any , a_ : Dict ): p.write(f'''{i}''' + '''\n''' ) p.write(batch['''prediction'''] + '''\n''' ) t.write(f'''{i}''' + '''\n''' ) t.write(batch['''target'''] + '''\n''' ) result.map(_A , with_indices=_A ) def __lowerCamelCase ( a_ : Any ) -> str: __SCREAMING_SNAKE_CASE :Optional[int] = '''[,?.!\-\;\:"“%‘”�—’…–]''' # noqa: W605 IMPORTANT: this should correspond to the chars that were ignored during training __SCREAMING_SNAKE_CASE :List[str] = re.sub(_A , '''''' , text.lower() ) # In addition, we can normalize the target text, e.g. removing new lines characters etc... # note that order is important here! __SCREAMING_SNAKE_CASE :List[Any] = ['''\n\n''', '''\n''', ''' ''', ''' '''] for t in token_sequences_to_ignore: __SCREAMING_SNAKE_CASE :Dict = ''' '''.join(text.split(_A ) ) return text def __lowerCamelCase ( a_ : Dict ) -> Optional[Any]: __SCREAMING_SNAKE_CASE :Dict = load_dataset(args.dataset , args.config , split=args.split , use_auth_token=_A ) # for testing: only process the first two examples as a test # dataset = dataset.select(range(10)) # load processor __SCREAMING_SNAKE_CASE :str = AutoFeatureExtractor.from_pretrained(args.model_id ) __SCREAMING_SNAKE_CASE :Tuple = feature_extractor.sampling_rate # resample audio __SCREAMING_SNAKE_CASE :str = dataset.cast_column('''audio''' , Audio(sampling_rate=_A ) ) # load eval pipeline if args.device is None: __SCREAMING_SNAKE_CASE :Any = 0 if torch.cuda.is_available() else -1 __SCREAMING_SNAKE_CASE :str = pipeline('''automatic-speech-recognition''' , model=args.model_id , device=args.device ) # map function to decode audio def map_to_pred(a_ : str ): __SCREAMING_SNAKE_CASE :Union[str, Any] = asr( batch['''audio''']['''array'''] , chunk_length_s=args.chunk_length_s , stride_length_s=args.stride_length_s ) __SCREAMING_SNAKE_CASE :Tuple = prediction['''text'''] __SCREAMING_SNAKE_CASE :Tuple = normalize_text(batch['''sentence'''] ) return batch # run inference on all examples __SCREAMING_SNAKE_CASE :int = dataset.map(_A , remove_columns=dataset.column_names ) # compute and log_results # do not change function below log_results(_A , _A ) if __name__ == "__main__": lowerCamelCase_ = argparse.ArgumentParser() parser.add_argument( "--model_id", type=str, required=True, help="Model identifier. Should be loadable with 🤗 Transformers" ) parser.add_argument( "--dataset", type=str, required=True, help="Dataset name to evaluate the `model_id`. Should be loadable with 🤗 Datasets", ) parser.add_argument( "--config", type=str, required=True, help="Config of the dataset. *E.g.* `'en'` for Common Voice" ) parser.add_argument("--split", type=str, required=True, help="Split of the dataset. *E.g.* `'test'`") parser.add_argument( "--chunk_length_s", type=float, default=None, help="Chunk length in seconds. Defaults to 5 seconds." ) parser.add_argument( "--stride_length_s", type=float, default=None, help="Stride of the audio chunks. Defaults to 1 second." ) parser.add_argument( "--log_outputs", action="store_true", help="If defined, write outputs to log file for analysis." ) parser.add_argument( "--device", type=int, default=None, help="The device to run the pipeline on. -1 for CPU (default), 0 for the first GPU and so on.", ) lowerCamelCase_ = parser.parse_args() main(args)
191
import os import string import sys __UpperCamelCase : List[Any] = 1 << 8 __UpperCamelCase : Union[str, Any] = { """tab""": ord("""\t"""), """newline""": ord("""\r"""), """esc""": 27, """up""": 65 + ARROW_KEY_FLAG, """down""": 66 + ARROW_KEY_FLAG, """right""": 67 + ARROW_KEY_FLAG, """left""": 68 + ARROW_KEY_FLAG, """mod_int""": 91, """undefined""": sys.maxsize, """interrupt""": 3, """insert""": 50, """delete""": 51, """pg_up""": 53, """pg_down""": 54, } __UpperCamelCase : Optional[Any] = KEYMAP["""up"""] __UpperCamelCase : Tuple = KEYMAP["""left"""] if sys.platform == "win32": __UpperCamelCase : List[Any] = [] __UpperCamelCase : int = { b"""\xe0H""": KEYMAP["""up"""] - ARROW_KEY_FLAG, b"""\x00H""": KEYMAP["""up"""] - ARROW_KEY_FLAG, b"""\xe0P""": KEYMAP["""down"""] - ARROW_KEY_FLAG, b"""\x00P""": KEYMAP["""down"""] - ARROW_KEY_FLAG, b"""\xe0M""": KEYMAP["""right"""] - ARROW_KEY_FLAG, b"""\x00M""": KEYMAP["""right"""] - ARROW_KEY_FLAG, b"""\xe0K""": KEYMAP["""left"""] - ARROW_KEY_FLAG, b"""\x00K""": KEYMAP["""left"""] - ARROW_KEY_FLAG, } for i in range(10): __UpperCamelCase : List[str] = ord(str(i)) def a_ ( ) -> Optional[int]: """simple docstring""" if os.name == "nt": import msvcrt snake_case__ = 'mbcs' # Flush the keyboard buffer while msvcrt.kbhit(): msvcrt.getch() if len(_A ) == 0: # Read the keystroke snake_case__ = msvcrt.getch() # If it is a prefix char, get second part if ch in (b"\x00", b"\xe0"): snake_case__ = ch + msvcrt.getch() # Translate actual Win chars to bullet char types try: snake_case__ = chr(WIN_KEYMAP[cha] ) WIN_CH_BUFFER.append(chr(KEYMAP['mod_int'] ) ) WIN_CH_BUFFER.append(_A ) if ord(_A ) in ( KEYMAP["insert"] - 1 << 9, KEYMAP["delete"] - 1 << 9, KEYMAP["pg_up"] - 1 << 9, KEYMAP["pg_down"] - 1 << 9, ): WIN_CH_BUFFER.append(chr(126 ) ) snake_case__ = chr(KEYMAP['esc'] ) except KeyError: snake_case__ = cha[1] else: snake_case__ = ch.decode(_A ) else: snake_case__ = WIN_CH_BUFFER.pop(0 ) elif os.name == "posix": import termios import tty snake_case__ = sys.stdin.fileno() snake_case__ = termios.tcgetattr(_A ) try: tty.setraw(_A ) snake_case__ = sys.stdin.read(1 ) finally: termios.tcsetattr(_A , termios.TCSADRAIN , _A ) return ch def a_ ( ) -> Union[str, Any]: """simple docstring""" snake_case__ = get_raw_chars() if ord(_A ) in [KEYMAP["interrupt"], KEYMAP["newline"]]: return char elif ord(_A ) == KEYMAP["esc"]: snake_case__ = get_raw_chars() if ord(_A ) == KEYMAP["mod_int"]: snake_case__ = get_raw_chars() if ord(_A ) >= KEYMAP["arrow_begin"] - ARROW_KEY_FLAG and ord(_A ) <= KEYMAP["arrow_end"] - ARROW_KEY_FLAG: return chr(ord(_A ) + ARROW_KEY_FLAG ) else: return KEYMAP["undefined"] else: return get_raw_chars() else: if char in string.printable: return char else: return KEYMAP["undefined"]
307
0
'''simple docstring''' import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } _lowerCAmelCase = { '''vocab_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json'''}, '''merges_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt'''}, '''tokenizer_config_file''': { '''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json''' }, } _lowerCAmelCase = {'''facebook/blenderbot-3B''': 128} @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def _SCREAMING_SNAKE_CASE ( ): """simple docstring""" lowerCAmelCase__ : Optional[Any] = ( list(range(ord("""!""" ) , ord("""~""" ) + 1 ) ) + list(range(ord("""¡""" ) , ord("""¬""" ) + 1 ) ) + list(range(ord("""®""" ) , ord("""ÿ""" ) + 1 ) ) ) lowerCAmelCase__ : Any = bs[:] lowerCAmelCase__ : Dict = 0 for b in range(2**8 ): if b not in bs: bs.append(UpperCamelCase ) cs.append(2**8 + n ) n += 1 lowerCAmelCase__ : Tuple = [chr(UpperCamelCase ) for n in cs] return dict(zip(UpperCamelCase , UpperCamelCase ) ) def _SCREAMING_SNAKE_CASE ( UpperCamelCase ): """simple docstring""" lowerCAmelCase__ : Any = set() lowerCAmelCase__ : str = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowerCAmelCase__ : Tuple = char return pairs class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' __lowercase : Any = VOCAB_FILES_NAMES __lowercase : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP __lowercase : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowercase : List[str] = ['''input_ids''', '''attention_mask'''] def __init__( self ,__UpperCAmelCase ,__UpperCAmelCase ,__UpperCAmelCase="replace" ,__UpperCAmelCase="<s>" ,__UpperCAmelCase="</s>" ,__UpperCAmelCase="</s>" ,__UpperCAmelCase="<s>" ,__UpperCAmelCase="<unk>" ,__UpperCAmelCase="<pad>" ,__UpperCAmelCase="<mask>" ,__UpperCAmelCase=False ,**__UpperCAmelCase ,) -> Union[str, Any]: lowerCAmelCase__ : Any = AddedToken(__UpperCAmelCase ,lstrip=__UpperCAmelCase ,rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase ,__UpperCAmelCase ) else bos_token lowerCAmelCase__ : List[str] = AddedToken(__UpperCAmelCase ,lstrip=__UpperCAmelCase ,rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase ,__UpperCAmelCase ) else eos_token lowerCAmelCase__ : int = AddedToken(__UpperCAmelCase ,lstrip=__UpperCAmelCase ,rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase ,__UpperCAmelCase ) else sep_token lowerCAmelCase__ : Tuple = AddedToken(__UpperCAmelCase ,lstrip=__UpperCAmelCase ,rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase ,__UpperCAmelCase ) else cls_token lowerCAmelCase__ : int = AddedToken(__UpperCAmelCase ,lstrip=__UpperCAmelCase ,rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase ,__UpperCAmelCase ) else unk_token lowerCAmelCase__ : Dict = AddedToken(__UpperCAmelCase ,lstrip=__UpperCAmelCase ,rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase ,__UpperCAmelCase ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase__ : str = AddedToken(__UpperCAmelCase ,lstrip=__UpperCAmelCase ,rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase ,__UpperCAmelCase ) else mask_token super().__init__( errors=__UpperCAmelCase ,bos_token=__UpperCAmelCase ,eos_token=__UpperCAmelCase ,unk_token=__UpperCAmelCase ,sep_token=__UpperCAmelCase ,cls_token=__UpperCAmelCase ,pad_token=__UpperCAmelCase ,mask_token=__UpperCAmelCase ,add_prefix_space=__UpperCAmelCase ,**__UpperCAmelCase ,) with open(__UpperCAmelCase ,encoding="""utf-8""" ) as vocab_handle: lowerCAmelCase__ : List[str] = json.load(__UpperCAmelCase ) lowerCAmelCase__ : Optional[int] = {v: k for k, v in self.encoder.items()} lowerCAmelCase__ : List[Any] = errors # how to handle errors in decoding lowerCAmelCase__ : Tuple = bytes_to_unicode() lowerCAmelCase__ : Optional[int] = {v: k for k, v in self.byte_encoder.items()} with open(__UpperCAmelCase ,encoding="""utf-8""" ) as merges_handle: lowerCAmelCase__ : int = merges_handle.read().split("""\n""" )[1:-1] lowerCAmelCase__ : Optional[int] = [tuple(merge.split() ) for merge in bpe_merges] lowerCAmelCase__ : Optional[int] = dict(zip(__UpperCAmelCase ,range(len(__UpperCAmelCase ) ) ) ) lowerCAmelCase__ : List[Any] = {} lowerCAmelCase__ : Optional[int] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowerCAmelCase__ : Optional[Any] = re.compile(R"""'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+""" ) @property # Copied from transformers.models.roberta.tokenization_roberta.RobertaTokenizer.vocab_size with Roberta->Blenderbot, RoBERTa->Blenderbot def UpperCAmelCase_ ( self ) -> str: return len(self.encoder ) def UpperCAmelCase_ ( self ) -> Tuple: return dict(self.encoder ,**self.added_tokens_encoder ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> Optional[Any]: if token in self.cache: return self.cache[token] lowerCAmelCase__ : int = tuple(__UpperCAmelCase ) lowerCAmelCase__ : Any = get_pairs(__UpperCAmelCase ) if not pairs: return token while True: lowerCAmelCase__ : int = min(__UpperCAmelCase ,key=lambda __UpperCAmelCase : self.bpe_ranks.get(__UpperCAmelCase ,float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = bigram lowerCAmelCase__ : Optional[Any] = [] lowerCAmelCase__ : Optional[int] = 0 while i < len(__UpperCAmelCase ): try: lowerCAmelCase__ : List[str] = word.index(__UpperCAmelCase ,__UpperCAmelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCAmelCase__ : Optional[int] = j if word[i] == first and i < len(__UpperCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCAmelCase__ : List[Any] = tuple(__UpperCAmelCase ) lowerCAmelCase__ : int = new_word if len(__UpperCAmelCase ) == 1: break else: lowerCAmelCase__ : Union[str, Any] = get_pairs(__UpperCAmelCase ) lowerCAmelCase__ : Any = """ """.join(__UpperCAmelCase ) lowerCAmelCase__ : int = word return word def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> Any: lowerCAmelCase__ : str = [] for token in re.findall(self.pat ,__UpperCAmelCase ): lowerCAmelCase__ : List[str] = """""".join( self.byte_encoder[b] for b in token.encode("""utf-8""" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(__UpperCAmelCase ).split(""" """ ) ) return bpe_tokens def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> int: return self.encoder.get(__UpperCAmelCase ,self.encoder.get(self.unk_token ) ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> List[Any]: return self.decoder.get(__UpperCAmelCase ) def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> Optional[int]: lowerCAmelCase__ : Dict = """""".join(__UpperCAmelCase ) lowerCAmelCase__ : Optional[int] = bytearray([self.byte_decoder[c] for c in text] ).decode("""utf-8""" ,errors=self.errors ) return text def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase = None ) -> Tuple[str]: if not os.path.isdir(__UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return lowerCAmelCase__ : Union[str, Any] = os.path.join( __UpperCAmelCase ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) lowerCAmelCase__ : str = os.path.join( __UpperCAmelCase ,(filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(__UpperCAmelCase ,"""w""" ,encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder ,indent=2 ,sort_keys=__UpperCAmelCase ,ensure_ascii=__UpperCAmelCase ) + """\n""" ) lowerCAmelCase__ : str = 0 with open(__UpperCAmelCase ,"""w""" ,encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() ,key=lambda __UpperCAmelCase : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" """ Please check that the tokenizer is not corrupted!""" ) lowerCAmelCase__ : Dict = token_index writer.write(""" """.join(__UpperCAmelCase ) + """\n""" ) index += 1 return vocab_file, merge_file def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase = None ,__UpperCAmelCase = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase ,token_ids_a=__UpperCAmelCase ,already_has_special_tokens=__UpperCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(__UpperCAmelCase )) + [1] return [1] + ([0] * len(__UpperCAmelCase )) + [1, 1] + ([0] * len(__UpperCAmelCase )) + [1] def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase = None ) -> List[int]: lowerCAmelCase__ : int = [self.sep_token_id] lowerCAmelCase__ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase=False ,**__UpperCAmelCase ) -> Union[str, Any]: lowerCAmelCase__ : int = kwargs.pop("""add_prefix_space""" ,self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(__UpperCAmelCase ) > 0 and not text[0].isspace()): lowerCAmelCase__ : Optional[Any] = """ """ + text return (text, kwargs) def UpperCAmelCase_ ( self ,__UpperCAmelCase ,__UpperCAmelCase = None ) -> int: return token_ids_a + [self.eos_token_id] def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> List[int]: lowerCAmelCase__ : Optional[int] = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(""" """ + text ) else: # Generated responses should contain them already. inputs.append(__UpperCAmelCase ) lowerCAmelCase__ : str = """ """.join(__UpperCAmelCase ) lowerCAmelCase__ : Optional[Any] = self.encode(__UpperCAmelCase ) if len(__UpperCAmelCase ) > self.model_max_length: lowerCAmelCase__ : Union[str, Any] = input_ids[-self.model_max_length :] logger.warning(F"""Trimmed input from conversation as it was longer than {self.model_max_length} tokens.""" ) return input_ids
184
'''simple docstring''' import unittest import numpy as np import requests from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: _lowerCAmelCase = False if is_vision_available(): from PIL import Image from transformers import PixaStructImageProcessor class lowerCAmelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self ,__UpperCAmelCase ,__UpperCAmelCase=7 ,__UpperCAmelCase=3 ,__UpperCAmelCase=18 ,__UpperCAmelCase=30 ,__UpperCAmelCase=400 ,__UpperCAmelCase=None ,__UpperCAmelCase=True ,__UpperCAmelCase=True ,__UpperCAmelCase=None ,) -> Optional[Any]: lowerCAmelCase__ : Optional[Any] = size if size is not None else {"""height""": 20, """width""": 20} lowerCAmelCase__ : List[Any] = parent lowerCAmelCase__ : List[str] = batch_size lowerCAmelCase__ : Tuple = num_channels lowerCAmelCase__ : Optional[int] = image_size lowerCAmelCase__ : Optional[Any] = min_resolution lowerCAmelCase__ : Tuple = max_resolution lowerCAmelCase__ : List[Any] = size lowerCAmelCase__ : List[str] = do_normalize lowerCAmelCase__ : Optional[int] = do_convert_rgb lowerCAmelCase__ : str = [512, 1024, 2048, 4096] lowerCAmelCase__ : int = patch_size if patch_size is not None else {"""height""": 16, """width""": 16} def UpperCAmelCase_ ( self ) -> Optional[int]: return {"do_normalize": self.do_normalize, "do_convert_rgb": self.do_convert_rgb} def UpperCAmelCase_ ( self ) -> Dict: lowerCAmelCase__ : Optional[Any] = """https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers/tasks/australia.jpg""" lowerCAmelCase__ : Union[str, Any] = Image.open(requests.get(__UpperCAmelCase ,stream=__UpperCAmelCase ).raw ).convert("""RGB""" ) return raw_image @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason='''`Pix2StructImageProcessor` requires `torch>=1.11.0`.''' , ) @require_torch @require_vision class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): '''simple docstring''' __lowercase : Union[str, Any] = PixaStructImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ) -> List[str]: lowerCAmelCase__ : Union[str, Any] = PixaStructImageProcessingTester(self ) @property def UpperCAmelCase_ ( self ) -> Dict: return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ) -> Optional[Any]: lowerCAmelCase__ : Dict = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__UpperCAmelCase ,"""do_normalize""" ) ) self.assertTrue(hasattr(__UpperCAmelCase ,"""do_convert_rgb""" ) ) def UpperCAmelCase_ ( self ) -> Optional[int]: lowerCAmelCase__ : List[Any] = self.image_processor_tester.prepare_dummy_image() lowerCAmelCase__ : List[Any] = self.image_processing_class(**self.image_processor_dict ) lowerCAmelCase__ : str = 2048 lowerCAmelCase__ : Tuple = image_processor(__UpperCAmelCase ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ) self.assertTrue(torch.allclose(inputs.flattened_patches.mean() ,torch.tensor(0.0_6_0_6 ) ,atol=1E-3 ,rtol=1E-3 ) ) def UpperCAmelCase_ ( self ) -> str: # Initialize image_processor lowerCAmelCase__ : int = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase__ : int = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase ,Image.Image ) # Test not batched input lowerCAmelCase__ : List[str] = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowerCAmelCase__ : str = image_processor( image_inputs[0] ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(1, max_patch, expected_hidden_dim) ,) # Test batched lowerCAmelCase__ : Any = image_processor( __UpperCAmelCase ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) ,) def UpperCAmelCase_ ( self ) -> Any: # Initialize image_processor lowerCAmelCase__ : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase__ : Optional[int] = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase ,Image.Image ) # Test not batched input lowerCAmelCase__ : Tuple = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * self.image_processor_tester.num_channels ) + 2 lowerCAmelCase__ : Optional[int] = True for max_patch in self.image_processor_tester.max_patches: # Test not batched input with self.assertRaises(__UpperCAmelCase ): lowerCAmelCase__ : Any = image_processor( image_inputs[0] ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches lowerCAmelCase__ : Optional[Any] = """Hello""" lowerCAmelCase__ : List[str] = image_processor( image_inputs[0] ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ,header_text=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(1, max_patch, expected_hidden_dim) ,) # Test batched lowerCAmelCase__ : str = image_processor( __UpperCAmelCase ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ,header_text=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) ,) def UpperCAmelCase_ ( self ) -> Tuple: # Initialize image_processor lowerCAmelCase__ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase__ : List[str] = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__UpperCAmelCase ,numpify=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase ,np.ndarray ) lowerCAmelCase__ : Tuple = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowerCAmelCase__ : Any = image_processor( image_inputs[0] ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(1, max_patch, expected_hidden_dim) ,) # Test batched lowerCAmelCase__ : int = image_processor( __UpperCAmelCase ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) ,) def UpperCAmelCase_ ( self ) -> Optional[int]: # Initialize image_processor lowerCAmelCase__ : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase__ : int = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__UpperCAmelCase ,torchify=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase ,torch.Tensor ) # Test not batched input lowerCAmelCase__ : Optional[int] = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * self.image_processor_tester.num_channels ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowerCAmelCase__ : Dict = image_processor( image_inputs[0] ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(1, max_patch, expected_hidden_dim) ,) # Test batched lowerCAmelCase__ : Union[str, Any] = image_processor( __UpperCAmelCase ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) ,) @unittest.skipIf( not is_torch_greater_or_equal_than_1_11 , reason='''`Pix2StructImageProcessor` requires `torch>=1.11.0`.''' , ) @require_torch @require_vision class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ , unittest.TestCase ): '''simple docstring''' __lowercase : Optional[int] = PixaStructImageProcessor if is_vision_available() else None def UpperCAmelCase_ ( self ) -> List[Any]: lowerCAmelCase__ : Dict = PixaStructImageProcessingTester(self ,num_channels=4 ) lowerCAmelCase__ : str = 3 @property def UpperCAmelCase_ ( self ) -> Tuple: return self.image_processor_tester.prepare_image_processor_dict() def UpperCAmelCase_ ( self ) -> Dict: lowerCAmelCase__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__UpperCAmelCase ,"""do_normalize""" ) ) self.assertTrue(hasattr(__UpperCAmelCase ,"""do_convert_rgb""" ) ) def UpperCAmelCase_ ( self ) -> Optional[Any]: # Initialize image_processor lowerCAmelCase__ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase__ : Dict = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase ,Image.Image ) # Test not batched input lowerCAmelCase__ : Dict = ( (self.image_processor_tester.patch_size["""height"""] * self.image_processor_tester.patch_size["""width"""]) * (self.image_processor_tester.num_channels - 1) ) + 2 for max_patch in self.image_processor_tester.max_patches: # Test not batched input lowerCAmelCase__ : int = image_processor( image_inputs[0] ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(1, max_patch, expected_hidden_dim) ,) # Test batched lowerCAmelCase__ : Dict = image_processor( __UpperCAmelCase ,return_tensors="""pt""" ,max_patches=__UpperCAmelCase ).flattened_patches self.assertEqual( encoded_images.shape ,(self.image_processor_tester.batch_size, max_patch, expected_hidden_dim) ,)
184
1
from __future__ import annotations from dataclasses import dataclass @dataclass class __SCREAMING_SNAKE_CASE : A : float A : TreeNode | None = None A : TreeNode | None = None def __lowercase ( _UpperCamelCase ) ->bool: """simple docstring""" def is_valid_tree(_UpperCamelCase ) -> bool: if node is None: return True if not isinstance(_UpperCamelCase, _UpperCamelCase ): return False try: float(node.data ) except (TypeError, ValueError): return False return is_valid_tree(node.left ) and is_valid_tree(node.right ) if not is_valid_tree(_UpperCamelCase ): raise ValueError( '''Each node should be type of TreeNode and data should be float.''' ) def is_binary_search_tree_recursive_check( _UpperCamelCase, _UpperCamelCase, _UpperCamelCase ) -> bool: if node is None: return True return ( left_bound < node.data < right_bound and is_binary_search_tree_recursive_check(node.left, _UpperCamelCase, node.data ) and is_binary_search_tree_recursive_check( node.right, node.data, _UpperCamelCase ) ) return is_binary_search_tree_recursive_check(_UpperCamelCase, -float('''inf''' ), float('''inf''' ) ) if __name__ == "__main__": import doctest doctest.testmod()
337
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import torch from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available @dataclass class __SCREAMING_SNAKE_CASE ( A__ ): A : Union[List[np.ndarray], torch.FloatTensor] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_text_to_video_synth import TextToVideoSDPipeline from .pipeline_text_to_video_synth_imgaimg import VideoToVideoSDPipeline # noqa: F401 from .pipeline_text_to_video_zero import TextToVideoZeroPipeline
337
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase_ = { 'configuration_deberta': ['DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'DebertaConfig', 'DebertaOnnxConfig'], 'tokenization_deberta': ['DebertaTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ['DebertaTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ 'DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'DebertaForMaskedLM', 'DebertaForQuestionAnswering', 'DebertaForSequenceClassification', 'DebertaForTokenClassification', 'DebertaModel', 'DebertaPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ 'TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFDebertaForMaskedLM', 'TFDebertaForQuestionAnswering', 'TFDebertaForSequenceClassification', 'TFDebertaForTokenClassification', 'TFDebertaModel', 'TFDebertaPreTrainedModel', ] if TYPE_CHECKING: from .configuration_deberta import DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, DebertaConfig, DebertaOnnxConfig from .tokenization_deberta import DebertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_deberta_fast import DebertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deberta import ( DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, DebertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deberta import ( TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFDebertaForMaskedLM, TFDebertaForQuestionAnswering, TFDebertaForSequenceClassification, TFDebertaForTokenClassification, TFDebertaModel, TFDebertaPreTrainedModel, ) else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
116
lowerCAmelCase_ = range(2, 20 + 1) lowerCAmelCase_ = [10**k for k in range(ks[-1] + 1)] lowerCAmelCase_ = {} def snake_case( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Any: '''simple docstring''' lowercase : str = sum(a_i[j] for j in range(__magic_name__ , len(__magic_name__ ) ) ) lowercase : Any = sum(a_i[j] * base[j] for j in range(min(len(__magic_name__ ) , __magic_name__ ) ) ) lowercase , lowercase : Optional[int] = 0, 0 lowercase : str = n - i lowercase : Optional[int] = memo.get(__magic_name__ ) if sub_memo is not None: lowercase : List[str] = sub_memo.get(__magic_name__ ) if jumps is not None and len(__magic_name__ ) > 0: # find and make the largest jump without going over lowercase : Dict = -1 for _k in range(len(__magic_name__ ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: lowercase : Any = _k break if max_jump >= 0: lowercase , lowercase , lowercase : List[str] = jumps[max_jump] # since the difference between jumps is cached, add c lowercase : str = diff + c for j in range(min(__magic_name__ , len(__magic_name__ ) ) ): lowercase , lowercase : Optional[Any] = divmod(__magic_name__ , 10 ) if new_c > 0: add(__magic_name__ , __magic_name__ , __magic_name__ ) else: lowercase : Dict = [] else: lowercase : Union[str, Any] = {c: []} lowercase : Optional[Any] = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps lowercase , lowercase : str = next_term(__magic_name__ , k - 1 , i + dn , __magic_name__ ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead lowercase , lowercase : Optional[Any] = compute(__magic_name__ , __magic_name__ , i + dn , __magic_name__ ) diff += _diff dn += terms_jumped lowercase : Optional[Any] = sub_memo[c] # keep jumps sorted by # of terms skipped lowercase : List[Any] = 0 while j < len(__magic_name__ ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(__magic_name__ , (diff, dn, k) ) return (diff, dn) def snake_case( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Optional[int]: '''simple docstring''' if i >= n: return 0, i if k > len(__magic_name__ ): a_i.extend([0 for _ in range(k - len(__magic_name__ ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) lowercase : Optional[Any] = i lowercase , lowercase , lowercase : List[str] = 0, 0, 0 for j in range(len(__magic_name__ ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 lowercase : List[str] = ds_c + ds_b diff += addend lowercase : Tuple = 0 for j in range(__magic_name__ ): lowercase : int = a_i[j] + addend lowercase , lowercase : Any = divmod(__magic_name__ , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(__magic_name__ , __magic_name__ , __magic_name__ ) return diff, i - start_i def snake_case( __magic_name__ , __magic_name__ , __magic_name__ ) -> Tuple: '''simple docstring''' for j in range(__magic_name__ , len(__magic_name__ ) ): lowercase : Any = digits[j] + addend if s >= 10: lowercase , lowercase : List[str] = divmod(__magic_name__ , 10 ) lowercase : List[str] = addend // 10 + quotient else: lowercase : Optional[Any] = s lowercase : Tuple = addend // 10 if addend == 0: break while addend > 0: lowercase , lowercase : str = divmod(__magic_name__ , 10 ) digits.append(__magic_name__ ) def snake_case( __magic_name__ = 10**15 ) -> int: '''simple docstring''' lowercase : List[Any] = [1] lowercase : List[Any] = 1 lowercase : str = 0 while True: lowercase , lowercase : str = next_term(__magic_name__ , 20 , i + dn , __magic_name__ ) dn += terms_jumped if dn == n - i: break lowercase : str = 0 for j in range(len(__magic_name__ ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(f'''{solution() = }''')
116
1
from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _snake_case ( A__ ): _lowercase : Any = CustomTokenizer pass
137
"""simple docstring""" def _snake_case ( lowercase__ : list[int] ) -> list[list[int]]: '''simple docstring''' lowerCAmelCase_ :Optional[Any] = [] if len(lowercase__ ) == 1: return [nums.copy()] for _ in range(len(lowercase__ ) ): lowerCAmelCase_ :Optional[Any] = nums.pop(0 ) lowerCAmelCase_ :str = permute(lowercase__ ) for perm in permutations: perm.append(lowercase__ ) result.extend(lowercase__ ) nums.append(lowercase__ ) return result def _snake_case ( lowercase__ : Tuple ) -> List[str]: '''simple docstring''' def backtrack(lowercase__ : str ): if start == len(lowercase__ ) - 1: output.append(nums[:] ) else: for i in range(lowercase__ , len(lowercase__ ) ): lowerCAmelCase_ , lowerCAmelCase_ :str = nums[i], nums[start] backtrack(start + 1 ) lowerCAmelCase_ , lowerCAmelCase_ :str = nums[i], nums[start] # backtrack lowerCAmelCase_ :int = [] backtrack(0 ) return output if __name__ == "__main__": import doctest # use res to print the data in permute2 function __UpperCAmelCase = permutea([1, 2, 3]) print(res) doctest.testmod()
84
0
"""simple docstring""" import dataclasses import re import string from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple import numpy as np from . import residue_constants lowerCamelCase_ : Any = Mapping[str, np.ndarray] lowerCamelCase_ : Tuple = Mapping[str, Any] # Is a nested dict. lowerCamelCase_ : List[Any] = 0.01 @dataclasses.dataclass(frozen=_SCREAMING_SNAKE_CASE ) class __A : """simple docstring""" __lowerCAmelCase = 42 # [num_res, num_atom_type, 3] # Amino-acid type for each residue represented as an integer between 0 and # 20, where 20 is 'X'. __lowerCAmelCase = 42 # [num_res] # Binary float mask to indicate presence of a particular atom. 1.0 if an atom # is present and 0.0 if not. This should be used for loss masking. __lowerCAmelCase = 42 # [num_res, num_atom_type] # Residue index as used in PDB. It is not necessarily continuous or 0-indexed. __lowerCAmelCase = 42 # [num_res] # B-factors, or temperature factors, of each residue (in sq. angstroms units), # representing the displacement of the residue from its ground truth mean # value. __lowerCAmelCase = 42 # [num_res, num_atom_type] # Chain indices for multi-chain predictions __lowerCAmelCase = None # Optional remark about the protein. Included as a comment in output PDB # files __lowerCAmelCase = None # Templates used to generate this protein (prediction-only) __lowerCAmelCase = None # Chain corresponding to each parent __lowerCAmelCase = None def _A ( lowercase ): """simple docstring""" a =R'''(\[[A-Z]+\]\n)''' a =[tag.strip() for tag in re.split(lowercase , lowercase ) if len(lowercase ) > 0] a =zip(tags[0::2] , [l.split('''\n''' ) for l in tags[1::2]] ) a =["N", "CA", "C"] a =None a =None a =None for g in groups: if "[PRIMARY]" == g[0]: a =g[1][0].strip() for i in range(len(lowercase ) ): if seq[i] not in residue_constants.restypes: a ='''X''' # FIXME: strings are immutable a =np.array( [residue_constants.restype_order.get(lowercase , residue_constants.restype_num ) for res_symbol in seq] ) elif "[TERTIARY]" == g[0]: a =[] for axis in range(3 ): tertiary.append(list(map(lowercase , g[1][axis].split() ) ) ) a =np.array(lowercase ) a =np.zeros((len(tertiary[0] ) // 3, residue_constants.atom_type_num, 3) ).astype(np.floataa ) for i, atom in enumerate(lowercase ): a =np.transpose(tertiary_np[:, i::3] ) atom_positions *= PICO_TO_ANGSTROM elif "[MASK]" == g[0]: a =np.array(list(map({'''-''': 0, '''+''': 1}.get , g[1][0].strip() ) ) ) a =np.zeros( ( len(lowercase ), residue_constants.atom_type_num, ) ).astype(np.floataa ) for i, atom in enumerate(lowercase ): a =1 atom_mask *= mask[..., None] assert aatype is not None return Protein( atom_positions=lowercase , atom_mask=lowercase , aatype=lowercase , residue_index=np.arange(len(lowercase ) ) , b_factors=lowercase , ) def _A ( lowercase , lowercase = 0 ): """simple docstring""" a =[] a =prot.remark if remark is not None: pdb_headers.append(f'''REMARK {remark}''' ) a =prot.parents a =prot.parents_chain_index if parents is not None and parents_chain_index is not None: a =[p for i, p in zip(lowercase , lowercase ) if i == chain_id] if parents is None or len(lowercase ) == 0: a =['''N/A'''] pdb_headers.append(f'''PARENT {" ".join(lowercase )}''' ) return pdb_headers def _A ( lowercase , lowercase ): """simple docstring""" a =[] a =pdb_str.split('''\n''' ) a =prot.remark if remark is not None: out_pdb_lines.append(f'''REMARK {remark}''' ) a =42 if prot.parents is not None and len(prot.parents ) > 0: a =[] if prot.parents_chain_index is not None: a ={} for p, i in zip(prot.parents , prot.parents_chain_index ): parent_dict.setdefault(str(lowercase ) , [] ) parent_dict[str(lowercase )].append(lowercase ) a =max([int(lowercase ) for chain_idx in parent_dict] ) for i in range(max_idx + 1 ): a =parent_dict.get(str(lowercase ) , ['''N/A'''] ) parents_per_chain.append(lowercase ) else: parents_per_chain.append(list(prot.parents ) ) else: a =[['''N/A''']] def make_parent_line(lowercase ) -> str: return f'''PARENT {" ".join(lowercase )}''' out_pdb_lines.append(make_parent_line(parents_per_chain[0] ) ) a =0 for i, l in enumerate(lowercase ): if "PARENT" not in l and "REMARK" not in l: out_pdb_lines.append(lowercase ) if "TER" in l and "END" not in lines[i + 1]: chain_counter += 1 if not chain_counter >= len(lowercase ): a =parents_per_chain[chain_counter] else: a =['''N/A'''] out_pdb_lines.append(make_parent_line(lowercase ) ) return "\n".join(lowercase ) def _A ( lowercase ): """simple docstring""" a =residue_constants.restypes + ['''X'''] def res_atoa(lowercase ) -> str: return residue_constants.restype_atoa.get(restypes[r] , '''UNK''' ) a =residue_constants.atom_types a =[] a =prot.atom_mask a =prot.aatype a =prot.atom_positions a =prot.residue_index.astype(np.intaa ) a =prot.b_factors a =prot.chain_index if np.any(aatype > residue_constants.restype_num ): raise ValueError('''Invalid aatypes.''' ) a =get_pdb_headers(lowercase ) if len(lowercase ) > 0: pdb_lines.extend(lowercase ) a =aatype.shape[0] a =1 a =0 a =string.ascii_uppercase a =None # Add all atom sites. for i in range(lowercase ): a =res_atoa(aatype[i] ) for atom_name, pos, mask, b_factor in zip(lowercase , atom_positions[i] , atom_mask[i] , b_factors[i] ): if mask < 0.5: continue a ='''ATOM''' a =atom_name if len(lowercase ) == 4 else f''' {atom_name}''' a ='''''' a ='''''' a =1.00 a =atom_name[0] # Protein supports only C, N, O, S, this works. a ='''''' a ='''A''' if chain_index is not None: a =chain_tags[chain_index[i]] # PDB is a columnar format, every space matters here! a =( f'''{record_type:<6}{atom_index:>5} {name:<4}{alt_loc:>1}''' f'''{res_name_a:>3} {chain_tag:>1}''' f'''{residue_index[i]:>4}{insertion_code:>1} ''' f'''{pos[0]:>8.3f}{pos[1]:>8.3f}{pos[2]:>8.3f}''' f'''{occupancy:>6.2f}{b_factor:>6.2f} ''' f'''{element:>2}{charge:>2}''' ) pdb_lines.append(lowercase ) atom_index += 1 a =i == n - 1 if chain_index is not None: if i != n - 1 and chain_index[i + 1] != prev_chain_index: a =True a =chain_index[i + 1] if should_terminate: # Close the chain. a ='''TER''' a =( f'''{chain_end:<6}{atom_index:>5} {res_atoa(aatype[i] ):>3} {chain_tag:>1}{residue_index[i]:>4}''' ) pdb_lines.append(lowercase ) atom_index += 1 if i != n - 1: # "prev" is a misnomer here. This happens at the beginning of # each new chain. pdb_lines.extend(get_pdb_headers(lowercase , lowercase ) ) pdb_lines.append('''END''' ) pdb_lines.append('''''' ) return "\n".join(lowercase ) def _A ( lowercase ): """simple docstring""" return residue_constants.STANDARD_ATOM_MASK[prot.aatype] def _A ( lowercase , lowercase , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = None , ): """simple docstring""" return Protein( aatype=features['''aatype'''] , atom_positions=result['''final_atom_positions'''] , atom_mask=result['''final_atom_mask'''] , residue_index=features['''residue_index'''] + 1 , b_factors=b_factors if b_factors is not None else np.zeros_like(result['''final_atom_mask'''] ) , chain_index=lowercase , remark=lowercase , parents=lowercase , parents_chain_index=lowercase , )
215
"""simple docstring""" import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, require_tf, slow if is_tf_available(): import tensorflow as tf if is_tensorflow_text_available(): from transformers.models.bert import TFBertTokenizer lowerCamelCase_ : str = ["""bert-base-uncased""", """bert-base-cased"""] lowerCamelCase_ : List[str] = """hf-internal-testing/tiny-bert-tf-only""" if is_tf_available(): class __A ( tf.keras.Model ): """simple docstring""" def __init__( self , __A ) -> Dict: super().__init__() a =tokenizer a =AutoConfig.from_pretrained(__A ) a =TFAutoModel.from_config(__A ) def SCREAMING_SNAKE_CASE ( self , __A ) -> int: a =self.tokenizer(__A ) a =self.bert(**__A ) return out["pooler_output"] @require_tf @require_tensorflow_text class __A ( unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self ) -> str: super().setUp() a =[ BertTokenizer.from_pretrained(__A ) for checkpoint in (TOKENIZER_CHECKPOINTS * 2) ] # repeat for when fast_bert_tokenizer=false a =[TFBertTokenizer.from_pretrained(__A ) for checkpoint in TOKENIZER_CHECKPOINTS] + [ TFBertTokenizer.from_pretrained(__A , use_fast_bert_tokenizer=__A ) for checkpoint in TOKENIZER_CHECKPOINTS ] assert len(self.tokenizers ) == len(self.tf_tokenizers ) a =[ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] a =list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in (self.test_sentences, self.paired_sentences): a =tokenizer(__A , return_tensors='''tf''' , padding='''longest''' ) a =tf_tokenizer(__A ) for key in python_outputs.keys(): self.assertTrue(tf.reduce_all(python_outputs[key].shape == tf_outputs[key].shape ) ) self.assertTrue(tf.reduce_all(tf.cast(python_outputs[key] , tf.intaa ) == tf_outputs[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self ) -> str: for tf_tokenizer in self.tf_tokenizers: a =tf_tokenizer(self.paired_sentences ) a =tf_tokenizer( text=[sentence[0] for sentence in self.paired_sentences] , text_pair=[sentence[1] for sentence in self.paired_sentences] , ) for key in merged_outputs.keys(): self.assertTrue(tf.reduce_all(tf.cast(merged_outputs[key] , tf.intaa ) == separated_outputs[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self ) -> List[str]: for tf_tokenizer in self.tf_tokenizers: a =tf.function(__A ) for test_inputs in (self.test_sentences, self.paired_sentences): a =tf.constant(__A ) a =compiled_tokenizer(__A ) a =tf_tokenizer(__A ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self ) -> Tuple: for tf_tokenizer in self.tf_tokenizers: a =ModelToSave(tokenizer=__A ) a =tf.convert_to_tensor(self.test_sentences ) a =model(__A ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: a =Path(__A ) / '''saved.model''' model.save(__A ) a =tf.keras.models.load_model(__A ) a =loaded_model(__A ) # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertLessEqual(tf.reduce_max(tf.abs(out - loaded_output ) ) , 1E-5 )
215
1
'''simple docstring''' import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMInverseScheduler, DDIMScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, StableDiffusionDiffEditPipeline, UNetaDConditionModel, ) from diffusers.utils import load_image, slow from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, require_torch_gpu, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class A_ ( lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): _lowerCamelCase : List[str] = StableDiffusionDiffEditPipeline _lowerCamelCase : Any = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""height""", """width""", """image"""} | {"""image_latents"""} _lowerCamelCase : str = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS - {"""image"""} | {"""image_latents"""} _lowerCamelCase : List[Any] = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _lowerCamelCase : Optional[Any] = frozenset([] ) def lowercase ( self : Optional[int] ): torch.manual_seed(0 ) _UpperCAmelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=3_2 , attention_head_dim=(2, 4) , use_linear_projection=snake_case_ , ) _UpperCAmelCase = DDIMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="scaled_linear" , clip_sample=snake_case_ , set_alpha_to_one=snake_case_ , ) _UpperCAmelCase = DDIMInverseScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="scaled_linear" , clip_sample=snake_case_ , set_alpha_to_zero=snake_case_ , ) torch.manual_seed(0 ) _UpperCAmelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) _UpperCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act="gelu" , projection_dim=5_1_2 , ) _UpperCAmelCase = CLIPTextModel(snake_case_ ) _UpperCAmelCase = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) _UpperCAmelCase = { "unet": unet, "scheduler": scheduler, "inverse_scheduler": inverse_scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def lowercase ( self : Optional[Any] , snake_case_ : str , snake_case_ : List[Any]=0 ): _UpperCAmelCase = floats_tensor((1, 1_6, 1_6) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) _UpperCAmelCase = floats_tensor((1, 2, 4, 1_6, 1_6) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) if str(snake_case_ ).startswith("mps" ): _UpperCAmelCase = torch.manual_seed(snake_case_ ) else: _UpperCAmelCase = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) _UpperCAmelCase = { "prompt": "a dog and a newt", "mask_image": mask, "image_latents": latents, "generator": generator, "num_inference_steps": 2, "inpaint_strength": 1.0, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def lowercase ( self : Dict , snake_case_ : Union[str, Any] , snake_case_ : Union[str, Any]=0 ): _UpperCAmelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) _UpperCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _UpperCAmelCase = Image.fromarray(np.uinta(snake_case_ ) ).convert("RGB" ) if str(snake_case_ ).startswith("mps" ): _UpperCAmelCase = torch.manual_seed(snake_case_ ) else: _UpperCAmelCase = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) _UpperCAmelCase = { "image": image, "source_prompt": "a cat and a frog", "target_prompt": "a dog and a newt", "generator": generator, "num_inference_steps": 2, "num_maps_per_mask": 2, "mask_encode_strength": 1.0, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def lowercase ( self : List[Any] , snake_case_ : List[str] , snake_case_ : Any=0 ): _UpperCAmelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) _UpperCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _UpperCAmelCase = Image.fromarray(np.uinta(snake_case_ ) ).convert("RGB" ) if str(snake_case_ ).startswith("mps" ): _UpperCAmelCase = torch.manual_seed(snake_case_ ) else: _UpperCAmelCase = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) _UpperCAmelCase = { "image": image, "prompt": "a cat and a frog", "generator": generator, "num_inference_steps": 2, "inpaint_strength": 1.0, "guidance_scale": 6.0, "decode_latents": True, "output_type": "numpy", } return inputs def lowercase ( self : Tuple ): if not hasattr(self.pipeline_class , "_optional_components" ): return _UpperCAmelCase = self.get_dummy_components() _UpperCAmelCase = self.pipeline_class(**snake_case_ ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) # set all optional components to None and update pipeline config accordingly for optional_component in pipe._optional_components: setattr(snake_case_ , snake_case_ , snake_case_ ) pipe.register_modules(**{optional_component: None for optional_component in pipe._optional_components} ) _UpperCAmelCase = self.get_dummy_inputs(snake_case_ ) _UpperCAmelCase = pipe(**snake_case_ )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(snake_case_ ) _UpperCAmelCase = self.pipeline_class.from_pretrained(snake_case_ ) pipe_loaded.to(snake_case_ ) pipe_loaded.set_progress_bar_config(disable=snake_case_ ) for optional_component in pipe._optional_components: self.assertTrue( getattr(snake_case_ , snake_case_ ) is None , f'`{optional_component}` did not stay set to None after loading.' , ) _UpperCAmelCase = self.get_dummy_inputs(snake_case_ ) _UpperCAmelCase = pipe_loaded(**snake_case_ )[0] _UpperCAmelCase = np.abs(output - output_loaded ).max() self.assertLess(snake_case_ , 1e-4 ) def lowercase ( self : Any ): _UpperCAmelCase = "cpu" _UpperCAmelCase = self.get_dummy_components() _UpperCAmelCase = self.pipeline_class(**snake_case_ ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) _UpperCAmelCase = self.get_dummy_mask_inputs(snake_case_ ) _UpperCAmelCase = pipe.generate_mask(**snake_case_ ) _UpperCAmelCase = mask[0, -3:, -3:] self.assertEqual(mask.shape , (1, 1_6, 1_6) ) _UpperCAmelCase = np.array([0] * 9 ) _UpperCAmelCase = np.abs(mask_slice.flatten() - expected_slice ).max() self.assertLessEqual(snake_case_ , 1e-3 ) self.assertEqual(mask[0, -3, -4] , 0 ) def lowercase ( self : Union[str, Any] ): _UpperCAmelCase = "cpu" _UpperCAmelCase = self.get_dummy_components() _UpperCAmelCase = self.pipeline_class(**snake_case_ ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) _UpperCAmelCase = self.get_dummy_inversion_inputs(snake_case_ ) _UpperCAmelCase = pipe.invert(**snake_case_ ).images _UpperCAmelCase = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 3_2, 3_2, 3) ) _UpperCAmelCase = np.array( [0.5_1_5_0, 0.5_1_3_4, 0.5_0_4_3, 0.5_3_7_6, 0.4_6_9_4, 0.5_1_0_5_0, 0.5_0_1_5, 0.4_4_0_7, 0.4_7_9_9] , ) _UpperCAmelCase = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(snake_case_ , 1e-3 ) def lowercase ( self : List[Any] ): super().test_inference_batch_single_identical(expected_max_diff=5e-3 ) def lowercase ( self : List[Any] ): _UpperCAmelCase = "cpu" _UpperCAmelCase = self.get_dummy_components() _UpperCAmelCase = {"beta_start": 0.0_0_0_8_5, "beta_end": 0.0_1_2, "beta_schedule": "scaled_linear"} _UpperCAmelCase = DPMSolverMultistepScheduler(**snake_case_ ) _UpperCAmelCase = DPMSolverMultistepInverseScheduler(**snake_case_ ) _UpperCAmelCase = self.pipeline_class(**snake_case_ ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) _UpperCAmelCase = self.get_dummy_inversion_inputs(snake_case_ ) _UpperCAmelCase = pipe.invert(**snake_case_ ).images _UpperCAmelCase = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 3_2, 3_2, 3) ) _UpperCAmelCase = np.array( [0.5_1_5_0, 0.5_1_3_4, 0.5_0_4_3, 0.5_3_7_6, 0.4_6_9_4, 0.5_1_0_5_0, 0.5_0_1_5, 0.4_4_0_7, 0.4_7_9_9] , ) _UpperCAmelCase = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(snake_case_ , 1e-3 ) @require_torch_gpu @slow class A_ ( unittest.TestCase ): def lowercase ( self : Any ): super().tearDown() gc.collect() torch.cuda.empty_cache() @classmethod def lowercase ( cls : Dict ): _UpperCAmelCase = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/diffedit/fruit.png" ) _UpperCAmelCase = raw_image.convert("RGB" ).resize((7_6_8, 7_6_8) ) _UpperCAmelCase = raw_image def lowercase ( self : Union[str, Any] ): _UpperCAmelCase = torch.manual_seed(0 ) _UpperCAmelCase = StableDiffusionDiffEditPipeline.from_pretrained( "stabilityai/stable-diffusion-2-1" , safety_checker=snake_case_ , torch_dtype=torch.floataa ) _UpperCAmelCase = DDIMScheduler.from_config(pipe.scheduler.config ) _UpperCAmelCase = DDIMInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=snake_case_ ) _UpperCAmelCase = "a bowl of fruit" _UpperCAmelCase = "a bowl of pears" _UpperCAmelCase = pipe.generate_mask( image=self.raw_image , source_prompt=snake_case_ , target_prompt=snake_case_ , generator=snake_case_ , ) _UpperCAmelCase = pipe.invert( prompt=snake_case_ , image=self.raw_image , inpaint_strength=0.7 , generator=snake_case_ ).latents _UpperCAmelCase = pipe( prompt=snake_case_ , mask_image=snake_case_ , image_latents=snake_case_ , generator=snake_case_ , negative_prompt=snake_case_ , inpaint_strength=0.7 , output_type="numpy" , ).images[0] _UpperCAmelCase = ( np.array( load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/diffedit/pears.png" ).resize((7_6_8, 7_6_8) ) ) / 2_5_5 ) assert np.abs((expected_image - image).max() ) < 5e-1 def lowercase ( self : Optional[int] ): _UpperCAmelCase = torch.manual_seed(0 ) _UpperCAmelCase = StableDiffusionDiffEditPipeline.from_pretrained( "stabilityai/stable-diffusion-2-1" , safety_checker=snake_case_ , torch_dtype=torch.floataa ) _UpperCAmelCase = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) _UpperCAmelCase = DPMSolverMultistepInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=snake_case_ ) _UpperCAmelCase = "a bowl of fruit" _UpperCAmelCase = "a bowl of pears" _UpperCAmelCase = pipe.generate_mask( image=self.raw_image , source_prompt=snake_case_ , target_prompt=snake_case_ , generator=snake_case_ , ) _UpperCAmelCase = pipe.invert( prompt=snake_case_ , image=self.raw_image , inpaint_strength=0.7 , generator=snake_case_ , num_inference_steps=2_5 , ).latents _UpperCAmelCase = pipe( prompt=snake_case_ , mask_image=snake_case_ , image_latents=snake_case_ , generator=snake_case_ , negative_prompt=snake_case_ , inpaint_strength=0.7 , num_inference_steps=2_5 , output_type="numpy" , ).images[0] _UpperCAmelCase = ( np.array( load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/diffedit/pears.png" ).resize((7_6_8, 7_6_8) ) ) / 2_5_5 ) assert np.abs((expected_image - image).max() ) < 5e-1
22
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def __magic_name__ ( __a : Optional[int] , __a : Union[str, Any] , __a : Union[str, Any]=1_024 , __a : str=1_024 , __a : Optional[Any]=False , **__a : Tuple ): '''simple docstring''' UpperCamelCase__ = AutoTokenizer.from_pretrained(__a ) UpperCamelCase__ = SeqaSeqDataset(__a , __a , __a , __a , type_path="""train""" , **__a ) UpperCamelCase__ = tok.pad_token_id def get_lens(__a : Optional[int] ): UpperCamelCase__ = tqdm( DataLoader(__a , batch_size=512 , num_workers=8 , shuffle=__a , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) UpperCamelCase__ = [] for batch in dl: UpperCamelCase__ = batch["""input_ids"""].ne(__a ).sum(1 ).tolist() UpperCamelCase__ = batch["""labels"""].ne(__a ).sum(1 ).tolist() if consider_target: for src, tgt in zip(__a , __a ): max_lens.append(max(__a , __a ) ) else: max_lens.extend(__a ) return max_lens UpperCamelCase__ = get_lens(__a ) UpperCamelCase__ = SeqaSeqDataset(__a , __a , __a , __a , type_path="""val""" , **__a ) UpperCamelCase__ = get_lens(__a ) pickle_save(__a , train_ds.len_file ) pickle_save(__a , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
244
0
"""simple docstring""" def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError('The given input must be positive' ) # get the generated string sequence lowerCamelCase__ : List[str] = gray_code_sequence_string(_UpperCAmelCase ) # # convert them to integers for i in range(len(_UpperCAmelCase ) ): lowerCamelCase__ : Any = int(sequence[i] , 2 ) return sequence def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] lowerCamelCase__ : List[str] = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits lowerCamelCase__ : int = gray_code_sequence_string(bit_count - 1 ) lowerCamelCase__ : int = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): lowerCamelCase__ : Tuple = '0' + smaller_sequence[i] sequence.append(_UpperCAmelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): lowerCamelCase__ : List[Any] = '1' + smaller_sequence[i] sequence.append(_UpperCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
363
def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> float: return 10 - x * x def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> float: # Bolzano theory in order to find if there is a root between a and b if equation(_UpperCAmelCase ) * equation(_UpperCAmelCase ) >= 0: raise ValueError('Wrong space!' ) lowerCamelCase__ : Tuple = a while (b - a) >= 0.01: # Find middle point lowerCamelCase__ : Optional[int] = (a + b) / 2 # Check if middle point is root if equation(_UpperCAmelCase ) == 0.0: break # Decide the side to repeat the steps if equation(_UpperCAmelCase ) * equation(_UpperCAmelCase ) < 0: lowerCamelCase__ : Tuple = c else: lowerCamelCase__ : Optional[int] = c return c if __name__ == "__main__": import doctest doctest.testmod() print(bisection(-2, 5)) print(bisection(0, 6))
45
0
import math def lowerCAmelCase_ ( snake_case_ = 100 ): _A : Optional[Any] = sum(i * i for i in range(1,n + 1 ) ) _A : Optional[Any] = int(math.pow(sum(range(1,n + 1 ) ),2 ) ) return square_of_sum - sum_of_squares if __name__ == "__main__": print(f"""{solution() = }""")
26
from __future__ import annotations from typing import Any class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self : Tuple , lowercase : int , lowercase : int , lowercase : float = 0 ): '''simple docstring''' _snake_case , _snake_case = row, column _snake_case = [[default_value for c in range(lowercase )] for r in range(lowercase )] def __str__( self : int ): '''simple docstring''' _snake_case = f'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier _snake_case = 0 for row_vector in self.array: for obj in row_vector: _snake_case = max(lowercase , len(str(lowercase ) ) ) _snake_case = f'''%{max_element_length}s''' # Make string and return def single_line(lowercase : list[float] ) -> str: nonlocal string_format_identifier _snake_case = '[' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase ) for row_vector in self.array ) return s def __repr__( self : Dict ): '''simple docstring''' return str(self ) def A ( self : str , lowercase : tuple[int, int] ): '''simple docstring''' if not (isinstance(lowercase , (list, tuple) ) and len(lowercase ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Dict , lowercase : tuple[int, int] ): '''simple docstring''' assert self.validate_indicies(lowercase ) return self.array[loc[0]][loc[1]] def __setitem__( self : str , lowercase : tuple[int, int] , lowercase : float ): '''simple docstring''' assert self.validate_indicies(lowercase ) _snake_case = value def __add__( self : str , lowercase : Matrix ): '''simple docstring''' assert isinstance(lowercase , lowercase ) assert self.row == another.row and self.column == another.column # Add _snake_case = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): _snake_case = self[r, c] + another[r, c] return result def __neg__( self : Tuple ): '''simple docstring''' _snake_case = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): _snake_case = -self[r, c] return result def __sub__( self : List[str] , lowercase : Matrix ): '''simple docstring''' return self + (-another) def __mul__( self : Dict , lowercase : int | float | Matrix ): '''simple docstring''' if isinstance(lowercase , (int, float) ): # Scalar multiplication _snake_case = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): _snake_case = self[r, c] * another return result elif isinstance(lowercase , lowercase ): # Matrix multiplication assert self.column == another.row _snake_case = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: _snake_case = f'''Unsupported type given for another ({type(lowercase )})''' raise TypeError(lowercase ) def A ( self : Optional[Any] ): '''simple docstring''' _snake_case = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): _snake_case = self[r, c] return result def A ( self : List[Any] , lowercase : Matrix , lowercase : Matrix ): '''simple docstring''' assert isinstance(lowercase , lowercase ) and isinstance(lowercase , lowercase ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate _snake_case = v.transpose() _snake_case = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def a_ ( ) -> None: # a^(-1) _snake_case = Matrix(3 , 3 , 0 ) for i in range(3 ): _snake_case = 1 print(f'''a^(-1) is {ainv}''' ) # u, v _snake_case = Matrix(3 , 1 , 0 ) _snake_case , _snake_case , _snake_case = 1, 2, -3 _snake_case = Matrix(3 , 1 , 0 ) _snake_case , _snake_case , _snake_case = 4, -2, 5 print(f'''u is {u}''' ) print(f'''v is {v}''' ) print(f'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(f'''(a + uv^T)^(-1) is {ainv.sherman_morrison(__lowercase , __lowercase )}''' ) def a_ ( ) -> None: import doctest doctest.testmod() testa()
282
0
'''simple docstring''' import argparse import re from flax.traverse_util import flatten_dict, unflatten_dict from tax import checkpoints from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model from transformers.utils import logging logging.set_verbosity_info() # should not include what is already done by the `from_pt` argument _lowercase : str = { "/attention/": "/0/SelfAttention/", "/self_attention/": "/0/SelfAttention/", "/encoder_decoder_attention/": "/1/EncDecAttention/", "value": "v", "query": "q", "key": "k", "out": "o", "pre_self_attention_layer_norm": "0/layer_norm", "pre_cross_attention_layer_norm": "1/layer_norm", "pre_attention_layer_norm": "0/layer_norm", # previously 1, but seems wrong "token_embedder": "shared", "encoder_norm": "final_layer_norm", "decoder_norm": "final_layer_norm", "relpos_bias/rel_embedding": "block/0/layer/0/SelfAttention/relative_attention_bias/weight", "router/router_weights/w/": "router/classifier/", "roer/roer_weights/w/": "router/classifier/", "logits_dense": "lm_head", } def snake_case_ ( __SCREAMING_SNAKE_CASE : int ): """simple docstring""" lowercase_ : List[str] = list(s_dict.keys() ) for key in keys: lowercase_ : List[Any] = r'.*/layers_(\d+)' lowercase_ : Union[str, Any] = key if re.match(lowerCAmelCase_ , lowerCAmelCase_ ): lowercase_ : int = re.sub(R'''layers_(\d+)''' , R'''block/\1/layer''' , lowerCAmelCase_ ) lowercase_ : Dict = r'(encoder|decoder)\/' if re.match(lowerCAmelCase_ , lowerCAmelCase_ ): lowercase_ : Union[str, Any] = re.match(lowerCAmelCase_ , lowerCAmelCase_ ).groups() if groups[0] == "encoder": lowercase_ : Dict = re.sub(R'''/mlp/''' , R'''/1/mlp/''' , lowerCAmelCase_ ) lowercase_ : Union[str, Any] = re.sub(R'''/pre_mlp_layer_norm/''' , R'''/1/layer_norm/''' , lowerCAmelCase_ ) elif groups[0] == "decoder": lowercase_ : Dict = re.sub(R'''/mlp/''' , R'''/2/mlp/''' , lowerCAmelCase_ ) lowercase_ : Dict = re.sub(R'''/pre_mlp_layer_norm/''' , R'''/2/layer_norm/''' , lowerCAmelCase_ ) # 2. Convert other classic mappings for old_key, temp_key in MOE_LAYER_NAME_MAPPING.items(): if old_key in new_key: lowercase_ : int = new_key.replace(lowerCAmelCase_ , lowerCAmelCase_ ) print(F'''{key} -> {new_key}''' ) lowercase_ : Optional[int] = s_dict.pop(lowerCAmelCase_ ) if "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: lowercase_ : int = s_dict[ 'encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight' ].T if "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: lowercase_ : Optional[int] = s_dict[ 'decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight' ].T # 3. Take extra care of the EXPERTS layer for key in list(s_dict.keys() ): if "expert" in key: lowercase_ : Any = s_dict[key].shape[0] lowercase_ : str = s_dict[key] for idx in range(lowerCAmelCase_ ): lowercase_ : Any = expert_weihts[idx] print(F'''{key} -> {key.replace('expert/' , 'nested fstring' )}''' ) s_dict.pop(lowerCAmelCase_ ) return s_dict _lowercase : List[str] = { "NUM_ENCODER_LAYERS": "num_layers", "NUM_DECODER_LAYERS": "num_decoder_layers", "NUM_HEADS": "num_heads", "HEAD_DIM": "d_kv", "EMBED_DIM": "d_model", "MLP_DIM": "d_ff", "NUM_SELECTED_EXPERTS": "num_selected_experts", "NUM_ENCODER_SPARSE_LAYERS": "num_sparse_encoder_layers", "NUM_DECODER_SPARSE_LAYERS": "num_sparse_decoder_layers", "dense.MlpBlock.activations": "feed_forward_proj", } def snake_case_ ( __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : Union[str, Any] ): """simple docstring""" import regex as re with open(lowerCAmelCase_ , '''r''' ) as f: lowercase_ : List[str] = f.read() lowercase_ : Union[str, Any] = re.findall(R'''(.*) = ([0-9.]*)''' , lowerCAmelCase_ ) lowercase_ : Tuple = {} for param, value in regex_match: if param in GIN_TO_CONFIG_MAPPING and value != "": lowercase_ : Union[str, Any] = float(lowerCAmelCase_ ) if '.' in value else int(lowerCAmelCase_ ) lowercase_ : Any = re.findall(R'''(.*activations) = \(\'(.*)\',\)''' , lowerCAmelCase_ )[0] lowercase_ : Tuple = str(activation[1] ) lowercase_ : Optional[Any] = num_experts lowercase_ : List[Any] = SwitchTransformersConfig(**lowerCAmelCase_ ) return config def snake_case_ ( __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : Optional[Any]=None , __SCREAMING_SNAKE_CASE : Union[str, Any]="./" , __SCREAMING_SNAKE_CASE : int=8 ): """simple docstring""" print(F'''Loading flax weights from : {flax_checkpoint_path}''' ) lowercase_ : Union[str, Any] = checkpoints.load_tax_checkpoint(lowerCAmelCase_ ) if gin_file is not None: lowercase_ : int = convert_gin_to_config(lowerCAmelCase_ , lowerCAmelCase_ ) else: lowercase_ : Dict = SwitchTransformersConfig.from_pretrained(lowerCAmelCase_ ) lowercase_ : Optional[int] = SwitchTransformersForConditionalGeneration(lowerCAmelCase_ ) lowercase_ : Union[str, Any] = flax_params['target'] lowercase_ : Optional[Any] = flatten_dict(lowerCAmelCase_ , sep='''/''' ) lowercase_ : List[Any] = rename_keys(lowerCAmelCase_ ) lowercase_ : Dict = unflatten_dict(lowerCAmelCase_ , sep='''/''' ) # Load the flax params in the PT model load_flax_weights_in_pytorch_model(lowerCAmelCase_ , lowerCAmelCase_ ) print(F'''Save PyTorch model to {pytorch_dump_path}''' ) pt_model.save_pretrained(lowerCAmelCase_ ) if __name__ == "__main__": _lowercase : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--switch_t5x_checkpoint_path", default=None, type=str, required=True, help=( "The config json file corresponding to the pre-trained SwitchTransformers model. \nThis specifies the" " model architecture. If not provided, a `gin_file` has to be provided." ), ) parser.add_argument( "--gin_file", default=None, type=str, required=False, help="Path to the gin config file. If not provided, a `config_file` has to be passed ", ) parser.add_argument( "--config_name", default=None, type=str, required=False, help="Config name of SwitchTransformers model." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output pytorch model." ) parser.add_argument("--num_experts", default=8, type=int, required=False, help="Number of experts") _lowercase : Any = parser.parse_args() convert_flax_checkpoint_to_pytorch( args.switch_tax_checkpoint_path, args.config_name, args.gin_file, args.pytorch_dump_folder_path, args.num_experts, )
369
'''simple docstring''' import warnings from ...utils import logging from .image_processing_donut import DonutImageProcessor _lowercase : Any = logging.get_logger(__name__) class lowerCAmelCase__ ( lowerCamelCase_ ): def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ): """simple docstring""" warnings.warn( '''The class DonutFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use DonutImageProcessor instead.''' , __SCREAMING_SNAKE_CASE , ) super().__init__(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE )
264
0
"""simple docstring""" from collections import namedtuple _lowercase = namedtuple('''from_to''', '''from_ to''') _lowercase = { '''cubicmeter''': from_to(1, 1), '''litre''': from_to(0.001, 10_00), '''kilolitre''': from_to(1, 1), '''gallon''': from_to(0.00_454, 264.172), '''cubicyard''': from_to(0.76_455, 1.30_795), '''cubicfoot''': from_to(0.028, 35.3_147), '''cup''': from_to(0.000_236_588, 4_226.75), } def _snake_case ( snake_case__ : float , snake_case__ : str , snake_case__ : str ): if from_type not in METRIC_CONVERSION: raise ValueError( F'Invalid \'from_type\' value: {from_type!r} Supported values are:\n' + ', '.join(snake_case__ ) ) if to_type not in METRIC_CONVERSION: raise ValueError( F'Invalid \'to_type\' value: {to_type!r}. Supported values are:\n' + ', '.join(snake_case__ ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
74
"""simple docstring""" from typing import List, Optional, Union import numpy as np import torch import torchaudio.compliance.kaldi as ta_kaldi from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging lowercase_ = logging.get_logger(__name__) class snake_case ( _lowerCAmelCase ): '''simple docstring''' A_ : int = ["input_features", "attention_mask"] def __init__( self : Optional[Any], _lowerCamelCase : Union[str, Any]=80, _lowerCamelCase : int=1_60_00, _lowerCamelCase : Any=80, _lowerCamelCase : List[str]=0.0, _lowerCamelCase : int=True, _lowerCamelCase : Optional[Any]=True, _lowerCamelCase : Optional[int]=True, **_lowerCamelCase : List[str], ): '''simple docstring''' super().__init__(feature_size=_lowerCamelCase, sampling_rate=_lowerCamelCase, padding_value=_lowerCamelCase, **_lowerCamelCase ) __A = num_mel_bins __A = do_ceptral_normalize __A = normalize_means __A = normalize_vars __A = True def _SCREAMING_SNAKE_CASE ( self : Dict, _lowerCamelCase : np.ndarray, ): '''simple docstring''' __A = waveform * (2**15) # Kaldi compliance: 16-bit signed integers __A = torch.from_numpy(_lowerCamelCase ).unsqueeze(0 ) __A = ta_kaldi.fbank(_lowerCamelCase, num_mel_bins=self.num_mel_bins, sample_frequency=self.sampling_rate ) return features.numpy() @staticmethod def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : np.ndarray, _lowerCamelCase : int, _lowerCamelCase : Optional[bool] = True, _lowerCamelCase : Optional[bool] = True, _lowerCamelCase : float = 0.0, ): '''simple docstring''' # make sure we normalize float32 arrays if normalize_means: __A = x[:input_length].mean(axis=0 ) __A = np.subtract(_lowerCamelCase, _lowerCamelCase ) if normalize_vars: __A = x[:input_length].std(axis=0 ) __A = np.divide(_lowerCamelCase, _lowerCamelCase ) if input_length < x.shape[0]: __A = padding_value # make sure array is in float32 __A = x.astype(np.floataa ) return x def _SCREAMING_SNAKE_CASE ( self : str, _lowerCamelCase : List[np.ndarray], _lowerCamelCase : Optional[np.ndarray] = None ): '''simple docstring''' __A = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [ self.utterance_cmvn(_lowerCamelCase, _lowerCamelCase, self.normalize_means, self.normalize_vars, self.padding_value ) for x, n in zip(_lowerCamelCase, _lowerCamelCase ) ] def __call__( self : Optional[Any], _lowerCamelCase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]], _lowerCamelCase : Union[bool, str, PaddingStrategy] = False, _lowerCamelCase : Optional[int] = None, _lowerCamelCase : bool = False, _lowerCamelCase : Optional[int] = None, _lowerCamelCase : Optional[Union[str, TensorType]] = None, _lowerCamelCase : Optional[int] = None, _lowerCamelCase : Optional[bool] = None, **_lowerCamelCase : Optional[Any], ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'The model corresponding to this feature extractor: {self} was trained using a sampling rate of' f' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with' f' {self.sampling_rate} and not {sampling_rate}.' ) else: logger.warning( '''It is strongly recommended to pass the `sampling_rate` argument to this function. ''' '''Failing to do so can result in silent errors that might be hard to debug.''' ) __A = isinstance(_lowerCamelCase, np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'Only mono-channel audio is supported for input to {self}' ) __A = is_batched_numpy or ( isinstance(_lowerCamelCase, (list, tuple) ) and (isinstance(raw_speech[0], (np.ndarray, tuple, list) )) ) if is_batched: __A = [np.asarray(_lowerCamelCase, dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(_lowerCamelCase, np.ndarray ): __A = np.asarray(_lowerCamelCase, dtype=np.floataa ) elif isinstance(_lowerCamelCase, np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __A = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __A = [raw_speech] # extract fbank features __A = [self._extract_fbank_features(_lowerCamelCase ) for waveform in raw_speech] # convert into correct format for padding __A = BatchFeature({'''input_features''': features} ) __A = self.pad( _lowerCamelCase, padding=_lowerCamelCase, max_length=_lowerCamelCase, truncation=_lowerCamelCase, pad_to_multiple_of=_lowerCamelCase, return_attention_mask=_lowerCamelCase, **_lowerCamelCase, ) # make sure list is in array format __A = padded_inputs.get('''input_features''' ) if isinstance(input_features[0], _lowerCamelCase ): __A = [np.asarray(_lowerCamelCase, dtype=np.floataa ) for feature in input_features] __A = padded_inputs.get('''attention_mask''' ) if attention_mask is not None: __A = [np.asarray(_lowerCamelCase, dtype=np.intaa ) for array in attention_mask] # Utterance-level cepstral mean and variance normalization if self.do_ceptral_normalize: __A = ( np.array(_lowerCamelCase, dtype=np.intaa ) if self._get_padding_strategies(_lowerCamelCase, max_length=_lowerCamelCase ) is not PaddingStrategy.DO_NOT_PAD else None ) __A = self.normalize( padded_inputs['''input_features'''], attention_mask=_lowerCamelCase ) if return_tensors is not None: __A = padded_inputs.convert_to_tensors(_lowerCamelCase ) return padded_inputs
266
0
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer import diffusers from diffusers import ( AutoencoderKL, EulerDiscreteScheduler, StableDiffusionLatentUpscalePipeline, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.schedulers import KarrasDiffusionSchedulers from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() def a__ ( A__ ): SCREAMING_SNAKE_CASE_ : Optional[int] = [tensor.shape for tensor in tensor_list] return all(shape == shapes[0] for shape in shapes[1:] ) class __lowercase (__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" _UpperCAmelCase = StableDiffusionLatentUpscalePipeline _UpperCAmelCase = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { """height""", """width""", """cross_attention_kwargs""", """negative_prompt_embeds""", """prompt_embeds""", } _UpperCAmelCase = PipelineTesterMixin.required_optional_params - {"""num_images_per_prompt"""} _UpperCAmelCase = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _UpperCAmelCase = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess _UpperCAmelCase = frozenset([] ) _UpperCAmelCase = True @property def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = 1 SCREAMING_SNAKE_CASE_ : Optional[Any] = 4 SCREAMING_SNAKE_CASE_ : int = (1_6, 1_6) SCREAMING_SNAKE_CASE_ : Optional[Any] = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(lowerCAmelCase__ ) return image def UpperCamelCase__ ( self ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ : str = UNetaDConditionModel( act_fn='gelu' , attention_head_dim=8 , norm_num_groups=lowerCAmelCase__ , block_out_channels=[3_2, 3_2, 6_4, 6_4] , time_cond_proj_dim=1_6_0 , conv_in_kernel=1 , conv_out_kernel=1 , cross_attention_dim=3_2 , down_block_types=( 'KDownBlock2D', 'KCrossAttnDownBlock2D', 'KCrossAttnDownBlock2D', 'KCrossAttnDownBlock2D', ) , in_channels=8 , mid_block_type=lowerCAmelCase__ , only_cross_attention=lowerCAmelCase__ , out_channels=5 , resnet_time_scale_shift='scale_shift' , time_embedding_type='fourier' , timestep_post_act='gelu' , up_block_types=('KCrossAttnUpBlock2D', 'KCrossAttnUpBlock2D', 'KCrossAttnUpBlock2D', 'KUpBlock2D') , ) SCREAMING_SNAKE_CASE_ : int = AutoencoderKL( block_out_channels=[3_2, 3_2, 6_4, 6_4] , in_channels=3 , out_channels=3 , down_block_types=[ 'DownEncoderBlock2D', 'DownEncoderBlock2D', 'DownEncoderBlock2D', 'DownEncoderBlock2D', ] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D', 'UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) SCREAMING_SNAKE_CASE_ : Optional[Any] = EulerDiscreteScheduler(prediction_type='sample' ) SCREAMING_SNAKE_CASE_ : List[str] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='quick_gelu' , projection_dim=5_1_2 , ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = CLIPTextModel(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Optional[int] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) SCREAMING_SNAKE_CASE_ : Tuple = { 'unet': model.eval(), 'vae': vae.eval(), 'scheduler': scheduler, 'text_encoder': text_encoder, 'tokenizer': tokenizer, } return components def UpperCamelCase__ ( self , lowerCAmelCase__ , lowerCAmelCase__=0 ): """simple docstring""" if str(lowerCAmelCase__ ).startswith('mps' ): SCREAMING_SNAKE_CASE_ : str = torch.manual_seed(lowerCAmelCase__ ) else: SCREAMING_SNAKE_CASE_ : Dict = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Optional[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': self.dummy_image.cpu(), 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = 'cpu' SCREAMING_SNAKE_CASE_ : List[str] = self.get_dummy_components() SCREAMING_SNAKE_CASE_ : Any = self.pipeline_class(**lowerCAmelCase__ ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Optional[int] = self.get_dummy_inputs(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = pipe(**lowerCAmelCase__ ).images SCREAMING_SNAKE_CASE_ : List[str] = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 2_5_6, 2_5_6, 3) ) SCREAMING_SNAKE_CASE_ : Optional[Any] = np.array( [0.47_222_412, 0.41_921_633, 0.44_717_434, 0.46_874_192, 0.42_588_258, 0.46_150_726, 0.4_677_534, 0.45_583_832, 0.48_579_055] ) SCREAMING_SNAKE_CASE_ : Tuple = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCAmelCase__ , 1E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" super().test_attention_slicing_forward_pass(expected_max_diff=7E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" super().test_cpu_offload_forward_pass(expected_max_diff=3E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=7E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" super().test_pt_np_pil_outputs_equivalent(expected_max_diff=3E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" super().test_save_load_local(expected_max_difference=3E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" super().test_save_load_optional_components(expected_max_difference=3E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = [ 'DDIMScheduler', 'DDPMScheduler', 'PNDMScheduler', 'HeunDiscreteScheduler', 'EulerAncestralDiscreteScheduler', 'KDPM2DiscreteScheduler', 'KDPM2AncestralDiscreteScheduler', 'DPMSolverSDEScheduler', ] SCREAMING_SNAKE_CASE_ : Dict = self.get_dummy_components() SCREAMING_SNAKE_CASE_ : List[Any] = self.pipeline_class(**lowerCAmelCase__ ) # make sure that PNDM does not need warm-up pipe.scheduler.register_to_config(skip_prk_steps=lowerCAmelCase__ ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : List[str] = self.get_dummy_inputs(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Any = 2 SCREAMING_SNAKE_CASE_ : List[str] = [] for scheduler_enum in KarrasDiffusionSchedulers: if scheduler_enum.name in skip_schedulers: # no sigma schedulers are not supported # no schedulers continue SCREAMING_SNAKE_CASE_ : str = getattr(lowerCAmelCase__ , scheduler_enum.name ) SCREAMING_SNAKE_CASE_ : Any = scheduler_cls.from_config(pipe.scheduler.config ) SCREAMING_SNAKE_CASE_ : Any = pipe(**lowerCAmelCase__ )[0] outputs.append(lowerCAmelCase__ ) assert check_same_shape(lowerCAmelCase__ ) @require_torch_gpu @slow class __lowercase (unittest.TestCase ): """simple docstring""" def UpperCamelCase__ ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = torch.manual_seed(3_3 ) SCREAMING_SNAKE_CASE_ : Any = StableDiffusionPipeline.from_pretrained('CompVis/stable-diffusion-v1-4' , torch_dtype=torch.floataa ) pipe.to('cuda' ) SCREAMING_SNAKE_CASE_ : Any = StableDiffusionLatentUpscalePipeline.from_pretrained( 'stabilityai/sd-x2-latent-upscaler' , torch_dtype=torch.floataa ) upscaler.to('cuda' ) SCREAMING_SNAKE_CASE_ : Optional[Any] = 'a photo of an astronaut high resolution, unreal engine, ultra realistic' SCREAMING_SNAKE_CASE_ : Optional[int] = pipe(lowerCAmelCase__ , generator=lowerCAmelCase__ , output_type='latent' ).images SCREAMING_SNAKE_CASE_ : List[Any] = upscaler( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , num_inference_steps=2_0 , guidance_scale=0 , generator=lowerCAmelCase__ , output_type='np' , ).images[0] SCREAMING_SNAKE_CASE_ : int = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/astronaut_1024.npy' ) assert np.abs((expected_image - image).mean() ) < 5E-2 def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = torch.manual_seed(3_3 ) SCREAMING_SNAKE_CASE_ : int = StableDiffusionLatentUpscalePipeline.from_pretrained( 'stabilityai/sd-x2-latent-upscaler' , torch_dtype=torch.floataa ) upscaler.to('cuda' ) SCREAMING_SNAKE_CASE_ : Any = 'the temple of fire by Ross Tran and Gerardo Dottori, oil on canvas' SCREAMING_SNAKE_CASE_ : List[Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_512.png' ) SCREAMING_SNAKE_CASE_ : Tuple = upscaler( prompt=lowerCAmelCase__ , image=lowerCAmelCase__ , num_inference_steps=2_0 , guidance_scale=0 , generator=lowerCAmelCase__ , output_type='np' , ).images[0] SCREAMING_SNAKE_CASE_ : List[str] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_1024.npy' ) assert np.abs((expected_image - image).max() ) < 5E-2
162
import math import unittest def a__ ( A__ ): assert isinstance(A__, A__ ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5, int(math.sqrt(A__ ) + 1 ), 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True class __lowercase (unittest.TestCase ): """simple docstring""" def UpperCamelCase__ ( self ): """simple docstring""" self.assertTrue(is_prime(2 ) ) self.assertTrue(is_prime(3 ) ) self.assertTrue(is_prime(5 ) ) self.assertTrue(is_prime(7 ) ) self.assertTrue(is_prime(1_1 ) ) self.assertTrue(is_prime(1_3 ) ) self.assertTrue(is_prime(1_7 ) ) self.assertTrue(is_prime(1_9 ) ) self.assertTrue(is_prime(2_3 ) ) self.assertTrue(is_prime(2_9 ) ) def UpperCamelCase__ ( self ): """simple docstring""" with self.assertRaises(lowerCAmelCase__ ): is_prime(-1_9 ) self.assertFalse( is_prime(0 ) , 'Zero doesn\'t have any positive factors, primes must have exactly two.' , ) self.assertFalse( is_prime(1 ) , 'One only has 1 positive factor, primes must have exactly two.' , ) self.assertFalse(is_prime(2 * 2 ) ) self.assertFalse(is_prime(2 * 3 ) ) self.assertFalse(is_prime(3 * 3 ) ) self.assertFalse(is_prime(3 * 5 ) ) self.assertFalse(is_prime(3 * 5 * 7 ) ) if __name__ == "__main__": unittest.main()
162
1
def _A ( SCREAMING_SNAKE_CASE : int = 1 , SCREAMING_SNAKE_CASE : int = 1_000 ): """simple docstring""" a__ : Union[str, Any] =1 a__ : Any =0 for divide_by_number in range(SCREAMING_SNAKE_CASE , digit + 1 ): a__ : list[int] =[] a__ : Tuple =numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(SCREAMING_SNAKE_CASE ): a__ : Tuple =len(SCREAMING_SNAKE_CASE ) a__ : str =divide_by_number else: has_been_divided.append(SCREAMING_SNAKE_CASE ) a__ : Dict =now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
95
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCAmelCase__ = { '''configuration_pix2struct''': [ '''PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Pix2StructConfig''', '''Pix2StructTextConfig''', '''Pix2StructVisionConfig''', ], '''processing_pix2struct''': ['''Pix2StructProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['''Pix2StructImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Pix2StructPreTrainedModel''', '''Pix2StructForConditionalGeneration''', '''Pix2StructVisionModel''', '''Pix2StructTextModel''', ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
108
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) lowerCAmelCase_ = { '''configuration_owlvit''': [ '''OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''OwlViTConfig''', '''OwlViTOnnxConfig''', '''OwlViTTextConfig''', '''OwlViTVisionConfig''', ], '''processing_owlvit''': ['''OwlViTProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ['''OwlViTFeatureExtractor'''] lowerCAmelCase_ = ['''OwlViTImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ '''OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''OwlViTModel''', '''OwlViTPreTrainedModel''', '''OwlViTTextModel''', '''OwlViTVisionModel''', '''OwlViTForObjectDetection''', ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
358
'''simple docstring''' import pytest lowerCAmelCase_ = "__dummy_dataset1__" lowerCAmelCase_ = "\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = \"https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/\"\nURLS = {\"train\": REPO_URL + \"wikiann-bn-train.jsonl\", \"validation\": REPO_URL + \"wikiann-bn-validation.jsonl\"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\n \"ner_tags\": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n \"O\",\n \"B-PER\",\n \"I-PER\",\n \"B-ORG\",\n \"I-ORG\",\n \"B-LOC\",\n \"I-LOC\",\n ]\n )\n ),\n \"langs\": datasets.Sequence(datasets.Value(\"string\")),\n \"spans\": datasets.Sequence(datasets.Value(\"string\")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={\"filepath\": dl_path[\"train\"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={\"filepath\": dl_path[\"validation\"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n" @pytest.fixture def __magic_name__ ( ) -> List[Any]: return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def __magic_name__ ( ) -> Union[str, Any]: return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def __magic_name__ ( A , A , A ) -> Optional[int]: snake_case = dataset_loading_script_name snake_case = tmp_path / 'datasets' / script_name script_dir.mkdir(parents=A ) snake_case = script_dir / F'''{script_name}.py''' with open(A , 'w' ) as f: f.write(A ) return str(A )
332
0
"""simple docstring""" import itertools import random import unittest import numpy as np from transformers import WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaConfig, WavaVecaFeatureExtractor from transformers.testing_utils import require_torch, slow from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin SCREAMING_SNAKE_CASE_ = random.Random() def lowercase (_lowerCAmelCase , _lowerCAmelCase=1.0 , _lowerCAmelCase=None , _lowerCAmelCase=None ): if rng is None: __lowerCAmelCase = global_rng __lowerCAmelCase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def __init__( self , snake_case_ , snake_case_=7 , snake_case_=400 , snake_case_=2_000 , snake_case_=1 , snake_case_=0.0 , snake_case_=16_000 , snake_case_=True , snake_case_=True , ) -> Tuple: __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = min_seq_length __lowerCAmelCase = max_seq_length __lowerCAmelCase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __lowerCAmelCase = feature_size __lowerCAmelCase = padding_value __lowerCAmelCase = sampling_rate __lowerCAmelCase = return_attention_mask __lowerCAmelCase = do_normalize def A__ ( self ) -> Optional[int]: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def A__ ( self , snake_case_=False , snake_case_=False ) -> Optional[int]: def _flatten(snake_case_ ): return list(itertools.chain(*snake_case_ ) ) if equal_length: __lowerCAmelCase = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size __lowerCAmelCase = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __lowerCAmelCase = [np.asarray(snake_case_ ) for x in speech_inputs] return speech_inputs class lowerCAmelCase_ ( A__ , unittest.TestCase ): '''simple docstring''' _snake_case = WavaVecaFeatureExtractor def A__ ( self ) -> Optional[Any]: __lowerCAmelCase = WavaVecaFeatureExtractionTester(self ) def A__ ( self , snake_case_ ) -> Dict: self.assertTrue(np.all(np.mean(snake_case_ , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(snake_case_ , axis=0 ) - 1 ) < 1e-3 ) ) def A__ ( self ) -> Optional[int]: # Tests that all call wrap to encode_plus and batch_encode_plus __lowerCAmelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __lowerCAmelCase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowerCAmelCase = [np.asarray(snake_case_ ) for speech_input in speech_inputs] # Test not batched input __lowerCAmelCase = feat_extract(speech_inputs[0] , return_tensors="""np""" ).input_values __lowerCAmelCase = feat_extract(np_speech_inputs[0] , return_tensors="""np""" ).input_values self.assertTrue(np.allclose(snake_case_ , snake_case_ , atol=1e-3 ) ) # Test batched __lowerCAmelCase = feat_extract(snake_case_ , return_tensors="""np""" ).input_values __lowerCAmelCase = feat_extract(snake_case_ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(snake_case_ , snake_case_ ): self.assertTrue(np.allclose(snake_case_ , snake_case_ , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __lowerCAmelCase = [floats_list((1, x) )[0] for x in (800, 800, 800)] __lowerCAmelCase = np.asarray(snake_case_ ) __lowerCAmelCase = feat_extract(snake_case_ , return_tensors="""np""" ).input_values __lowerCAmelCase = feat_extract(snake_case_ , return_tensors="""np""" ).input_values for enc_seq_a, enc_seq_a in zip(snake_case_ , snake_case_ ): self.assertTrue(np.allclose(snake_case_ , snake_case_ , atol=1e-3 ) ) def A__ ( self ) -> Dict: __lowerCAmelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowerCAmelCase = ["""longest""", """max_length""", """do_not_pad"""] __lowerCAmelCase = [None, 1_600, None] for max_length, padding in zip(snake_case_ , snake_case_ ): __lowerCAmelCase = feat_extract(snake_case_ , padding=snake_case_ , max_length=snake_case_ , return_tensors="""np""" ) __lowerCAmelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[1][:1_000] ) self.assertTrue(input_values[0][1_000:].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_values[2][:1_200] ) def A__ ( self ) -> List[str]: __lowerCAmelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase = range(800 , 1_400 , 200 ) __lowerCAmelCase = [floats_list((1, x) )[0] for x in lengths] __lowerCAmelCase = ["""longest""", """max_length""", """do_not_pad"""] __lowerCAmelCase = [None, 1_600, None] for max_length, padding in zip(snake_case_ , snake_case_ ): __lowerCAmelCase = feat_extract(snake_case_ , max_length=snake_case_ , padding=snake_case_ ) __lowerCAmelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1_000] ) self._check_zero_mean_unit_variance(input_values[2][:1_200] ) def A__ ( self ) -> Optional[int]: __lowerCAmelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowerCAmelCase = feat_extract( snake_case_ , truncation=snake_case_ , max_length=1_000 , padding="""max_length""" , return_tensors="""np""" ) __lowerCAmelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def A__ ( self ) -> Union[str, Any]: __lowerCAmelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowerCAmelCase = feat_extract( snake_case_ , truncation=snake_case_ , max_length=1_000 , padding="""longest""" , return_tensors="""np""" ) __lowerCAmelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1_000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1_000) ) __lowerCAmelCase = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] __lowerCAmelCase = feat_extract( snake_case_ , truncation=snake_case_ , max_length=2_000 , padding="""longest""" , return_tensors="""np""" ) __lowerCAmelCase = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1_000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1_200) ) @require_torch def A__ ( self ) -> Optional[int]: import torch __lowerCAmelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __lowerCAmelCase = np.random.rand(100 ).astype(np.floataa ) __lowerCAmelCase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __lowerCAmelCase = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) __lowerCAmelCase = feature_extractor.pad([{"""input_values""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) @slow @require_torch def A__ ( self ) -> List[Any]: # this test makes sure that models that are using # group norm don't have their feature extractor return the # attention_mask for model_id in WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST: __lowerCAmelCase = WavaVecaConfig.from_pretrained(snake_case_ ) __lowerCAmelCase = WavaVecaFeatureExtractor.from_pretrained(snake_case_ ) # only "layer" feature extraction norm should make use of # attention_mask self.assertEqual(feat_extract.return_attention_mask , config.feat_extract_norm == """layer""" )
301
"""simple docstring""" import inspect import tempfile import unittest from huggingface_hub import hf_hub_download from transformers import is_torch_available from transformers.testing_utils import is_flaky, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin SCREAMING_SNAKE_CASE_ = 1E-4 if is_torch_available(): import torch from transformers import AutoformerConfig, AutoformerForPrediction, AutoformerModel from transformers.models.autoformer.modeling_autoformer import AutoformerDecoder, AutoformerEncoder @require_torch class lowerCAmelCase_ : '''simple docstring''' def __init__( self , snake_case_ , snake_case_=16 , snake_case_=13 , snake_case_=7 , snake_case_=14 , snake_case_=10 , snake_case_=19 , snake_case_=5 , snake_case_=4 , snake_case_=True , snake_case_=16 , snake_case_=2 , snake_case_=4 , snake_case_=4 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=[1, 2, 3, 4, 5] , snake_case_=25 , snake_case_=5 , ) -> Tuple: __lowerCAmelCase = d_model __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = prediction_length __lowerCAmelCase = context_length __lowerCAmelCase = cardinality __lowerCAmelCase = num_time_features __lowerCAmelCase = lags_sequence __lowerCAmelCase = embedding_dimension __lowerCAmelCase = is_training __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = context_length __lowerCAmelCase = prediction_length + label_length __lowerCAmelCase = label_length __lowerCAmelCase = moving_average __lowerCAmelCase = autocorrelation_factor def A__ ( self ) -> List[Any]: return AutoformerConfig( d_model=self.d_model , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , prediction_length=self.prediction_length , context_length=self.context_length , label_length=self.label_length , lags_sequence=self.lags_sequence , num_time_features=self.num_time_features , num_static_categorical_features=1 , cardinality=[self.cardinality] , embedding_dimension=[self.embedding_dimension] , moving_average=self.moving_average , ) def A__ ( self , snake_case_ ) -> Any: __lowerCAmelCase = config.context_length + max(config.lags_sequence ) __lowerCAmelCase = ids_tensor([self.batch_size, 1] , config.cardinality[0] ) __lowerCAmelCase = floats_tensor([self.batch_size, _past_length, config.num_time_features] ) __lowerCAmelCase = floats_tensor([self.batch_size, _past_length] ) __lowerCAmelCase = floats_tensor([self.batch_size, _past_length] ) > 0.5 # decoder inputs __lowerCAmelCase = floats_tensor([self.batch_size, config.prediction_length, config.num_time_features] ) __lowerCAmelCase = floats_tensor([self.batch_size, config.prediction_length] ) __lowerCAmelCase = { """past_values""": past_values, """static_categorical_features""": static_categorical_features, """past_time_features""": past_time_features, """past_observed_mask""": past_observed_mask, """future_time_features""": future_time_features, """future_values""": future_values, } return inputs_dict def A__ ( self ) -> Union[str, Any]: __lowerCAmelCase = self.get_config() __lowerCAmelCase = self.prepare_autoformer_inputs_dict(snake_case_ ) return config, inputs_dict def A__ ( self ) -> int: __lowerCAmelCase , __lowerCAmelCase = self.prepare_config_and_inputs() return config, inputs_dict def A__ ( self , snake_case_ , snake_case_ ) -> int: __lowerCAmelCase = AutoformerModel(config=snake_case_ ).to(snake_case_ ).eval() __lowerCAmelCase = model(**snake_case_ ) __lowerCAmelCase = outputs.encoder_last_hidden_state __lowerCAmelCase = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase = model.get_encoder() encoder.save_pretrained(snake_case_ ) __lowerCAmelCase = AutoformerEncoder.from_pretrained(snake_case_ ).to(snake_case_ ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = model.create_network_inputs(**snake_case_ ) __lowerCAmelCase , __lowerCAmelCase = model.decomposition_layer(transformer_inputs[:, : config.context_length, ...] ) __lowerCAmelCase = torch.cat( (transformer_inputs[:, : config.context_length, ...], feature[:, : config.context_length, ...]) , dim=-1 , ) __lowerCAmelCase = encoder(inputs_embeds=snake_case_ )[0] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1e-3 ) __lowerCAmelCase = ( torch.mean(transformer_inputs[:, : config.context_length, ...] , dim=1 ) .unsqueeze(1 ) .repeat(1 , config.prediction_length , 1 ) ) __lowerCAmelCase = torch.zeros( [transformer_inputs.shape[0], config.prediction_length, transformer_inputs.shape[2]] , device=enc_input.device , ) __lowerCAmelCase = torch.cat( ( torch.cat((seasonal_input[:, -config.label_length :, ...], zeros) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) __lowerCAmelCase = torch.cat( ( torch.cat((trend_input[:, -config.label_length :, ...], mean) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) with tempfile.TemporaryDirectory() as tmpdirname: __lowerCAmelCase = model.get_decoder() decoder.save_pretrained(snake_case_ ) __lowerCAmelCase = AutoformerDecoder.from_pretrained(snake_case_ ).to(snake_case_ ) __lowerCAmelCase = decoder( trend=snake_case_ , inputs_embeds=snake_case_ , encoder_hidden_states=snake_case_ , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1e-3 ) @require_torch class lowerCAmelCase_ ( A__ , A__ , unittest.TestCase ): '''simple docstring''' _snake_case = (AutoformerModel, AutoformerForPrediction) if is_torch_available() else () _snake_case = (AutoformerForPrediction,) if is_torch_available() else () _snake_case = {'''feature-extraction''': AutoformerModel} if is_torch_available() else {} _snake_case = False _snake_case = False _snake_case = False _snake_case = False _snake_case = False _snake_case = False def A__ ( self ) -> Optional[int]: __lowerCAmelCase = AutoformerModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_ ) def A__ ( self ) -> Optional[int]: self.config_tester.run_common_tests() def A__ ( self ) -> Dict: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: __lowerCAmelCase = model_class(snake_case_ ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(snake_case_ ) __lowerCAmelCase , __lowerCAmelCase = model_class.from_pretrained(snake_case_ , output_loading_info=snake_case_ ) self.assertEqual(info["""missing_keys"""] , [] ) def A__ ( self ) -> Optional[Any]: __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*snake_case_ ) @unittest.skip(reason="""Model has no tokens embeddings""" ) def A__ ( self ) -> Any: pass def A__ ( self ) -> str: __lowerCAmelCase = inspect.signature(getattr(snake_case_ , """forward""" ) ) # The main input is the name of the argument after `self` __lowerCAmelCase = list(model_signature.parameters.keys() )[1] self.assertEqual(AutoformerModel.main_input_name , snake_case_ ) def A__ ( self ) -> Any: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowerCAmelCase = model_class(snake_case_ ) __lowerCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowerCAmelCase = [*signature.parameters.keys()] __lowerCAmelCase = [ """past_values""", """past_time_features""", """past_observed_mask""", """static_categorical_features""", """static_real_features""", """future_values""", """future_time_features""", ] if model.__class__.__name__ in ["AutoformerForPrediction"]: expected_arg_names.append("""future_observed_mask""" ) expected_arg_names.extend( [ """decoder_attention_mask""", """head_mask""", """decoder_head_mask""", """cross_attn_head_mask""", """encoder_outputs""", """past_key_values""", """output_hidden_states""", """output_attentions""", """use_cache""", """return_dict""", ] ) self.assertListEqual(arg_names[: len(snake_case_ )] , snake_case_ ) def A__ ( self ) -> Optional[int]: __lowerCAmelCase , __lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() __lowerCAmelCase = True __lowerCAmelCase = getattr(self.model_tester , """seq_length""" , snake_case_ ) __lowerCAmelCase = getattr(self.model_tester , """decoder_seq_length""" , snake_case_ ) __lowerCAmelCase = getattr(self.model_tester , """encoder_seq_length""" , snake_case_ ) __lowerCAmelCase = getattr(self.model_tester , """d_model""" , snake_case_ ) __lowerCAmelCase = getattr(self.model_tester , """num_attention_heads""" , snake_case_ ) __lowerCAmelCase = d_model // num_attention_heads for model_class in self.all_model_classes: __lowerCAmelCase = True __lowerCAmelCase = False __lowerCAmelCase = True __lowerCAmelCase = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): __lowerCAmelCase = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) __lowerCAmelCase = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __lowerCAmelCase = True __lowerCAmelCase = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): __lowerCAmelCase = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) __lowerCAmelCase = outputs.encoder_attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) __lowerCAmelCase = len(snake_case_ ) __lowerCAmelCase = 7 if "last_hidden_state" in outputs: correct_outlen += 1 if "trend" in outputs: correct_outlen += 1 if "past_key_values" in outputs: correct_outlen += 1 # past_key_values have been returned if "loss" in outputs: correct_outlen += 1 if "params" in outputs: correct_outlen += 1 self.assertEqual(snake_case_ , snake_case_ ) # decoder attentions __lowerCAmelCase = outputs.decoder_attentions self.assertIsInstance(snake_case_ , (list, tuple) ) self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # cross attentions __lowerCAmelCase = outputs.cross_attentions self.assertIsInstance(snake_case_ , (list, tuple) ) self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(cross_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # Check attention is always last and order is fine __lowerCAmelCase = True __lowerCAmelCase = True __lowerCAmelCase = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): __lowerCAmelCase = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(out_len + 2 , len(snake_case_ ) ) __lowerCAmelCase = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) @is_flaky() def A__ ( self ) -> int: super().test_retain_grad_hidden_states_attentions() def lowercase (_lowerCAmelCase="train-batch.pt" ): __lowerCAmelCase = hf_hub_download(repo_id="""hf-internal-testing/tourism-monthly-batch""" , filename=_lowerCAmelCase , repo_type="""dataset""" ) __lowerCAmelCase = torch.load(_lowerCAmelCase , map_location=_lowerCAmelCase ) return batch @require_torch @slow class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def A__ ( self ) -> int: __lowerCAmelCase = AutoformerModel.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(snake_case_ ) __lowerCAmelCase = prepare_batch() with torch.no_grad(): __lowerCAmelCase = model( past_values=batch["""past_values"""] , past_time_features=batch["""past_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , static_categorical_features=batch["""static_categorical_features"""] , future_values=batch["""future_values"""] , future_time_features=batch["""future_time_features"""] , )[0] __lowerCAmelCase = torch.Size( (64, model.config.prediction_length + model.config.label_length, model.config.feature_size) ) self.assertEqual(output.shape , snake_case_ ) __lowerCAmelCase = torch.tensor( [[0.3_593, -1.3_398, 0.6_330], [0.2_279, 1.5_396, -0.1_792], [0.0_450, 1.3_225, -0.2_335]] , device=snake_case_ ) self.assertTrue(torch.allclose(output[0, :3, :3] , snake_case_ , atol=snake_case_ ) ) def A__ ( self ) -> List[str]: __lowerCAmelCase = AutoformerForPrediction.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(snake_case_ ) __lowerCAmelCase = prepare_batch("""val-batch.pt""" ) with torch.no_grad(): __lowerCAmelCase = model( past_values=batch["""past_values"""] , past_time_features=batch["""past_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , static_categorical_features=batch["""static_categorical_features"""] , ).encoder_last_hidden_state __lowerCAmelCase = torch.Size((64, model.config.context_length, model.config.d_model) ) self.assertEqual(output.shape , snake_case_ ) __lowerCAmelCase = torch.tensor( [[-0.0_734, -0.9_036, 0.8_358], [4.7_186, 2.4_113, 1.9_581], [1.7_953, 2.3_558, 1.2_970]] , device=snake_case_ ) self.assertTrue(torch.allclose(output[0, :3, :3] , snake_case_ , atol=snake_case_ ) ) def A__ ( self ) -> Any: __lowerCAmelCase = AutoformerForPrediction.from_pretrained("""huggingface/autoformer-tourism-monthly""" ).to(snake_case_ ) __lowerCAmelCase = prepare_batch("""val-batch.pt""" ) with torch.no_grad(): __lowerCAmelCase = model.generate( static_categorical_features=batch["""static_categorical_features"""] , past_time_features=batch["""past_time_features"""] , past_values=batch["""past_values"""] , future_time_features=batch["""future_time_features"""] , past_observed_mask=batch["""past_observed_mask"""] , ) __lowerCAmelCase = torch.Size((64, model.config.num_parallel_samples, model.config.prediction_length) ) self.assertEqual(outputs.sequences.shape , snake_case_ ) __lowerCAmelCase = torch.tensor([3_130.6_763, 4_056.5_293, 7_053.0_786] , device=snake_case_ ) __lowerCAmelCase = outputs.sequences.mean(dim=1 ) self.assertTrue(torch.allclose(mean_prediction[0, -3:] , snake_case_ , rtol=1e-1 ) )
301
1
import multiprocessing from typing import TYPE_CHECKING, Optional, Union from .. import Dataset, Features, config from ..formatting import query_table from ..packaged_modules.sql.sql import Sql from ..utils import logging from .abc import AbstractDatasetInputStream if TYPE_CHECKING: import sqlitea import sqlalchemy class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , lowerCamelCase : Union[str, "sqlalchemy.sql.Selectable"] , lowerCamelCase : Union[str, "sqlalchemy.engine.Connection", "sqlalchemy.engine.Engine", "sqlite3.Connection"] , lowerCamelCase : Optional[Features] = None , lowerCamelCase : str = None , lowerCamelCase : bool = False , **lowerCamelCase : Optional[int] , ) -> List[Any]: super().__init__(features=lowerCamelCase , cache_dir=lowerCamelCase , keep_in_memory=lowerCamelCase , **lowerCamelCase ) __snake_case : Union[str, Any] = Sql( cache_dir=lowerCamelCase , features=lowerCamelCase , sql=lowerCamelCase , con=lowerCamelCase , **lowerCamelCase , ) def __snake_case ( self : Dict ) -> List[Any]: __snake_case : Dict = None __snake_case : List[str] = None __snake_case : List[str] = None __snake_case : str = None self.builder.download_and_prepare( download_config=lowerCamelCase , download_mode=lowerCamelCase , verification_mode=lowerCamelCase , base_path=lowerCamelCase , ) # Build dataset for splits __snake_case : List[str] = self.builder.as_dataset( split="train" , verification_mode=lowerCamelCase , in_memory=self.keep_in_memory ) return dataset class a : """simple docstring""" def __init__( self : Optional[Any] , lowerCamelCase : Dataset , lowerCamelCase : str , lowerCamelCase : Union[str, "sqlalchemy.engine.Connection", "sqlalchemy.engine.Engine", "sqlite3.Connection"] , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[int] = None , **lowerCamelCase : str , ) -> Optional[int]: if num_proc is not None and num_proc <= 0: raise ValueError(F'num_proc {num_proc} must be an integer > 0.' ) __snake_case : List[Any] = dataset __snake_case : int = name __snake_case : int = con __snake_case : Dict = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE __snake_case : Optional[Any] = num_proc __snake_case : str = to_sql_kwargs def __snake_case ( self : Any ) -> int: __snake_case : Tuple = self.to_sql_kwargs.pop("sql" , lowerCamelCase ) __snake_case : Any = self.to_sql_kwargs.pop("con" , lowerCamelCase ) __snake_case : Optional[Any] = self.to_sql_kwargs.pop("index" , lowerCamelCase ) __snake_case : List[Any] = self._write(index=lowerCamelCase , **self.to_sql_kwargs ) return written def __snake_case ( self : Optional[int] , lowerCamelCase : Any ) -> List[Any]: __snake_case , __snake_case , __snake_case : List[str] = args __snake_case : Any = {**to_sql_kwargs, "if_exists": "append"} if offset > 0 else to_sql_kwargs __snake_case : str = query_table( table=self.dataset.data , key=slice(lowerCamelCase , offset + self.batch_size ) , indices=self.dataset._indices , ) __snake_case : Tuple = batch.to_pandas() __snake_case : Dict = df.to_sql(self.name , self.con , index=lowerCamelCase , **lowerCamelCase ) return num_rows or len(lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : List[str] , **lowerCamelCase : Any ) -> int: __snake_case : Dict = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 , len(self.dataset ) , self.batch_size ) , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating SQL from Arrow format" , ): written += self._batch_sql((offset, index, to_sql_kwargs) ) else: __snake_case , __snake_case : Optional[Any] = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for num_rows in logging.tqdm( pool.imap( self._batch_sql , [(offset, index, to_sql_kwargs) for offset in range(0 , lowerCamelCase , lowerCamelCase )] , ) , total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size , unit="ba" , disable=not logging.is_progress_bar_enabled() , desc="Creating SQL from Arrow format" , ): written += num_rows return written
134
from typing import List, Optional, Union import torch from transformers import ( XLMRobertaTokenizer, ) from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) from .text_encoder import MultilingualCLIP _snake_case : Optional[int] = logging.get_logger(__name__) # pylint: disable=invalid-name _snake_case : Union[str, Any] = "\n Examples:\n ```py\n >>> from diffusers import KandinskyPipeline, KandinskyPriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyPriorPipeline.from_pretrained(\"kandinsky-community/Kandinsky-2-1-prior\")\n >>> pipe_prior.to(\"cuda\")\n\n >>> prompt = \"red cat, 4k photo\"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> negative_image_emb = out.negative_image_embeds\n\n >>> pipe = KandinskyPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-1\")\n >>> pipe.to(\"cuda\")\n\n >>> image = pipe(\n ... prompt,\n ... image_embeds=image_emb,\n ... negative_image_embeds=negative_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=100,\n ... ).images\n\n >>> image[0].save(\"cat.png\")\n ```\n" def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=8 ): __snake_case : List[Any] = h // scale_factor**2 if h % scale_factor**2 != 0: new_h += 1 __snake_case : Optional[int] = w // scale_factor**2 if w % scale_factor**2 != 0: new_w += 1 return new_h * scale_factor, new_w * scale_factor class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Union[str, Any] , lowerCamelCase : MultilingualCLIP , lowerCamelCase : XLMRobertaTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, DDPMScheduler] , lowerCamelCase : VQModel , ) -> Optional[int]: super().__init__() self.register_modules( text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , movq=lowerCamelCase , ) __snake_case : List[Any] = 2 ** (len(self.movq.config.block_out_channels ) - 1) def __snake_case ( self : Any , lowerCamelCase : Dict , lowerCamelCase : List[Any] , lowerCamelCase : Optional[int] , lowerCamelCase : Union[str, Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : int ) -> Any: if latents is None: __snake_case : str = randn_tensor(lowerCamelCase , generator=lowerCamelCase , device=lowerCamelCase , dtype=lowerCamelCase ) else: if latents.shape != shape: raise ValueError(F'Unexpected latents shape, got {latents.shape}, expected {shape}' ) __snake_case : Optional[int] = latents.to(lowerCamelCase ) __snake_case : List[Any] = latents * scheduler.init_noise_sigma return latents def __snake_case ( self : Optional[int] , lowerCamelCase : int , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : List[Any] , lowerCamelCase : str=None , ) -> List[str]: __snake_case : Tuple = len(lowerCamelCase ) if isinstance(lowerCamelCase , lowerCamelCase ) else 1 # get prompt text embeddings __snake_case : Optional[int] = self.tokenizer( lowerCamelCase , padding="max_length" , truncation=lowerCamelCase , max_length=77 , return_attention_mask=lowerCamelCase , add_special_tokens=lowerCamelCase , return_tensors="pt" , ) __snake_case : List[str] = text_inputs.input_ids __snake_case : List[Any] = self.tokenizer(lowerCamelCase , padding="longest" , return_tensors="pt" ).input_ids if untruncated_ids.shape[-1] >= text_input_ids.shape[-1] and not torch.equal(lowerCamelCase , lowerCamelCase ): __snake_case : Optional[Any] = self.tokenizer.batch_decode(untruncated_ids[:, self.tokenizer.model_max_length - 1 : -1] ) logger.warning( "The following part of your input was truncated because CLIP can only handle sequences up to" F' {self.tokenizer.model_max_length} tokens: {removed_text}' ) __snake_case : Any = text_input_ids.to(lowerCamelCase ) __snake_case : List[str] = text_inputs.attention_mask.to(lowerCamelCase ) __snake_case , __snake_case : List[str] = self.text_encoder( input_ids=lowerCamelCase , attention_mask=lowerCamelCase ) __snake_case : List[Any] = prompt_embeds.repeat_interleave(lowerCamelCase , dim=0 ) __snake_case : List[str] = text_encoder_hidden_states.repeat_interleave(lowerCamelCase , dim=0 ) __snake_case : Optional[int] = text_mask.repeat_interleave(lowerCamelCase , dim=0 ) if do_classifier_free_guidance: __snake_case : List[str] if negative_prompt is None: __snake_case : Any = [""] * batch_size elif type(lowerCamelCase ) is not type(lowerCamelCase ): raise TypeError( F'`negative_prompt` should be the same type to `prompt`, but got {type(lowerCamelCase )} !=' F' {type(lowerCamelCase )}.' ) elif isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : List[Any] = [negative_prompt] elif batch_size != len(lowerCamelCase ): raise ValueError( F'`negative_prompt`: {negative_prompt} has batch size {len(lowerCamelCase )}, but `prompt`:' F' {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches' " the batch size of `prompt`." ) else: __snake_case : int = negative_prompt __snake_case : Dict = self.tokenizer( lowerCamelCase , padding="max_length" , max_length=77 , truncation=lowerCamelCase , return_attention_mask=lowerCamelCase , add_special_tokens=lowerCamelCase , return_tensors="pt" , ) __snake_case : Dict = uncond_input.input_ids.to(lowerCamelCase ) __snake_case : List[Any] = uncond_input.attention_mask.to(lowerCamelCase ) __snake_case , __snake_case : Tuple = self.text_encoder( input_ids=lowerCamelCase , attention_mask=lowerCamelCase ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method __snake_case : Dict = negative_prompt_embeds.shape[1] __snake_case : int = negative_prompt_embeds.repeat(1 , lowerCamelCase ) __snake_case : List[str] = negative_prompt_embeds.view(batch_size * num_images_per_prompt , lowerCamelCase ) __snake_case : Union[str, Any] = uncond_text_encoder_hidden_states.shape[1] __snake_case : Tuple = uncond_text_encoder_hidden_states.repeat(1 , lowerCamelCase , 1 ) __snake_case : str = uncond_text_encoder_hidden_states.view( batch_size * num_images_per_prompt , lowerCamelCase , -1 ) __snake_case : Optional[int] = uncond_text_mask.repeat_interleave(lowerCamelCase , dim=0 ) # done duplicates # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes __snake_case : Optional[int] = torch.cat([negative_prompt_embeds, prompt_embeds] ) __snake_case : List[Any] = torch.cat([uncond_text_encoder_hidden_states, text_encoder_hidden_states] ) __snake_case : Any = torch.cat([uncond_text_mask, text_mask] ) return prompt_embeds, text_encoder_hidden_states, text_mask def __snake_case ( self : List[str] , lowerCamelCase : Dict=0 ) -> Tuple: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device(F'cuda:{gpu_id}' ) __snake_case : Optional[Any] = [ self.unet, self.text_encoder, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : List[Any] , lowerCamelCase : int=0 ) -> Optional[int]: if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) __snake_case : Optional[Any] = torch.device(F'cuda:{gpu_id}' ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=lowerCamelCase ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __snake_case : List[str] = None for cpu_offloaded_model in [self.text_encoder, self.unet, self.movq]: __snake_case , __snake_case : List[Any] = cpu_offload_with_hook(lowerCamelCase , lowerCamelCase , prev_module_hook=lowerCamelCase ) if self.safety_checker is not None: __snake_case , __snake_case : Optional[int] = cpu_offload_with_hook(self.safety_checker , lowerCamelCase , prev_module_hook=lowerCamelCase ) # We'll offload the last model manually. __snake_case : str = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : List[Any] ) -> Optional[int]: if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(lowerCamelCase ) def __call__( self : Dict , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, List[torch.FloatTensor]] , lowerCamelCase : Union[torch.FloatTensor, List[torch.FloatTensor]] , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 100 , lowerCamelCase : float = 4.0 , lowerCamelCase : int = 1 , lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Optional[int] = 1 elif isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : List[Any] = len(lowerCamelCase ) else: raise ValueError(F'`prompt` has to be of type `str` or `list` but is {type(lowerCamelCase )}' ) __snake_case : Any = self._execution_device __snake_case : Any = batch_size * num_images_per_prompt __snake_case : Any = guidance_scale > 1.0 __snake_case , __snake_case , __snake_case : Optional[Any] = self._encode_prompt( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : List[Any] = torch.cat(lowerCamelCase , dim=0 ) if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : str = torch.cat(lowerCamelCase , dim=0 ) if do_classifier_free_guidance: __snake_case : Dict = image_embeds.repeat_interleave(lowerCamelCase , dim=0 ) __snake_case : Optional[Any] = negative_image_embeds.repeat_interleave(lowerCamelCase , dim=0 ) __snake_case : str = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to( dtype=prompt_embeds.dtype , device=lowerCamelCase ) self.scheduler.set_timesteps(lowerCamelCase , device=lowerCamelCase ) __snake_case : Tuple = self.scheduler.timesteps __snake_case : Union[str, Any] = self.unet.config.in_channels __snake_case , __snake_case : Tuple = get_new_h_w(lowerCamelCase , lowerCamelCase , self.movq_scale_factor ) # create initial latent __snake_case : Any = self.prepare_latents( (batch_size, num_channels_latents, height, width) , text_encoder_hidden_states.dtype , lowerCamelCase , lowerCamelCase , lowerCamelCase , self.scheduler , ) for i, t in enumerate(self.progress_bar(lowerCamelCase ) ): # expand the latents if we are doing classifier free guidance __snake_case : Any = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __snake_case : int = {"text_embeds": prompt_embeds, "image_embeds": image_embeds} __snake_case : Optional[Any] = self.unet( sample=lowerCamelCase , timestep=lowerCamelCase , encoder_hidden_states=lowerCamelCase , added_cond_kwargs=lowerCamelCase , return_dict=lowerCamelCase , )[0] if do_classifier_free_guidance: __snake_case , __snake_case : Any = noise_pred.split(latents.shape[1] , dim=1 ) __snake_case , __snake_case : Union[str, Any] = noise_pred.chunk(2 ) __snake_case , __snake_case : str = variance_pred.chunk(2 ) __snake_case : Optional[Any] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __snake_case : Union[str, Any] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __snake_case , __snake_case : Union[str, Any] = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __snake_case : str = self.scheduler.step( lowerCamelCase , lowerCamelCase , lowerCamelCase , generator=lowerCamelCase , ).prev_sample # post-processing __snake_case : str = self.movq.decode(lowerCamelCase , force_not_quantize=lowerCamelCase )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(F'Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}' ) if output_type in ["np", "pil"]: __snake_case : Union[str, Any] = image * 0.5 + 0.5 __snake_case : Union[str, Any] = image.clamp(0 , 1 ) __snake_case : Optional[int] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __snake_case : str = self.numpy_to_pil(lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase )
134
1
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING A__ : str =logging.get_logger(__name__) A__ : int ={ '''microsoft/table-transformer-detection''': ( '''https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json''' ), } class UpperCAmelCase ( snake_case_ ): _lowercase: Union[str, Any] = '''table-transformer''' _lowercase: str = ['''past_key_values'''] _lowercase: Dict = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', } def __init__( self : Dict , __snake_case : int=True , __snake_case : Optional[Any]=None , __snake_case : Optional[int]=3 , __snake_case : int=1_00 , __snake_case : Optional[Any]=6 , __snake_case : Any=20_48 , __snake_case : Optional[Any]=8 , __snake_case : int=6 , __snake_case : int=20_48 , __snake_case : Optional[int]=8 , __snake_case : Dict=0.0 , __snake_case : Optional[Any]=0.0 , __snake_case : Union[str, Any]=True , __snake_case : Any="relu" , __snake_case : Union[str, Any]=2_56 , __snake_case : Optional[int]=0.1 , __snake_case : List[str]=0.0 , __snake_case : List[str]=0.0 , __snake_case : Optional[int]=0.02 , __snake_case : Dict=1.0 , __snake_case : Optional[Any]=False , __snake_case : Tuple="sine" , __snake_case : List[Any]="resnet50" , __snake_case : int=True , __snake_case : int=False , __snake_case : str=1 , __snake_case : Optional[Any]=5 , __snake_case : List[str]=2 , __snake_case : Optional[Any]=1 , __snake_case : Union[str, Any]=1 , __snake_case : Optional[int]=5 , __snake_case : int=2 , __snake_case : int=0.1 , **__snake_case : int , ) -> Dict: if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) _lowerCAmelCase = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(__snake_case , __snake_case ): _lowerCAmelCase = backbone_config.get("""model_type""" ) _lowerCAmelCase = CONFIG_MAPPING[backbone_model_type] _lowerCAmelCase = config_class.from_dict(__snake_case ) # set timm attributes to None _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = None, None, None _lowerCAmelCase = use_timm_backbone _lowerCAmelCase = backbone_config _lowerCAmelCase = num_channels _lowerCAmelCase = num_queries _lowerCAmelCase = d_model _lowerCAmelCase = encoder_ffn_dim _lowerCAmelCase = encoder_layers _lowerCAmelCase = encoder_attention_heads _lowerCAmelCase = decoder_ffn_dim _lowerCAmelCase = decoder_layers _lowerCAmelCase = decoder_attention_heads _lowerCAmelCase = dropout _lowerCAmelCase = attention_dropout _lowerCAmelCase = activation_dropout _lowerCAmelCase = activation_function _lowerCAmelCase = init_std _lowerCAmelCase = init_xavier_std _lowerCAmelCase = encoder_layerdrop _lowerCAmelCase = decoder_layerdrop _lowerCAmelCase = encoder_layers _lowerCAmelCase = auxiliary_loss _lowerCAmelCase = position_embedding_type _lowerCAmelCase = backbone _lowerCAmelCase = use_pretrained_backbone _lowerCAmelCase = dilation # Hungarian matcher _lowerCAmelCase = class_cost _lowerCAmelCase = bbox_cost _lowerCAmelCase = giou_cost # Loss coefficients _lowerCAmelCase = mask_loss_coefficient _lowerCAmelCase = dice_loss_coefficient _lowerCAmelCase = bbox_loss_coefficient _lowerCAmelCase = giou_loss_coefficient _lowerCAmelCase = eos_coefficient super().__init__(is_encoder_decoder=__snake_case , **__snake_case ) @property def lowercase__ ( self : Optional[int] ) -> int: return self.encoder_attention_heads @property def lowercase__ ( self : Dict ) -> int: return self.d_model class UpperCAmelCase ( snake_case_ ): _lowercase: List[Any] = version.parse('''1.11''' ) @property def lowercase__ ( self : Tuple ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def lowercase__ ( self : Optional[Any] ) -> float: return 1E-5 @property def lowercase__ ( self : Optional[int] ) -> int: return 12
70
'''simple docstring''' import heapq as hq import math from collections.abc import Iterator class UpperCAmelCase : def __init__( self : str , __snake_case : Any ) -> str: _lowerCAmelCase = str(id_ ) _lowerCAmelCase = None _lowerCAmelCase = None _lowerCAmelCase = [] _lowerCAmelCase = {} # {vertex:distance} def __lt__( self : List[str] , __snake_case : Union[str, Any] ) -> Any: return self.key < other.key def __repr__( self : Optional[Any] ) -> Optional[Any]: return self.id def lowercase__ ( self : Union[str, Any] , __snake_case : Tuple ) -> Optional[Any]: self.neighbors.append(__snake_case ) def lowercase__ ( self : Tuple , __snake_case : List[str] , __snake_case : Tuple ) -> Any: _lowerCAmelCase = weight def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , lowerCAmelCase ) graph[b - 1].add_edge(graph[a - 1] , lowerCAmelCase ) def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = [] for u in graph: _lowerCAmelCase = math.inf _lowerCAmelCase = None _lowerCAmelCase = 0 _lowerCAmelCase = graph[:] while q: _lowerCAmelCase = min(lowerCAmelCase ) q.remove(lowerCAmelCase ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): _lowerCAmelCase = u _lowerCAmelCase = u.edges[v.id] for i in range(1 , len(lowerCAmelCase ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase ): """simple docstring""" for u in graph: _lowerCAmelCase = math.inf _lowerCAmelCase = None _lowerCAmelCase = 0 _lowerCAmelCase = list(lowerCAmelCase ) hq.heapify(lowerCAmelCase ) while h: _lowerCAmelCase = hq.heappop(lowerCAmelCase ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): _lowerCAmelCase = u _lowerCAmelCase = u.edges[v.id] hq.heapify(lowerCAmelCase ) for i in range(1 , len(lowerCAmelCase ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def UpperCamelCase__ ( ): """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
70
1
"""simple docstring""" import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class a ( a__ ): snake_case__ = (DEISMultistepScheduler,) snake_case__ = (('''num_inference_steps''', 2_5),) def UpperCamelCase__ ( self , **_snake_case ): """simple docstring""" lowerCAmelCase = { 'num_train_timesteps': 10_00, 'beta_start': 0.0_001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'solver_order': 2, } config.update(**_snake_case ) return config def UpperCamelCase__ ( self , _snake_case=0 , **_snake_case ): """simple docstring""" lowerCAmelCase = dict(self.forward_default_kwargs ) lowerCAmelCase = kwargs.pop('num_inference_steps' , _snake_case ) lowerCAmelCase = self.dummy_sample lowerCAmelCase = 0.1 * sample lowerCAmelCase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowerCAmelCase = self.get_scheduler_config(**_snake_case ) lowerCAmelCase = scheduler_class(**_snake_case ) scheduler.set_timesteps(_snake_case ) # copy over dummy past residuals lowerCAmelCase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_snake_case ) lowerCAmelCase = scheduler_class.from_pretrained(_snake_case ) new_scheduler.set_timesteps(_snake_case ) # copy over dummy past residuals lowerCAmelCase = dummy_past_residuals[: new_scheduler.config.solver_order] lowerCAmelCase ,lowerCAmelCase = sample, sample for t in range(_snake_case , time_step + scheduler.config.solver_order + 1 ): lowerCAmelCase = scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample lowerCAmelCase = new_scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self , _snake_case=0 , **_snake_case ): """simple docstring""" lowerCAmelCase = dict(self.forward_default_kwargs ) lowerCAmelCase = kwargs.pop('num_inference_steps' , _snake_case ) lowerCAmelCase = self.dummy_sample lowerCAmelCase = 0.1 * sample lowerCAmelCase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: lowerCAmelCase = self.get_scheduler_config() lowerCAmelCase = scheduler_class(**_snake_case ) scheduler.set_timesteps(_snake_case ) # copy over dummy past residuals (must be after setting timesteps) lowerCAmelCase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_snake_case ) lowerCAmelCase = scheduler_class.from_pretrained(_snake_case ) # copy over dummy past residuals new_scheduler.set_timesteps(_snake_case ) # copy over dummy past residual (must be after setting timesteps) lowerCAmelCase = dummy_past_residuals[: new_scheduler.config.solver_order] lowerCAmelCase = scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample lowerCAmelCase = new_scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCamelCase__ ( self , _snake_case=None , **_snake_case ): """simple docstring""" if scheduler is None: lowerCAmelCase = self.scheduler_classes[0] lowerCAmelCase = self.get_scheduler_config(**_snake_case ) lowerCAmelCase = scheduler_class(**_snake_case ) lowerCAmelCase = self.scheduler_classes[0] lowerCAmelCase = self.get_scheduler_config(**_snake_case ) lowerCAmelCase = scheduler_class(**_snake_case ) lowerCAmelCase = 10 lowerCAmelCase = self.dummy_model() lowerCAmelCase = self.dummy_sample_deter scheduler.set_timesteps(_snake_case ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase = model(_snake_case , _snake_case ) lowerCAmelCase = scheduler.step(_snake_case , _snake_case , _snake_case ).prev_sample return sample def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = dict(self.forward_default_kwargs ) lowerCAmelCase = kwargs.pop('num_inference_steps' , _snake_case ) for scheduler_class in self.scheduler_classes: lowerCAmelCase = self.get_scheduler_config() lowerCAmelCase = scheduler_class(**_snake_case ) lowerCAmelCase = self.dummy_sample lowerCAmelCase = 0.1 * sample if num_inference_steps is not None and hasattr(_snake_case , 'set_timesteps' ): scheduler.set_timesteps(_snake_case ) elif num_inference_steps is not None and not hasattr(_snake_case , 'set_timesteps' ): lowerCAmelCase = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) lowerCAmelCase = [residual + 0.2, residual + 0.15, residual + 0.10] lowerCAmelCase = dummy_past_residuals[: scheduler.config.solver_order] lowerCAmelCase = scheduler.timesteps[5] lowerCAmelCase = scheduler.timesteps[6] lowerCAmelCase = scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample lowerCAmelCase = scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = DEISMultistepScheduler(**self.get_scheduler_config() ) lowerCAmelCase = self.full_loop(scheduler=_snake_case ) lowerCAmelCase = torch.mean(torch.abs(_snake_case ) ) assert abs(result_mean.item() - 0.23_916 ) < 1E-3 lowerCAmelCase = DPMSolverSinglestepScheduler.from_config(scheduler.config ) lowerCAmelCase = DPMSolverMultistepScheduler.from_config(scheduler.config ) lowerCAmelCase = UniPCMultistepScheduler.from_config(scheduler.config ) lowerCAmelCase = DEISMultistepScheduler.from_config(scheduler.config ) lowerCAmelCase = self.full_loop(scheduler=_snake_case ) lowerCAmelCase = torch.mean(torch.abs(_snake_case ) ) assert abs(result_mean.item() - 0.23_916 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" for timesteps in [25, 50, 1_00, 9_99, 10_00]: self.check_over_configs(num_train_timesteps=_snake_case ) def UpperCamelCase__ ( self ): """simple docstring""" self.check_over_configs(thresholding=_snake_case ) for order in [1, 2, 3]: for solver_type in ["logrho"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=_snake_case , prediction_type=_snake_case , sample_max_value=_snake_case , algorithm_type='deis' , solver_order=_snake_case , solver_type=_snake_case , ) def UpperCamelCase__ ( self ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_snake_case ) def UpperCamelCase__ ( self ): """simple docstring""" for algorithm_type in ["deis"]: for solver_type in ["logrho"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=_snake_case , solver_type=_snake_case , prediction_type=_snake_case , algorithm_type=_snake_case , ) lowerCAmelCase = self.full_loop( solver_order=_snake_case , solver_type=_snake_case , prediction_type=_snake_case , algorithm_type=_snake_case , ) assert not torch.isnan(_snake_case ).any(), "Samples have nan numbers" def UpperCamelCase__ ( self ): """simple docstring""" self.check_over_configs(lower_order_final=_snake_case ) self.check_over_configs(lower_order_final=_snake_case ) def UpperCamelCase__ ( self ): """simple docstring""" for num_inference_steps in [1, 2, 3, 5, 10, 50, 1_00, 9_99, 10_00]: self.check_over_forward(num_inference_steps=_snake_case , time_step=0 ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = self.full_loop() lowerCAmelCase = torch.mean(torch.abs(_snake_case ) ) assert abs(result_mean.item() - 0.23_916 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = self.full_loop(prediction_type='v_prediction' ) lowerCAmelCase = torch.mean(torch.abs(_snake_case ) ) assert abs(result_mean.item() - 0.091 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = self.scheduler_classes[0] lowerCAmelCase = self.get_scheduler_config(thresholding=_snake_case , dynamic_thresholding_ratio=0 ) lowerCAmelCase = scheduler_class(**_snake_case ) lowerCAmelCase = 10 lowerCAmelCase = self.dummy_model() lowerCAmelCase = self.dummy_sample_deter.half() scheduler.set_timesteps(_snake_case ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase = model(_snake_case , _snake_case ) lowerCAmelCase = scheduler.step(_snake_case , _snake_case , _snake_case ).prev_sample assert sample.dtype == torch.floataa
309
"""simple docstring""" from __future__ import annotations import unittest from transformers import EsmConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers.models.esm.modeling_tf_esm import ( TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, TFEsmModel, ) class a : def __init__( self , _snake_case , ): """simple docstring""" lowerCAmelCase = parent lowerCAmelCase = 13 lowerCAmelCase = 7 lowerCAmelCase = True lowerCAmelCase = True lowerCAmelCase = True lowerCAmelCase = 99 lowerCAmelCase = 32 lowerCAmelCase = 2 lowerCAmelCase = 4 lowerCAmelCase = 37 lowerCAmelCase = 'gelu' lowerCAmelCase = 0.1 lowerCAmelCase = 0.1 lowerCAmelCase = 5_12 lowerCAmelCase = 16 lowerCAmelCase = 2 lowerCAmelCase = 0.02 lowerCAmelCase = 3 lowerCAmelCase = 4 lowerCAmelCase = None def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase = None if self.use_input_mask: lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = None if self.use_labels: lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase = EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , pad_token_id=1 , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase__ ( self ): """simple docstring""" ( ( lowerCAmelCase ) ,( lowerCAmelCase ) ,( lowerCAmelCase ) ,( lowerCAmelCase ) ,( lowerCAmelCase ) ,( lowerCAmelCase ) , ) = self.prepare_config_and_inputs() lowerCAmelCase = True lowerCAmelCase = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def UpperCamelCase__ ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ): """simple docstring""" lowerCAmelCase = TFEsmModel(config=_snake_case ) lowerCAmelCase = {'input_ids': input_ids, 'attention_mask': input_mask} lowerCAmelCase = model(_snake_case ) lowerCAmelCase = [input_ids, input_mask] lowerCAmelCase = model(_snake_case ) lowerCAmelCase = model(_snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase__ ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , ): """simple docstring""" lowerCAmelCase = True lowerCAmelCase = TFEsmModel(config=_snake_case ) lowerCAmelCase = { 'input_ids': input_ids, 'attention_mask': input_mask, 'encoder_hidden_states': encoder_hidden_states, 'encoder_attention_mask': encoder_attention_mask, } lowerCAmelCase = model(_snake_case ) lowerCAmelCase = [input_ids, input_mask] lowerCAmelCase = model(_snake_case , encoder_hidden_states=_snake_case ) # Also check the case where encoder outputs are not passed lowerCAmelCase = model(_snake_case , attention_mask=_snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase__ ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ): """simple docstring""" lowerCAmelCase = TFEsmForMaskedLM(config=_snake_case ) lowerCAmelCase = model([input_ids, input_mask] ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase__ ( self , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ): """simple docstring""" lowerCAmelCase = self.num_labels lowerCAmelCase = TFEsmForTokenClassification(config=_snake_case ) lowerCAmelCase = {'input_ids': input_ids, 'attention_mask': input_mask} lowerCAmelCase = model(_snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = self.prepare_config_and_inputs() ( ( lowerCAmelCase ) ,( lowerCAmelCase ) ,( lowerCAmelCase ) ,( lowerCAmelCase ) ,( lowerCAmelCase ) ,( lowerCAmelCase ) , ) = config_and_inputs lowerCAmelCase = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class a ( a__ , a__ , unittest.TestCase ): snake_case__ = ( ( TFEsmModel, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, ) if is_tf_available() else () ) snake_case__ = ( { '''feature-extraction''': TFEsmModel, '''fill-mask''': TFEsmForMaskedLM, '''text-classification''': TFEsmForSequenceClassification, '''token-classification''': TFEsmForTokenClassification, '''zero-shot''': TFEsmForSequenceClassification, } if is_tf_available() else {} ) snake_case__ = False snake_case__ = False def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = TFEsmModelTester(self ) lowerCAmelCase = ConfigTester(self , config_class=_snake_case , hidden_size=37 ) def UpperCamelCase__ ( self ): """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_snake_case ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*_snake_case ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_snake_case ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_snake_case ) @slow def UpperCamelCase__ ( self ): """simple docstring""" for model_name in TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase = TFEsmModel.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) @unittest.skip('Protein models do not support embedding resizing.' ) def UpperCamelCase__ ( self ): """simple docstring""" pass @unittest.skip('Protein models do not support embedding resizing.' ) def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase ,lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = model_class(_snake_case ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class is TFEsmForMaskedLM: # Output embedding test differs from the main test because they're a matrix, not a layer lowerCAmelCase = model.get_bias() assert isinstance(_snake_case , _snake_case ) for k, v in name.items(): assert isinstance(_snake_case , tf.Variable ) else: lowerCAmelCase = model.get_output_embeddings() assert x is None lowerCAmelCase = model.get_bias() assert name is None @require_tf class a ( unittest.TestCase ): @slow def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = TFEsmForMaskedLM.from_pretrained('facebook/esm2_t6_8M_UR50D' ) lowerCAmelCase = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowerCAmelCase = model(_snake_case )[0] lowerCAmelCase = [1, 6, 33] self.assertEqual(list(output.numpy().shape ) , _snake_case ) # compare the actual values for a slice. lowerCAmelCase = tf.constant( [ [ [8.921_518, -10.589_814, -6.4_671_307], [-6.3_967_156, -13.911_377, -1.1_211_915], [-7.781_247, -13.951_557, -3.740_592], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-2 ) ) @slow def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = TFEsmModel.from_pretrained('facebook/esm2_t6_8M_UR50D' ) lowerCAmelCase = tf.constant([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) lowerCAmelCase = model(_snake_case )[0] # compare the actual values for a slice. lowerCAmelCase = tf.constant( [ [ [0.14_443_092, 0.54_125_327, 0.3_247_739], [0.30_340_484, 0.00_526_676, 0.31_077_722], [0.32_278_043, -0.24_987_096, 0.3_414_628], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
309
1
'''simple docstring''' import fire from utils import calculate_rouge, save_json def lowercase__ ( __lowercase : List[str] , __lowercase : Union[str, Any] , __lowercase : Union[str, Any]=None , **__lowercase : Dict ) -> Optional[int]: """simple docstring""" __UpperCamelCase = [x.strip() for x in open(__lowercase ).readlines()] __UpperCamelCase = [x.strip() for x in open(__lowercase ).readlines()][: len(__lowercase )] __UpperCamelCase = calculate_rouge(__lowercase , __lowercase , **__lowercase ) if save_path is not None: save_json(__lowercase , __lowercase , indent=__lowercase ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
53
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = { # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class __lowerCAmelCase ( _a ): lowerCamelCase_ : Any = '''megatron-bert''' def __init__(self , __magic_name__=2_9056 , __magic_name__=1024 , __magic_name__=24 , __magic_name__=16 , __magic_name__=4096 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=512 , __magic_name__=2 , __magic_name__=0.02 , __magic_name__=1e-12 , __magic_name__=0 , __magic_name__="absolute" , __magic_name__=True , **__magic_name__ , ) -> Any: '''simple docstring''' super().__init__(pad_token_id=__magic_name__ , **__magic_name__ ) snake_case_ : Union[str, Any] = vocab_size snake_case_ : Dict = hidden_size snake_case_ : Dict = num_hidden_layers snake_case_ : Optional[int] = num_attention_heads snake_case_ : int = hidden_act snake_case_ : List[str] = intermediate_size snake_case_ : Dict = hidden_dropout_prob snake_case_ : str = attention_probs_dropout_prob snake_case_ : str = max_position_embeddings snake_case_ : Any = type_vocab_size snake_case_ : int = initializer_range snake_case_ : int = layer_norm_eps snake_case_ : List[str] = position_embedding_type snake_case_ : Dict = use_cache
279
0
"""simple docstring""" import os import sys import unittest a : Union[str, Any] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, create_dummy_object, find_backend, read_init # noqa: E402 # Align TRANSFORMERS_PATH in check_dummies with the current path a : List[Any] = os.path.join(git_repo_path, """src""", """transformers""") a : Tuple = """ {0} = None """ a : List[str] = """ class {0}(metaclass=DummyObject): _backends = {1} def __init__(self, *args, **kwargs): requires_backends(self, {1}) """ a : Optional[int] = """ def {0}(*args, **kwargs): requires_backends({0}, {1}) """ class __UpperCAmelCase( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self ): '''simple docstring''' lowercase__ : List[Any]= find_backend(" _import_structure[\"models.albert\"].append(\"AlbertTokenizerFast\")" ) self.assertIsNone(snake_case__ ) lowercase__ : List[str]= find_backend(" if not is_tokenizers_available():" ) self.assertEqual(snake_case__ , "tokenizers" ) lowercase__ : List[str]= find_backend(" if not is_tensorflow_text_available():" ) self.assertEqual(snake_case__ , "tensorflow_text" ) lowercase__ : List[str]= find_backend(" if not (is_sentencepiece_available() and is_tokenizers_available()):" ) self.assertEqual(snake_case__ , "sentencepiece_and_tokenizers" ) lowercase__ : Optional[Any]= find_backend( " if not (is_sentencepiece_available() and is_tensorflow_text_available()):" ) self.assertEqual(snake_case__ , "sentencepiece_and_tensorflow_text" ) lowercase__ : Any= find_backend( " if not (is_sentencepiece_available() and is_tokenizers_available() and is_vision_available()):" ) self.assertEqual(snake_case__ , "sentencepiece_and_tokenizers_and_vision" ) def UpperCAmelCase_ ( self ): '''simple docstring''' lowercase__ : List[str]= read_init() # We don't assert on the exact list of keys to allow for smooth grow of backend-specific objects self.assertIn("torch" , snake_case__ ) self.assertIn("tensorflow_text" , snake_case__ ) self.assertIn("sentencepiece_and_tokenizers" , snake_case__ ) # Likewise, we can't assert on the exact content of a key self.assertIn("BertModel" , objects["torch"] ) self.assertIn("TFBertModel" , objects["tf"] ) self.assertIn("FlaxBertModel" , objects["flax"] ) self.assertIn("BertModel" , objects["torch"] ) self.assertIn("TFBertTokenizer" , objects["tensorflow_text"] ) self.assertIn("convert_slow_tokenizer" , objects["sentencepiece_and_tokenizers"] ) def UpperCAmelCase_ ( self ): '''simple docstring''' lowercase__ : Optional[int]= create_dummy_object("CONSTANT" , "'torch'" ) self.assertEqual(snake_case__ , "\nCONSTANT = None\n" ) lowercase__ : Optional[Any]= create_dummy_object("function" , "'torch'" ) self.assertEqual( snake_case__ , "\ndef function(*args, **kwargs):\n requires_backends(function, 'torch')\n" ) lowercase__ : str= "\nclass FakeClass(metaclass=DummyObject):\n _backends = 'torch'\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, 'torch')\n" lowercase__ : Dict= create_dummy_object("FakeClass" , "'torch'" ) self.assertEqual(snake_case__ , snake_case__ ) def UpperCAmelCase_ ( self ): '''simple docstring''' lowercase__ : int= "# This file is autogenerated by the command `make fix-copies`, do not edit.\nfrom ..utils import DummyObject, requires_backends\n\n\nCONSTANT = None\n\n\ndef function(*args, **kwargs):\n requires_backends(function, [\"torch\"])\n\n\nclass FakeClass(metaclass=DummyObject):\n _backends = [\"torch\"]\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, [\"torch\"])\n" lowercase__ : Optional[Any]= create_dummy_files({"torch": ["CONSTANT", "function", "FakeClass"]} ) self.assertEqual(dummy_files["torch"] , snake_case__ )
367
"""simple docstring""" from __future__ import annotations def lowercase__(A ) ->int: """simple docstring""" for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(A ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 , len(A ) ): for j in range(1 , len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] , matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
150
0
from torch import nn class A ( nn.Module ): """simple docstring""" def __init__( self : str,lowercase_ : List[Any],lowercase_ : List[Any] )-> Optional[Any]: '''simple docstring''' super().__init__() A__ = class_size A__ = embed_size # self.mlp1 = nn.Linear(embed_size, embed_size) # self.mlp2 = (nn.Linear(embed_size, class_size)) A__ = nn.Linear(lowercase_,lowercase_ ) def snake_case__ ( self : Union[str, Any],lowercase_ : List[str] )-> List[str]: '''simple docstring''' A__ = self.mlp(lowercase_ ) return logits
7
def _snake_case( SCREAMING_SNAKE_CASE__ : float , SCREAMING_SNAKE_CASE__ : float , SCREAMING_SNAKE_CASE__ : float , SCREAMING_SNAKE_CASE__ : float , SCREAMING_SNAKE_CASE__ : float , ) -> float: '''simple docstring''' A__ = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError('All input parameters must be positive' ) if any(p > 1 for p in parameters[1:4] ): raise ValueError('Relative densities cannot be greater than one' ) else: A__ = 1 - (matter_density + radiation_density + dark_energy) A__ = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) A__ = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation lowercase_ = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1e-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
7
1
import unittest from transformers import TrOCRConfig from transformers.testing_utils import is_torch_available, require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.trocr.modeling_trocr import TrOCRDecoder, TrOCRForCausalLM @require_torch class UpperCAmelCase__ : """simple docstring""" def __init__( self , A_ , A_=99 , A_=13 , A_=16 , A_=7 , A_=True , A_=True , A_=True , A_=False , A_=True , A_=2 , A_=32 , A_=4 , A_=4 , A_=30 , A_=0 , A_=1 , A_=2 , A_=None , ) -> List[str]: __UpperCamelCase =parent __UpperCamelCase =batch_size __UpperCamelCase =decoder_seq_length # For common tests __UpperCamelCase =self.decoder_seq_length __UpperCamelCase =is_training __UpperCamelCase =use_attention_mask __UpperCamelCase =use_labels __UpperCamelCase =vocab_size __UpperCamelCase =d_model __UpperCamelCase =d_model __UpperCamelCase =decoder_layers __UpperCamelCase =decoder_layers __UpperCamelCase =decoder_ffn_dim __UpperCamelCase =decoder_attention_heads __UpperCamelCase =decoder_attention_heads __UpperCamelCase =eos_token_id __UpperCamelCase =bos_token_id __UpperCamelCase =pad_token_id __UpperCamelCase =decoder_start_token_id __UpperCamelCase =use_cache __UpperCamelCase =max_position_embeddings __UpperCamelCase =None __UpperCamelCase =decoder_seq_length __UpperCamelCase =2 __UpperCamelCase =1 def _a ( self ) -> List[Any]: __UpperCamelCase =ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) __UpperCamelCase =None if self.use_attention_mask: __UpperCamelCase =ids_tensor([self.batch_size, self.decoder_seq_length] , vocab_size=2 ) __UpperCamelCase =None if self.use_labels: __UpperCamelCase =ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) __UpperCamelCase =TrOCRConfig( vocab_size=self.vocab_size , d_model=self.d_model , decoder_layers=self.decoder_layers , decoder_ffn_dim=self.decoder_ffn_dim , decoder_attention_heads=self.decoder_attention_heads , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , use_cache=self.use_cache , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , max_position_embeddings=self.max_position_embeddings , ) return (config, input_ids, attention_mask, lm_labels) def _a ( self , A_ , A_ , A_ , A_ , ) -> Dict: __UpperCamelCase =True __UpperCamelCase =TrOCRDecoder(config=A_ ).to(A_ ).eval() __UpperCamelCase =input_ids[:2] input_ids[input_ids == 0] += 1 # first forward pass __UpperCamelCase =model(A_ , use_cache=A_ ) __UpperCamelCase =model(A_ ) __UpperCamelCase =model(A_ , use_cache=A_ ) self.parent.assertTrue(len(A_ ) == len(A_ ) ) self.parent.assertTrue(len(A_ ) == len(A_ ) + 1 ) __UpperCamelCase =outputs['past_key_values'] # create hypothetical next token and extent to next_input_ids __UpperCamelCase =ids_tensor((2, 1) , config.vocab_size - 1 ) + 1 # append to next input_ids and __UpperCamelCase =torch.cat([input_ids, next_tokens] , dim=-1 ) __UpperCamelCase =model(A_ )['last_hidden_state'] __UpperCamelCase =model(A_ , past_key_values=A_ )['last_hidden_state'] # select random slice __UpperCamelCase =ids_tensor((1,) , output_from_past.shape[-1] ).item() __UpperCamelCase =output_from_no_past[:, next_input_ids.shape[-1] - 1, random_slice_idx].detach() __UpperCamelCase =output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice assert torch.allclose(A_ , A_ , atol=1E-3 ) def _a ( self ) -> Union[str, Any]: __UpperCamelCase =self.prepare_config_and_inputs() __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase =config_and_inputs __UpperCamelCase ={'input_ids': input_ids, 'attention_mask': attention_mask} return config, inputs_dict @require_torch class UpperCAmelCase__ ( A_ , A_ , A_ , unittest.TestCase ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = (TrOCRDecoder, TrOCRForCausalLM) if is_torch_available() else () UpperCAmelCase__ : str = (TrOCRForCausalLM,) if is_torch_available() else () UpperCAmelCase__ : Dict = {"text-generation": TrOCRForCausalLM} if is_torch_available() else {} UpperCAmelCase__ : Union[str, Any] = True UpperCAmelCase__ : Tuple = False def _a ( self ) -> Optional[Any]: __UpperCamelCase =TrOCRStandaloneDecoderModelTester(self , is_training=A_ ) __UpperCamelCase =ConfigTester(self , config_class=A_ ) def _a ( self ) -> List[str]: pass def _a ( self ) -> List[Any]: pass def _a ( self ) -> Union[str, Any]: pass def _a ( self ) -> Tuple: self.config_tester.run_common_tests() def _a ( self ) -> Optional[int]: __UpperCamelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past(*A_ ) def _a ( self ) -> int: return @unittest.skip('The model doesn\'t support left padding' ) # and it's not used enough to be worth fixing :) def _a ( self ) -> List[Any]: pass
117
from __future__ import annotations def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ): if b == 0: return (1, 0) ((__UpperCamelCase) , (__UpperCamelCase)) =extended_euclid(SCREAMING_SNAKE_CASE__ , a % b ) __UpperCamelCase =a // b return (y, x - k * y) def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ): ((__UpperCamelCase) , (__UpperCamelCase)) =extended_euclid(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =na * na __UpperCamelCase =ra * x * na + ra * y * na return (n % m + m) % m def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ): ((__UpperCamelCase) , (__UpperCamelCase)) =extended_euclid(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if b < 0: __UpperCamelCase =(b % n + n) % n return b def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ): __UpperCamelCase , __UpperCamelCase =invert_modulo(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ), invert_modulo(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =na * na __UpperCamelCase =ra * x * na + ra * y * na return (n % m + m) % m if __name__ == "__main__": from doctest import testmod testmod(name='chinese_remainder_theorem', verbose=True) testmod(name='chinese_remainder_theorem2', verbose=True) testmod(name='invert_modulo', verbose=True) testmod(name='extended_euclid', verbose=True)
117
1
def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' if number > 0: raise ValueError('''input must be a negative integer''' ) __UpperCamelCase :str = len(bin(SCREAMING_SNAKE_CASE )[3:] ) __UpperCamelCase :Optional[Any] = bin(abs(SCREAMING_SNAKE_CASE ) - (1 << binary_number_length) )[3:] __UpperCamelCase :Optional[Any] = ( ( '''1''' + '''0''' * (binary_number_length - len(SCREAMING_SNAKE_CASE )) + twos_complement_number ) if number < 0 else '''0''' ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
43
import argparse import torch from ...utils import logging from . import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert logging.set_verbosity_info() def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = AlbertConfig.from_json_file(SCREAMING_SNAKE_CASE ) print(f"""Building PyTorch model from configuration: {config}""" ) __UpperCamelCase :List[str] = AlbertForPreTraining(SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint load_tf_weights_in_albert(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--albert_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained ALBERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __lowercase = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.albert_config_file, args.pytorch_dump_path)
43
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { 'openai/whisper-base': 'https://huggingface.co/openai/whisper-base/resolve/main/config.json', } # fmt: off __UpperCAmelCase = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 3_57, 3_66, 4_38, 5_32, 6_85, 7_05, 7_96, 9_30, 10_58, 12_20, 12_67, 12_79, 13_03, 13_43, 13_77, 13_91, 16_35, 17_82, 18_75, 21_62, 23_61, 24_88, 34_67, 40_08, 42_11, 46_00, 48_08, 52_99, 58_55, 63_29, 72_03, 96_09, 99_59, 1_05_63, 1_07_86, 1_14_20, 1_17_09, 1_19_07, 1_31_63, 1_36_97, 1_37_00, 1_48_08, 1_53_06, 1_64_10, 1_67_91, 1_79_92, 1_92_03, 1_95_10, 2_07_24, 2_23_05, 2_29_35, 2_70_07, 3_01_09, 3_04_20, 3_34_09, 3_49_49, 4_02_83, 4_04_93, 4_05_49, 4_72_82, 4_91_46, 5_02_57, 5_03_59, 5_03_60, 5_03_61 ] __UpperCAmelCase = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 3_59, 5_03, 5_22, 5_42, 8_73, 8_93, 9_02, 9_18, 9_22, 9_31, 13_50, 18_53, 19_82, 24_60, 26_27, 32_46, 32_53, 32_68, 35_36, 38_46, 39_61, 41_83, 46_67, 65_85, 66_47, 72_73, 90_61, 93_83, 1_04_28, 1_09_29, 1_19_38, 1_20_33, 1_23_31, 1_25_62, 1_37_93, 1_41_57, 1_46_35, 1_52_65, 1_56_18, 1_65_53, 1_66_04, 1_83_62, 1_89_56, 2_00_75, 2_16_75, 2_25_20, 2_61_30, 2_61_61, 2_64_35, 2_82_79, 2_94_64, 3_16_50, 3_23_02, 3_24_70, 3_68_65, 4_28_63, 4_74_25, 4_98_70, 5_02_54, 5_02_58, 5_03_60, 5_03_61, 5_03_62 ] class __a ( __UpperCamelCase ): __snake_case : Optional[Any] = """whisper""" __snake_case : str = ["""past_key_values"""] __snake_case : List[str] = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self : Optional[int] , UpperCAmelCase : Any=5_18_65 , UpperCAmelCase : Optional[int]=80 , UpperCAmelCase : Union[str, Any]=6 , UpperCAmelCase : int=4 , UpperCAmelCase : Tuple=6 , UpperCAmelCase : Union[str, Any]=4 , UpperCAmelCase : Any=15_36 , UpperCAmelCase : Optional[Any]=15_36 , UpperCAmelCase : Tuple=0.0 , UpperCAmelCase : Optional[Any]=0.0 , UpperCAmelCase : List[str]=5_02_57 , UpperCAmelCase : int=True , UpperCAmelCase : str=True , UpperCAmelCase : Tuple="gelu" , UpperCAmelCase : Tuple=2_56 , UpperCAmelCase : List[str]=0.0 , UpperCAmelCase : int=0.0 , UpperCAmelCase : List[Any]=0.0 , UpperCAmelCase : List[str]=0.02 , UpperCAmelCase : Tuple=False , UpperCAmelCase : Union[str, Any]=15_00 , UpperCAmelCase : Optional[int]=4_48 , UpperCAmelCase : Optional[int]=5_02_56 , UpperCAmelCase : str=5_02_56 , UpperCAmelCase : str=5_02_56 , UpperCAmelCase : Union[str, Any]=None , UpperCAmelCase : List[Any]=[2_20, 5_02_56] , UpperCAmelCase : Optional[Any]=False , UpperCAmelCase : Any=2_56 , UpperCAmelCase : Optional[int]=False , UpperCAmelCase : Optional[Any]=0.05 , UpperCAmelCase : int=10 , UpperCAmelCase : Any=2 , UpperCAmelCase : Any=0.0 , UpperCAmelCase : List[Any]=10 , UpperCAmelCase : Tuple=0 , UpperCAmelCase : Optional[int]=7 , **UpperCAmelCase : Union[str, Any] , ): lowerCAmelCase_ : int = vocab_size lowerCAmelCase_ : Optional[Any] = num_mel_bins lowerCAmelCase_ : int = d_model lowerCAmelCase_ : int = encoder_layers lowerCAmelCase_ : int = encoder_attention_heads lowerCAmelCase_ : Any = decoder_layers lowerCAmelCase_ : str = decoder_attention_heads lowerCAmelCase_ : Optional[Any] = decoder_ffn_dim lowerCAmelCase_ : Tuple = encoder_ffn_dim lowerCAmelCase_ : int = dropout lowerCAmelCase_ : Dict = attention_dropout lowerCAmelCase_ : Union[str, Any] = activation_dropout lowerCAmelCase_ : List[Any] = activation_function lowerCAmelCase_ : str = init_std lowerCAmelCase_ : Tuple = encoder_layerdrop lowerCAmelCase_ : List[str] = decoder_layerdrop lowerCAmelCase_ : List[Any] = use_cache lowerCAmelCase_ : Optional[int] = encoder_layers lowerCAmelCase_ : Optional[int] = scale_embedding # scale factor will be sqrt(d_model) if True lowerCAmelCase_ : int = max_source_positions lowerCAmelCase_ : Any = max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. lowerCAmelCase_ : Optional[int] = classifier_proj_size lowerCAmelCase_ : Dict = use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 lowerCAmelCase_ : List[Any] = apply_spec_augment lowerCAmelCase_ : Any = mask_time_prob lowerCAmelCase_ : Dict = mask_time_length lowerCAmelCase_ : List[Any] = mask_time_min_masks lowerCAmelCase_ : str = mask_feature_prob lowerCAmelCase_ : Dict = mask_feature_length lowerCAmelCase_ : Dict = mask_feature_min_masks lowerCAmelCase_ : Any = median_filter_width super().__init__( pad_token_id=UpperCAmelCase , bos_token_id=UpperCAmelCase , eos_token_id=UpperCAmelCase , is_encoder_decoder=UpperCAmelCase , decoder_start_token_id=UpperCAmelCase , suppress_tokens=UpperCAmelCase , begin_suppress_tokens=UpperCAmelCase , **UpperCAmelCase , ) class __a ( __UpperCamelCase ): @property def A ( self : Any ): lowerCAmelCase_ : List[str] = OrderedDict( [ ("""input_features""", {0: """batch""", 1: """feature_size""", 2: """encoder_sequence"""}), ] ) if self.use_past: lowerCAmelCase_ : int = {0: """batch"""} else: lowerCAmelCase_ : Optional[Any] = {0: """batch""", 1: """decoder_sequence"""} if self.use_past: self.fill_with_past_key_values_(UpperCAmelCase , direction="""inputs""" ) return common_inputs def A ( self : str , UpperCAmelCase : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , UpperCAmelCase : int = -1 , UpperCAmelCase : int = -1 , UpperCAmelCase : bool = False , UpperCAmelCase : Optional["TensorType"] = None , UpperCAmelCase : int = 2_20_50 , UpperCAmelCase : float = 5.0 , UpperCAmelCase : int = 2_20 , ): lowerCAmelCase_ : Tuple = OrderedDict() lowerCAmelCase_ : List[str] = OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=UpperCAmelCase , framework=UpperCAmelCase , sampling_rate=UpperCAmelCase , time_duration=UpperCAmelCase , frequency=UpperCAmelCase , ) lowerCAmelCase_ : Optional[Any] = encoder_inputs["""input_features"""].shape[2] lowerCAmelCase_ : List[str] = encoder_sequence_length // 2 if self.use_past else seq_length lowerCAmelCase_ : Optional[Any] = super().generate_dummy_inputs( preprocessor.tokenizer , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) lowerCAmelCase_ : int = encoder_inputs.pop("""input_features""" ) lowerCAmelCase_ : Any = decoder_inputs.pop("""decoder_input_ids""" ) if "past_key_values" in decoder_inputs: lowerCAmelCase_ : Optional[Any] = decoder_inputs.pop("""past_key_values""" ) return dummy_inputs @property def A ( self : List[Any] ): return 1e-3
359
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() __UpperCAmelCase = logging.get_logger(__name__) def __UpperCamelCase ( lowercase__ : Optional[Any] ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : Optional[int] = """huggingface/label-files""" lowerCAmelCase_ : int = """imagenet-1k-id2label.json""" lowerCAmelCase_ : List[str] = json.load(open(hf_hub_download(lowercase__ , lowercase__ , repo_type="""dataset""" ) , """r""" ) ) lowerCAmelCase_ : Tuple = {int(lowercase__ ): v for k, v in idalabel.items()} lowerCAmelCase_ : Optional[int] = {v: k for k, v in idalabel.items()} lowerCAmelCase_ : Optional[Any] = """std_conv""" if """bit""" in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" lowerCAmelCase_ : Tuple = BitConfig( conv_layer=lowercase__ , num_labels=1000 , idalabel=lowercase__ , labelaid=lowercase__ , ) return config def __UpperCamelCase ( lowercase__ : List[Any] ) -> Optional[int]: '''simple docstring''' if "stem.conv" in name: lowerCAmelCase_ : str = name.replace("""stem.conv""" , """bit.embedder.convolution""" ) if "blocks" in name: lowerCAmelCase_ : Tuple = name.replace("""blocks""" , """layers""" ) if "head.fc" in name: lowerCAmelCase_ : Dict = name.replace("""head.fc""" , """classifier.1""" ) if name.startswith("""norm""" ): lowerCAmelCase_ : List[str] = """bit.""" + name if "bit" not in name and "classifier" not in name: lowerCAmelCase_ : Any = """bit.encoder.""" + name return name def __UpperCamelCase ( ) -> str: '''simple docstring''' lowerCAmelCase_ : List[Any] = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowerCAmelCase_ : List[Any] = Image.open(requests.get(lowercase__ , stream=lowercase__ ).raw ) return im @torch.no_grad() def __UpperCamelCase ( lowercase__ : List[str] , lowercase__ : Any , lowercase__ : Any=False ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = get_config(lowercase__ ) # load original model from timm lowerCAmelCase_ : str = create_model(lowercase__ , pretrained=lowercase__ ) timm_model.eval() # load state_dict of original model lowerCAmelCase_ : Any = timm_model.state_dict() for key in state_dict.copy().keys(): lowerCAmelCase_ : List[str] = state_dict.pop(lowercase__ ) lowerCAmelCase_ : Dict = val.squeeze() if """head""" in key else val # load HuggingFace model lowerCAmelCase_ : Tuple = BitForImageClassification(lowercase__ ) model.eval() model.load_state_dict(lowercase__ ) # create image processor lowerCAmelCase_ : Tuple = create_transform(**resolve_data_config({} , model=lowercase__ ) ) lowerCAmelCase_ : Union[str, Any] = transform.transforms lowerCAmelCase_ : str = { """bilinear""": PILImageResampling.BILINEAR, """bicubic""": PILImageResampling.BICUBIC, """nearest""": PILImageResampling.NEAREST, } lowerCAmelCase_ : List[str] = BitImageProcessor( do_resize=lowercase__ , size={"""shortest_edge""": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=lowercase__ , crop_size={"""height""": timm_transforms[1].size[0], """width""": timm_transforms[1].size[1]} , do_normalize=lowercase__ , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) lowerCAmelCase_ : int = prepare_img() lowerCAmelCase_ : Tuple = transform(lowercase__ ).unsqueeze(0 ) lowerCAmelCase_ : List[str] = processor(lowercase__ , return_tensors="""pt""" ).pixel_values # verify pixel values assert torch.allclose(lowercase__ , lowercase__ ) # verify logits with torch.no_grad(): lowerCAmelCase_ : Tuple = model(lowercase__ ) lowerCAmelCase_ : List[str] = outputs.logits print("""Logits:""" , logits[0, :3] ) print("""Predicted class:""" , model.config.idalabel[logits.argmax(-1 ).item()] ) lowerCAmelCase_ : Optional[Any] = timm_model(lowercase__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowercase__ , outputs.logits , atol=1E-3 ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: Path(lowercase__ ).mkdir(exist_ok=lowercase__ ) print(f'Saving model {model_name} and processor to {pytorch_dump_folder_path}' ) model.save_pretrained(lowercase__ ) processor.save_pretrained(lowercase__ ) if push_to_hub: print(f'Pushing model {model_name} and processor to the hub' ) model.push_to_hub(f'ybelkada/{model_name}' ) processor.push_to_hub(f'ybelkada/{model_name}' ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='resnetv2_50x1_bitm', type=str, help='Name of the BiT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model to the hub.', ) __UpperCAmelCase = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
28
0
'''simple docstring''' import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class _a ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ), 1 ) self.assertEqual(x.component(2 ), 3 ) SCREAMING_SNAKE_CASE : List[str] = Vector() def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(lowerCamelCase__ ), '(0,0,0,0,0,1)' ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = Vector([1, 2, 3, 4] ) self.assertEqual(len(lowerCamelCase__ ), 4 ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Dict = Vector([1, 2] ) SCREAMING_SNAKE_CASE : Union[str, Any] = Vector([1, 2, 3, 4, 5] ) SCREAMING_SNAKE_CASE : int = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) SCREAMING_SNAKE_CASE : List[Any] = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length(), 2.2_36, 3 ) self.assertAlmostEqual(y.euclidean_length(), 7.4_16, 3 ) self.assertEqual(z.euclidean_length(), 0 ) self.assertAlmostEqual(w.euclidean_length(), 7.6_16, 3 ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = Vector([1, 2, 3] ) SCREAMING_SNAKE_CASE : Optional[int] = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ), 2 ) self.assertEqual((x + y).component(1 ), 3 ) self.assertEqual((x + y).component(2 ), 4 ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Dict = Vector([1, 2, 3] ) SCREAMING_SNAKE_CASE : Tuple = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ), 0 ) self.assertEqual((x - y).component(1 ), 1 ) self.assertEqual((x - y).component(2 ), 2 ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[Any] = Vector([1, 2, 3] ) SCREAMING_SNAKE_CASE : Any = Vector([2, -1, 4] ) # for test of dot product SCREAMING_SNAKE_CASE : Any = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ), '(3.0,6.0,9.0)' ) self.assertEqual((a * b), 0 ) def UpperCamelCase_ ( self ): '''simple docstring''' self.assertEqual(str(zero_vector(10 ) ).count('0' ), 10 ) def UpperCamelCase_ ( self ): '''simple docstring''' self.assertEqual(str(unit_basis_vector(3, 1 ) ), '(0,1,0)' ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = Vector([1, 2, 3] ) SCREAMING_SNAKE_CASE : Optional[int] = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2, lowerCamelCase__, lowerCamelCase__ ) ), '(3,4,7)' ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = Vector([1, 0, 0, 0, 0, 0] ) SCREAMING_SNAKE_CASE : Any = x.copy() self.assertEqual(str(lowerCamelCase__ ), str(lowerCamelCase__ ) ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = Vector([1, 0, 0] ) x.change_component(0, 0 ) x.change_component(1, 1 ) self.assertEqual(str(lowerCamelCase__ ), '(0,1,0)' ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3 ) self.assertEqual('|1,2,3|\n|2,4,5|\n|6,7,8|\n', str(lowerCamelCase__ ) ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3 ) SCREAMING_SNAKE_CASE : Any = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y], a.minor(lowerCamelCase__, lowerCamelCase__ ) ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3 ) SCREAMING_SNAKE_CASE : Tuple = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y], a.cofactor(lowerCamelCase__, lowerCamelCase__ ) ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3 ) self.assertEqual(-5, a.determinant() ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]], 3, 3 ) SCREAMING_SNAKE_CASE : Union[str, Any] = Vector([1, 2, 3] ) self.assertEqual('(14,32,50)', str(a * x ) ) self.assertEqual('|2,4,6|\n|8,10,12|\n|14,16,18|\n', str(a * 2 ) ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3 ) a.change_component(0, 2, 5 ) self.assertEqual('|1,2,5|\n|2,4,5|\n|6,7,8|\n', str(lowerCamelCase__ ) ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3 ) self.assertEqual(7, a.component(2, 1 ), 0.01 ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3 ) SCREAMING_SNAKE_CASE : Optional[Any] = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]], 3, 3 ) self.assertEqual('|2,4,10|\n|4,8,10|\n|12,14,18|\n', str(a + b ) ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]], 3, 3 ) SCREAMING_SNAKE_CASE : Optional[Any] = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]], 3, 3 ) self.assertEqual('|0,0,-4|\n|0,0,0|\n|0,0,-2|\n', str(a - b ) ) def UpperCamelCase_ ( self ): '''simple docstring''' self.assertEqual( '|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n', str(square_zero_matrix(5 ) ), ) if __name__ == "__main__": unittest.main()
251
'''simple docstring''' from dataclasses import dataclass, field from typing import Optional from transformers import AutoConfig, AutoImageProcessor, AutoTokenizer, FlaxVisionEncoderDecoderModel, HfArgumentParser @dataclass class lowercase__ : lowercase__ = field( metadata={"""help""": """The output directory where the model will be written."""} , ) lowercase__ = field( metadata={ """help""": ( """The encoder model checkpoint for weights initialization.""" """Don't set if you want to train an encoder model from scratch.""" ) } , ) lowercase__ = field( metadata={ """help""": ( """The decoder model checkpoint for weights initialization.""" """Don't set if you want to train a decoder model from scratch.""" ) } , ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained encoder config name or path if not the same as encoder_model_name"""} ) lowercase__ = field( default=lowercase , metadata={"""help""": """Pretrained decoder config name or path if not the same as decoder_model_name"""} ) def A__ ( ): _UpperCamelCase : Optional[Any] = HfArgumentParser((ModelArguments,) ) ((_UpperCamelCase) , ) : Optional[int] = parser.parse_args_into_dataclasses() # Load pretrained model and tokenizer # Use explicit specified encoder config if model_args.encoder_config_name: _UpperCamelCase : Any = AutoConfig.from_pretrained(model_args.encoder_config_name ) # Use pretrained encoder model's config else: _UpperCamelCase : Union[str, Any] = AutoConfig.from_pretrained(model_args.encoder_model_name_or_path ) # Use explicit specified decoder config if model_args.decoder_config_name: _UpperCamelCase : str = AutoConfig.from_pretrained(model_args.decoder_config_name ) # Use pretrained decoder model's config else: _UpperCamelCase : str = AutoConfig.from_pretrained(model_args.decoder_model_name_or_path ) # necessary for `from_encoder_decoder_pretrained` when `decoder_config` is passed _UpperCamelCase : List[Any] = True _UpperCamelCase : Union[str, Any] = True _UpperCamelCase : str = FlaxVisionEncoderDecoderModel.from_encoder_decoder_pretrained( encoder_pretrained_model_name_or_path=model_args.encoder_model_name_or_path , decoder_pretrained_model_name_or_path=model_args.decoder_model_name_or_path , encoder_config=UpperCAmelCase_ , decoder_config=UpperCAmelCase_ , ) # GPT2 only has bos/eos tokens but not decoder_start/pad tokens _UpperCamelCase : str = decoder_config.decoder_start_token_id _UpperCamelCase : Optional[int] = decoder_config.pad_token_id if decoder_start_token_id is None: _UpperCamelCase : int = decoder_config.bos_token_id if pad_token_id is None: _UpperCamelCase : Dict = decoder_config.eos_token_id # This is necessary to make Flax's generate() work _UpperCamelCase : List[Any] = decoder_config.eos_token_id _UpperCamelCase : Dict = decoder_start_token_id _UpperCamelCase : int = pad_token_id _UpperCamelCase : List[str] = AutoImageProcessor.from_pretrained(model_args.encoder_model_name_or_path ) _UpperCamelCase : List[Any] = AutoTokenizer.from_pretrained(model_args.decoder_model_name_or_path ) _UpperCamelCase : List[Any] = tokenizer.convert_ids_to_tokens(model.config.pad_token_id ) model.save_pretrained(model_args.output_dir ) image_processor.save_pretrained(model_args.output_dir ) tokenizer.save_pretrained(model_args.output_dir ) if __name__ == "__main__": main()
83
0
"""simple docstring""" # this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys _A = subprocess.check_output('git merge-base main HEAD'.split()).decode('utf-8') _A = ( subprocess.check_output(F"""git diff --diff-filter=d --name-only {fork_point_sha}""".split()).decode('utf-8').split() ) _A = '|'.join(sys.argv[1:]) _A = re.compile(rF"""^({joined_dirs}).*?\.py$""") _A = [x for x in modified_files if regex.match(x)] print(' '.join(relevant_modified_files), end='')
205
"""simple docstring""" import os try: from .build_directory_md import good_file_paths except ImportError: from build_directory_md import good_file_paths # type: ignore _A = list(good_file_paths()) assert filepaths, "good_file_paths() failed!" _A = [file for file in filepaths if file != file.lower()] if upper_files: print(F"""{len(upper_files)} files contain uppercase characters:""") print('\n'.join(upper_files) + '\n') _A = [file for file in filepaths if ' ' in file] if space_files: print(F"""{len(space_files)} files contain space characters:""") print('\n'.join(space_files) + '\n') _A = [file for file in filepaths if '-' in file] if hyphen_files: print(F"""{len(hyphen_files)} files contain hyphen characters:""") print('\n'.join(hyphen_files) + '\n') _A = [file for file in filepaths if os.sep not in file] if nodir_files: print(F"""{len(nodir_files)} files are not in a directory:""") print('\n'.join(nodir_files) + '\n') _A = len(upper_files + space_files + hyphen_files + nodir_files) if bad_files: import sys sys.exit(bad_files)
205
1
import os from typing import List, Optional, Union from ...tokenization_utils import PreTrainedTokenizer from ...tokenization_utils_base import AddedToken from ...utils import logging A : Optional[int] = logging.get_logger(__name__) A : Dict = {'vocab_file': 'vocab.txt'} A : Union[str, Any] = { 'vocab_file': { 'facebook/esm2_t6_8M_UR50D': 'https://huggingface.co/facebook/esm2_t6_8M_UR50D/resolve/main/vocab.txt', 'facebook/esm2_t12_35M_UR50D': 'https://huggingface.co/facebook/esm2_t12_35M_UR50D/resolve/main/vocab.txt', }, } A : int = { 'facebook/esm2_t6_8M_UR50D': 1_0_2_4, 'facebook/esm2_t12_35M_UR50D': 1_0_2_4, } def __lowerCAmelCase ( a__ ) -> Dict: with open(a__ , '''r''' ) as f: __a = f.read().splitlines() return [l.strip() for l in lines] class __A( a ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = ['''input_ids''', '''attention_mask'''] def __init__( self , _snake_case , _snake_case="<unk>" , _snake_case="<cls>" , _snake_case="<pad>" , _snake_case="<mask>" , _snake_case="<eos>" , **_snake_case , ) -> str: '''simple docstring''' super().__init__(**_snake_case ) __a = load_vocab_file(_snake_case ) __a = dict(enumerate(self.all_tokens ) ) __a = {tok: ind for ind, tok in enumerate(self.all_tokens )} __a = unk_token __a = cls_token __a = pad_token __a = mask_token __a = eos_token __a = self.all_tokens self._create_trie(self.unique_no_split_tokens ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> str: '''simple docstring''' return self._id_to_token.get(_snake_case , self.unk_token ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> int: '''simple docstring''' return self._token_to_id.get(_snake_case , self._token_to_id.get(self.unk_token ) ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , **_snake_case ) -> Optional[int]: '''simple docstring''' return text.split() def SCREAMING_SNAKE_CASE_ ( self , _snake_case=False ) -> str: '''simple docstring''' return len(self._id_to_token ) def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[Any]: '''simple docstring''' return {token: i for i, token in enumerate(self.all_tokens )} def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> int: '''simple docstring''' return self._token_to_id.get(_snake_case , self._token_to_id.get(self.unk_token ) ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> str: '''simple docstring''' return self._id_to_token.get(_snake_case , self.unk_token ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case = None ) -> List[int]: '''simple docstring''' __a = [self.cls_token_id] __a = [self.eos_token_id] # No sep token in ESM vocabulary if token_ids_a is None: if self.eos_token_id is None: return cls + token_ids_a else: return cls + token_ids_a + sep elif self.eos_token_id is None: raise ValueError('''Cannot tokenize multiple sequences when EOS token is not set!''' ) return cls + token_ids_a + sep + token_ids_a + sep # Multiple inputs always have an EOS token def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case = None , _snake_case = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if token in self.all_special_ids else 0 for token in token_ids_a] __a = [1] + ([0] * len(_snake_case )) + [1] if token_ids_a is not None: mask += [0] * len(_snake_case ) + [1] return mask def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> Optional[Any]: '''simple docstring''' __a = os.path.join(_snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + '''vocab.txt''' ) with open(_snake_case , '''w''' ) as f: f.write('''\n'''.join(self.all_tokens ) ) return (vocab_file,) @property def SCREAMING_SNAKE_CASE_ ( self ) -> int: '''simple docstring''' return self.get_vocab_size(with_added_tokens=_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case = False ) -> int: '''simple docstring''' return super()._add_tokens(_snake_case , special_tokens=_snake_case )
6
import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class __A( a , unittest.TestCase ): # FIXME: add fast tests pass @nightly @require_onnxruntime @require_torch_gpu class __A( unittest.TestCase ): @property def SCREAMING_SNAKE_CASE_ ( self ) -> List[str]: '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def SCREAMING_SNAKE_CASE_ ( self ) -> Union[str, Any]: '''simple docstring''' __a = ort.SessionOptions() __a = False return options def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[Any]: '''simple docstring''' __a = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/in_paint/overture-creations-5sI6fQgYIuo.png''' ) __a = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/in_paint/overture-creations-5sI6fQgYIuo_mask.png''' ) __a = OnnxStableDiffusionInpaintPipeline.from_pretrained( '''runwayml/stable-diffusion-inpainting''' , revision='''onnx''' , safety_checker=_snake_case , feature_extractor=_snake_case , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_snake_case ) __a = '''A red cat sitting on a park bench''' __a = np.random.RandomState(0 ) __a = pipe( prompt=_snake_case , image=_snake_case , mask_image=_snake_case , guidance_scale=7.5 , num_inference_steps=10 , generator=_snake_case , output_type='''np''' , ) __a = output.images __a = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) __a = np.array([0.2514, 0.3007, 0.3517, 0.1790, 0.2382, 0.3167, 0.1944, 0.2273, 0.2464] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def SCREAMING_SNAKE_CASE_ ( self ) -> Tuple: '''simple docstring''' __a = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/in_paint/overture-creations-5sI6fQgYIuo.png''' ) __a = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/in_paint/overture-creations-5sI6fQgYIuo_mask.png''' ) __a = LMSDiscreteScheduler.from_pretrained( '''runwayml/stable-diffusion-inpainting''' , subfolder='''scheduler''' , revision='''onnx''' ) __a = OnnxStableDiffusionInpaintPipeline.from_pretrained( '''runwayml/stable-diffusion-inpainting''' , revision='''onnx''' , scheduler=_snake_case , safety_checker=_snake_case , feature_extractor=_snake_case , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=_snake_case ) __a = '''A red cat sitting on a park bench''' __a = np.random.RandomState(0 ) __a = pipe( prompt=_snake_case , image=_snake_case , mask_image=_snake_case , guidance_scale=7.5 , num_inference_steps=20 , generator=_snake_case , output_type='''np''' , ) __a = output.images __a = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) __a = np.array([0.0086, 0.0077, 0.0083, 0.0093, 0.0107, 0.0139, 0.0094, 0.0097, 0.0125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
6
1
"""simple docstring""" import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType __snake_case = logging.get_logger(__name__) class __lowerCamelCase ( a__ ): '''simple docstring''' A_ : Optional[Any] = 'vision-encoder-decoder' A_ : Tuple = True def __init__( self , **__UpperCAmelCase ) -> List[str]: super().__init__(**__UpperCAmelCase ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( F'A configuraton of type {self.model_type} cannot be instantiated because ' F'not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}' ) _a = kwargs.pop('''encoder''' ) _a = encoder_config.pop('''model_type''' ) _a = kwargs.pop('''decoder''' ) _a = decoder_config.pop('''model_type''' ) _a = AutoConfig.for_model(__UpperCAmelCase , **__UpperCAmelCase ) _a = AutoConfig.for_model(__UpperCAmelCase , **__UpperCAmelCase ) _a = True @classmethod def _UpperCAmelCase ( cls , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) -> PretrainedConfig: logger.info('''Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config''' ) _a = True _a = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **__UpperCAmelCase ) def _UpperCAmelCase ( self ) -> Dict: _a = copy.deepcopy(self.__dict__ ) _a = self.encoder.to_dict() _a = self.decoder.to_dict() _a = self.__class__.model_type return output class __lowerCamelCase ( a__ ): '''simple docstring''' A_ : Tuple = version.parse('1.11' ) @property def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def _UpperCAmelCase ( self ) -> float: return 1e-4 @property def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict({'''last_hidden_state''': {0: '''batch''', 1: '''encoder_sequence'''}} ) class __lowerCamelCase ( a__ ): '''simple docstring''' @property def _UpperCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: _a = OrderedDict() _a = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} _a = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} _a = {0: '''batch''', 1: '''encoder_sequence'''} return common_inputs def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase = -1 , __UpperCAmelCase = -1 , __UpperCAmelCase = False , __UpperCAmelCase = None , ) -> Mapping[str, Any]: import torch _a = OrderedDict() _a = super().generate_dummy_inputs( __UpperCAmelCase , batch_size=__UpperCAmelCase , seq_length=__UpperCAmelCase , is_pair=__UpperCAmelCase , framework=__UpperCAmelCase ) _a , _a = dummy_input['''input_ids'''].shape _a = (batch, encoder_sequence, self._config.encoder_hidden_size) _a = dummy_input.pop('''input_ids''' ) _a = dummy_input.pop('''attention_mask''' ) _a = torch.zeros(__UpperCAmelCase ) return common_inputs class __lowerCamelCase ( a__ ): '''simple docstring''' @property def _UpperCAmelCase ( self ) -> None: pass def _UpperCAmelCase ( self , __UpperCAmelCase ) -> OnnxConfig: return VisionEncoderDecoderEncoderOnnxConfig(__UpperCAmelCase ) def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = "default" ) -> OnnxConfig: _a = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(__UpperCAmelCase , __UpperCAmelCase )
153
"""simple docstring""" import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class __lowerCamelCase : '''simple docstring''' @staticmethod def _UpperCAmelCase ( *__UpperCAmelCase , **__UpperCAmelCase ) -> Tuple: pass def A_ ( _lowerCAmelCase : Image ): """simple docstring""" _a = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' A_ : List[Any] = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> str: _a = DepthEstimationPipeline(model=__UpperCAmelCase , image_processor=__UpperCAmelCase ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase ) -> Dict: _a = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} , __UpperCAmelCase ) import datasets _a = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' , '''image''' , split='''test''' ) _a = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] , __UpperCAmelCase , ) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def _UpperCAmelCase ( self ) -> Tuple: pass @slow @require_torch def _UpperCAmelCase ( self ) -> List[str]: _a = '''Intel/dpt-large''' _a = pipeline('''depth-estimation''' , model=__UpperCAmelCase ) _a = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) _a = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) , 29.304 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) , 2.662 ) @require_torch def _UpperCAmelCase ( self ) -> List[Any]: # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
153
1
"""simple docstring""" import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging SCREAMING_SNAKE_CASE_ : Dict = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ : Dict = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt'} # See all LED models at https://huggingface.co/models?filter=LED SCREAMING_SNAKE_CASE_ : Tuple = { 'vocab_file': { 'allenai/led-base-16384': 'https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json', }, 'merges_file': { 'allenai/led-base-16384': 'https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt', }, 'tokenizer_file': { 'allenai/led-base-16384': 'https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json', }, } SCREAMING_SNAKE_CASE_ : Optional[int] = { 'allenai/led-base-16384': 1_6_3_8_4, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def _snake_case ( ): A__ = ( list(range(ord("""!""" ) , ord("""~""" ) + 1 ) ) + list(range(ord("""¡""" ) , ord("""¬""" ) + 1 ) ) + list(range(ord("""®""" ) , ord("""ÿ""" ) + 1 ) ) ) A__ = bs[:] A__ = 0 for b in range(2**8 ): if b not in bs: bs.append(UpperCAmelCase_ ) cs.append(2**8 + n ) n += 1 A__ = [chr(UpperCAmelCase_ ) for n in cs] return dict(zip(UpperCAmelCase_ , UpperCAmelCase_ ) ) def _snake_case ( UpperCAmelCase_ : List[Any] ): A__ = set() A__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) A__ = char return pairs class a ( _lowerCamelCase ): """simple docstring""" UpperCAmelCase = VOCAB_FILES_NAMES UpperCAmelCase = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase = ["input_ids", "attention_mask"] def __init__( self: Dict , UpperCamelCase: Dict , UpperCamelCase: str , UpperCamelCase: Dict="replace" , UpperCamelCase: List[Any]="<s>" , UpperCamelCase: int="</s>" , UpperCamelCase: Optional[int]="</s>" , UpperCamelCase: Optional[int]="<s>" , UpperCamelCase: Any="<unk>" , UpperCamelCase: Union[str, Any]="<pad>" , UpperCamelCase: List[Any]="<mask>" , UpperCamelCase: str=False , **UpperCamelCase: List[str] , ): """simple docstring""" A__ = AddedToken(UpperCamelCase , lstrip=UpperCamelCase , rstrip=UpperCamelCase ) if isinstance(UpperCamelCase , UpperCamelCase ) else bos_token A__ = AddedToken(UpperCamelCase , lstrip=UpperCamelCase , rstrip=UpperCamelCase ) if isinstance(UpperCamelCase , UpperCamelCase ) else eos_token A__ = AddedToken(UpperCamelCase , lstrip=UpperCamelCase , rstrip=UpperCamelCase ) if isinstance(UpperCamelCase , UpperCamelCase ) else sep_token A__ = AddedToken(UpperCamelCase , lstrip=UpperCamelCase , rstrip=UpperCamelCase ) if isinstance(UpperCamelCase , UpperCamelCase ) else cls_token A__ = AddedToken(UpperCamelCase , lstrip=UpperCamelCase , rstrip=UpperCamelCase ) if isinstance(UpperCamelCase , UpperCamelCase ) else unk_token A__ = AddedToken(UpperCamelCase , lstrip=UpperCamelCase , rstrip=UpperCamelCase ) if isinstance(UpperCamelCase , UpperCamelCase ) else pad_token # Mask token behave like a normal word, i.e. include the space before it A__ = AddedToken(UpperCamelCase , lstrip=UpperCamelCase , rstrip=UpperCamelCase ) if isinstance(UpperCamelCase , UpperCamelCase ) else mask_token super().__init__( errors=UpperCamelCase , bos_token=UpperCamelCase , eos_token=UpperCamelCase , unk_token=UpperCamelCase , sep_token=UpperCamelCase , cls_token=UpperCamelCase , pad_token=UpperCamelCase , mask_token=UpperCamelCase , add_prefix_space=UpperCamelCase , **UpperCamelCase , ) with open(UpperCamelCase , encoding="""utf-8""" ) as vocab_handle: A__ = json.load(UpperCamelCase ) A__ = {v: k for k, v in self.encoder.items()} A__ = errors # how to handle errors in decoding A__ = bytes_to_unicode() A__ = {v: k for k, v in self.byte_encoder.items()} with open(UpperCamelCase , encoding="""utf-8""" ) as merges_handle: A__ = merges_handle.read().split("""\n""" )[1:-1] A__ = [tuple(merge.split() ) for merge in bpe_merges] A__ = dict(zip(UpperCamelCase , range(len(UpperCamelCase ) ) ) ) A__ = {} A__ = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions A__ = re.compile(r"""'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+""" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def UpperCamelCase ( self: Dict ): """simple docstring""" return len(self.encoder ) def UpperCamelCase ( self: int ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def UpperCamelCase ( self: Dict , UpperCamelCase: Tuple ): """simple docstring""" if token in self.cache: return self.cache[token] A__ = tuple(UpperCamelCase ) A__ = get_pairs(UpperCamelCase ) if not pairs: return token while True: A__ = min(UpperCamelCase , key=lambda UpperCamelCase : self.bpe_ranks.get(UpperCamelCase , float("""inf""" ) ) ) if bigram not in self.bpe_ranks: break A__ , A__ = bigram A__ = [] A__ = 0 while i < len(UpperCamelCase ): try: A__ = word.index(UpperCamelCase , UpperCamelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) A__ = j if word[i] == first and i < len(UpperCamelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 A__ = tuple(UpperCamelCase ) A__ = new_word if len(UpperCamelCase ) == 1: break else: A__ = get_pairs(UpperCamelCase ) A__ = """ """.join(UpperCamelCase ) A__ = word return word def UpperCamelCase ( self: Optional[int] , UpperCamelCase: Union[str, Any] ): """simple docstring""" A__ = [] for token in re.findall(self.pat , UpperCamelCase ): A__ = """""".join( self.byte_encoder[b] for b in token.encode("""utf-8""" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(UpperCamelCase ).split(""" """ ) ) return bpe_tokens def UpperCamelCase ( self: Dict , UpperCamelCase: int ): """simple docstring""" return self.encoder.get(UpperCamelCase , self.encoder.get(self.unk_token ) ) def UpperCamelCase ( self: Any , UpperCamelCase: Any ): """simple docstring""" return self.decoder.get(UpperCamelCase ) def UpperCamelCase ( self: Any , UpperCamelCase: Tuple ): """simple docstring""" A__ = """""".join(UpperCamelCase ) A__ = bytearray([self.byte_decoder[c] for c in text] ).decode("""utf-8""" , errors=self.errors ) return text def UpperCamelCase ( self: List[str] , UpperCamelCase: str , UpperCamelCase: Optional[str] = None ): """simple docstring""" if not os.path.isdir(UpperCamelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return A__ = os.path.join( UpperCamelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) A__ = os.path.join( UpperCamelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""merges_file"""] ) with open(UpperCamelCase , """w""" , encoding="""utf-8""" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=UpperCamelCase , ensure_ascii=UpperCamelCase ) + """\n""" ) A__ = 0 with open(UpperCamelCase , """w""" , encoding="""utf-8""" ) as writer: writer.write("""#version: 0.2\n""" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda UpperCamelCase : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" """ Please check that the tokenizer is not corrupted!""" ) A__ = token_index writer.write(""" """.join(UpperCamelCase ) + """\n""" ) index += 1 return vocab_file, merge_file def UpperCamelCase ( self: int , UpperCamelCase: List[int] , UpperCamelCase: Optional[List[int]] = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] A__ = [self.cls_token_id] A__ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCamelCase ( self: Union[str, Any] , UpperCamelCase: List[int] , UpperCamelCase: Optional[List[int]] = None , UpperCamelCase: bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase , token_ids_a=UpperCamelCase , already_has_special_tokens=UpperCamelCase ) if token_ids_a is None: return [1] + ([0] * len(UpperCamelCase )) + [1] return [1] + ([0] * len(UpperCamelCase )) + [1, 1] + ([0] * len(UpperCamelCase )) + [1] def UpperCamelCase ( self: List[str] , UpperCamelCase: List[int] , UpperCamelCase: Optional[List[int]] = None ): """simple docstring""" A__ = [self.sep_token_id] A__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase ( self: int , UpperCamelCase: Dict , UpperCamelCase: str=False , **UpperCamelCase: Union[str, Any] ): """simple docstring""" A__ = kwargs.pop("""add_prefix_space""" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(UpperCamelCase ) > 0 and not text[0].isspace()): A__ = """ """ + text return (text, kwargs) def UpperCamelCase ( self: str , UpperCamelCase: Union[Dict[str, EncodedInput], BatchEncoding] , UpperCamelCase: Optional[int] = None , UpperCamelCase: PaddingStrategy = PaddingStrategy.DO_NOT_PAD , UpperCamelCase: Optional[int] = None , UpperCamelCase: Optional[bool] = None , ): """simple docstring""" A__ = super()._pad( encoded_inputs=UpperCamelCase , max_length=UpperCamelCase , padding_strategy=UpperCamelCase , pad_to_multiple_of=UpperCamelCase , return_attention_mask=UpperCamelCase , ) # Load from model defaults if return_attention_mask is None: A__ = """attention_mask""" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: A__ = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. A__ = len(encoded_inputs["""global_attention_mask"""] ) != len(UpperCamelCase ) if needs_to_be_padded: A__ = len(UpperCamelCase ) - len(encoded_inputs["""global_attention_mask"""] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` A__ = ( encoded_inputs["""global_attention_mask"""] + [-1] * difference ) elif self.padding_side == "left": A__ = [-1] * difference + encoded_inputs[ """global_attention_mask""" ] else: raise ValueError("""Invalid padding strategy:""" + str(self.padding_side ) ) return encoded_inputs
335
"""simple docstring""" import math class a : """simple docstring""" def __init__( self: List[Any] , UpperCamelCase: List[str]=0 ): # a graph with Node 0,1,...,N-1 """simple docstring""" A__ = n A__ = [ [math.inf for j in range(0 , UpperCamelCase )] for i in range(0 , UpperCamelCase ) ] # adjacency matrix for weight A__ = [ [math.inf for j in range(0 , UpperCamelCase )] for i in range(0 , UpperCamelCase ) ] # dp[i][j] stores minimum distance from i to j def UpperCamelCase ( self: Union[str, Any] , UpperCamelCase: Optional[int] , UpperCamelCase: Union[str, Any] , UpperCamelCase: Tuple ): """simple docstring""" A__ = w def UpperCamelCase ( self: int ): """simple docstring""" for k in range(0 , self.n ): for i in range(0 , self.n ): for j in range(0 , self.n ): A__ = min(self.dp[i][j] , self.dp[i][k] + self.dp[k][j] ) def UpperCamelCase ( self: int , UpperCamelCase: List[str] , UpperCamelCase: Dict ): """simple docstring""" return self.dp[u][v] if __name__ == "__main__": SCREAMING_SNAKE_CASE_ : List[Any] = Graph(5) graph.add_edge(0, 2, 9) graph.add_edge(0, 4, 1_0) graph.add_edge(1, 3, 5) graph.add_edge(2, 3, 7) graph.add_edge(3, 0, 1_0) graph.add_edge(3, 1, 2) graph.add_edge(3, 2, 1) graph.add_edge(3, 4, 6) graph.add_edge(4, 1, 3) graph.add_edge(4, 2, 4) graph.add_edge(4, 3, 9) graph.floyd_warshall() graph.show_min(1, 4) graph.show_min(0, 3)
335
1
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _UpperCamelCase : str = logging.get_logger(__name__) _UpperCamelCase : str = {'vocab_file': 'vocab.json'} _UpperCamelCase : Optional[Any] = { 'vocab_file': { 'mgp-str': 'https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json', } } _UpperCamelCase : List[str] = {'mgp-str': 27} class snake_case ( UpperCAmelCase ): __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : List[Any] , A : Tuple , A : str="[GO]" , A : List[str]="[GO]" , A : Optional[Any]="[s]" , A : Any="[GO]" , **A : str ): '''simple docstring''' super().__init__( unk_token=A , bos_token=A , eos_token=A , pad_token=A , **A , ) with open(A , encoding='utf-8' ) as vocab_handle: a : Optional[Any] = json.load(A ) a : Union[str, Any] = {v: k for k, v in self.vocab.items()} @property def lowerCamelCase__ ( self : int ): '''simple docstring''' return len(self.vocab ) def lowerCamelCase__ ( self : Union[str, Any] ): '''simple docstring''' return dict(self.vocab , **self.added_tokens_encoder ) def lowerCamelCase__ ( self : List[Any] , A : Tuple ): '''simple docstring''' a : Any = [] for s in text: char_tokens.extend(A ) return char_tokens def lowerCamelCase__ ( self : List[Any] , A : Union[str, Any] ): '''simple docstring''' return self.vocab.get(A , self.vocab.get(self.unk_token ) ) def lowerCamelCase__ ( self : Optional[int] , A : Optional[int] ): '''simple docstring''' return self.decoder.get(A ) def lowerCamelCase__ ( self : List[Any] , A : str , A : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(A ): logger.error('Vocabulary path ({}) should be a directory'.format(A ) ) return a : List[str] = os.path.join( A , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) with open(A , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=A , ensure_ascii=A ) + '\n' ) return (vocab_file,)
186
"""simple docstring""" import inspect import unittest import numpy as np from transformers import ViTConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax from transformers.models.vit.modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel class snake_case ( unittest.TestCase ): def __init__( self : List[str] , A : Union[str, Any] , A : Optional[Any]=1_3 , A : List[Any]=3_0 , A : List[Any]=2 , A : Optional[Any]=3 , A : Union[str, Any]=True , A : Union[str, Any]=True , A : Optional[int]=3_2 , A : Tuple=5 , A : List[str]=4 , A : List[Any]=3_7 , A : Optional[Any]="gelu" , A : Any=0.1 , A : Tuple=0.1 , A : Optional[int]=1_0 , A : Union[str, Any]=0.02 , ): '''simple docstring''' a : Optional[Any] = parent a : Tuple = batch_size a : int = image_size a : str = patch_size a : List[str] = num_channels a : List[str] = is_training a : List[str] = use_labels a : Optional[int] = hidden_size a : Optional[Any] = num_hidden_layers a : Optional[int] = num_attention_heads a : str = intermediate_size a : List[str] = hidden_act a : List[str] = hidden_dropout_prob a : Union[str, Any] = attention_probs_dropout_prob a : List[Any] = type_sequence_label_size a : Optional[Any] = initializer_range # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) a : Optional[int] = (image_size // patch_size) ** 2 a : List[Any] = num_patches + 1 def lowerCamelCase__ ( self : Tuple ): '''simple docstring''' a : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) a : str = ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A , initializer_range=self.initializer_range , ) return config, pixel_values def lowerCamelCase__ ( self : Union[str, Any] , A : str , A : Union[str, Any] ): '''simple docstring''' a : Tuple = FlaxViTModel(config=A ) a : int = model(A ) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) a : Optional[Any] = (self.image_size, self.image_size) a : List[str] = (self.patch_size, self.patch_size) a : Optional[Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, num_patches + 1, self.hidden_size) ) def lowerCamelCase__ ( self : Tuple , A : Dict , A : Optional[int] ): '''simple docstring''' a : Optional[Any] = self.type_sequence_label_size a : List[Any] = FlaxViTForImageClassification(config=A ) a : Tuple = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images a : Dict = 1 a : Tuple = FlaxViTForImageClassification(A ) a : List[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) a : Optional[int] = model(A ) def lowerCamelCase__ ( self : Optional[int] ): '''simple docstring''' a : Optional[int] = self.prepare_config_and_inputs() ( ( a ), ( a ), ) : Dict = config_and_inputs a : List[Any] = {'pixel_values': pixel_values} return config, inputs_dict @require_flax class snake_case ( UpperCAmelCase , unittest.TestCase ): __magic_name__ = (FlaxViTModel, FlaxViTForImageClassification) if is_flax_available() else () def lowerCamelCase__ ( self : Any ): '''simple docstring''' a : Any = FlaxViTModelTester(self ) a : List[str] = ConfigTester(self , config_class=A , has_text_modality=A , hidden_size=3_7 ) def lowerCamelCase__ ( self : Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase__ ( self : List[Any] ): '''simple docstring''' a : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def lowerCamelCase__ ( self : List[Any] ): '''simple docstring''' a : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A ) def lowerCamelCase__ ( self : Dict ): '''simple docstring''' a, a : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a : Tuple = model_class(A ) a : str = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic a : List[str] = [*signature.parameters.keys()] a : Dict = ['pixel_values'] self.assertListEqual(arg_names[:1] , A ) def lowerCamelCase__ ( self : Any ): '''simple docstring''' a, a : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): a : List[Any] = self._prepare_for_class(A , A ) a : Tuple = model_class(A ) @jax.jit def model_jitted(A : Tuple , **A : int ): return model(pixel_values=A , **A ) with self.subTest('JIT Enabled' ): a : List[str] = model_jitted(**A ).to_tuple() with self.subTest('JIT Disabled' ): with jax.disable_jit(): a : List[str] = model_jitted(**A ).to_tuple() self.assertEqual(len(A ) , len(A ) ) for jitted_output, output in zip(A , A ): self.assertEqual(jitted_output.shape , output.shape ) @slow def lowerCamelCase__ ( self : List[str] ): '''simple docstring''' for model_class_name in self.all_model_classes: a : List[str] = model_class_name.from_pretrained('google/vit-base-patch16-224' ) a : Optional[Any] = model(np.ones((1, 3, 2_2_4, 2_2_4) ) ) self.assertIsNotNone(A )
186
1
'''simple docstring''' from collections import defaultdict def _A ( lowercase__ ): lowercase__ = 1 lowercase__ = True for v in tree[start]: if v not in visited: ret += dfs(lowercase__ ) if ret % 2 == 0: cuts.append(lowercase__ ) return ret def _A ( ): dfs(1 ) if __name__ == "__main__": __A , __A = 10, 9 __A = defaultdict(list) __A = {} __A = [] __A = 0 __A = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
164
'''simple docstring''' def _A ( ): lowercase__ = [31, 28, 31, 30, 31, 30, 31, 31, 30, 31, 30, 31] lowercase__ = 6 lowercase__ = 1 lowercase__ = 1901 lowercase__ = 0 while year < 2001: day += 7 if (year % 4 == 0 and year % 100 != 0) or (year % 400 == 0): if day > days_per_month[month - 1] and month != 2: month += 1 lowercase__ = day - days_per_month[month - 2] elif day > 29 and month == 2: month += 1 lowercase__ = day - 29 else: if day > days_per_month[month - 1]: month += 1 lowercase__ = day - days_per_month[month - 2] if month > 12: year += 1 lowercase__ = 1 if year < 2001 and day == 1: sundays += 1 return sundays if __name__ == "__main__": print(solution())
164
1
"""simple docstring""" import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand lowerCAmelCase__ = ( '''4S 3H 2C 7S 5H''', '''9D 8H 2C 6S 7H''', '''2D 6D 9D TH 7D''', '''TC 8C 2S JH 6C''', '''JH 8S TH AH QH''', '''TS KS 5S 9S AC''', '''KD 6S 9D TH AD''', '''KS 8D 4D 9S 4S''', # pair '''8C 4S KH JS 4D''', # pair '''QH 8H KD JH 8S''', # pair '''KC 4H KS 2H 8D''', # pair '''KD 4S KC 3H 8S''', # pair '''AH 8S AS KC JH''', # pair '''3H 4C 4H 3S 2H''', # 2 pairs '''5S 5D 2C KH KH''', # 2 pairs '''3C KH 5D 5S KH''', # 2 pairs '''AS 3C KH AD KH''', # 2 pairs '''7C 7S 3S 7H 5S''', # 3 of a kind '''7C 7S KH 2H 7H''', # 3 of a kind '''AC KH QH AH AS''', # 3 of a kind '''2H 4D 3C AS 5S''', # straight (low ace) '''3C 5C 4C 2C 6H''', # straight '''6S 8S 7S 5H 9H''', # straight '''JS QS 9H TS KH''', # straight '''QC KH TS JS AH''', # straight (high ace) '''8C 9C 5C 3C TC''', # flush '''3S 8S 9S 5S KS''', # flush '''4C 5C 9C 8C KC''', # flush '''JH 8H AH KH QH''', # flush '''3D 2H 3H 2C 2D''', # full house '''2H 2C 3S 3H 3D''', # full house '''KH KC 3S 3H 3D''', # full house '''JC 6H JS JD JH''', # 4 of a kind '''JC 7H JS JD JH''', # 4 of a kind '''JC KH JS JD JH''', # 4 of a kind '''2S AS 4S 5S 3S''', # straight flush (low ace) '''2D 6D 3D 4D 5D''', # straight flush '''5C 6C 3C 7C 4C''', # straight flush '''JH 9H TH KH QH''', # straight flush '''JH AH TH KH QH''', # royal flush (high ace straight flush) ) lowerCAmelCase__ = ( ('''2H 3H 4H 5H 6H''', '''KS AS TS QS JS''', '''Loss'''), ('''2H 3H 4H 5H 6H''', '''AS AD AC AH JD''', '''Win'''), ('''AS AH 2H AD AC''', '''JS JD JC JH 3D''', '''Win'''), ('''2S AH 2H AS AC''', '''JS JD JC JH AD''', '''Loss'''), ('''2S AH 2H AS AC''', '''2H 3H 5H 6H 7H''', '''Win'''), ('''AS 3S 4S 8S 2S''', '''2H 3H 5H 6H 7H''', '''Win'''), ('''2H 3H 5H 6H 7H''', '''2S 3H 4H 5S 6C''', '''Win'''), ('''2S 3H 4H 5S 6C''', '''3D 4C 5H 6H 2S''', '''Tie'''), ('''2S 3H 4H 5S 6C''', '''AH AC 5H 6H AS''', '''Win'''), ('''2S 2H 4H 5S 4C''', '''AH AC 5H 6H AS''', '''Loss'''), ('''2S 2H 4H 5S 4C''', '''AH AC 5H 6H 7S''', '''Win'''), ('''6S AD 7H 4S AS''', '''AH AC 5H 6H 7S''', '''Loss'''), ('''2S AH 4H 5S KC''', '''AH AC 5H 6H 7S''', '''Loss'''), ('''2S 3H 6H 7S 9C''', '''7H 3C TH 6H 9S''', '''Loss'''), ('''4S 5H 6H TS AC''', '''3S 5H 6H TS AC''', '''Win'''), ('''2S AH 4H 5S 6C''', '''AD 4C 5H 6H 2C''', '''Tie'''), ('''AS AH 3H AD AC''', '''AS AH 2H AD AC''', '''Win'''), ('''AH AC 5H 5C QS''', '''AH AC 5H 5C KS''', '''Loss'''), ('''AH AC 5H 5C QS''', '''KH KC 5H 5C QS''', '''Win'''), ('''7C 7S KH 2H 7H''', '''3C 3S AH 2H 3H''', '''Win'''), ('''3C 3S AH 2H 3H''', '''7C 7S KH 2H 7H''', '''Loss'''), ('''6H 5H 4H 3H 2H''', '''5H 4H 3H 2H AH''', '''Win'''), ('''5H 4H 3H 2H AH''', '''5H 4H 3H 2H AH''', '''Tie'''), ('''5H 4H 3H 2H AH''', '''6H 5H 4H 3H 2H''', '''Loss'''), ('''AH AD KS KC AC''', '''AH KD KH AC KC''', '''Win'''), ('''2H 4D 3C AS 5S''', '''2H 4D 3C 6S 5S''', '''Loss'''), ('''2H 3S 3C 3H 2S''', '''3S 3C 2S 2H 2D''', '''Win'''), ('''4D 6D 5D 2D JH''', '''3S 8S 3H TC KH''', '''Loss'''), ('''4S 6C 8S 3S 7S''', '''AD KS 2D 7D 7C''', '''Loss'''), ('''6S 4C 7H 8C 3H''', '''5H JC AH 9D 9C''', '''Loss'''), ('''9D 9H JH TC QH''', '''3C 2S JS 5C 7H''', '''Win'''), ('''2H TC 8S AD 9S''', '''4H TS 7H 2C 5C''', '''Win'''), ('''9D 3S 2C 7S 7C''', '''JC TD 3C TC 9H''', '''Loss'''), ) lowerCAmelCase__ = ( ('''2H 3H 4H 5H 6H''', True), ('''AS AH 2H AD AC''', False), ('''2H 3H 5H 6H 7H''', True), ('''KS AS TS QS JS''', True), ('''8H 9H QS JS TH''', False), ('''AS 3S 4S 8S 2S''', True), ) lowerCAmelCase__ = ( ('''2H 3H 4H 5H 6H''', True), ('''AS AH 2H AD AC''', False), ('''2H 3H 5H 6H 7H''', False), ('''KS AS TS QS JS''', True), ('''8H 9H QS JS TH''', True), ) lowerCAmelCase__ = ( ('''2H 4D 3C AS 5S''', True, [5, 4, 3, 2, 14]), ('''2H 5D 3C AS 5S''', False, [14, 5, 5, 3, 2]), ('''JH QD KC AS TS''', False, [14, 13, 12, 11, 10]), ('''9D 3S 2C 7S 7C''', False, [9, 7, 7, 3, 2]), ) lowerCAmelCase__ = ( ('''JH AH TH KH QH''', 0), ('''JH 9H TH KH QH''', 0), ('''JC KH JS JD JH''', 7), ('''KH KC 3S 3H 3D''', 6), ('''8C 9C 5C 3C TC''', 0), ('''JS QS 9H TS KH''', 0), ('''7C 7S KH 2H 7H''', 3), ('''3C KH 5D 5S KH''', 2), ('''QH 8H KD JH 8S''', 1), ('''2D 6D 9D TH 7D''', 0), ) lowerCAmelCase__ = ( ('''JH AH TH KH QH''', 23), ('''JH 9H TH KH QH''', 22), ('''JC KH JS JD JH''', 21), ('''KH KC 3S 3H 3D''', 20), ('''8C 9C 5C 3C TC''', 19), ('''JS QS 9H TS KH''', 18), ('''7C 7S KH 2H 7H''', 17), ('''3C KH 5D 5S KH''', 16), ('''QH 8H KD JH 8S''', 15), ('''2D 6D 9D TH 7D''', 14), ) def a__ ( ): '''simple docstring''' lowerCAmelCase : str = randrange(len(SCREAMING_SNAKE_CASE ) ), randrange(len(SCREAMING_SNAKE_CASE ) ) lowerCAmelCase : Any = ["Loss", "Tie", "Win"][(play >= oppo) + (play > oppo)] lowerCAmelCase : Union[str, Any] = SORTED_HANDS[play], SORTED_HANDS[oppo] return hand, other, expected def a__ ( SCREAMING_SNAKE_CASE : int = 1_0_0 ): '''simple docstring''' return (generate_random_hand() for _ in range(SCREAMING_SNAKE_CASE )) @pytest.mark.parametrize("hand, expected" , SCREAMING_SNAKE_CASE ) def a__ ( SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' assert PokerHand(SCREAMING_SNAKE_CASE )._is_flush() == expected @pytest.mark.parametrize("hand, expected" , SCREAMING_SNAKE_CASE ) def a__ ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' assert PokerHand(SCREAMING_SNAKE_CASE )._is_straight() == expected @pytest.mark.parametrize("hand, expected, card_values" , SCREAMING_SNAKE_CASE ) def a__ ( SCREAMING_SNAKE_CASE : Optional[Any] , SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' lowerCAmelCase : Dict = PokerHand(SCREAMING_SNAKE_CASE ) assert player._is_five_high_straight() == expected assert player._card_values == card_values @pytest.mark.parametrize("hand, expected" , SCREAMING_SNAKE_CASE ) def a__ ( SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' assert PokerHand(SCREAMING_SNAKE_CASE )._is_same_kind() == expected @pytest.mark.parametrize("hand, expected" , SCREAMING_SNAKE_CASE ) def a__ ( SCREAMING_SNAKE_CASE : List[str] , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' assert PokerHand(SCREAMING_SNAKE_CASE )._hand_type == expected @pytest.mark.parametrize("hand, other, expected" , SCREAMING_SNAKE_CASE ) def a__ ( SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' assert PokerHand(SCREAMING_SNAKE_CASE ).compare_with(PokerHand(SCREAMING_SNAKE_CASE ) ) == expected @pytest.mark.parametrize("hand, other, expected" , generate_random_hands() ) def a__ ( SCREAMING_SNAKE_CASE : List[str] , SCREAMING_SNAKE_CASE : Optional[int] , SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' assert PokerHand(SCREAMING_SNAKE_CASE ).compare_with(PokerHand(SCREAMING_SNAKE_CASE ) ) == expected def a__ ( ): '''simple docstring''' lowerCAmelCase : str = [PokerHand(SCREAMING_SNAKE_CASE ) for hand in SORTED_HANDS] lowerCAmelCase : Optional[int] = poker_hands.copy() shuffle(SCREAMING_SNAKE_CASE ) lowerCAmelCase : Optional[Any] = chain(sorted(SCREAMING_SNAKE_CASE ) ) for index, hand in enumerate(SCREAMING_SNAKE_CASE ): assert hand == poker_hands[index] def a__ ( ): '''simple docstring''' lowerCAmelCase : Optional[Any] = [PokerHand("2D AC 3H 4H 5S" ), PokerHand("2S 3H 4H 5S 6C" )] pokerhands.sort(reverse=SCREAMING_SNAKE_CASE ) assert pokerhands[0].__str__() == "2S 3H 4H 5S 6C" def a__ ( ): '''simple docstring''' lowerCAmelCase : List[str] = PokerHand("2C 4S AS 3D 5C" ) lowerCAmelCase : int = True lowerCAmelCase : str = [5, 4, 3, 2, 1_4] for _ in range(1_0 ): assert pokerhand._is_five_high_straight() == expected assert pokerhand._card_values == expected_card_values def a__ ( ): '''simple docstring''' lowerCAmelCase : Optional[Any] = 0 lowerCAmelCase : List[Any] = os.path.abspath(os.path.dirname(SCREAMING_SNAKE_CASE ) ) lowerCAmelCase : int = os.path.join(SCREAMING_SNAKE_CASE , "poker_hands.txt" ) with open(SCREAMING_SNAKE_CASE ) as file_hand: for line in file_hand: lowerCAmelCase : List[Any] = line[:1_4].strip() lowerCAmelCase : Tuple = line[1_5:].strip() lowerCAmelCase : int = PokerHand(SCREAMING_SNAKE_CASE ), PokerHand(SCREAMING_SNAKE_CASE ) lowerCAmelCase : Optional[int] = player.compare_with(SCREAMING_SNAKE_CASE ) if output == "Win": answer += 1 assert answer == 3_7_6
356
"""simple docstring""" from ...configuration_utils import PretrainedConfig lowerCAmelCase__ = { '''google/tapas-base-finetuned-sqa''': ( '''https://huggingface.co/google/tapas-base-finetuned-sqa/resolve/main/config.json''' ), '''google/tapas-base-finetuned-wtq''': ( '''https://huggingface.co/google/tapas-base-finetuned-wtq/resolve/main/config.json''' ), '''google/tapas-base-finetuned-wikisql-supervised''': ( '''https://huggingface.co/google/tapas-base-finetuned-wikisql-supervised/resolve/main/config.json''' ), '''google/tapas-base-finetuned-tabfact''': ( '''https://huggingface.co/google/tapas-base-finetuned-tabfact/resolve/main/config.json''' ), } class SCREAMING_SNAKE_CASE__ ( lowercase ): """simple docstring""" a : Any ="tapas" def __init__( self , snake_case__=30_522 , snake_case__=768 , snake_case__=12 , snake_case__=12 , snake_case__=3_072 , snake_case__="gelu" , snake_case__=0.1 , snake_case__=0.1 , snake_case__=1_024 , snake_case__=[3, 256, 256, 2, 256, 256, 10] , snake_case__=0.02 , snake_case__=1e-12 , snake_case__=0 , snake_case__=10.0 , snake_case__=0 , snake_case__=1.0 , snake_case__=None , snake_case__=1.0 , snake_case__=False , snake_case__=None , snake_case__=1.0 , snake_case__=1.0 , snake_case__=False , snake_case__=False , snake_case__="ratio" , snake_case__=None , snake_case__=None , snake_case__=64 , snake_case__=32 , snake_case__=False , snake_case__=True , snake_case__=False , snake_case__=False , snake_case__=True , snake_case__=False , snake_case__=None , snake_case__=None , **snake_case__ , ): """simple docstring""" super().__init__(pad_token_id=snake_case__ , **snake_case__ ) # BERT hyperparameters (with updated max_position_embeddings and type_vocab_sizes) lowerCAmelCase : List[Any] = vocab_size lowerCAmelCase : List[str] = hidden_size lowerCAmelCase : Optional[Any] = num_hidden_layers lowerCAmelCase : List[str] = num_attention_heads lowerCAmelCase : Optional[Any] = hidden_act lowerCAmelCase : List[Any] = intermediate_size lowerCAmelCase : Optional[Any] = hidden_dropout_prob lowerCAmelCase : str = attention_probs_dropout_prob lowerCAmelCase : Any = max_position_embeddings lowerCAmelCase : Dict = type_vocab_sizes lowerCAmelCase : Union[str, Any] = initializer_range lowerCAmelCase : str = layer_norm_eps # Fine-tuning task hyperparameters lowerCAmelCase : Dict = positive_label_weight lowerCAmelCase : Union[str, Any] = num_aggregation_labels lowerCAmelCase : Optional[Any] = aggregation_loss_weight lowerCAmelCase : List[Any] = use_answer_as_supervision lowerCAmelCase : Dict = answer_loss_importance lowerCAmelCase : List[Any] = use_normalized_answer_loss lowerCAmelCase : List[str] = huber_loss_delta lowerCAmelCase : Optional[int] = temperature lowerCAmelCase : Optional[int] = aggregation_temperature lowerCAmelCase : Any = use_gumbel_for_cells lowerCAmelCase : Union[str, Any] = use_gumbel_for_aggregation lowerCAmelCase : Union[str, Any] = average_approximation_function lowerCAmelCase : int = cell_selection_preference lowerCAmelCase : Dict = answer_loss_cutoff lowerCAmelCase : Optional[int] = max_num_rows lowerCAmelCase : Union[str, Any] = max_num_columns lowerCAmelCase : Any = average_logits_per_cell lowerCAmelCase : List[Any] = select_one_column lowerCAmelCase : Tuple = allow_empty_column_selection lowerCAmelCase : str = init_cell_selection_weights_to_zero lowerCAmelCase : List[Any] = reset_position_index_per_cell lowerCAmelCase : Optional[Any] = disable_per_token_loss # Aggregation hyperparameters lowerCAmelCase : List[str] = aggregation_labels lowerCAmelCase : List[str] = no_aggregation_label_index if isinstance(self.aggregation_labels , snake_case__ ): lowerCAmelCase : Union[str, Any] = {int(snake_case__ ): v for k, v in aggregation_labels.items()}
133
0
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __UpperCAmelCase ( A__ ): '''simple docstring''' __lowerCAmelCase = ['''image_processor''', '''tokenizer'''] __lowerCAmelCase = '''CLIPImageProcessor''' __lowerCAmelCase = ('''CLIPTokenizer''', '''CLIPTokenizerFast''') def __init__(self : List[Any] , _lowerCAmelCase : Union[str, Any]=None , _lowerCAmelCase : Dict=None , **_lowerCAmelCase : Optional[Any] ): A = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , __UpperCamelCase , ) A = kwargs.pop("""feature_extractor""" ) A = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(__UpperCamelCase , __UpperCamelCase ) def __call__(self : Dict , _lowerCAmelCase : List[Any]=None , _lowerCAmelCase : Any=None , _lowerCAmelCase : Optional[Any]=None , **_lowerCAmelCase : Optional[Any] ): if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: A = self.tokenizer(__UpperCamelCase , return_tensors=__UpperCamelCase , **__UpperCamelCase ) if images is not None: A = self.image_processor(__UpperCamelCase , return_tensors=__UpperCamelCase , **__UpperCamelCase ) if text is not None and images is not None: A = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__UpperCamelCase ) , tensor_type=__UpperCamelCase ) def A (self : Tuple , *_lowerCAmelCase : Tuple , **_lowerCAmelCase : Optional[int] ): return self.tokenizer.batch_decode(*__UpperCamelCase , **__UpperCamelCase ) def A (self : Any , *_lowerCAmelCase : Optional[int] , **_lowerCAmelCase : Optional[Any] ): return self.tokenizer.decode(*__UpperCamelCase , **__UpperCamelCase ) @property def A (self : Tuple ): A = self.tokenizer.model_input_names A = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def A (self : int ): warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" , __UpperCamelCase , ) return self.image_processor_class @property def A (self : int ): warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" , __UpperCamelCase , ) return self.image_processor
258
"""simple docstring""" def lowercase ( _SCREAMING_SNAKE_CASE : list[list[int]] , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : set ): '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase = len(_SCREAMING_SNAKE_CASE ), len(grid[0] ) if ( min(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) _UpperCAmelCase = 0 count += depth_first_search(_SCREAMING_SNAKE_CASE , row + 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) count += depth_first_search(_SCREAMING_SNAKE_CASE , row - 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) count += depth_first_search(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , col + 1 , _SCREAMING_SNAKE_CASE ) count += depth_first_search(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , col - 1 , _SCREAMING_SNAKE_CASE ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
260
0
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetrImageProcessor class __snake_case ( unittest.TestCase ): def __init__( self , lowercase , lowercase=7 , lowercase=3 , lowercase=30 , lowercase=4_00 , lowercase=True , lowercase=None , lowercase=True , lowercase=1 / 2_55 , lowercase=True , lowercase=[0.5, 0.5, 0.5] , lowercase=[0.5, 0.5, 0.5] , lowercase=True , ) -> Any: '''simple docstring''' a__: str = size if size is not None else {'shortest_edge': 18, 'longest_edge': 13_33} a__: List[str] = parent a__: Optional[int] = batch_size a__: int = num_channels a__: Any = min_resolution a__: Tuple = max_resolution a__: Optional[Any] = do_resize a__: Any = size a__: int = do_rescale a__: List[Any] = rescale_factor a__: Union[str, Any] = do_normalize a__: List[Any] = image_mean a__: Tuple = image_std a__: Any = do_pad def lowerCamelCase_ ( self) -> int: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_pad": self.do_pad, } def lowerCamelCase_ ( self , lowercase , lowercase=False) -> Tuple: '''simple docstring''' if not batched: a__: List[str] = image_inputs[0] if isinstance(lowercase , Image.Image): a__ , a__: Dict = image.size else: a__ , a__: List[str] = image.shape[1], image.shape[2] if w < h: a__: Tuple = int(self.size['shortest_edge'] * h / w) a__: Tuple = self.size['shortest_edge'] elif w > h: a__: Optional[Any] = self.size['shortest_edge'] a__: List[str] = int(self.size['shortest_edge'] * w / h) else: a__: Optional[Any] = self.size['shortest_edge'] a__: int = self.size['shortest_edge'] else: a__: int = [] for image in image_inputs: a__ , a__: Any = self.get_expected_values([image]) expected_values.append((expected_height, expected_width)) a__: List[Any] = max(lowercase , key=lambda lowercase: item[0])[0] a__: Any = max(lowercase , key=lambda lowercase: item[1])[1] return expected_height, expected_width @require_torch @require_vision class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = DetrImageProcessor if is_vision_available() else None def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: int = DetrImageProcessingTester(self) @property def lowerCamelCase_ ( self) -> int: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def lowerCamelCase_ ( self) -> int: '''simple docstring''' a__: Union[str, Any] = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(lowercase , 'image_mean')) self.assertTrue(hasattr(lowercase , 'image_std')) self.assertTrue(hasattr(lowercase , 'do_normalize')) self.assertTrue(hasattr(lowercase , 'do_rescale')) self.assertTrue(hasattr(lowercase , 'rescale_factor')) self.assertTrue(hasattr(lowercase , 'do_resize')) self.assertTrue(hasattr(lowercase , 'size')) self.assertTrue(hasattr(lowercase , 'do_pad')) def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: Any = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 13_33}) self.assertEqual(image_processor.do_pad , lowercase) a__: int = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowercase) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84}) self.assertEqual(image_processor.do_pad , lowercase) def lowerCamelCase_ ( self) -> Any: '''simple docstring''' pass def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: Any = self.image_processing_class(**self.image_processor_dict) # create random PIL images a__: List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase) for image in image_inputs: self.assertIsInstance(lowercase , Image.Image) # Test not batched input a__: str = image_processing(image_inputs[0] , return_tensors='pt').pixel_values a__ , a__: Dict = self.image_processor_tester.get_expected_values(lowercase) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched a__ , a__: Tuple = self.image_processor_tester.get_expected_values(lowercase , batched=lowercase) a__: Any = image_processing(lowercase , return_tensors='pt').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Optional[int] = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors a__: str = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase , numpify=lowercase) for image in image_inputs: self.assertIsInstance(lowercase , np.ndarray) # Test not batched input a__: Tuple = image_processing(image_inputs[0] , return_tensors='pt').pixel_values a__ , a__: Any = self.image_processor_tester.get_expected_values(lowercase) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched a__: Any = image_processing(lowercase , return_tensors='pt').pixel_values a__ , a__: Optional[Any] = self.image_processor_tester.get_expected_values(lowercase , batched=lowercase) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: int = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors a__: Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase , torchify=lowercase) for image in image_inputs: self.assertIsInstance(lowercase , torch.Tensor) # Test not batched input a__: Optional[Any] = image_processing(image_inputs[0] , return_tensors='pt').pixel_values a__ , a__: Dict = self.image_processor_tester.get_expected_values(lowercase) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched a__: Optional[int] = image_processing(lowercase , return_tensors='pt').pixel_values a__ , a__: Union[str, Any] = self.image_processor_tester.get_expected_values(lowercase , batched=lowercase) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' a__: Optional[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png') with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r') as f: a__: Tuple = json.loads(f.read()) a__: Tuple = {'image_id': 3_97_69, 'annotations': target} # encode them a__: Optional[Any] = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50') a__: List[Any] = image_processing(images=lowercase , annotations=lowercase , return_tensors='pt') # verify pixel values a__: Optional[Any] = torch.Size([1, 3, 8_00, 10_66]) self.assertEqual(encoding['pixel_values'].shape , lowercase) a__: int = torch.tensor([0.2796, 0.3138, 0.3481]) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , lowercase , atol=1e-4)) # verify area a__: Optional[int] = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438]) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , lowercase)) # verify boxes a__: Optional[int] = torch.Size([6, 4]) self.assertEqual(encoding['labels'][0]['boxes'].shape , lowercase) a__: List[str] = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215]) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , lowercase , atol=1e-3)) # verify image_id a__: Dict = torch.tensor([3_97_69]) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , lowercase)) # verify is_crowd a__: str = torch.tensor([0, 0, 0, 0, 0, 0]) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , lowercase)) # verify class_labels a__: Union[str, Any] = torch.tensor([75, 75, 63, 65, 17, 17]) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , lowercase)) # verify orig_size a__: List[Any] = torch.tensor([4_80, 6_40]) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , lowercase)) # verify size a__: str = torch.tensor([8_00, 10_66]) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , lowercase)) @slow def lowerCamelCase_ ( self) -> Optional[Any]: '''simple docstring''' a__: List[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png') with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r') as f: a__: List[str] = json.loads(f.read()) a__: int = {'file_name': '000000039769.png', 'image_id': 3_97_69, 'segments_info': target} a__: str = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic') # encode them a__: int = DetrImageProcessor.from_pretrained('facebook/detr-resnet-50-panoptic') a__: Union[str, Any] = image_processing(images=lowercase , annotations=lowercase , masks_path=lowercase , return_tensors='pt') # verify pixel values a__: Any = torch.Size([1, 3, 8_00, 10_66]) self.assertEqual(encoding['pixel_values'].shape , lowercase) a__: Any = torch.tensor([0.2796, 0.3138, 0.3481]) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , lowercase , atol=1e-4)) # verify area a__: List[str] = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147]) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , lowercase)) # verify boxes a__: Union[str, Any] = torch.Size([6, 4]) self.assertEqual(encoding['labels'][0]['boxes'].shape , lowercase) a__: Union[str, Any] = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625]) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , lowercase , atol=1e-3)) # verify image_id a__: List[Any] = torch.tensor([3_97_69]) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , lowercase)) # verify is_crowd a__: Optional[Any] = torch.tensor([0, 0, 0, 0, 0, 0]) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , lowercase)) # verify class_labels a__: Dict = torch.tensor([17, 17, 63, 75, 75, 93]) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , lowercase)) # verify masks a__: Any = 82_28_73 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , lowercase) # verify orig_size a__: Union[str, Any] = torch.tensor([4_80, 6_40]) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , lowercase)) # verify size a__: Optional[int] = torch.tensor([8_00, 10_66]) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , lowercase))
203
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: lowercase__ = None lowercase__ = logging.get_logger(__name__) lowercase__ = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} lowercase__ = { 'vocab_file': { 'facebook/mbart-large-en-ro': ( 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model' ), 'facebook/mbart-large-cc25': ( 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model' ), }, 'tokenizer_file': { 'facebook/mbart-large-en-ro': 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json', 'facebook/mbart-large-cc25': 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json', }, } lowercase__ = { 'facebook/mbart-large-en-ro': 1024, 'facebook/mbart-large-cc25': 1024, } # fmt: off lowercase__ = ['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN'] class __snake_case ( __lowerCAmelCase ): a__ = VOCAB_FILES_NAMES a__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ = PRETRAINED_VOCAB_FILES_MAP a__ = ["""input_ids""", """attention_mask"""] a__ = MBartTokenizer a__ = [] a__ = [] def __init__( self , lowercase=None , lowercase=None , lowercase="<s>" , lowercase="</s>" , lowercase="</s>" , lowercase="<s>" , lowercase="<unk>" , lowercase="<pad>" , lowercase="<mask>" , lowercase=None , lowercase=None , lowercase=None , **lowercase , ) -> Tuple: '''simple docstring''' a__: List[str] = AddedToken(lowercase , lstrip=lowercase , rstrip=lowercase) if isinstance(lowercase , lowercase) else mask_token super().__init__( vocab_file=lowercase , tokenizer_file=lowercase , bos_token=lowercase , eos_token=lowercase , sep_token=lowercase , cls_token=lowercase , unk_token=lowercase , pad_token=lowercase , mask_token=lowercase , src_lang=lowercase , tgt_lang=lowercase , additional_special_tokens=lowercase , **lowercase , ) a__: Tuple = vocab_file a__: Union[str, Any] = False if not self.vocab_file else True a__: Union[str, Any] = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens]) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens}) a__: int = { lang_code: self.convert_tokens_to_ids(lowercase) for lang_code in FAIRSEQ_LANGUAGE_CODES } a__: List[Any] = src_lang if src_lang is not None else 'en_XX' a__: Tuple = self.convert_tokens_to_ids(self._src_lang) a__: str = tgt_lang self.set_src_lang_special_tokens(self._src_lang) @property def lowerCamelCase_ ( self) -> str: '''simple docstring''' return self._src_lang @src_lang.setter def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: Tuple = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def lowerCamelCase_ ( self , lowercase , lowercase = None) -> List[int]: '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowerCamelCase_ ( self , lowercase , lowercase = None) -> List[int]: '''simple docstring''' a__: Any = [self.sep_token_id] a__: List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def lowerCamelCase_ ( self , lowercase , lowercase , lowercase , lowercase , **lowercase) -> Union[str, Any]: '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model') a__: Union[str, Any] = src_lang a__: Any = self(lowercase , add_special_tokens=lowercase , return_tensors=lowercase , **lowercase) a__: str = self.convert_tokens_to_ids(lowercase) a__: Any = tgt_lang_id return inputs def lowerCamelCase_ ( self , lowercase , lowercase = "en_XX" , lowercase = None , lowercase = "ro_RO" , **lowercase , ) -> BatchEncoding: '''simple docstring''' a__: Any = src_lang a__: List[Any] = tgt_lang return super().prepare_seqaseq_batch(lowercase , lowercase , **lowercase) def lowerCamelCase_ ( self) -> Dict: '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: int = self.convert_tokens_to_ids(lowercase) a__: List[Any] = [] a__: List[str] = [self.eos_token_id, self.cur_lang_code] a__: Dict = self.convert_ids_to_tokens(self.prefix_tokens) a__: Any = self.convert_ids_to_tokens(self.suffix_tokens) a__: int = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens)) , ) def lowerCamelCase_ ( self , lowercase) -> None: '''simple docstring''' a__: str = self.convert_tokens_to_ids(lowercase) a__: List[Any] = [] a__: Dict = [self.eos_token_id, self.cur_lang_code] a__: Any = self.convert_ids_to_tokens(self.prefix_tokens) a__: Optional[Any] = self.convert_ids_to_tokens(self.suffix_tokens) a__: str = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens)) , ) def lowerCamelCase_ ( self , lowercase , lowercase = None) -> Tuple[str]: '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.') if not os.path.isdir(lowercase): logger.error(f'Vocabulary path ({save_directory}) should be a directory.') return a__: Any = os.path.join( lowercase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowercase): copyfile(self.vocab_file , lowercase) return (out_vocab_file,)
203
1
'''simple docstring''' import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class UpperCAmelCase__ ( unittest.TestCase): def __lowerCamelCase ( self ) -> Union[str, Any]: __UpperCamelCase = 0 def __lowerCamelCase ( self ) -> Tuple: __UpperCamelCase = AutoImageProcessor.from_pretrained("""openai/clip-vit-base-patch32""" ) self.assertIsInstance(lowercase , lowercase ) def __lowerCamelCase ( self ) -> List[Any]: with tempfile.TemporaryDirectory() as tmpdirname: __UpperCamelCase = Path(lowercase ) / """preprocessor_config.json""" __UpperCamelCase = Path(lowercase ) / """config.json""" json.dump( {"""image_processor_type""": """CLIPImageProcessor""", """processor_class""": """CLIPProcessor"""} , open(lowercase , """w""" ) , ) json.dump({"""model_type""": """clip"""} , open(lowercase , """w""" ) ) __UpperCamelCase = AutoImageProcessor.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) def __lowerCamelCase ( self ) -> List[str]: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: __UpperCamelCase = Path(lowercase ) / """preprocessor_config.json""" __UpperCamelCase = Path(lowercase ) / """config.json""" json.dump( {"""feature_extractor_type""": """CLIPFeatureExtractor""", """processor_class""": """CLIPProcessor"""} , open(lowercase , """w""" ) , ) json.dump({"""model_type""": """clip"""} , open(lowercase , """w""" ) ) __UpperCamelCase = AutoImageProcessor.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) def __lowerCamelCase ( self ) -> List[Any]: with tempfile.TemporaryDirectory() as tmpdirname: __UpperCamelCase = CLIPConfig() # Create a dummy config file with image_proceesor_type __UpperCamelCase = Path(lowercase ) / """preprocessor_config.json""" __UpperCamelCase = Path(lowercase ) / """config.json""" json.dump( {"""image_processor_type""": """CLIPImageProcessor""", """processor_class""": """CLIPProcessor"""} , open(lowercase , """w""" ) , ) json.dump({"""model_type""": """clip"""} , open(lowercase , """w""" ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally __UpperCamelCase = AutoImageProcessor.from_pretrained(lowercase ).to_dict() config_dict.pop("""image_processor_type""" ) __UpperCamelCase = CLIPImageProcessor(**lowercase ) # save in new folder model_config.save_pretrained(lowercase ) config.save_pretrained(lowercase ) __UpperCamelCase = AutoImageProcessor.from_pretrained(lowercase ) # make sure private variable is not incorrectly saved __UpperCamelCase = json.loads(config.to_json_string() ) self.assertTrue("""_processor_class""" not in dict_as_saved ) self.assertIsInstance(lowercase , lowercase ) def __lowerCamelCase ( self ) -> int: with tempfile.TemporaryDirectory() as tmpdirname: __UpperCamelCase = Path(lowercase ) / """preprocessor_config.json""" json.dump( {"""image_processor_type""": """CLIPImageProcessor""", """processor_class""": """CLIPProcessor"""} , open(lowercase , """w""" ) , ) __UpperCamelCase = AutoImageProcessor.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) def __lowerCamelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( lowercase , """clip-base is not a local folder and is not a valid model identifier""" ): __UpperCamelCase = AutoImageProcessor.from_pretrained("""clip-base""" ) def __lowerCamelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( lowercase , r"""aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)""" ): __UpperCamelCase = AutoImageProcessor.from_pretrained(lowercase , revision="""aaaaaa""" ) def __lowerCamelCase ( self ) -> Tuple: with self.assertRaisesRegex( lowercase , """hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.""" , ): __UpperCamelCase = AutoImageProcessor.from_pretrained("""hf-internal-testing/config-no-model""" ) def __lowerCamelCase ( self ) -> Union[str, Any]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(lowercase ): __UpperCamelCase = AutoImageProcessor.from_pretrained("""hf-internal-testing/test_dynamic_image_processor""" ) # If remote code is disabled, we can't load this config. with self.assertRaises(lowercase ): __UpperCamelCase = AutoImageProcessor.from_pretrained( """hf-internal-testing/test_dynamic_image_processor""" , trust_remote_code=lowercase ) __UpperCamelCase = AutoImageProcessor.from_pretrained( """hf-internal-testing/test_dynamic_image_processor""" , trust_remote_code=lowercase ) self.assertEqual(image_processor.__class__.__name__ , """NewImageProcessor""" ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(lowercase ) __UpperCamelCase = AutoImageProcessor.from_pretrained(lowercase , trust_remote_code=lowercase ) self.assertEqual(reloaded_image_processor.__class__.__name__ , """NewImageProcessor""" ) def __lowerCamelCase ( self ) -> Any: try: AutoConfig.register("""custom""" , lowercase ) AutoImageProcessor.register(lowercase , lowercase ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowercase ): AutoImageProcessor.register(lowercase , lowercase ) with tempfile.TemporaryDirectory() as tmpdirname: __UpperCamelCase = Path(lowercase ) / """preprocessor_config.json""" __UpperCamelCase = Path(lowercase ) / """config.json""" json.dump( {"""feature_extractor_type""": """CLIPFeatureExtractor""", """processor_class""": """CLIPProcessor"""} , open(lowercase , """w""" ) , ) json.dump({"""model_type""": """clip"""} , open(lowercase , """w""" ) ) __UpperCamelCase = CustomImageProcessor.from_pretrained(lowercase ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(lowercase ) __UpperCamelCase = AutoImageProcessor.from_pretrained(lowercase ) self.assertIsInstance(lowercase , lowercase ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def __lowerCamelCase ( self ) -> List[str]: class UpperCAmelCase__ ( UpperCAmelCase_): __SCREAMING_SNAKE_CASE = True try: AutoConfig.register("""custom""" , lowercase ) AutoImageProcessor.register(lowercase , lowercase ) # If remote code is not set, the default is to use local __UpperCamelCase = AutoImageProcessor.from_pretrained("""hf-internal-testing/test_dynamic_image_processor""" ) self.assertEqual(image_processor.__class__.__name__ , """NewImageProcessor""" ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. __UpperCamelCase = AutoImageProcessor.from_pretrained( """hf-internal-testing/test_dynamic_image_processor""" , trust_remote_code=lowercase ) self.assertEqual(image_processor.__class__.__name__ , """NewImageProcessor""" ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub __UpperCamelCase = AutoImageProcessor.from_pretrained( """hf-internal-testing/test_dynamic_image_processor""" , trust_remote_code=lowercase ) self.assertEqual(image_processor.__class__.__name__ , """NewImageProcessor""" ) self.assertTrue(not hasattr(lowercase , """is_local""" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
349
'''simple docstring''' import gc import random import tempfile import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline from diffusers.utils import floats_tensor, nightly, torch_device from diffusers.utils.testing_utils import require_torch_gpu class UpperCAmelCase__ ( unittest.TestCase): def __lowerCamelCase ( self ) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def __lowerCamelCase ( self ) -> int: __UpperCamelCase = 1 __UpperCamelCase = 3 __UpperCamelCase = (3_2, 3_2) __UpperCamelCase = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(lowercase ) return image @property def __lowerCamelCase ( self ) -> Dict: torch.manual_seed(0 ) __UpperCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=3_2 , ) return model @property def __lowerCamelCase ( self ) -> List[str]: torch.manual_seed(0 ) __UpperCamelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) return model @property def __lowerCamelCase ( self ) -> Optional[int]: torch.manual_seed(0 ) __UpperCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModel(lowercase ) @property def __lowerCamelCase ( self ) -> Tuple: def extract(*lowercase , **lowercase ): class UpperCAmelCase__ : def __init__( self ) -> Tuple: __UpperCamelCase = torch.ones([0] ) def __lowerCamelCase ( self , lowercase ) -> List[str]: self.pixel_values.to(lowercase ) return self return Out() return extract def __lowerCamelCase ( self ) -> Any: __UpperCamelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator __UpperCamelCase = self.dummy_cond_unet __UpperCamelCase = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=lowercase , set_alpha_to_one=lowercase , ) __UpperCamelCase = self.dummy_vae __UpperCamelCase = self.dummy_text_encoder __UpperCamelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # make sure here that pndm scheduler skips prk __UpperCamelCase = StableDiffusionPipeline( unet=lowercase , scheduler=lowercase , vae=lowercase , text_encoder=lowercase , tokenizer=lowercase , safety_checker=lowercase , feature_extractor=self.dummy_extractor , ) __UpperCamelCase = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) __UpperCamelCase = """A painting of a squirrel eating a burger""" __UpperCamelCase = torch.Generator(device=lowercase ).manual_seed(0 ) __UpperCamelCase = sd_pipe([prompt] , generator=lowercase , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" ) __UpperCamelCase = output.images __UpperCamelCase = torch.Generator(device=lowercase ).manual_seed(0 ) __UpperCamelCase = sd_pipe( [prompt] , generator=lowercase , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" , return_dict=lowercase , )[0] __UpperCamelCase = image[0, -3:, -3:, -1] __UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) __UpperCamelCase = np.array([0.5_756, 0.6_118, 0.5_005, 0.5_041, 0.5_471, 0.4_726, 0.4_976, 0.4_865, 0.4_864] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def __lowerCamelCase ( self ) -> Tuple: __UpperCamelCase = """cpu""" # ensure determinism for the device-dependent torch.Generator __UpperCamelCase = self.dummy_cond_unet __UpperCamelCase = PNDMScheduler(skip_prk_steps=lowercase ) __UpperCamelCase = self.dummy_vae __UpperCamelCase = self.dummy_text_encoder __UpperCamelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # make sure here that pndm scheduler skips prk __UpperCamelCase = StableDiffusionPipeline( unet=lowercase , scheduler=lowercase , vae=lowercase , text_encoder=lowercase , tokenizer=lowercase , safety_checker=lowercase , feature_extractor=self.dummy_extractor , ) __UpperCamelCase = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) __UpperCamelCase = """A painting of a squirrel eating a burger""" __UpperCamelCase = torch.Generator(device=lowercase ).manual_seed(0 ) __UpperCamelCase = sd_pipe([prompt] , generator=lowercase , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" ) __UpperCamelCase = output.images __UpperCamelCase = torch.Generator(device=lowercase ).manual_seed(0 ) __UpperCamelCase = sd_pipe( [prompt] , generator=lowercase , guidance_scale=6.0 , num_inference_steps=2 , output_type="""np""" , return_dict=lowercase , )[0] __UpperCamelCase = image[0, -3:, -3:, -1] __UpperCamelCase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) __UpperCamelCase = np.array([0.5_125, 0.5_716, 0.4_828, 0.5_060, 0.5_650, 0.4_768, 0.5_185, 0.4_895, 0.4_993] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def __lowerCamelCase ( self ) -> Union[str, Any]: __UpperCamelCase = StableDiffusionPipeline.from_pretrained( """hf-internal-testing/tiny-stable-diffusion-lms-pipe""" , safety_checker=lowercase ) assert isinstance(lowercase , lowercase ) assert isinstance(pipe.scheduler , lowercase ) assert pipe.safety_checker is None __UpperCamelCase = pipe("""example prompt""" , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowercase ) __UpperCamelCase = StableDiffusionPipeline.from_pretrained(lowercase ) # sanity check that the pipeline still works assert pipe.safety_checker is None __UpperCamelCase = pipe("""example prompt""" , num_inference_steps=2 ).images[0] assert image is not None @unittest.skipIf(torch_device != """cuda""" , """This test requires a GPU""" ) def __lowerCamelCase ( self ) -> Optional[int]: __UpperCamelCase = self.dummy_cond_unet __UpperCamelCase = PNDMScheduler(skip_prk_steps=lowercase ) __UpperCamelCase = self.dummy_vae __UpperCamelCase = self.dummy_text_encoder __UpperCamelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) # put models in fp16 __UpperCamelCase = unet.half() __UpperCamelCase = vae.half() __UpperCamelCase = bert.half() # make sure here that pndm scheduler skips prk __UpperCamelCase = StableDiffusionPipeline( unet=lowercase , scheduler=lowercase , vae=lowercase , text_encoder=lowercase , tokenizer=lowercase , safety_checker=lowercase , feature_extractor=self.dummy_extractor , ) __UpperCamelCase = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) __UpperCamelCase = """A painting of a squirrel eating a burger""" __UpperCamelCase = sd_pipe([prompt] , num_inference_steps=2 , output_type="""np""" ).images assert image.shape == (1, 6_4, 6_4, 3) @nightly @require_torch_gpu class UpperCAmelCase__ ( unittest.TestCase): def __lowerCamelCase ( self ) -> Tuple: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCamelCase ( self ) -> Dict: __UpperCamelCase = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" , safety_checker=lowercase ) __UpperCamelCase = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) __UpperCamelCase = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) __UpperCamelCase = ( """portrait of girl with smokey eyes makeup in abandoned hotel, grange clothes, redshift, wide high angle""" """ coloured polaroid photograph with flash, kodak film, hyper real, stunning moody cinematography, with""" """ anamorphic lenses, by maripol, fallen angels by wong kar - wai, style of suspiria and neon demon and""" """ children from bahnhof zoo, detailed """ ) __UpperCamelCase = 4_0_0_3_6_6_0_3_4_6 __UpperCamelCase = 7 # without safety guidance (sld_guidance_scale = 0) __UpperCamelCase = torch.manual_seed(lowercase ) __UpperCamelCase = sd_pipe( [prompt] , generator=lowercase , guidance_scale=lowercase , num_inference_steps=5_0 , output_type="""np""" , width=5_1_2 , height=5_1_2 , sld_guidance_scale=0 , ) __UpperCamelCase = output.images __UpperCamelCase = image[0, -3:, -3:, -1] __UpperCamelCase = [0.2_278, 0.2_231, 0.2_249, 0.2_333, 0.2_303, 0.1_885, 0.2_273, 0.2_144, 0.2_176] assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 # without safety guidance (strong configuration) __UpperCamelCase = torch.manual_seed(lowercase ) __UpperCamelCase = sd_pipe( [prompt] , generator=lowercase , guidance_scale=lowercase , num_inference_steps=5_0 , output_type="""np""" , width=5_1_2 , height=5_1_2 , sld_guidance_scale=2_0_0_0 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) __UpperCamelCase = output.images __UpperCamelCase = image[0, -3:, -3:, -1] __UpperCamelCase = [0.2_383, 0.2_276, 0.236, 0.2_192, 0.2_186, 0.2_053, 0.1_971, 0.1_901, 0.1_719] assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __lowerCamelCase ( self ) -> Optional[Any]: __UpperCamelCase = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" , safety_checker=lowercase ) __UpperCamelCase = LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) __UpperCamelCase = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) __UpperCamelCase = """padme amidala taking a bath artwork, safe for work, no nudity""" __UpperCamelCase = 2_7_3_4_9_7_1_7_5_5 __UpperCamelCase = 7 __UpperCamelCase = torch.manual_seed(lowercase ) __UpperCamelCase = sd_pipe( [prompt] , generator=lowercase , guidance_scale=lowercase , num_inference_steps=5_0 , output_type="""np""" , width=5_1_2 , height=5_1_2 , sld_guidance_scale=0 , ) __UpperCamelCase = output.images __UpperCamelCase = image[0, -3:, -3:, -1] __UpperCamelCase = [0.3_502, 0.3_622, 0.3_396, 0.3_642, 0.3_478, 0.3_318, 0.35, 0.3_348, 0.3_297] assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 __UpperCamelCase = torch.manual_seed(lowercase ) __UpperCamelCase = sd_pipe( [prompt] , generator=lowercase , guidance_scale=lowercase , num_inference_steps=5_0 , output_type="""np""" , width=5_1_2 , height=5_1_2 , sld_guidance_scale=2_0_0_0 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) __UpperCamelCase = output.images __UpperCamelCase = image[0, -3:, -3:, -1] __UpperCamelCase = [0.5_531, 0.5_206, 0.4_895, 0.5_156, 0.5_182, 0.4_751, 0.4_802, 0.4_803, 0.4_443] assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __lowerCamelCase ( self ) -> Optional[Any]: __UpperCamelCase = StableDiffusionPipeline.from_pretrained("""runwayml/stable-diffusion-v1-5""" ) __UpperCamelCase = sd_pipe.to(lowercase ) sd_pipe.set_progress_bar_config(disable=lowercase ) __UpperCamelCase = ( """the four horsewomen of the apocalypse, painting by tom of finland, gaston bussiere, craig mullins, j. c.""" """ leyendecker""" ) __UpperCamelCase = 1_0_4_4_3_5_5_2_3_4 __UpperCamelCase = 1_2 __UpperCamelCase = torch.manual_seed(lowercase ) __UpperCamelCase = sd_pipe( [prompt] , generator=lowercase , guidance_scale=lowercase , num_inference_steps=5_0 , output_type="""np""" , width=5_1_2 , height=5_1_2 , sld_guidance_scale=0 , ) __UpperCamelCase = output.images __UpperCamelCase = image[0, -3:, -3:, -1] __UpperCamelCase = np.array([0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0] ) assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-7 __UpperCamelCase = torch.manual_seed(lowercase ) __UpperCamelCase = sd_pipe( [prompt] , generator=lowercase , guidance_scale=lowercase , num_inference_steps=5_0 , output_type="""np""" , width=5_1_2 , height=5_1_2 , sld_guidance_scale=2_0_0_0 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) __UpperCamelCase = output.images __UpperCamelCase = image[0, -3:, -3:, -1] __UpperCamelCase = np.array([0.5_818, 0.6_285, 0.6_835, 0.6_019, 0.625, 0.6_754, 0.6_096, 0.6_334, 0.6_561] ) assert image.shape == (1, 5_1_2, 5_1_2, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
349
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : Dict = logging.get_logger(__name__) A : str = { 'facebook/data2vec-text-base': 'https://huggingface.co/data2vec/resolve/main/config.json', } class A ( UpperCAmelCase__ ): '''simple docstring''' A__ = '''data2vec-text''' def __init__(self : Union[str, Any] , _UpperCAmelCase : Any=3_0522 , _UpperCAmelCase : Dict=768 , _UpperCAmelCase : List[str]=12 , _UpperCAmelCase : Optional[int]=12 , _UpperCAmelCase : Union[str, Any]=3072 , _UpperCAmelCase : str="gelu" , _UpperCAmelCase : Optional[int]=0.1 , _UpperCAmelCase : List[str]=0.1 , _UpperCAmelCase : Tuple=512 , _UpperCAmelCase : str=2 , _UpperCAmelCase : Optional[int]=0.02 , _UpperCAmelCase : Tuple=1E-1_2 , _UpperCAmelCase : Any=1 , _UpperCAmelCase : Optional[int]=0 , _UpperCAmelCase : Optional[Any]=2 , _UpperCAmelCase : Tuple="absolute" , _UpperCAmelCase : Any=True , _UpperCAmelCase : int=None , **_UpperCAmelCase : List[str] , ) -> Tuple: """simple docstring""" super().__init__(pad_token_id=_UpperCAmelCase , bos_token_id=_UpperCAmelCase , eos_token_id=_UpperCAmelCase , **_UpperCAmelCase ) lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = hidden_act lowercase__ = intermediate_size lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = type_vocab_size lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = position_embedding_type lowercase__ = use_cache lowercase__ = classifier_dropout class A ( UpperCAmelCase__ ): '''simple docstring''' @property def lowerCamelCase__ (self : int ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": lowercase__ = {0: """batch""", 1: """choice""", 2: """sequence"""} else: lowercase__ = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
146
A : Tuple = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def UpperCamelCase ( ) -> None: """simple docstring""" lowercase__ = input("""Enter message: """ ) lowercase__ = input("""Enter key [alphanumeric]: """ ) lowercase__ = input("""Encrypt/Decrypt [e/d]: """ ) if mode.lower().startswith("""e""" ): lowercase__ = """encrypt""" lowercase__ = encrypt_message(__magic_name__ , __magic_name__ ) elif mode.lower().startswith("""d""" ): lowercase__ = """decrypt""" lowercase__ = decrypt_message(__magic_name__ , __magic_name__ ) print(f'''\n{mode.title()}ed message:''' ) print(__magic_name__ ) def UpperCamelCase ( __magic_name__ : str , __magic_name__ : str ) -> str: """simple docstring""" return translate_message(__magic_name__ , __magic_name__ , """encrypt""" ) def UpperCamelCase ( __magic_name__ : str , __magic_name__ : str ) -> str: """simple docstring""" return translate_message(__magic_name__ , __magic_name__ , """decrypt""" ) def UpperCamelCase ( __magic_name__ : str , __magic_name__ : str , __magic_name__ : str ) -> str: """simple docstring""" lowercase__ = [] lowercase__ = 0 lowercase__ = key.upper() for symbol in message: lowercase__ = LETTERS.find(symbol.upper() ) if num != -1: if mode == "encrypt": num += LETTERS.find(key[key_index] ) elif mode == "decrypt": num -= LETTERS.find(key[key_index] ) num %= len(__magic_name__ ) if symbol.isupper(): translated.append(LETTERS[num] ) elif symbol.islower(): translated.append(LETTERS[num].lower() ) key_index += 1 if key_index == len(__magic_name__ ): lowercase__ = 0 else: translated.append(__magic_name__ ) return "".join(__magic_name__ ) if __name__ == "__main__": main()
146
1
'''simple docstring''' import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all feature extractors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...feature_extraction_utils import FeatureExtractionMixin from ...utils import CONFIG_NAME, FEATURE_EXTRACTOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) lowerCAmelCase :Optional[Any] = logging.get_logger(__name__) lowerCAmelCase :Optional[int] = OrderedDict( [ ('''audio-spectrogram-transformer''', '''ASTFeatureExtractor'''), ('''beit''', '''BeitFeatureExtractor'''), ('''chinese_clip''', '''ChineseCLIPFeatureExtractor'''), ('''clap''', '''ClapFeatureExtractor'''), ('''clip''', '''CLIPFeatureExtractor'''), ('''clipseg''', '''ViTFeatureExtractor'''), ('''conditional_detr''', '''ConditionalDetrFeatureExtractor'''), ('''convnext''', '''ConvNextFeatureExtractor'''), ('''cvt''', '''ConvNextFeatureExtractor'''), ('''data2vec-audio''', '''Wav2Vec2FeatureExtractor'''), ('''data2vec-vision''', '''BeitFeatureExtractor'''), ('''deformable_detr''', '''DeformableDetrFeatureExtractor'''), ('''deit''', '''DeiTFeatureExtractor'''), ('''detr''', '''DetrFeatureExtractor'''), ('''dinat''', '''ViTFeatureExtractor'''), ('''donut-swin''', '''DonutFeatureExtractor'''), ('''dpt''', '''DPTFeatureExtractor'''), ('''encodec''', '''EncodecFeatureExtractor'''), ('''flava''', '''FlavaFeatureExtractor'''), ('''glpn''', '''GLPNFeatureExtractor'''), ('''groupvit''', '''CLIPFeatureExtractor'''), ('''hubert''', '''Wav2Vec2FeatureExtractor'''), ('''imagegpt''', '''ImageGPTFeatureExtractor'''), ('''layoutlmv2''', '''LayoutLMv2FeatureExtractor'''), ('''layoutlmv3''', '''LayoutLMv3FeatureExtractor'''), ('''levit''', '''LevitFeatureExtractor'''), ('''maskformer''', '''MaskFormerFeatureExtractor'''), ('''mctct''', '''MCTCTFeatureExtractor'''), ('''mobilenet_v1''', '''MobileNetV1FeatureExtractor'''), ('''mobilenet_v2''', '''MobileNetV2FeatureExtractor'''), ('''mobilevit''', '''MobileViTFeatureExtractor'''), ('''nat''', '''ViTFeatureExtractor'''), ('''owlvit''', '''OwlViTFeatureExtractor'''), ('''perceiver''', '''PerceiverFeatureExtractor'''), ('''poolformer''', '''PoolFormerFeatureExtractor'''), ('''regnet''', '''ConvNextFeatureExtractor'''), ('''resnet''', '''ConvNextFeatureExtractor'''), ('''segformer''', '''SegformerFeatureExtractor'''), ('''sew''', '''Wav2Vec2FeatureExtractor'''), ('''sew-d''', '''Wav2Vec2FeatureExtractor'''), ('''speech_to_text''', '''Speech2TextFeatureExtractor'''), ('''speecht5''', '''SpeechT5FeatureExtractor'''), ('''swiftformer''', '''ViTFeatureExtractor'''), ('''swin''', '''ViTFeatureExtractor'''), ('''swinv2''', '''ViTFeatureExtractor'''), ('''table-transformer''', '''DetrFeatureExtractor'''), ('''timesformer''', '''VideoMAEFeatureExtractor'''), ('''tvlt''', '''TvltFeatureExtractor'''), ('''unispeech''', '''Wav2Vec2FeatureExtractor'''), ('''unispeech-sat''', '''Wav2Vec2FeatureExtractor'''), ('''van''', '''ConvNextFeatureExtractor'''), ('''videomae''', '''VideoMAEFeatureExtractor'''), ('''vilt''', '''ViltFeatureExtractor'''), ('''vit''', '''ViTFeatureExtractor'''), ('''vit_mae''', '''ViTFeatureExtractor'''), ('''vit_msn''', '''ViTFeatureExtractor'''), ('''wav2vec2''', '''Wav2Vec2FeatureExtractor'''), ('''wav2vec2-conformer''', '''Wav2Vec2FeatureExtractor'''), ('''wavlm''', '''Wav2Vec2FeatureExtractor'''), ('''whisper''', '''WhisperFeatureExtractor'''), ('''xclip''', '''CLIPFeatureExtractor'''), ('''yolos''', '''YolosFeatureExtractor'''), ] ) lowerCAmelCase :Union[str, Any] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FEATURE_EXTRACTOR_MAPPING_NAMES) def lowerCamelCase ( lowerCAmelCase : str ): """simple docstring""" for module_name, extractors in FEATURE_EXTRACTOR_MAPPING_NAMES.items(): if class_name in extractors: __magic_name__ : int = model_type_to_module_name(lowerCAmelCase ) __magic_name__ : str = importlib.import_module(f'.{module_name}' , 'transformers.models' ) try: return getattr(lowerCAmelCase , lowerCAmelCase ) except AttributeError: continue for _, extractor in FEATURE_EXTRACTOR_MAPPING._extra_content.items(): if getattr(lowerCAmelCase , '__name__' , lowerCAmelCase ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. __magic_name__ : List[str] = importlib.import_module('transformers' ) if hasattr(lowerCAmelCase , lowerCAmelCase ): return getattr(lowerCAmelCase , lowerCAmelCase ) return None def lowerCamelCase ( lowerCAmelCase : Union[str, os.PathLike] , lowerCAmelCase : Optional[Union[str, os.PathLike]] = None , lowerCAmelCase : bool = False , lowerCAmelCase : bool = False , lowerCAmelCase : Optional[Dict[str, str]] = None , lowerCAmelCase : Optional[Union[bool, str]] = None , lowerCAmelCase : Optional[str] = None , lowerCAmelCase : bool = False , **lowerCAmelCase : Optional[Any] , ): """simple docstring""" __magic_name__ : str = get_file_from_repo( lowerCAmelCase , lowerCAmelCase , cache_dir=lowerCAmelCase , force_download=lowerCAmelCase , resume_download=lowerCAmelCase , proxies=lowerCAmelCase , use_auth_token=lowerCAmelCase , revision=lowerCAmelCase , local_files_only=lowerCAmelCase , ) if resolved_config_file is None: logger.info( 'Could not locate the feature extractor configuration file, will try to use the model config instead.' ) return {} with open(lowerCAmelCase , encoding='utf-8' ) as reader: return json.load(lowerCAmelCase ) class _lowerCamelCase : '''simple docstring''' def __init__( self : str ) -> Optional[int]: raise EnvironmentError( 'AutoFeatureExtractor is designed to be instantiated ' 'using the `AutoFeatureExtractor.from_pretrained(pretrained_model_name_or_path)` method.' ) @classmethod @replace_list_option_in_docstrings(_A ) def __lowerCAmelCase ( cls : List[str] , _A : List[Any] , **_A : Tuple ) -> Optional[Any]: __magic_name__ : Tuple = kwargs.pop('config' , _A ) __magic_name__ : Dict = kwargs.pop('trust_remote_code' , _A ) __magic_name__ : str = True __magic_name__ , __magic_name__ : List[str] = FeatureExtractionMixin.get_feature_extractor_dict(_A , **_A ) __magic_name__ : List[str] = config_dict.get('feature_extractor_type' , _A ) __magic_name__ : Dict = None if "AutoFeatureExtractor" in config_dict.get('auto_map' , {} ): __magic_name__ : str = config_dict['auto_map']['AutoFeatureExtractor'] # If we don't find the feature extractor class in the feature extractor config, let's try the model config. if feature_extractor_class is None and feature_extractor_auto_map is None: if not isinstance(_A , _A ): __magic_name__ : int = AutoConfig.from_pretrained(_A , **_A ) # It could be in `config.feature_extractor_type`` __magic_name__ : Tuple = getattr(_A , 'feature_extractor_type' , _A ) if hasattr(_A , 'auto_map' ) and "AutoFeatureExtractor" in config.auto_map: __magic_name__ : List[str] = config.auto_map['AutoFeatureExtractor'] if feature_extractor_class is not None: __magic_name__ : Optional[int] = feature_extractor_class_from_name(_A ) __magic_name__ : Optional[int] = feature_extractor_auto_map is not None __magic_name__ : List[Any] = feature_extractor_class is not None or type(_A ) in FEATURE_EXTRACTOR_MAPPING __magic_name__ : List[str] = resolve_trust_remote_code( _A , _A , _A , _A ) if has_remote_code and trust_remote_code: __magic_name__ : Tuple = get_class_from_dynamic_module( _A , _A , **_A ) __magic_name__ : Dict = kwargs.pop('code_revision' , _A ) if os.path.isdir(_A ): feature_extractor_class.register_for_auto_class() return feature_extractor_class.from_dict(_A , **_A ) elif feature_extractor_class is not None: return feature_extractor_class.from_dict(_A , **_A ) # Last try: we use the FEATURE_EXTRACTOR_MAPPING. elif type(_A ) in FEATURE_EXTRACTOR_MAPPING: __magic_name__ : int = FEATURE_EXTRACTOR_MAPPING[type(_A )] return feature_extractor_class.from_dict(_A , **_A ) raise ValueError( F'Unrecognized feature extractor in {pretrained_model_name_or_path}. Should have a ' F'`feature_extractor_type` key in its {FEATURE_EXTRACTOR_NAME} of {CONFIG_NAME}, or one of the following ' F'`model_type` keys in its {CONFIG_NAME}: {", ".join(c for c in FEATURE_EXTRACTOR_MAPPING_NAMES.keys() )}' ) @staticmethod def __lowerCAmelCase ( _A : str , _A : int ) -> str: FEATURE_EXTRACTOR_MAPPING.register(_A , _A )
331
'''simple docstring''' import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def lowerCamelCase ( lowerCAmelCase : Tuple ): """simple docstring""" __magic_name__ : List[Any] = filter(lambda lowerCAmelCase : p.requires_grad , model.parameters() ) __magic_name__ : Tuple = sum([np.prod(p.size() ) for p in model_parameters] ) return params lowerCAmelCase :Union[str, Any] = logging.getLogger(__name__) def lowerCamelCase ( lowerCAmelCase : List[Any] , lowerCAmelCase : int ): """simple docstring""" if metric == "rouge2": __magic_name__ : Any = '{val_avg_rouge2:.4f}-{step_count}' elif metric == "bleu": __magic_name__ : Optional[Any] = '{val_avg_bleu:.4f}-{step_count}' elif metric == "em": __magic_name__ : Dict = '{val_avg_em:.4f}-{step_count}' elif metric == "loss": __magic_name__ : int = '{val_avg_loss:.4f}-{step_count}' else: raise NotImplementedError( f'seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this' ' function.' ) __magic_name__ : List[Any] = ModelCheckpoint( dirpath=lowerCAmelCase , filename=lowerCAmelCase , monitor=f'val_{metric}' , mode='max' , save_top_k=1 , every_n_epochs=1 , ) return checkpoint_callback def lowerCamelCase ( lowerCAmelCase : Optional[int] , lowerCAmelCase : Optional[Any] ): """simple docstring""" return EarlyStopping( monitor=f'val_{metric}' , mode='min' if 'loss' in metric else 'max' , patience=lowerCAmelCase , verbose=lowerCAmelCase , ) class _lowerCamelCase ( pl.Callback ): '''simple docstring''' def __lowerCAmelCase ( self : List[str] , _A : Optional[Any] , _A : List[str] ) -> int: __magic_name__ : Optional[Any] = {F'lr_group_{i}': param['lr'] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(_A ) @rank_zero_only def __lowerCAmelCase ( self : Any , _A : pl.Trainer , _A : pl.LightningModule , _A : str , _A : Dict=True ) -> None: logger.info(F'***** {type_path} results at step {trainer.global_step:05d} *****' ) __magic_name__ : List[str] = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['log', 'progress_bar', 'preds']} ) # Log results __magic_name__ : Optional[Any] = Path(pl_module.hparams.output_dir ) if type_path == "test": __magic_name__ : List[Any] = od / 'test_results.txt' __magic_name__ : Dict = od / 'test_generations.txt' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __magic_name__ : Dict = od / F'{type_path}_results/{trainer.global_step:05d}.txt' __magic_name__ : Optional[Any] = od / F'{type_path}_generations/{trainer.global_step:05d}.txt' results_file.parent.mkdir(exist_ok=_A ) generations_file.parent.mkdir(exist_ok=_A ) with open(_A , 'a+' ) as writer: for key in sorted(_A ): if key in ["log", "progress_bar", "preds"]: continue __magic_name__ : Optional[Any] = metrics[key] if isinstance(_A , torch.Tensor ): __magic_name__ : Tuple = val.item() __magic_name__ : int = F'{key}: {val:.6f}\n' writer.write(_A ) if not save_generations: return if "preds" in metrics: __magic_name__ : str = '\n'.join(metrics['preds'] ) generations_file.open('w+' ).write(_A ) @rank_zero_only def __lowerCAmelCase ( self : List[str] , _A : Union[str, Any] , _A : Tuple ) -> Tuple: try: __magic_name__ : str = pl_module.model.model.num_parameters() except AttributeError: __magic_name__ : List[str] = pl_module.model.num_parameters() __magic_name__ : List[Any] = count_trainable_parameters(_A ) # mp stands for million parameters trainer.logger.log_metrics({'n_params': npars, 'mp': npars / 1E6, 'grad_mp': n_trainable_pars / 1E6} ) @rank_zero_only def __lowerCAmelCase ( self : Union[str, Any] , _A : pl.Trainer , _A : pl.LightningModule ) -> List[Any]: save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(_A , _A , 'test' ) @rank_zero_only def __lowerCAmelCase ( self : Tuple , _A : pl.Trainer , _A : Any ) -> List[Any]: save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
331
1
from __future__ import annotations SCREAMING_SNAKE_CASE__ = [] def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: list[list[int]] , __lowerCamelCase: int , __lowerCamelCase: int ): '''simple docstring''' for i in range(len(lowerCAmelCase__ ) ): if board[row][i] == 1: return False for i in range(len(lowerCAmelCase__ ) ): if board[i][column] == 1: return False for i, j in zip(range(lowerCAmelCase__ , -1 , -1 ) , range(lowerCAmelCase__ , -1 , -1 ) ): if board[i][j] == 1: return False for i, j in zip(range(lowerCAmelCase__ , -1 , -1 ) , range(lowerCAmelCase__ , len(lowerCAmelCase__ ) ) ): if board[i][j] == 1: return False return True def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: list[list[int]] , __lowerCamelCase: int ): '''simple docstring''' if row >= len(lowerCAmelCase__ ): solution.append(lowerCAmelCase__ ) printboard(lowerCAmelCase__ ) print() return True for i in range(len(lowerCAmelCase__ ) ): if is_safe(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): lowercase_ = 1 solve(lowerCAmelCase__ , row + 1 ) lowercase_ = 0 return False def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: list[list[int]] ): '''simple docstring''' for i in range(len(lowerCAmelCase__ ) ): for j in range(len(lowerCAmelCase__ ) ): if board[i][j] == 1: print("Q" , end=" " ) else: print("." , end=" " ) print() # n=int(input("The no. of queens")) SCREAMING_SNAKE_CASE__ = 8 SCREAMING_SNAKE_CASE__ = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print("""The total no. of solutions are :""", len(solution))
360
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __lowerCamelCase ( snake_case_ , snake_case_ ): """simple docstring""" lowerCAmelCase__ = 1 @register_to_config def __init__( self , UpperCAmelCase = 1000 , UpperCAmelCase = None ) -> List[Any]: '''simple docstring''' self.set_timesteps(UpperCAmelCase ) # standard deviation of the initial noise distribution lowercase_ = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. lowercase_ = 4 # running values lowercase_ = [] def A__ ( self , UpperCAmelCase , UpperCAmelCase = None ) -> Optional[int]: '''simple docstring''' lowercase_ = num_inference_steps lowercase_ = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] lowercase_ = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: lowercase_ = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: lowercase_ = torch.sin(steps * math.pi / 2 ) ** 2 lowercase_ = (1.0 - self.betas**2) ** 0.5 lowercase_ = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] lowercase_ = timesteps.to(UpperCAmelCase ) lowercase_ = [] def A__ ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = True , ) -> Union[SchedulerOutput, Tuple]: '''simple docstring''' if self.num_inference_steps is None: raise ValueError( "Number of inference steps is 'None', you need to run 'set_timesteps' after creating the scheduler" ) lowercase_ = (self.timesteps == timestep).nonzero().item() lowercase_ = timestep_index + 1 lowercase_ = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(UpperCAmelCase ) if len(self.ets ) == 1: lowercase_ = self.ets[-1] elif len(self.ets ) == 2: lowercase_ = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: lowercase_ = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: lowercase_ = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) lowercase_ = self._get_prev_sample(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=UpperCAmelCase ) def A__ ( self , UpperCAmelCase , *UpperCAmelCase , **UpperCAmelCase ) -> torch.FloatTensor: '''simple docstring''' return sample def A__ ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Dict: '''simple docstring''' lowercase_ = self.alphas[timestep_index] lowercase_ = self.betas[timestep_index] lowercase_ = self.alphas[prev_timestep_index] lowercase_ = self.betas[prev_timestep_index] lowercase_ = (sample - sigma * ets) / max(UpperCAmelCase , 1e-8 ) lowercase_ = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self ) -> List[str]: '''simple docstring''' return self.config.num_train_timesteps
297
0
'''simple docstring''' from __future__ import annotations import os from typing import Any import requests UpperCamelCase__: List[Any] = "https://api.github.com" # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user UpperCamelCase__: int = BASE_URL + "/user" # https://github.com/settings/tokens UpperCamelCase__: List[str] = os.environ.get("USER_TOKEN", "") def snake_case_ ( _lowerCAmelCase : List[Any] ) -> Union[str, Any]: UpperCAmelCase : List[str] = { """Authorization""": f"""token {auth_token}""", """Accept""": """application/vnd.github.v3+json""", } return requests.get(_lowerCAmelCase , headers=_lowerCAmelCase ).json() if __name__ == "__main__": # pragma: no cover if USER_TOKEN: for key, value in fetch_github_info(USER_TOKEN).items(): print(F"{key}: {value}") else: raise ValueError("'USER_TOKEN' field cannot be empty.")
23
def _a ( lowerCamelCase ): if p < 2: raise ValueError("""p should not be less than 2!""" ) elif p == 2: return True lowerCamelCase : Any = 4 lowerCamelCase : List[str] = (1 << p) - 1 for _ in range(p - 2 ): lowerCamelCase : List[Any] = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(1_1))
287
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from ...utils.dataclasses import ( ComputeEnvironment, DistributedType, DynamoBackend, PrecisionType, SageMakerDistributedType, ) from ..menu import BulletMenu lowercase : int = [ "EAGER", "AOT_EAGER", "INDUCTOR", "NVFUSER", "AOT_NVFUSER", "AOT_CUDAGRAPHS", "OFI", "FX2TRT", "ONNXRT", "IPEX", ] def UpperCAmelCase_ (_lowerCAmelCase : str , _lowerCAmelCase : Union[str, Any]=None , _lowerCAmelCase : Dict=None , _lowerCAmelCase : Optional[int]=None ): __UpperCamelCase : str = True while ask_again: __UpperCamelCase : Tuple = input(_lowerCAmelCase ) try: if default is not None and len(_lowerCAmelCase ) == 0: return default return convert_value(_lowerCAmelCase ) if convert_value is not None else result except Exception: if error_message is not None: print(_lowerCAmelCase ) def UpperCAmelCase_ (_lowerCAmelCase : int , _lowerCAmelCase : Any=[] , _lowerCAmelCase : List[str]=None , _lowerCAmelCase : List[str]=0 ): __UpperCamelCase : Union[str, Any] = BulletMenu(_lowerCAmelCase , _lowerCAmelCase ) __UpperCamelCase : str = menu.run(default_choice=_lowerCAmelCase ) return convert_value(_lowerCAmelCase ) if convert_value is not None else result def UpperCAmelCase_ (_lowerCAmelCase : List[Any] ): __UpperCamelCase : Dict = int(_lowerCAmelCase ) return ComputeEnvironment(["LOCAL_MACHINE", "AMAZON_SAGEMAKER"][value] ) def UpperCAmelCase_ (_lowerCAmelCase : List[str] ): __UpperCamelCase : Tuple = int(_lowerCAmelCase ) return DistributedType(["NO", "MULTI_CPU", "MULTI_XPU", "MULTI_GPU", "MULTI_NPU", "TPU"][value] ) def UpperCAmelCase_ (_lowerCAmelCase : List[str] ): __UpperCamelCase : Any = int(_lowerCAmelCase ) return DynamoBackend(DYNAMO_BACKENDS[value] ).value def UpperCAmelCase_ (_lowerCAmelCase : List[str] ): __UpperCamelCase : Dict = int(_lowerCAmelCase ) return PrecisionType(["no", "fp16", "bf16", "fp8"][value] ) def UpperCAmelCase_ (_lowerCAmelCase : Any ): __UpperCamelCase : Any = int(_lowerCAmelCase ) return SageMakerDistributedType(["NO", "DATA_PARALLEL", "MODEL_PARALLEL"][value] ) def UpperCAmelCase_ (_lowerCAmelCase : Union[str, Any] ): return {"yes": True, "no": False}[value.lower()] class SCREAMING_SNAKE_CASE__ ( argparse.RawDescriptionHelpFormatter ): """simple docstring""" def __lowerCamelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) -> Optional[Any]: '''simple docstring''' __UpperCamelCase : int = super()._format_usage(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) __UpperCamelCase : int = usage.replace("<command> [<args>] " , "" ) return usage
365
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging lowercase : List[Any] = logging.get_logger(__name__) lowercase : Optional[Any] = { "Helsinki-NLP/opus-mt-en-de": "https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json", # See all Marian models at https://huggingface.co/models?filter=marian } class SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ): """simple docstring""" lowercase : List[str] = 'marian' lowercase : int = ['past_key_values'] lowercase : Optional[Any] = {'num_attention_heads': 'encoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self , __UpperCamelCase=5_81_01 , __UpperCamelCase=None , __UpperCamelCase=10_24 , __UpperCamelCase=12 , __UpperCamelCase=40_96 , __UpperCamelCase=16 , __UpperCamelCase=12 , __UpperCamelCase=40_96 , __UpperCamelCase=16 , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase="gelu" , __UpperCamelCase=10_24 , __UpperCamelCase=0.1 , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=0.02 , __UpperCamelCase=5_81_00 , __UpperCamelCase=False , __UpperCamelCase=5_81_00 , __UpperCamelCase=0 , __UpperCamelCase=0 , __UpperCamelCase=True , **__UpperCamelCase , ) -> List[Any]: '''simple docstring''' __UpperCamelCase : Any = vocab_size __UpperCamelCase : str = decoder_vocab_size or vocab_size __UpperCamelCase : Any = max_position_embeddings __UpperCamelCase : List[Any] = d_model __UpperCamelCase : Optional[int] = encoder_ffn_dim __UpperCamelCase : Union[str, Any] = encoder_layers __UpperCamelCase : Tuple = encoder_attention_heads __UpperCamelCase : Dict = decoder_ffn_dim __UpperCamelCase : Optional[Any] = decoder_layers __UpperCamelCase : Optional[int] = decoder_attention_heads __UpperCamelCase : Union[str, Any] = dropout __UpperCamelCase : List[str] = attention_dropout __UpperCamelCase : int = activation_dropout __UpperCamelCase : Tuple = activation_function __UpperCamelCase : List[str] = init_std __UpperCamelCase : int = encoder_layerdrop __UpperCamelCase : List[Any] = decoder_layerdrop __UpperCamelCase : Dict = use_cache __UpperCamelCase : str = encoder_layers __UpperCamelCase : Dict = scale_embedding # scale factor will be sqrt(d_model) if True __UpperCamelCase : List[str] = share_encoder_decoder_embeddings super().__init__( pad_token_id=__UpperCamelCase , eos_token_id=__UpperCamelCase , is_encoder_decoder=__UpperCamelCase , decoder_start_token_id=__UpperCamelCase , forced_eos_token_id=__UpperCamelCase , **__UpperCamelCase , ) class SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ): """simple docstring""" @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def __lowerCamelCase ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: __UpperCamelCase : Union[str, Any] = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: __UpperCamelCase : str = {0: "batch"} __UpperCamelCase : Optional[int] = {0: "batch", 1: "past_decoder_sequence + sequence"} else: __UpperCamelCase : Optional[Any] = {0: "batch", 1: "decoder_sequence"} __UpperCamelCase : List[Any] = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(__UpperCamelCase , direction="inputs" ) elif self.task == "causal-lm": # TODO: figure this case out. __UpperCamelCase : str = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: __UpperCamelCase , __UpperCamelCase : Any = self.num_layers for i in range(__UpperCamelCase ): __UpperCamelCase : Any = {0: "batch", 2: "past_sequence + sequence"} __UpperCamelCase : List[Any] = {0: "batch", 2: "past_sequence + sequence"} else: __UpperCamelCase : int = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ("decoder_input_ids", {0: "batch", 1: "decoder_sequence"}), ("decoder_attention_mask", {0: "batch", 1: "decoder_sequence"}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def __lowerCamelCase ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: __UpperCamelCase : List[Any] = super().outputs else: __UpperCamelCase : Optional[Any] = super(__UpperCamelCase , self ).outputs if self.use_past: __UpperCamelCase , __UpperCamelCase : int = self.num_layers for i in range(__UpperCamelCase ): __UpperCamelCase : List[str] = {0: "batch", 2: "past_sequence + sequence"} __UpperCamelCase : str = {0: "batch", 2: "past_sequence + sequence"} return common_outputs def __lowerCamelCase ( self , __UpperCamelCase , __UpperCamelCase = -1 , __UpperCamelCase = -1 , __UpperCamelCase = False , __UpperCamelCase = None , ) -> Mapping[str, Any]: '''simple docstring''' __UpperCamelCase : str = self._generate_dummy_inputs_for_encoder_and_decoder( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # Generate decoder inputs __UpperCamelCase : Any = seq_length if not self.use_past else 1 __UpperCamelCase : int = self._generate_dummy_inputs_for_encoder_and_decoder( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) __UpperCamelCase : Any = {f'''decoder_{name}''': tensor for name, tensor in decoder_inputs.items()} __UpperCamelCase : List[Any] = dict(**__UpperCamelCase , **__UpperCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch __UpperCamelCase , __UpperCamelCase : Dict = common_inputs["input_ids"].shape __UpperCamelCase : Dict = common_inputs["decoder_input_ids"].shape[1] __UpperCamelCase , __UpperCamelCase : Any = self.num_attention_heads __UpperCamelCase : str = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) __UpperCamelCase : List[str] = decoder_seq_length + 3 __UpperCamelCase : Optional[Any] = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) __UpperCamelCase : List[str] = torch.cat( [common_inputs["decoder_attention_mask"], torch.ones(__UpperCamelCase , __UpperCamelCase )] , dim=1 ) __UpperCamelCase : int = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered __UpperCamelCase , __UpperCamelCase : List[str] = self.num_layers __UpperCamelCase : Optional[int] = min(__UpperCamelCase , __UpperCamelCase ) __UpperCamelCase : Optional[int] = max(__UpperCamelCase , __UpperCamelCase ) - min_num_layers __UpperCamelCase : Dict = "encoder" if num_encoder_layers > num_decoder_layers else "decoder" for _ in range(__UpperCamelCase ): common_inputs["past_key_values"].append( ( torch.zeros(__UpperCamelCase ), torch.zeros(__UpperCamelCase ), torch.zeros(__UpperCamelCase ), torch.zeros(__UpperCamelCase ), ) ) # TODO: test this. __UpperCamelCase : Any = encoder_shape if remaining_side_name == "encoder" else decoder_shape for _ in range(__UpperCamelCase , __UpperCamelCase ): common_inputs["past_key_values"].append((torch.zeros(__UpperCamelCase ), torch.zeros(__UpperCamelCase )) ) return common_inputs def __lowerCamelCase ( self , __UpperCamelCase , __UpperCamelCase = -1 , __UpperCamelCase = -1 , __UpperCamelCase = False , __UpperCamelCase = None , ) -> Mapping[str, Any]: '''simple docstring''' __UpperCamelCase : int = self._generate_dummy_inputs_for_encoder_and_decoder( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch __UpperCamelCase , __UpperCamelCase : str = common_inputs["input_ids"].shape # Not using the same length for past_key_values __UpperCamelCase : int = seqlen + 2 __UpperCamelCase , __UpperCamelCase : str = self.num_layers __UpperCamelCase , __UpperCamelCase : List[str] = self.num_attention_heads __UpperCamelCase : str = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) __UpperCamelCase : Any = common_inputs["attention_mask"].dtype __UpperCamelCase : Optional[Any] = torch.cat( [common_inputs["attention_mask"], torch.ones(__UpperCamelCase , __UpperCamelCase , dtype=__UpperCamelCase )] , dim=1 ) __UpperCamelCase : int = [ (torch.zeros(__UpperCamelCase ), torch.zeros(__UpperCamelCase )) for _ in range(__UpperCamelCase ) ] return common_inputs def __lowerCamelCase ( self , __UpperCamelCase , __UpperCamelCase = -1 , __UpperCamelCase = -1 , __UpperCamelCase = False , __UpperCamelCase = None , ) -> Mapping[str, Any]: '''simple docstring''' __UpperCamelCase : Any = compute_effective_axis_dimension( __UpperCamelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX __UpperCamelCase : List[Any] = tokenizer.num_special_tokens_to_add(__UpperCamelCase ) __UpperCamelCase : Union[str, Any] = compute_effective_axis_dimension( __UpperCamelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__UpperCamelCase ) # Generate dummy inputs according to compute batch and sequence __UpperCamelCase : Tuple = [" ".join([tokenizer.unk_token] ) * seq_length] * batch_size __UpperCamelCase : Tuple = dict(tokenizer(__UpperCamelCase , return_tensors=__UpperCamelCase ) ) return common_inputs def __lowerCamelCase ( self , __UpperCamelCase , __UpperCamelCase = -1 , __UpperCamelCase = -1 , __UpperCamelCase = False , __UpperCamelCase = None , ) -> Mapping[str, Any]: '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: __UpperCamelCase : int = self._generate_dummy_inputs_for_default_and_seqaseq_lm( __UpperCamelCase , batch_size=__UpperCamelCase , seq_length=__UpperCamelCase , is_pair=__UpperCamelCase , framework=__UpperCamelCase ) else: __UpperCamelCase : int = self._generate_dummy_inputs_for_causal_lm( __UpperCamelCase , batch_size=__UpperCamelCase , seq_length=__UpperCamelCase , is_pair=__UpperCamelCase , framework=__UpperCamelCase ) return common_inputs def __lowerCamelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) -> Any: '''simple docstring''' if self.task in ["default", "seq2seq-lm"]: __UpperCamelCase : List[Any] = super()._flatten_past_key_values_(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) else: __UpperCamelCase : str = super(__UpperCamelCase , self )._flatten_past_key_values_( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) @property def __lowerCamelCase ( self ) -> float: '''simple docstring''' return 1E-4
171
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging lowerCAmelCase = logging.get_logger(__name__) if is_vision_available(): import PIL class A ( a_ ): UpperCamelCase_ : Dict =["pixel_values"] def __init__(self , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = PILImageResampling.BICUBIC , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = True , lowerCAmelCase = 1 / 2_5_5 , lowerCAmelCase = True , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = True , **lowerCAmelCase , ): super().__init__(**_lowerCamelCase ) __lowercase= size if size is not None else {'shortest_edge': 2_2_4} __lowercase= get_size_dict(_lowerCamelCase , default_to_square=_lowerCamelCase ) __lowercase= crop_size if crop_size is not None else {'height': 2_2_4, 'width': 2_2_4} __lowercase= get_size_dict(_lowerCamelCase , default_to_square=_lowerCamelCase , param_name='crop_size' ) __lowercase= do_resize __lowercase= size __lowercase= resample __lowercase= do_center_crop __lowercase= crop_size __lowercase= do_rescale __lowercase= rescale_factor __lowercase= do_normalize __lowercase= image_mean if image_mean is not None else OPENAI_CLIP_MEAN __lowercase= image_std if image_std is not None else OPENAI_CLIP_STD __lowercase= do_convert_rgb def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = PILImageResampling.BICUBIC , lowerCAmelCase = None , **lowerCAmelCase , ): __lowercase= get_size_dict(_lowerCamelCase , default_to_square=_lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(f'The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}' ) __lowercase= get_resize_output_image_size(_lowerCamelCase , size=size['shortest_edge'] , default_to_square=_lowerCamelCase ) return resize(_lowerCamelCase , size=_lowerCamelCase , resample=_lowerCamelCase , data_format=_lowerCamelCase , **_lowerCamelCase ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): __lowercase= get_size_dict(_lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(f'The `size` parameter must contain the keys (height, width). Got {size.keys()}' ) return center_crop(_lowerCamelCase , size=(size['height'], size['width']) , data_format=_lowerCamelCase , **_lowerCamelCase ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): return rescale(_lowerCamelCase , scale=_lowerCamelCase , data_format=_lowerCamelCase , **_lowerCamelCase ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = None , **lowerCAmelCase , ): return normalize(_lowerCamelCase , mean=_lowerCamelCase , std=_lowerCamelCase , data_format=_lowerCamelCase , **_lowerCamelCase ) def _A (self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = None , lowerCAmelCase = ChannelDimension.FIRST , **lowerCAmelCase , ): __lowercase= do_resize if do_resize is not None else self.do_resize __lowercase= size if size is not None else self.size __lowercase= get_size_dict(_lowerCamelCase , param_name='size' , default_to_square=_lowerCamelCase ) __lowercase= resample if resample is not None else self.resample __lowercase= do_center_crop if do_center_crop is not None else self.do_center_crop __lowercase= crop_size if crop_size is not None else self.crop_size __lowercase= get_size_dict(_lowerCamelCase , param_name='crop_size' , default_to_square=_lowerCamelCase ) __lowercase= do_rescale if do_rescale is not None else self.do_rescale __lowercase= rescale_factor if rescale_factor is not None else self.rescale_factor __lowercase= do_normalize if do_normalize is not None else self.do_normalize __lowercase= image_mean if image_mean is not None else self.image_mean __lowercase= image_std if image_std is not None else self.image_std __lowercase= do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb __lowercase= make_list_of_images(_lowerCamelCase ) if not valid_images(_lowerCamelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # PIL RGBA images are converted to RGB if do_convert_rgb: __lowercase= [convert_to_rgb(_lowerCamelCase ) for image in images] # All transformations expect numpy arrays. __lowercase= [to_numpy_array(_lowerCamelCase ) for image in images] if do_resize: __lowercase= [self.resize(image=_lowerCamelCase , size=_lowerCamelCase , resample=_lowerCamelCase ) for image in images] if do_center_crop: __lowercase= [self.center_crop(image=_lowerCamelCase , size=_lowerCamelCase ) for image in images] if do_rescale: __lowercase= [self.rescale(image=_lowerCamelCase , scale=_lowerCamelCase ) for image in images] if do_normalize: __lowercase= [self.normalize(image=_lowerCamelCase , mean=_lowerCamelCase , std=_lowerCamelCase ) for image in images] __lowercase= [to_channel_dimension_format(_lowerCamelCase , _lowerCamelCase ) for image in images] __lowercase= {'pixel_values': images} return BatchFeature(data=_lowerCamelCase , tensor_type=_lowerCamelCase )
295
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( __snake_case : int = 4_00_00_00 ): '''simple docstring''' lowercase = [] lowercase , lowercase = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(__snake_case ) lowercase , lowercase = b, a + b return sum(__snake_case ) if __name__ == "__main__": print(F'''{solution() = }''')
220
0
"""simple docstring""" import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class lowerCamelCase__ ( snake_case ): SCREAMING_SNAKE_CASE = (DEISMultistepScheduler,) SCREAMING_SNAKE_CASE = (('''num_inference_steps''', 25),) def _UpperCamelCase ( self ,**A ): UpperCAmelCase = { """num_train_timesteps""": 1_000, """beta_start""": 0.0001, """beta_end""": 0.02, """beta_schedule""": """linear""", """solver_order""": 2, } config.update(**A ) return config def _UpperCamelCase ( self ,A=0 ,**A ): UpperCAmelCase = dict(self.forward_default_kwargs ) UpperCAmelCase = kwargs.pop("""num_inference_steps""" ,A ) UpperCAmelCase = self.dummy_sample UpperCAmelCase = 0.1 * sample UpperCAmelCase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: UpperCAmelCase = self.get_scheduler_config(**A ) UpperCAmelCase = scheduler_class(**A ) scheduler.set_timesteps(A ) # copy over dummy past residuals UpperCAmelCase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A ) UpperCAmelCase = scheduler_class.from_pretrained(A ) new_scheduler.set_timesteps(A ) # copy over dummy past residuals UpperCAmelCase = dummy_past_residuals[: new_scheduler.config.solver_order] UpperCAmelCase , UpperCAmelCase = sample, sample for t in range(A ,time_step + scheduler.config.solver_order + 1 ): UpperCAmelCase = scheduler.step(A ,A ,A ,**A ).prev_sample UpperCAmelCase = new_scheduler.step(A ,A ,A ,**A ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def _UpperCamelCase ( self ): pass def _UpperCamelCase ( self ,A=0 ,**A ): UpperCAmelCase = dict(self.forward_default_kwargs ) UpperCAmelCase = kwargs.pop("""num_inference_steps""" ,A ) UpperCAmelCase = self.dummy_sample UpperCAmelCase = 0.1 * sample UpperCAmelCase = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: UpperCAmelCase = self.get_scheduler_config() UpperCAmelCase = scheduler_class(**A ) scheduler.set_timesteps(A ) # copy over dummy past residuals (must be after setting timesteps) UpperCAmelCase = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(A ) UpperCAmelCase = scheduler_class.from_pretrained(A ) # copy over dummy past residuals new_scheduler.set_timesteps(A ) # copy over dummy past residual (must be after setting timesteps) UpperCAmelCase = dummy_past_residuals[: new_scheduler.config.solver_order] UpperCAmelCase = scheduler.step(A ,A ,A ,**A ).prev_sample UpperCAmelCase = new_scheduler.step(A ,A ,A ,**A ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def _UpperCamelCase ( self ,A=None ,**A ): if scheduler is None: UpperCAmelCase = self.scheduler_classes[0] UpperCAmelCase = self.get_scheduler_config(**A ) UpperCAmelCase = scheduler_class(**A ) UpperCAmelCase = self.scheduler_classes[0] UpperCAmelCase = self.get_scheduler_config(**A ) UpperCAmelCase = scheduler_class(**A ) UpperCAmelCase = 10 UpperCAmelCase = self.dummy_model() UpperCAmelCase = self.dummy_sample_deter scheduler.set_timesteps(A ) for i, t in enumerate(scheduler.timesteps ): UpperCAmelCase = model(A ,A ) UpperCAmelCase = scheduler.step(A ,A ,A ).prev_sample return sample def _UpperCamelCase ( self ): UpperCAmelCase = dict(self.forward_default_kwargs ) UpperCAmelCase = kwargs.pop("""num_inference_steps""" ,A ) for scheduler_class in self.scheduler_classes: UpperCAmelCase = self.get_scheduler_config() UpperCAmelCase = scheduler_class(**A ) UpperCAmelCase = self.dummy_sample UpperCAmelCase = 0.1 * sample if num_inference_steps is not None and hasattr(A ,"""set_timesteps""" ): scheduler.set_timesteps(A ) elif num_inference_steps is not None and not hasattr(A ,"""set_timesteps""" ): UpperCAmelCase = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) UpperCAmelCase = [residual + 0.2, residual + 0.15, residual + 0.10] UpperCAmelCase = dummy_past_residuals[: scheduler.config.solver_order] UpperCAmelCase = scheduler.timesteps[5] UpperCAmelCase = scheduler.timesteps[6] UpperCAmelCase = scheduler.step(A ,A ,A ,**A ).prev_sample UpperCAmelCase = scheduler.step(A ,A ,A ,**A ).prev_sample self.assertEqual(output_a.shape ,sample.shape ) self.assertEqual(output_a.shape ,output_a.shape ) def _UpperCamelCase ( self ): # make sure that iterating over schedulers with same config names gives same results # for defaults UpperCAmelCase = DEISMultistepScheduler(**self.get_scheduler_config() ) UpperCAmelCase = self.full_loop(scheduler=A ) UpperCAmelCase = torch.mean(torch.abs(A ) ) assert abs(result_mean.item() - 0.23916 ) < 1e-3 UpperCAmelCase = DPMSolverSinglestepScheduler.from_config(scheduler.config ) UpperCAmelCase = DPMSolverMultistepScheduler.from_config(scheduler.config ) UpperCAmelCase = UniPCMultistepScheduler.from_config(scheduler.config ) UpperCAmelCase = DEISMultistepScheduler.from_config(scheduler.config ) UpperCAmelCase = self.full_loop(scheduler=A ) UpperCAmelCase = torch.mean(torch.abs(A ) ) assert abs(result_mean.item() - 0.23916 ) < 1e-3 def _UpperCamelCase ( self ): for timesteps in [25, 50, 100, 999, 1_000]: self.check_over_configs(num_train_timesteps=A ) def _UpperCamelCase ( self ): self.check_over_configs(thresholding=A ) for order in [1, 2, 3]: for solver_type in ["logrho"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=A ,prediction_type=A ,sample_max_value=A ,algorithm_type="""deis""" ,solver_order=A ,solver_type=A ,) def _UpperCamelCase ( self ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A ) def _UpperCamelCase ( self ): for algorithm_type in ["deis"]: for solver_type in ["logrho"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=A ,solver_type=A ,prediction_type=A ,algorithm_type=A ,) UpperCAmelCase = self.full_loop( solver_order=A ,solver_type=A ,prediction_type=A ,algorithm_type=A ,) assert not torch.isnan(A ).any(), "Samples have nan numbers" def _UpperCamelCase ( self ): self.check_over_configs(lower_order_final=A ) self.check_over_configs(lower_order_final=A ) def _UpperCamelCase ( self ): for num_inference_steps in [1, 2, 3, 5, 10, 50, 100, 999, 1_000]: self.check_over_forward(num_inference_steps=A ,time_step=0 ) def _UpperCamelCase ( self ): UpperCAmelCase = self.full_loop() UpperCAmelCase = torch.mean(torch.abs(A ) ) assert abs(result_mean.item() - 0.23916 ) < 1e-3 def _UpperCamelCase ( self ): UpperCAmelCase = self.full_loop(prediction_type="""v_prediction""" ) UpperCAmelCase = torch.mean(torch.abs(A ) ) assert abs(result_mean.item() - 0.091 ) < 1e-3 def _UpperCamelCase ( self ): UpperCAmelCase = self.scheduler_classes[0] UpperCAmelCase = self.get_scheduler_config(thresholding=A ,dynamic_thresholding_ratio=0 ) UpperCAmelCase = scheduler_class(**A ) UpperCAmelCase = 10 UpperCAmelCase = self.dummy_model() UpperCAmelCase = self.dummy_sample_deter.half() scheduler.set_timesteps(A ) for i, t in enumerate(scheduler.timesteps ): UpperCAmelCase = model(A ,A ) UpperCAmelCase = scheduler.step(A ,A ,A ).prev_sample assert sample.dtype == torch.floataa
234
"""simple docstring""" import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu _UpperCamelCase = get_tests_dir() + """/test_data/fsmt/fsmt_val_data.json""" with io.open(filename, """r""", encoding="""utf-8""") as f: _UpperCamelCase = json.load(f) @require_torch class lowerCamelCase__ ( unittest.TestCase ): def _UpperCamelCase ( self ,A ): return FSMTTokenizer.from_pretrained(A ) def _UpperCamelCase ( self ,A ): UpperCAmelCase = FSMTForConditionalGeneration.from_pretrained(A ).to(A ) if torch_device == "cuda": model.half() return model @parameterized.expand( [ ["""en-ru""", 26.0], ["""ru-en""", 22.0], ["""en-de""", 22.0], ["""de-en""", 29.0], ] ) @slow def _UpperCamelCase ( self ,A ,A ): # note: this test is not testing the best performance since it only evals a small batch # but it should be enough to detect a regression in the output quality UpperCAmelCase = F'''facebook/wmt19-{pair}''' UpperCAmelCase = self.get_tokenizer(A ) UpperCAmelCase = self.get_model(A ) UpperCAmelCase = bleu_data[pair]["""src"""] UpperCAmelCase = bleu_data[pair]["""tgt"""] UpperCAmelCase = tokenizer(A ,return_tensors="""pt""" ,truncation=A ,padding="""longest""" ).to(A ) UpperCAmelCase = model.generate( input_ids=batch.input_ids ,num_beams=8 ,) UpperCAmelCase = tokenizer.batch_decode( A ,skip_special_tokens=A ,clean_up_tokenization_spaces=A ) UpperCAmelCase = calculate_bleu(A ,A ) print(A ) self.assertGreaterEqual(scores["""bleu"""] ,A )
234
1
def lowerCamelCase__ ( A__ : str ): '''simple docstring''' return credit_card_number.startswith(("""34""", """35""", """37""", """4""", """5""", """6""") ) def lowerCamelCase__ ( A__ : str ): '''simple docstring''' __lowerCamelCase = credit_card_number __lowerCamelCase = 0 __lowerCamelCase = len(A__ ) - 2 for i in range(A__ , -1 , -2 ): # double the value of every second digit __lowerCamelCase = int(cc_number[i] ) digit *= 2 # If doubling of a number results in a two digit number # i.e greater than 9(e.g., 6 × 2 = 12), # then add the digits of the product (e.g., 12: 1 + 2 = 3, 15: 1 + 5 = 6), # to get a single digit number. if digit > 9: digit %= 10 digit += 1 __lowerCamelCase = cc_number[:i] + str(A__ ) + cc_number[i + 1 :] total += digit # Sum up the remaining digits for i in range(len(A__ ) - 1 , -1 , -2 ): total += int(cc_number[i] ) return total % 10 == 0 def lowerCamelCase__ ( A__ : str ): '''simple docstring''' __lowerCamelCase = f'{credit_card_number} is an invalid credit card number because' if not credit_card_number.isdigit(): print(f'{error_message} it has nonnumerical characters.' ) return False if not 13 <= len(A__ ) <= 16: print(f'{error_message} of its length.' ) return False if not validate_initial_digits(A__ ): print(f'{error_message} of its first two digits.' ) return False if not luhn_validation(A__ ): print(f'{error_message} it fails the Luhn check.' ) return False print(f'{credit_card_number} is a valid credit card number.' ) return True if __name__ == "__main__": import doctest doctest.testmod() validate_credit_card_number('4111111111111111') validate_credit_card_number('32323')
12
import os import pytest import yaml from datasets.features.features import Features, Value from datasets.info import DatasetInfo, DatasetInfosDict @pytest.mark.parametrize( """files""" , [ ["""full:README.md""", """dataset_infos.json"""], ["""empty:README.md""", """dataset_infos.json"""], ["""dataset_infos.json"""], ["""full:README.md"""], ] , ) def UpperCamelCase (lowercase_: List[str] , lowercase_: Optional[int] ) -> List[str]: A__ : Tuple = tmp_path_factory.mktemp("""dset_infos_dir""" ) if "full:README.md" in files: with open(dataset_infos_dir / """README.md""" , """w""" ) as f: f.write("""---\ndataset_info:\n dataset_size: 42\n---""" ) if "empty:README.md" in files: with open(dataset_infos_dir / """README.md""" , """w""" ) as f: f.write("""""" ) # we want to support dataset_infos.json for backward compatibility if "dataset_infos.json" in files: with open(dataset_infos_dir / """dataset_infos.json""" , """w""" ) as f: f.write("""{\"default\": {\"dataset_size\": 42}}""" ) A__ : List[Any] = DatasetInfosDict.from_directory(lowercase_ ) assert dataset_infos assert dataset_infos["default"].dataset_size == 42 @pytest.mark.parametrize( """dataset_info""" , [ DatasetInfo(), DatasetInfo( description="""foo""" , features=Features({"""a""": Value("""int32""" )} ) , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train"""}] , download_size=42 , ), ] , ) def UpperCamelCase (lowercase_: str , lowercase_: DatasetInfo ) -> List[Any]: A__ : Union[str, Any] = str(lowercase_ ) dataset_info.write_to_directory(lowercase_ ) A__ : List[Any] = DatasetInfo.from_directory(lowercase_ ) assert dataset_info == reloaded assert os.path.exists(os.path.join(lowercase_ , """dataset_info.json""" ) ) def UpperCamelCase () -> List[Any]: A__ : Union[str, Any] = DatasetInfo( description="""foo""" , citation="""bar""" , homepage="""https://foo.bar""" , license="""CC0""" , features=Features({"""a""": Value("""int32""" )} ) , post_processed={} , supervised_keys=() , task_templates=[] , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train""", """num_examples""": 42}] , download_checksums={} , download_size=1337 , post_processing_size=442 , dataset_size=1234 , size_in_bytes=1337 + 442 + 1234 , ) A__ : Dict = dataset_info._to_yaml_dict() assert sorted(lowercase_ ) == sorted(DatasetInfo._INCLUDED_INFO_IN_YAML ) for key in DatasetInfo._INCLUDED_INFO_IN_YAML: assert key in dataset_info_yaml_dict assert isinstance(dataset_info_yaml_dict[key] , (list, dict, int, str) ) A__ : Union[str, Any] = yaml.safe_dump(lowercase_ ) A__ : List[Any] = yaml.safe_load(lowercase_ ) assert dataset_info_yaml_dict == reloaded def UpperCamelCase () -> List[str]: A__ : Optional[int] = DatasetInfo() A__ : List[Any] = dataset_info._to_yaml_dict() assert dataset_info_yaml_dict == {} @pytest.mark.parametrize( """dataset_infos_dict""" , [ DatasetInfosDict(), DatasetInfosDict({"""default""": DatasetInfo()} ), DatasetInfosDict({"""my_config_name""": DatasetInfo()} ), DatasetInfosDict( { """default""": DatasetInfo( description="""foo""" , features=Features({"""a""": Value("""int32""" )} ) , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train"""}] , download_size=42 , ) } ), DatasetInfosDict( { """v1""": DatasetInfo(dataset_size=42 ), """v2""": DatasetInfo(dataset_size=1337 ), } ), ] , ) def UpperCamelCase (lowercase_: Tuple , lowercase_: DatasetInfosDict ) -> Optional[Any]: A__ : List[Any] = str(lowercase_ ) dataset_infos_dict.write_to_directory(lowercase_ ) A__ : Dict = DatasetInfosDict.from_directory(lowercase_ ) # the config_name of the dataset_infos_dict take over the attribute for config_name, dataset_info in dataset_infos_dict.items(): A__ : Optional[int] = config_name # the yaml representation doesn't include fields like description or citation # so we just test that we can recover what we can from the yaml A__ : List[str] = DatasetInfo._from_yaml_dict(dataset_info._to_yaml_dict() ) assert dataset_infos_dict == reloaded if dataset_infos_dict: assert os.path.exists(os.path.join(lowercase_ , """README.md""" ) )
192
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import DistilBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.distilbert.modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertModel, ) class lowerCamelCase : def __init__( self , SCREAMING_SNAKE_CASE_ , ): UpperCamelCase : List[str] = parent UpperCamelCase : Tuple = 13 UpperCamelCase : Optional[Any] = 7 UpperCamelCase : List[Any] = True UpperCamelCase : List[str] = True UpperCamelCase : str = False UpperCamelCase : Dict = True UpperCamelCase : Tuple = 99 UpperCamelCase : int = 32 UpperCamelCase : Any = 2 UpperCamelCase : Union[str, Any] = 4 UpperCamelCase : Optional[int] = 37 UpperCamelCase : Any = """gelu""" UpperCamelCase : Any = 0.1 UpperCamelCase : Any = 0.1 UpperCamelCase : Optional[int] = 512 UpperCamelCase : List[Any] = 16 UpperCamelCase : Any = 2 UpperCamelCase : List[Any] = 0.02 UpperCamelCase : List[str] = 3 UpperCamelCase : int = 4 UpperCamelCase : str = None def a_ ( self ): UpperCamelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase : Any = None if self.use_input_mask: UpperCamelCase : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase : Tuple = None UpperCamelCase : List[str] = None UpperCamelCase : Any = None if self.use_labels: UpperCamelCase : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCamelCase : str = ids_tensor([self.batch_size] , self.num_choices ) UpperCamelCase : List[Any] = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def a_ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): UpperCamelCase : Dict = TFDistilBertModel(config=SCREAMING_SNAKE_CASE_ ) UpperCamelCase : str = {"""input_ids""": input_ids, """attention_mask""": input_mask} UpperCamelCase : Tuple = model(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Optional[int] = [input_ids, input_mask] UpperCamelCase : List[str] = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a_ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): UpperCamelCase : List[Any] = TFDistilBertForMaskedLM(config=SCREAMING_SNAKE_CASE_ ) UpperCamelCase : int = {"""input_ids""": input_ids, """attention_mask""": input_mask} UpperCamelCase : Tuple = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a_ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): UpperCamelCase : Union[str, Any] = TFDistilBertForQuestionAnswering(config=SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Optional[int] = { """input_ids""": input_ids, """attention_mask""": input_mask, } UpperCamelCase : List[str] = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def a_ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): UpperCamelCase : List[Any] = self.num_labels UpperCamelCase : Tuple = TFDistilBertForSequenceClassification(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Tuple = {"""input_ids""": input_ids, """attention_mask""": input_mask} UpperCamelCase : Optional[Any] = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a_ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): UpperCamelCase : str = self.num_choices UpperCamelCase : Tuple = TFDistilBertForMultipleChoice(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : List[Any] = tf.tile(tf.expand_dims(SCREAMING_SNAKE_CASE_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : Tuple = tf.tile(tf.expand_dims(SCREAMING_SNAKE_CASE_ , 1 ) , (1, self.num_choices, 1) ) UpperCamelCase : Union[str, Any] = { """input_ids""": multiple_choice_inputs_ids, """attention_mask""": multiple_choice_input_mask, } UpperCamelCase : str = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def a_ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): UpperCamelCase : List[Any] = self.num_labels UpperCamelCase : str = TFDistilBertForTokenClassification(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Union[str, Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} UpperCamelCase : Union[str, Any] = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def a_ ( self ): UpperCamelCase : int = self.prepare_config_and_inputs() (UpperCamelCase) : int = config_and_inputs UpperCamelCase : str = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_tf class lowerCamelCase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): lowercase : Tuple = ( ( TFDistilBertModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertForMultipleChoice, ) if is_tf_available() else None ) lowercase : Tuple = ( { 'feature-extraction': TFDistilBertModel, 'fill-mask': TFDistilBertForMaskedLM, 'question-answering': TFDistilBertForQuestionAnswering, 'text-classification': TFDistilBertForSequenceClassification, 'token-classification': TFDistilBertForTokenClassification, 'zero-shot': TFDistilBertForSequenceClassification, } if is_tf_available() else {} ) lowercase : Any = False lowercase : int = False def a_ ( self ): UpperCamelCase : Any = TFDistilBertModelTester(self ) UpperCamelCase : Optional[Any] = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE_ , dim=37 ) def a_ ( self ): self.config_tester.run_common_tests() def a_ ( self ): UpperCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*SCREAMING_SNAKE_CASE_ ) def a_ ( self ): UpperCamelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*SCREAMING_SNAKE_CASE_ ) def a_ ( self ): UpperCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*SCREAMING_SNAKE_CASE_ ) def a_ ( self ): UpperCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*SCREAMING_SNAKE_CASE_ ) def a_ ( self ): UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*SCREAMING_SNAKE_CASE_ ) def a_ ( self ): UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*SCREAMING_SNAKE_CASE_ ) @slow def a_ ( self ): for model_name in list(TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1] ): UpperCamelCase : Optional[int] = TFDistilBertModel.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE_ ) @require_tf class lowerCamelCase ( unittest.TestCase ): @slow def a_ ( self ): UpperCamelCase : List[str] = TFDistilBertModel.from_pretrained("""distilbert-base-uncased""" ) UpperCamelCase : Union[str, Any] = tf.constant([[0, 1, 2, 3, 4, 5]] ) UpperCamelCase : Dict = model(SCREAMING_SNAKE_CASE_ )[0] UpperCamelCase : List[Any] = [1, 6, 768] self.assertEqual(output.shape , SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Any = tf.constant( [ [ [0.19261885, -0.13732955, 0.4119799], [0.22150156, -0.07422661, 0.39037204], [0.22756018, -0.0896414, 0.3701467], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , SCREAMING_SNAKE_CASE_ , atol=1e-4 )
358
"""simple docstring""" import requests from bsa import BeautifulSoup def A_ ( snake_case_ : str = "https://www.worldometers.info/coronavirus" ): '''simple docstring''' UpperCamelCase : Any = BeautifulSoup(requests.get(snake_case_ ).text ,"""html.parser""" ) UpperCamelCase : Optional[int] = soup.findAll("""h1""" ) UpperCamelCase : List[Any] = soup.findAll("""div""" ,{"""class""": """maincounter-number"""} ) keys += soup.findAll("""span""" ,{"""class""": """panel-title"""} ) values += soup.findAll("""div""" ,{"""class""": """number-table-main"""} ) return {key.text.strip(): value.text.strip() for key, value in zip(snake_case_ ,snake_case_ )} if __name__ == "__main__": print('''\033[1m''' + '''COVID-19 Status of the World''' + '''\033[0m\n''') for key, value in world_covidaa_stats().items(): print(F'''{key}\n{value}\n''')
27
0
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("""TEST_SAGEMAKER""", """False""" ) ) is not True, reason="""Skipping test because should only be run when releasing minor transformers version""", ) @pytest.mark.usefixtures("""sm_env""" ) @parameterized_class( [ { """framework""": """pytorch""", """script""": """run_glue.py""", """model_name_or_path""": """distilbert-base-cased""", """instance_type""": """ml.p3.16xlarge""", """results""": {"""train_runtime""": 6_50, """eval_accuracy""": 0.7, """eval_loss""": 0.6}, }, { """framework""": """pytorch""", """script""": """run_ddp.py""", """model_name_or_path""": """distilbert-base-cased""", """instance_type""": """ml.p3.16xlarge""", """results""": {"""train_runtime""": 6_00, """eval_accuracy""": 0.7, """eval_loss""": 0.6}, }, { """framework""": """tensorflow""", """script""": """run_tf_dist.py""", """model_name_or_path""": """distilbert-base-cased""", """instance_type""": """ml.p3.16xlarge""", """results""": {"""train_runtime""": 6_00, """eval_accuracy""": 0.6, """eval_loss""": 0.7}, }, ] ) class snake_case ( unittest.TestCase ): def lowercase_ ( self : int)-> Tuple: '''simple docstring''' if self.framework == "pytorch": subprocess.run( f"cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py".split() , encoding="utf-8" , check=_snake_case , ) assert hasattr(self , "env") def lowercase_ ( self : List[str] , UpperCamelCase__ : Tuple)-> Tuple: '''simple docstring''' __lowerCAmelCase: int = f"{self.env.base_job_name}-{instance_count}-{'ddp' if 'ddp' in self.script else 'smd'}" # distributed data settings __lowerCAmelCase: str = {'''smdistributed''': {'''dataparallel''': {'''enabled''': True}}} if self.script != '''run_ddp.py''' else None # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=_snake_case , instance_count=_snake_case , instance_type=self.instance_type , debugger_hook_config=_snake_case , hyperparameters={**self.env.distributed_hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , distribution=_snake_case , py_version="py36" , ) def lowercase_ ( self : str , UpperCamelCase__ : Union[str, Any])-> int: '''simple docstring''' TrainingJobAnalytics(_snake_case).export_csv(f"{self.env.test_path}/{job_name}_metrics.csv") @parameterized.expand([(2,)]) def lowercase_ ( self : Optional[Any] , UpperCamelCase__ : Optional[int])-> List[str]: '''simple docstring''' __lowerCAmelCase: int = self.create_estimator(_snake_case) # run training estimator.fit() # result dataframe __lowerCAmelCase: Any = TrainingJobAnalytics(estimator.latest_training_job.name).dataframe() # extract kpis __lowerCAmelCase: Optional[int] = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"]) __lowerCAmelCase: Optional[int] = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"]) # get train time from SageMaker job, this includes starting, preprocessing, stopping __lowerCAmelCase: Tuple = ( Session().describe_training_job(estimator.latest_training_job.name).get("TrainingTimeInSeconds" , 9_9_9_9_9_9) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy) assert all(t <= self.results["eval_loss"] for t in eval_loss) # dump tests result into json file to share in PR with open(f"{estimator.latest_training_job.name}.json" , "w") as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , _snake_case)
217
import numpy as np from PIL import Image def UpperCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> np.ndarray: __lowercase : Optional[int] = np.array(__lowerCAmelCase ) if arr.shape[0] != arr.shape[1]: raise ValueError('''The input array is not a square matrix''' ) __lowercase : Optional[int] = 0 __lowercase : Union[str, Any] = 0 __lowercase : Optional[Any] = 0 __lowercase : str = 0 # compute the shape of the output matrix __lowercase : Optional[Any] = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape maxpool_shape __lowercase : List[str] = np.zeros((maxpool_shape, maxpool_shape) ) while i < arr.shape[0]: if i + size > arr.shape[0]: # if the end of the matrix is reached, break break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the maximum of the pooling matrix __lowercase : Optional[int] = np.max(arr[i : i + size, j : j + size] ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __lowercase : Any = 0 __lowercase : List[Any] = 0 return updated_arr def UpperCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> np.ndarray: __lowercase : Optional[Any] = np.array(__lowerCAmelCase ) if arr.shape[0] != arr.shape[1]: raise ValueError('''The input array is not a square matrix''' ) __lowercase : int = 0 __lowercase : str = 0 __lowercase : List[str] = 0 __lowercase : Dict = 0 # compute the shape of the output matrix __lowercase : List[Any] = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape avgpool_shape __lowercase : Union[str, Any] = np.zeros((avgpool_shape, avgpool_shape) ) while i < arr.shape[0]: # if the end of the matrix is reached, break if i + size > arr.shape[0]: break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the average of the pooling matrix __lowercase : str = int(np.average(arr[i : i + size, j : j + size] ) ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __lowercase : int = 0 __lowercase : Tuple = 0 return updated_arr # Main Function if __name__ == "__main__": from doctest import testmod testmod(name="avgpooling", verbose=True) # Loading the image __lowerCAmelCase : List[Any] = Image.open("path_to_image") # Converting the image to numpy array and maxpooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(maxpooling(np.array(image), size=3, stride=2)).show() # Converting the image to numpy array and averagepooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(avgpooling(np.array(image), size=3, stride=2)).show()
156
0
import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class A__ ( unittest.TestCase ): """simple docstring""" __A : int = MODEL_FOR_MASKED_LM_MAPPING __A : int = TF_MODEL_FOR_MASKED_LM_MAPPING def __lowercase ( self) -> Dict: '''simple docstring''' super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def __lowercase ( self) -> int: '''simple docstring''' a__ : List[Any] = pipeline(task='fill-mask' , model='sshleifer/tiny-distilroberta-base' , top_k=2 , framework='tf') a__ : int = unmasker('My name is <mask>') self.assertEqual( nested_simplify(lowercase , decimals=6) , [ {'sequence': 'My name is grouped', 'score': 2.1e-05, 'token': 3_8015, 'token_str': ' grouped'}, {'sequence': 'My name is accuser', 'score': 2.1e-05, 'token': 2_5506, 'token_str': ' accuser'}, ] , ) a__ : List[Any] = unmasker('The largest city in France is <mask>') self.assertEqual( nested_simplify(lowercase , decimals=6) , [ { 'sequence': 'The largest city in France is grouped', 'score': 2.1e-05, 'token': 3_8015, 'token_str': ' grouped', }, { 'sequence': 'The largest city in France is accuser', 'score': 2.1e-05, 'token': 2_5506, 'token_str': ' accuser', }, ] , ) a__ : List[Any] = unmasker('My name is <mask>' , targets=[' Patrick', ' Clara', ' Teven'] , top_k=3) self.assertEqual( nested_simplify(lowercase , decimals=6) , [ {'sequence': 'My name is Clara', 'score': 2e-05, 'token': 1_3606, 'token_str': ' Clara'}, {'sequence': 'My name is Patrick', 'score': 2e-05, 'token': 3499, 'token_str': ' Patrick'}, {'sequence': 'My name is Te', 'score': 1.9e-05, 'token': 2941, 'token_str': ' Te'}, ] , ) @require_torch def __lowercase ( self) -> List[Any]: '''simple docstring''' a__ : str = pipeline(task='fill-mask' , model='sshleifer/tiny-distilroberta-base' , top_k=2 , framework='pt') a__ : int = unmasker('My name is <mask>') self.assertEqual( nested_simplify(lowercase , decimals=6) , [ {'sequence': 'My name is Maul', 'score': 2.2e-05, 'token': 3_5676, 'token_str': ' Maul'}, {'sequence': 'My name isELS', 'score': 2.2e-05, 'token': 1_6416, 'token_str': 'ELS'}, ] , ) a__ : List[str] = unmasker('The largest city in France is <mask>') self.assertEqual( nested_simplify(lowercase , decimals=6) , [ { 'sequence': 'The largest city in France is Maul', 'score': 2.2e-05, 'token': 3_5676, 'token_str': ' Maul', }, {'sequence': 'The largest city in France isELS', 'score': 2.2e-05, 'token': 1_6416, 'token_str': 'ELS'}, ] , ) a__ : int = unmasker('My name is <mask>' , targets=[' Patrick', ' Clara', ' Teven'] , top_k=3) self.assertEqual( nested_simplify(lowercase , decimals=6) , [ {'sequence': 'My name is Patrick', 'score': 2.1e-05, 'token': 3499, 'token_str': ' Patrick'}, {'sequence': 'My name is Te', 'score': 2e-05, 'token': 2941, 'token_str': ' Te'}, {'sequence': 'My name is Clara', 'score': 2e-05, 'token': 1_3606, 'token_str': ' Clara'}, ] , ) a__ : int = unmasker('My name is <mask> <mask>' , top_k=2) self.assertEqual( nested_simplify(lowercase , decimals=6) , [ [ { 'score': 2.2e-05, 'token': 3_5676, 'token_str': ' Maul', 'sequence': '<s>My name is Maul<mask></s>', }, {'score': 2.2e-05, 'token': 1_6416, 'token_str': 'ELS', 'sequence': '<s>My name isELS<mask></s>'}, ], [ { 'score': 2.2e-05, 'token': 3_5676, 'token_str': ' Maul', 'sequence': '<s>My name is<mask> Maul</s>', }, {'score': 2.2e-05, 'token': 1_6416, 'token_str': 'ELS', 'sequence': '<s>My name is<mask>ELS</s>'}, ], ] , ) @require_torch_gpu def __lowercase ( self) -> Optional[int]: '''simple docstring''' a__ : Union[str, Any] = pipeline('fill-mask' , model='hf-internal-testing/tiny-random-distilbert' , device=0 , framework='pt') # convert model to fp16 pipe.model.half() a__ : Tuple = pipe('Paris is the [MASK] of France.') # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(lowercase , lowercase) @slow @require_torch def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' a__ : List[Any] = pipeline(task='fill-mask' , model='distilroberta-base' , top_k=2 , framework='pt') self.run_large_test(lowercase) @slow @require_tf def __lowercase ( self) -> Optional[int]: '''simple docstring''' a__ : List[str] = pipeline(task='fill-mask' , model='distilroberta-base' , top_k=2 , framework='tf') self.run_large_test(lowercase) def __lowercase ( self , lowercase) -> Optional[Any]: '''simple docstring''' a__ : List[str] = unmasker('My name is <mask>') self.assertEqual( nested_simplify(lowercase) , [ {'sequence': 'My name is John', 'score': 0.0_08, 'token': 610, 'token_str': ' John'}, {'sequence': 'My name is Chris', 'score': 0.0_07, 'token': 1573, 'token_str': ' Chris'}, ] , ) a__ : Optional[Any] = unmasker('The largest city in France is <mask>') self.assertEqual( nested_simplify(lowercase) , [ { 'sequence': 'The largest city in France is Paris', 'score': 0.2_51, 'token': 2201, 'token_str': ' Paris', }, { 'sequence': 'The largest city in France is Lyon', 'score': 0.2_14, 'token': 1_2790, 'token_str': ' Lyon', }, ] , ) a__ : Tuple = unmasker('My name is <mask>' , targets=[' Patrick', ' Clara', ' Teven'] , top_k=3) self.assertEqual( nested_simplify(lowercase) , [ {'sequence': 'My name is Patrick', 'score': 0.0_05, 'token': 3499, 'token_str': ' Patrick'}, {'sequence': 'My name is Clara', 'score': 0.0_00, 'token': 1_3606, 'token_str': ' Clara'}, {'sequence': 'My name is Te', 'score': 0.0_00, 'token': 2941, 'token_str': ' Te'}, ] , ) @require_torch def __lowercase ( self) -> Tuple: '''simple docstring''' a__ : Any = pipeline(task='fill-mask' , model='sshleifer/tiny-distilroberta-base' , framework='pt') a__ : Union[str, Any] = None a__ : str = None self.run_pipeline_test(lowercase , []) @require_tf def __lowercase ( self) -> Tuple: '''simple docstring''' a__ : List[Any] = pipeline(task='fill-mask' , model='sshleifer/tiny-distilroberta-base' , framework='tf') a__ : Optional[Any] = None a__ : Any = None self.run_pipeline_test(lowercase , []) def __lowercase ( self , lowercase , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest('The provided tokenizer has no mask token, (probably reformer or wav2vec2)') a__ : Dict = FillMaskPipeline(model=lowercase , tokenizer=lowercase) a__ : str = [ F'This is another {tokenizer.mask_token} test', ] return fill_masker, examples def __lowercase ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' a__ : List[str] = fill_masker.tokenizer a__ : Optional[int] = fill_masker.model a__ : Optional[int] = fill_masker( F'This is a {tokenizer.mask_token}' , ) self.assertEqual( lowercase , [ {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, ] , ) a__ : Tuple = fill_masker([F'This is a {tokenizer.mask_token}']) self.assertEqual( lowercase , [ {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, ] , ) a__ : List[str] = fill_masker([F'This is a {tokenizer.mask_token}', F'Another {tokenizer.mask_token} great test.']) self.assertEqual( lowercase , [ [ {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, ], [ {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, ], ] , ) with self.assertRaises(lowercase): fill_masker([None]) # No mask_token is not supported with self.assertRaises(lowercase): fill_masker('This is') self.run_test_top_k(lowercase , lowercase) self.run_test_targets(lowercase , lowercase) self.run_test_top_k_targets(lowercase , lowercase) self.fill_mask_with_duplicate_targets_and_top_k(lowercase , lowercase) self.fill_mask_with_multiple_masks(lowercase , lowercase) def __lowercase ( self , lowercase , lowercase) -> str: '''simple docstring''' a__ : List[str] = tokenizer.get_vocab() a__ : Optional[Any] = sorted(vocab.keys())[:2] # Pipeline argument a__ : List[Any] = FillMaskPipeline(model=lowercase , tokenizer=lowercase , targets=lowercase) a__ : Optional[int] = fill_masker(F'This is a {tokenizer.mask_token}') self.assertEqual( lowercase , [ {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, ] , ) a__ : Optional[int] = {vocab[el] for el in targets} self.assertEqual({el['token'] for el in outputs} , lowercase) a__ : Optional[Any] = [tokenizer.decode([x]) for x in target_ids] self.assertEqual({el['token_str'] for el in outputs} , set(lowercase)) # Call argument a__ : List[str] = FillMaskPipeline(model=lowercase , tokenizer=lowercase) a__ : Union[str, Any] = fill_masker(F'This is a {tokenizer.mask_token}' , targets=lowercase) self.assertEqual( lowercase , [ {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, ] , ) a__ : Optional[int] = {vocab[el] for el in targets} self.assertEqual({el['token'] for el in outputs} , lowercase) a__ : str = [tokenizer.decode([x]) for x in target_ids] self.assertEqual({el['token_str'] for el in outputs} , set(lowercase)) # Score equivalence a__ : int = fill_masker(F'This is a {tokenizer.mask_token}' , targets=lowercase) a__ : Optional[int] = [top_mask['token_str'] for top_mask in outputs] a__ : Optional[Any] = [top_mask['score'] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowercase) == set(lowercase): a__ : List[Any] = fill_masker(F'This is a {tokenizer.mask_token}' , targets=lowercase) a__ : Tuple = [top_mask['score'] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(lowercase) , nested_simplify(lowercase)) # Raises with invalid with self.assertRaises(lowercase): a__ : Tuple = fill_masker(F'This is a {tokenizer.mask_token}' , targets=[]) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(lowercase): a__ : Tuple = fill_masker(F'This is a {tokenizer.mask_token}' , targets=['']) with self.assertRaises(lowercase): a__ : int = fill_masker(F'This is a {tokenizer.mask_token}' , targets='') def __lowercase ( self , lowercase , lowercase) -> Tuple: '''simple docstring''' a__ : str = FillMaskPipeline(model=lowercase , tokenizer=lowercase , top_k=2) a__ : Tuple = fill_masker(F'This is a {tokenizer.mask_token}') self.assertEqual( lowercase , [ {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, ] , ) a__ : Optional[Any] = FillMaskPipeline(model=lowercase , tokenizer=lowercase) a__ : Optional[Any] = fill_masker(F'This is a {tokenizer.mask_token}' , top_k=2) self.assertEqual( lowercase , [ {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, ] , ) self.assertEqual(nested_simplify(lowercase) , nested_simplify(lowercase)) def __lowercase ( self , lowercase , lowercase) -> Tuple: '''simple docstring''' a__ : List[str] = tokenizer.get_vocab() a__ : Tuple = FillMaskPipeline(model=lowercase , tokenizer=lowercase) # top_k=2, ntargets=3 a__ : int = sorted(vocab.keys())[:3] a__ : Optional[int] = fill_masker(F'This is a {tokenizer.mask_token}' , top_k=2 , targets=lowercase) # If we use the most probably targets, and filter differently, we should still # have the same results a__ : Union[str, Any] = [el['token_str'] for el in sorted(lowercase , key=lambda lowercase: x["score"] , reverse=lowercase)] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(lowercase).issubset(lowercase): a__ : Union[str, Any] = fill_masker(F'This is a {tokenizer.mask_token}' , top_k=3 , targets=lowercase) # They should yield exactly the same result self.assertEqual(nested_simplify(lowercase) , nested_simplify(lowercase)) def __lowercase ( self , lowercase , lowercase) -> Union[str, Any]: '''simple docstring''' a__ : Optional[Any] = FillMaskPipeline(model=lowercase , tokenizer=lowercase) a__ : int = tokenizer.get_vocab() # String duplicates + id duplicates a__ : int = sorted(vocab.keys())[:3] a__ : Optional[Any] = [targets[0], targets[1], targets[0], targets[2], targets[1]] a__ : Union[str, Any] = fill_masker(F'My name is {tokenizer.mask_token}' , targets=lowercase , top_k=10) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(lowercase) , 3) def __lowercase ( self , lowercase , lowercase) -> Tuple: '''simple docstring''' a__ : Dict = FillMaskPipeline(model=lowercase , tokenizer=lowercase) a__ : List[Any] = fill_masker( F'This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}' , top_k=2) self.assertEqual( lowercase , [ [ {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, ], [ {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, ], [ {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, {'sequence': ANY(lowercase), 'score': ANY(lowercase), 'token': ANY(lowercase), 'token_str': ANY(lowercase)}, ], ] , )
225
import argparse import torch from transformers import GPTaConfig, GPTaModel, load_tf_weights_in_gpta from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def A_ ( A__ , A__ , A__ ) -> Any: # Construct model if gpta_config_file == "": a__ : Optional[int] = GPTaConfig() else: a__ : List[str] = GPTaConfig.from_json_file(A__ ) a__ : List[str] = GPTaModel(A__ ) # Load weights from numpy load_tf_weights_in_gpta(A__ , A__ , A__ ) # Save pytorch-model a__ : Optional[Any] = pytorch_dump_folder_path + '/' + WEIGHTS_NAME a__ : Any = pytorch_dump_folder_path + '/' + CONFIG_NAME print(F'Save PyTorch model to {pytorch_weights_dump_path}' ) torch.save(model.state_dict() , A__ ) print(F'Save configuration file to {pytorch_config_dump_path}' ) with open(A__ , 'w' , encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowercase : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( """--gpt2_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--gpt2_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained OpenAI model. \n""" """This specifies the model architecture.""" ), ) lowercase : int = parser.parse_args() convert_gpta_checkpoint_to_pytorch(args.gpta_checkpoint_path, args.gpta_config_file, args.pytorch_dump_folder_path)
225
1