code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ : Optional[int] = logging.get_logger(__name__) lowerCAmelCase__ : int = {'vocab_file': 'spiece.model'} lowerCAmelCase__ : Dict = { 'vocab_file': { 'bert_for_seq_generation': ( 'https://huggingface.co/google/bert_for_seq_generation_L-24_bbc_encoder/resolve/main/spiece.model' ), } } lowerCAmelCase__ : List[str] = {'bert_for_seq_generation': 512} class snake_case ( A_ ): """simple docstring""" snake_case__ = VOCAB_FILES_NAMES snake_case__ = PRETRAINED_VOCAB_FILES_MAP snake_case__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case__ = [] snake_case__ = ["input_ids", "attention_mask"] def __init__( self : List[Any] ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[int]="<s>" ,lowerCamelCase__ : List[Any]="</s>" ,lowerCamelCase__ : Tuple="<unk>" ,lowerCamelCase__ : List[Any]="<pad>" ,lowerCamelCase__ : List[str]="<::::>" ,lowerCamelCase__ : Optional[Dict[str, Any]] = None ,**lowerCamelCase__ : Dict ,): UpperCAmelCase__ = {} if sp_model_kwargs is None else sp_model_kwargs # Add extra_ids to the special token list super().__init__( bos_token=snake_case__ ,eos_token=snake_case__ ,unk_token=snake_case__ ,pad_token=snake_case__ ,sep_token=snake_case__ ,sp_model_kwargs=self.sp_model_kwargs ,**snake_case__ ,) UpperCAmelCase__ = vocab_file UpperCAmelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case__ ) @property def __lowerCAmelCase ( self : List[str] ): return self.sp_model.get_piece_size() def __lowerCAmelCase ( self : Union[str, Any] ): UpperCAmelCase__ = {self.convert_ids_to_tokens(snake_case__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Optional[Any] ): UpperCAmelCase__ = self.__dict__.copy() UpperCAmelCase__ = None return state def __setstate__( self : List[Any] ,lowerCamelCase__ : Dict ): UpperCAmelCase__ = d # for backward compatibility if not hasattr(self ,'sp_model_kwargs' ): UpperCAmelCase__ = {} UpperCAmelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __lowerCAmelCase ( self : List[str] ,lowerCamelCase__ : str ): return self.sp_model.encode(snake_case__ ,out_type=snake_case__ ) def __lowerCAmelCase ( self : Union[str, Any] ,lowerCamelCase__ : List[str] ): return self.sp_model.piece_to_id(snake_case__ ) def __lowerCAmelCase ( self : Optional[Any] ,lowerCamelCase__ : Dict ): UpperCAmelCase__ = self.sp_model.IdToPiece(snake_case__ ) return token def __lowerCAmelCase ( self : Optional[int] ,lowerCamelCase__ : Optional[Any] ): UpperCAmelCase__ = [] UpperCAmelCase__ = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(snake_case__ ) + token UpperCAmelCase__ = [] else: current_sub_tokens.append(snake_case__ ) out_string += self.sp_model.decode(snake_case__ ) return out_string.strip() def __lowerCAmelCase ( self : str ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): if not os.path.isdir(snake_case__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase__ = os.path.join( snake_case__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,snake_case__ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case__ ,'wb' ) as fi: UpperCAmelCase__ = self.sp_model.serialized_model_proto() fi.write(snake_case__ ) return (out_vocab_file,)
98
import gc import unittest from parameterized import parameterized from diffusers import FlaxUNetaDConditionModel from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp @slow @require_flax class UpperCAmelCase ( unittest.TestCase ): def _SCREAMING_SNAKE_CASE (self : str , snake_case__ : Union[str, Any] , snake_case__ : List[str] ) -> List[str]: '''simple docstring''' return f"""gaussian_noise_s={seed}_shape={'_'.join([str(snake_case__ ) for s in shape] )}.npy""" def _SCREAMING_SNAKE_CASE (self : Tuple ) -> int: '''simple docstring''' super().tearDown() gc.collect() def _SCREAMING_SNAKE_CASE (self : str , snake_case__ : Optional[Any]=0 , snake_case__ : Any=(4, 4, 64, 64) , snake_case__ : List[Any]=False ) -> int: '''simple docstring''' snake_case : Optional[Any] = jnp.bfloataa if fpaa else jnp.floataa snake_case : Optional[int] = jnp.array(load_hf_numpy(self.get_file_format(snake_case__ , snake_case__ ) ) , dtype=snake_case__ ) return image def _SCREAMING_SNAKE_CASE (self : Optional[int] , snake_case__ : Tuple=False , snake_case__ : List[Any]="CompVis/stable-diffusion-v1-4" ) -> List[Any]: '''simple docstring''' snake_case : List[str] = jnp.bfloataa if fpaa else jnp.floataa snake_case : str = "bf16" if fpaa else None snake_case , snake_case : Optional[int] = FlaxUNetaDConditionModel.from_pretrained( snake_case__ , subfolder="unet" , dtype=snake_case__ , revision=snake_case__ ) return model, params def _SCREAMING_SNAKE_CASE (self : List[str] , snake_case__ : Union[str, Any]=0 , snake_case__ : Union[str, Any]=(4, 77, 7_68) , snake_case__ : Dict=False ) -> List[str]: '''simple docstring''' snake_case : Any = jnp.bfloataa if fpaa else jnp.floataa snake_case : Any = jnp.array(load_hf_numpy(self.get_file_format(snake_case__ , snake_case__ ) ) , dtype=snake_case__ ) return hidden_states @parameterized.expand( [ # fmt: off [83, 4, [-0.2323, -0.1304, 0.0813, -0.3093, -0.0919, -0.1571, -0.1125, -0.5806]], [17, 0.55, [-0.0831, -0.2443, 0.0901, -0.0919, 0.3396, 0.0103, -0.3743, 0.0701]], [8, 0.89, [-0.4863, 0.0859, 0.0875, -0.1658, 0.9199, -0.0114, 0.4839, 0.4639]], [3, 10_00, [-0.5649, 0.2402, -0.5518, 0.1248, 1.1328, -0.2443, -0.0325, -1.0078]], # fmt: on ] ) def _SCREAMING_SNAKE_CASE (self : Optional[int] , snake_case__ : int , snake_case__ : Optional[int] , snake_case__ : Dict ) -> List[str]: '''simple docstring''' snake_case , snake_case : List[str] = self.get_unet_model(model_id="CompVis/stable-diffusion-v1-4" , fpaa=snake_case__ ) snake_case : Union[str, Any] = self.get_latents(snake_case__ , fpaa=snake_case__ ) snake_case : List[str] = self.get_encoder_hidden_states(snake_case__ , fpaa=snake_case__ ) snake_case : Dict = model.apply( {"params": params} , snake_case__ , jnp.array(snake_case__ , dtype=jnp.intaa ) , encoder_hidden_states=snake_case__ , ).sample assert sample.shape == latents.shape snake_case : Optional[Any] = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) snake_case : Optional[int] = jnp.array(snake_case__ , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, in the same hardware assert jnp.allclose(snake_case__ , snake_case__ , atol=1e-2 ) @parameterized.expand( [ # fmt: off [83, 4, [0.1514, 0.0807, 0.1624, 0.1016, -0.1896, 0.0263, 0.0677, 0.2310]], [17, 0.55, [0.1164, -0.0216, 0.0170, 0.1589, -0.3120, 0.1005, -0.0581, -0.1458]], [8, 0.89, [-0.1758, -0.0169, 0.1004, -0.1411, 0.1312, 0.1103, -0.1996, 0.2139]], [3, 10_00, [0.1214, 0.0352, -0.0731, -0.1562, -0.0994, -0.0906, -0.2340, -0.0539]], # fmt: on ] ) def _SCREAMING_SNAKE_CASE (self : Optional[int] , snake_case__ : Optional[int] , snake_case__ : str , snake_case__ : Tuple ) -> str: '''simple docstring''' snake_case , snake_case : List[Any] = self.get_unet_model(model_id="stabilityai/stable-diffusion-2" , fpaa=snake_case__ ) snake_case : List[str] = self.get_latents(snake_case__ , shape=(4, 4, 96, 96) , fpaa=snake_case__ ) snake_case : Union[str, Any] = self.get_encoder_hidden_states(snake_case__ , shape=(4, 77, 10_24) , fpaa=snake_case__ ) snake_case : Optional[int] = model.apply( {"params": params} , snake_case__ , jnp.array(snake_case__ , dtype=jnp.intaa ) , encoder_hidden_states=snake_case__ , ).sample assert sample.shape == latents.shape snake_case : int = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) snake_case : Dict = jnp.array(snake_case__ , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, on the same hardware assert jnp.allclose(snake_case__ , snake_case__ , atol=1e-2 )
59
0
"""simple docstring""" from datetime import datetime import requests def lowercase ( _SCREAMING_SNAKE_CASE : str ): '''simple docstring''' _UpperCAmelCase = '''https://downloadgram.net/wp-json/wppress/video-downloader/video?url=''' _UpperCAmelCase = requests.get(base_url + url ).json()[0]['''urls'''][0]['''src'''] return requests.get(_SCREAMING_SNAKE_CASE ).content if __name__ == "__main__": __A : str = input("Enter Video/IGTV url: ").strip() __A : str = f'''{datetime.now():%Y-%m-%d_%H:%M:%S}.mp4''' with open(file_name, "wb") as fp: fp.write(download_video(url)) print(f'''Done. Video saved to disk as {file_name}.''')
326
"""simple docstring""" import importlib import inspect import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __A : Union[str, Any] = "src/transformers" # This is to make sure the transformers module imported is the one in the repo. __A : Tuple = importlib.util.spec_from_file_location( "transformers", os.path.join(PATH_TO_TRANSFORMERS, "__init__.py"), submodule_search_locations=[PATH_TO_TRANSFORMERS], ) __A : List[str] = spec.loader.load_module() __A : Any = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` __A : Optional[int] = re.compile("\[(.+?)\]\((https://huggingface\.co/.+?)\)") __A : List[str] = { "CLIPConfigMixin", "DecisionTransformerConfigMixin", "EncoderDecoderConfigMixin", "RagConfigMixin", "SpeechEncoderDecoderConfigMixin", "VisionEncoderDecoderConfigMixin", "VisionTextDualEncoderConfigMixin", } def lowercase ( ): '''simple docstring''' _UpperCAmelCase = [] for config_class in list(CONFIG_MAPPING.values() ): _UpperCAmelCase = False # source code of `config_class` _UpperCAmelCase = inspect.getsource(_SCREAMING_SNAKE_CASE ) _UpperCAmelCase = _re_checkpoint.findall(_SCREAMING_SNAKE_CASE ) for checkpoint in checkpoints: # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` _UpperCAmelCase , _UpperCAmelCase = checkpoint # verify the checkpoint name corresponds to the checkpoint link _UpperCAmelCase = f'https://huggingface.co/{ckpt_name}' if ckpt_link == ckpt_link_from_name: _UpperCAmelCase = True break _UpperCAmelCase = config_class.__name__ if not checkpoint_found and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(_SCREAMING_SNAKE_CASE ) if len(_SCREAMING_SNAKE_CASE ) > 0: _UpperCAmelCase = '''\n'''.join(sorted(_SCREAMING_SNAKE_CASE ) ) raise ValueError(f'The following configurations don\'t contain any valid checkpoint:\n{message}' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
326
1
"""simple docstring""" import cva import numpy as np class __lowerCamelCase : '''simple docstring''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase ) -> List[str]: if k in (0.04, 0.06): _a = k _a = window_size else: raise ValueError('''invalid k value''' ) def __str__( self ) -> str: return str(self.k ) def _UpperCAmelCase ( self , __UpperCAmelCase ) -> tuple[cva.Mat, list[list[int]]]: _a = cva.imread(lowerCamelCase__ , 0 ) _a , _a = img.shape _a = [] _a = img.copy() _a = cva.cvtColor(lowerCamelCase__ , cva.COLOR_GRAY2RGB ) _a , _a = np.gradient(lowerCamelCase__ ) _a = dx**2 _a = dy**2 _a = dx * dy _a = 0.04 _a = self.window_size // 2 for y in range(lowerCamelCase__ , h - offset ): for x in range(lowerCamelCase__ , w - offset ): _a = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() _a = (wxx * wyy) - (wxy**2) _a = wxx + wyy _a = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": __snake_case = HarrisCorner(0.04, 3) __snake_case ,__snake_case = edge_detect.detect('''path_to_image''') cva.imwrite('''detect.png''', color_img)
320
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device __A = False class A ( unittest.TestCase ): pass @slow @require_torch_gpu class A ( unittest.TestCase ): def A__ ( self ) -> Any: '''simple docstring''' lowercase__ = VersatileDiffusionImageVariationPipeline.from_pretrained("""shi-labs/versatile-diffusion""" ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) lowercase__ = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg""" ) lowercase__ = torch.manual_seed(0 ) lowercase__ = pipe( image=lowerCamelCase__ , generator=lowerCamelCase__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="""numpy""" , ).images lowercase__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) lowercase__ = np.array([0.04_41, 0.04_69, 0.05_07, 0.05_75, 0.06_32, 0.06_50, 0.08_65, 0.09_09, 0.09_45] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
164
0
"""simple docstring""" from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def lowercase_ ( ) -> tuple[list[int], int]: lowerCAmelCase__ : List[str] = [randint(-1000 , 1000 ) for i in range(10 )] lowerCAmelCase__ : int = randint(-5000 , 5000 ) return (arr, r) _A = make_dataset() def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase ) -> tuple[int, ...]: for triplet in permutations(__UpperCAmelCase , 3 ): if sum(__UpperCAmelCase ) == target: return tuple(sorted(__UpperCAmelCase ) ) return (0, 0, 0) def lowercase_ ( __UpperCAmelCase , __UpperCAmelCase ) -> tuple[int, int, int]: arr.sort() lowerCAmelCase__ : List[str] = len(__UpperCAmelCase ) for i in range(n - 1 ): lowerCAmelCase__ , lowerCAmelCase__ : str = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def lowercase_ ( ) -> tuple[float, float]: lowerCAmelCase__ : Union[str, Any] = """ from __main__ import dataset, triplet_sum1, triplet_sum2 """ lowerCAmelCase__ : Tuple = """ triplet_sum1(*dataset) """ lowerCAmelCase__ : Tuple = """ triplet_sum2(*dataset) """ lowerCAmelCase__ : int = repeat(setup=__UpperCAmelCase , stmt=__UpperCAmelCase , repeat=5 , number=1_0000 ) lowerCAmelCase__ : str = repeat(setup=__UpperCAmelCase , stmt=__UpperCAmelCase , repeat=5 , number=1_0000 ) return (min(__UpperCAmelCase ), min(__UpperCAmelCase )) if __name__ == "__main__": from doctest import testmod testmod() _A = solution_times() print(f"""The time for naive implementation is {times[0]}.""") print(f"""The time for optimized implementation is {times[1]}.""")
212
"""simple docstring""" import shutil import tempfile import unittest from transformers import ( SPIECE_UNDERLINE, AddedToken, BatchEncoding, NllbTokenizer, NllbTokenizerFast, is_torch_available, ) from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin _A = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right _A = 2_5_6_0_4_7 _A = 2_5_6_1_4_5 @require_sentencepiece @require_tokenizers class _lowerCamelCase ( a_ , unittest.TestCase ): _lowerCamelCase :Any = NllbTokenizer _lowerCamelCase :Dict = NllbTokenizerFast _lowerCamelCase :str = True _lowerCamelCase :Optional[Any] = True _lowerCamelCase :Union[str, Any] = {} def _lowerCAmelCase ( self : List[str] ) -> int: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase__ : Optional[int] = NllbTokenizer(UpperCamelCase , keep_accents=UpperCamelCase ) tokenizer.save_pretrained(self.tmpdirname ) def _lowerCAmelCase ( self : Optional[Any] ) -> List[Any]: """simple docstring""" lowerCAmelCase__ : Dict = NllbTokenizer(UpperCamelCase , keep_accents=UpperCamelCase ) lowerCAmelCase__ : List[str] = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(UpperCamelCase , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCamelCase ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) lowerCAmelCase__ : Dict = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( UpperCamelCase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """9""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """é""", """.""", ] , ) lowerCAmelCase__ : Optional[Any] = tokenizer.convert_tokens_to_ids(UpperCamelCase ) self.assertListEqual( UpperCamelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) lowerCAmelCase__ : List[str] = tokenizer.convert_ids_to_tokens(UpperCamelCase ) self.assertListEqual( UpperCamelCase , [ SPIECE_UNDERLINE + """I""", SPIECE_UNDERLINE + """was""", SPIECE_UNDERLINE + """b""", """or""", """n""", SPIECE_UNDERLINE + """in""", SPIECE_UNDERLINE + """""", """<unk>""", """2""", """0""", """0""", """0""", """,""", SPIECE_UNDERLINE + """and""", SPIECE_UNDERLINE + """this""", SPIECE_UNDERLINE + """is""", SPIECE_UNDERLINE + """f""", """al""", """s""", """<unk>""", """.""", ] , ) def _lowerCAmelCase ( self : Any ) -> Any: """simple docstring""" lowerCAmelCase__ : str = (self.rust_tokenizer_class, """hf-internal-testing/tiny-random-nllb""", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase__ : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(UpperCamelCase , **UpperCamelCase ) lowerCAmelCase__ : str = self.tokenizer_class.from_pretrained(UpperCamelCase , **UpperCamelCase ) lowerCAmelCase__ : int = tempfile.mkdtemp() lowerCAmelCase__ : Tuple = tokenizer_r.save_pretrained(UpperCamelCase ) lowerCAmelCase__ : Optional[int] = tokenizer_p.save_pretrained(UpperCamelCase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) lowerCAmelCase__ : Dict = tuple(f for f in tokenizer_r_files if """tokenizer.json""" not in f ) self.assertSequenceEqual(UpperCamelCase , UpperCamelCase ) # Checks everything loads correctly in the same way lowerCAmelCase__ : int = tokenizer_r.from_pretrained(UpperCamelCase ) lowerCAmelCase__ : Union[str, Any] = tokenizer_p.from_pretrained(UpperCamelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase , UpperCamelCase ) ) shutil.rmtree(UpperCamelCase ) # Save tokenizer rust, legacy_format=True lowerCAmelCase__ : List[str] = tempfile.mkdtemp() lowerCAmelCase__ : Optional[Any] = tokenizer_r.save_pretrained(UpperCamelCase , legacy_format=UpperCamelCase ) lowerCAmelCase__ : List[str] = tokenizer_p.save_pretrained(UpperCamelCase ) # Checks it save with the same files self.assertSequenceEqual(UpperCamelCase , UpperCamelCase ) # Checks everything loads correctly in the same way lowerCAmelCase__ : List[str] = tokenizer_r.from_pretrained(UpperCamelCase ) lowerCAmelCase__ : Optional[Any] = tokenizer_p.from_pretrained(UpperCamelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase , UpperCamelCase ) ) shutil.rmtree(UpperCamelCase ) # Save tokenizer rust, legacy_format=False lowerCAmelCase__ : List[Any] = tempfile.mkdtemp() lowerCAmelCase__ : int = tokenizer_r.save_pretrained(UpperCamelCase , legacy_format=UpperCamelCase ) lowerCAmelCase__ : str = tokenizer_p.save_pretrained(UpperCamelCase ) # Checks it saved the tokenizer.json file self.assertTrue(any("""tokenizer.json""" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way lowerCAmelCase__ : Dict = tokenizer_r.from_pretrained(UpperCamelCase ) lowerCAmelCase__ : Optional[int] = tokenizer_p.from_pretrained(UpperCamelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase , UpperCamelCase ) ) shutil.rmtree(UpperCamelCase ) @require_torch def _lowerCAmelCase ( self : Tuple ) -> List[str]: """simple docstring""" if not self.test_seqaseq: return lowerCAmelCase__ : int = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Longer text that will definitely require truncation. lowerCAmelCase__ : Any = [ """ UN Chief Says There Is No Military Solution in Syria""", """ Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for""" """ Syria is that 'there is no military solution' to the nearly five-year conflict and more weapons""" """ will only worsen the violence and misery for millions of people.""", ] lowerCAmelCase__ : Optional[int] = [ """Şeful ONU declară că nu există o soluţie militară în Siria""", """Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al""" """ Rusiei pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi""" """ că noi arme nu vor face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.""", ] try: lowerCAmelCase__ : Dict = tokenizer.prepare_seqaseq_batch( src_texts=UpperCamelCase , tgt_texts=UpperCamelCase , max_length=3 , max_target_length=10 , return_tensors="""pt""" , src_lang="""eng_Latn""" , tgt_lang="""ron_Latn""" , ) except NotImplementedError: return self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.labels.shape[1] , 10 ) # max_target_length will default to max_length if not specified lowerCAmelCase__ : str = tokenizer.prepare_seqaseq_batch( UpperCamelCase , tgt_texts=UpperCamelCase , max_length=3 , return_tensors="""pt""" ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.labels.shape[1] , 3 ) lowerCAmelCase__ : int = tokenizer.prepare_seqaseq_batch( src_texts=UpperCamelCase , max_length=3 , max_target_length=10 , return_tensors="""pt""" ) self.assertEqual(batch_encoder_only.input_ids.shape[1] , 3 ) self.assertEqual(batch_encoder_only.attention_mask.shape[1] , 3 ) self.assertNotIn("""decoder_input_ids""" , UpperCamelCase ) @unittest.skip("""Unfortunately way too slow to build a BPE with SentencePiece.""" ) def _lowerCAmelCase ( self : List[str] ) -> Dict: """simple docstring""" pass def _lowerCAmelCase ( self : str ) -> Union[str, Any]: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCAmelCase__ : str = [AddedToken("""<special>""" , lstrip=UpperCamelCase )] lowerCAmelCase__ : Optional[Any] = self.rust_tokenizer_class.from_pretrained( UpperCamelCase , additional_special_tokens=UpperCamelCase , **UpperCamelCase ) lowerCAmelCase__ : Dict = tokenizer_r.encode("""Hey this is a <special> token""" ) lowerCAmelCase__ : Dict = tokenizer_r.encode("""<special>""" , add_special_tokens=UpperCamelCase )[0] self.assertTrue(special_token_id in r_output ) if self.test_slow_tokenizer: lowerCAmelCase__ : List[Any] = self.rust_tokenizer_class.from_pretrained( UpperCamelCase , additional_special_tokens=UpperCamelCase , **UpperCamelCase , ) lowerCAmelCase__ : Dict = self.tokenizer_class.from_pretrained( UpperCamelCase , additional_special_tokens=UpperCamelCase , **UpperCamelCase ) lowerCAmelCase__ : Optional[int] = tokenizer_p.encode("""Hey this is a <special> token""" ) lowerCAmelCase__ : Dict = tokenizer_cr.encode("""Hey this is a <special> token""" ) self.assertEqual(UpperCamelCase , UpperCamelCase ) self.assertEqual(UpperCamelCase , UpperCamelCase ) self.assertTrue(special_token_id in p_output ) self.assertTrue(special_token_id in cr_output ) @require_torch @require_sentencepiece @require_tokenizers class _lowerCamelCase ( unittest.TestCase ): _lowerCamelCase :int = "facebook/nllb-200-distilled-600M" _lowerCamelCase :List[str] = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] _lowerCamelCase :Optional[Any] = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] _lowerCamelCase :Tuple = [ 256047, 16297, 134408, 8165, 248066, 14734, 950, 1135, 105721, 3573, 83, 27352, 108, 49486, 2, ] @classmethod def _lowerCAmelCase ( cls : Optional[Any] ) -> Any: """simple docstring""" lowerCAmelCase__ : NllbTokenizer = NllbTokenizer.from_pretrained( cls.checkpoint_name , src_lang="""eng_Latn""" , tgt_lang="""ron_Latn""" ) lowerCAmelCase__ : Optional[Any] = 1 return cls def _lowerCAmelCase ( self : Any ) -> int: """simple docstring""" self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""ace_Arab"""] , 25_60_01 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""ace_Latn"""] , 25_60_02 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["""fra_Latn"""] , 25_60_57 ) def _lowerCAmelCase ( self : List[str] ) -> Optional[Any]: """simple docstring""" lowerCAmelCase__ : Union[str, Any] = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase ) def _lowerCAmelCase ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" self.assertIn(UpperCamelCase , self.tokenizer.all_special_ids ) # fmt: off lowerCAmelCase__ : str = [RO_CODE, 42_54, 9_80_68, 11_29_23, 3_90_72, 39_09, 7_13, 10_27_67, 26, 1_73_14, 3_56_42, 1_46_83, 3_31_18, 20_22, 6_69_87, 2, 25_60_47] # fmt: on lowerCAmelCase__ : Any = self.tokenizer.decode(UpperCamelCase , skip_special_tokens=UpperCamelCase ) lowerCAmelCase__ : Tuple = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=UpperCamelCase ) self.assertEqual(UpperCamelCase , UpperCamelCase ) self.assertNotIn(self.tokenizer.eos_token , UpperCamelCase ) def _lowerCAmelCase ( self : Optional[int] ) -> List[str]: """simple docstring""" lowerCAmelCase__ : List[Any] = ["""this is gunna be a long sentence """ * 20] assert isinstance(src_text[0] , UpperCamelCase ) lowerCAmelCase__ : int = 10 lowerCAmelCase__ : Any = self.tokenizer(UpperCamelCase , max_length=UpperCamelCase , truncation=UpperCamelCase ).input_ids[0] self.assertEqual(ids[-1] , 2 ) self.assertEqual(ids[0] , UpperCamelCase ) self.assertEqual(len(UpperCamelCase ) , UpperCamelCase ) def _lowerCAmelCase ( self : Any ) -> Any: """simple docstring""" self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["""<mask>""", """ar_AR"""] ) , [25_62_03, 3] ) def _lowerCAmelCase ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" lowerCAmelCase__ : List[Any] = tempfile.mkdtemp() lowerCAmelCase__ : int = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(UpperCamelCase ) lowerCAmelCase__ : Union[str, Any] = NllbTokenizer.from_pretrained(UpperCamelCase ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , UpperCamelCase ) @require_torch def _lowerCAmelCase ( self : int ) -> List[str]: """simple docstring""" lowerCAmelCase__ : Optional[int] = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=UpperCamelCase , truncation=UpperCamelCase , max_length=len(self.expected_src_tokens ) , return_tensors="""pt""" , ) lowerCAmelCase__ : int = shift_tokens_right( batch["""labels"""] , self.tokenizer.pad_token_id , self.tokenizer.lang_code_to_id["""ron_Latn"""] ) self.assertIsInstance(UpperCamelCase , UpperCamelCase ) self.assertEqual((2, 15) , batch.input_ids.shape ) self.assertEqual((2, 15) , batch.attention_mask.shape ) lowerCAmelCase__ : Dict = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase ) self.assertEqual(UpperCamelCase , batch.decoder_input_ids[0, 0] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) def _lowerCAmelCase ( self : Optional[int] ) -> int: """simple docstring""" lowerCAmelCase__ : str = self.tokenizer(self.src_text , padding=UpperCamelCase , truncation=UpperCamelCase , max_length=3 , return_tensors="""pt""" ) lowerCAmelCase__ : Any = self.tokenizer( text_target=self.tgt_text , padding=UpperCamelCase , truncation=UpperCamelCase , max_length=10 , return_tensors="""pt""" ) lowerCAmelCase__ : str = targets["""input_ids"""] lowerCAmelCase__ : Any = shift_tokens_right( UpperCamelCase , self.tokenizer.pad_token_id , decoder_start_token_id=self.tokenizer.lang_code_to_id[self.tokenizer.tgt_lang] , ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def _lowerCAmelCase ( self : Dict ) -> Optional[Any]: """simple docstring""" lowerCAmelCase__ : Any = self.tokenizer._build_translation_inputs( """A test""" , return_tensors="""pt""" , src_lang="""eng_Latn""" , tgt_lang="""fra_Latn""" ) self.assertEqual( nested_simplify(UpperCamelCase ) , { # A, test, EOS, en_XX """input_ids""": [[25_60_47, 70, 73_56, 2]], """attention_mask""": [[1, 1, 1, 1]], # ar_AR """forced_bos_token_id""": 25_60_57, } , ) @require_torch def _lowerCAmelCase ( self : Dict ) -> Tuple: """simple docstring""" lowerCAmelCase__ : Union[str, Any] = True lowerCAmelCase__ : str = self.tokenizer( """UN Chief says there is no military solution in Syria""" , src_lang="""eng_Latn""" , tgt_lang="""fra_Latn""" ) self.assertEqual( inputs.input_ids , [1_62_97, 13_44_08, 2_56_53, 63_70, 2_48, 2_54, 10_39_29, 9_49_95, 1_08, 4_94_86, 2, 25_60_47] ) lowerCAmelCase__ : List[str] = False lowerCAmelCase__ : Union[str, Any] = self.tokenizer( """UN Chief says there is no military solution in Syria""" , src_lang="""eng_Latn""" , tgt_lang="""fra_Latn""" ) self.assertEqual( inputs.input_ids , [25_60_47, 1_62_97, 13_44_08, 2_56_53, 63_70, 2_48, 2_54, 10_39_29, 9_49_95, 1_08, 4_94_86, 2] )
212
1
'''simple docstring''' import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def __lowerCamelCase ( A__ ) -> Dict: """simple docstring""" return EnvironmentCommand() class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" @staticmethod def A ( UpperCamelCase__ : ArgumentParser ): """simple docstring""" UpperCamelCase = parser.add_parser('env' ) download_parser.set_defaults(func=UpperCamelCase__ ) def A ( self : Tuple ): """simple docstring""" UpperCamelCase = huggingface_hub.__version__ UpperCamelCase = 'not installed' UpperCamelCase = 'NA' if is_torch_available(): import torch UpperCamelCase = torch.__version__ UpperCamelCase = torch.cuda.is_available() UpperCamelCase = 'not installed' if is_transformers_available(): import transformers UpperCamelCase = transformers.__version__ UpperCamelCase = 'not installed' if is_accelerate_available(): import accelerate UpperCamelCase = accelerate.__version__ UpperCamelCase = 'not installed' if is_xformers_available(): import xformers UpperCamelCase = xformers.__version__ UpperCamelCase = { '`diffusers` version': version, 'Platform': platform.platform(), 'Python version': platform.python_version(), 'PyTorch version (GPU?)': f"""{pt_version} ({pt_cuda_available})""", 'Huggingface_hub version': hub_version, 'Transformers version': transformers_version, 'Accelerate version': accelerate_version, 'xFormers version': xformers_version, 'Using GPU in script?': '<fill in>', 'Using distributed or parallel set-up in script?': '<fill in>', } print('\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n' ) print(self.format_dict(UpperCamelCase__ ) ) return info @staticmethod def A ( UpperCamelCase__ : Dict ): """simple docstring""" return "\n".join([f"""- {prop}: {val}""" for prop, val in d.items()] ) + "\n"
28
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=a ) class __A( a ): snake_case_ = field(default='''language-modeling''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) snake_case_ = Features({'''text''': Value('''string''' )} ) snake_case_ = Features({} ) snake_case_ = "text" @property def SCREAMING_SNAKE_CASE_ ( self ) -> Dict[str, str]: '''simple docstring''' return {self.text_column: "text"}
6
0
import copy import random from transformers import CLIPTokenizer class SCREAMING_SNAKE_CASE ( a__ ): '''simple docstring''' def __init__( self : str , *UpperCAmelCase_ : List[str] , **UpperCAmelCase_ : List[Any] ): super().__init__(*_lowerCamelCase , **_lowerCamelCase ) SCREAMING_SNAKE_CASE : Any = {} def _A ( self : List[Any] , UpperCAmelCase_ : List[str] , *UpperCAmelCase_ : Union[str, Any] , **UpperCAmelCase_ : Dict ): SCREAMING_SNAKE_CASE : int = super().add_tokens(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) if num_added_tokens == 0: raise ValueError( f'''The tokenizer already contains the token {placeholder_token}. Please pass a different''' " `placeholder_token` that is not already in the tokenizer." ) def _A ( self : str , UpperCAmelCase_ : List[Any] , *UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : int=1 , **UpperCAmelCase_ : List[str] ): SCREAMING_SNAKE_CASE : Dict = [] if num_vec_per_token == 1: self.try_adding_tokens(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) output.append(_lowerCamelCase ) else: SCREAMING_SNAKE_CASE : Union[str, Any] = [] for i in range(_lowerCamelCase ): SCREAMING_SNAKE_CASE : Optional[int] = placeholder_token + f'''_{i}''' self.try_adding_tokens(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) output.append(_lowerCamelCase ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( f'''The tokenizer already has placeholder token {token} that can get confused with''' f''' {placeholder_token}keep placeholder tokens independent''' ) SCREAMING_SNAKE_CASE : int = output def _A ( self : List[str] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Optional[int]=False , UpperCAmelCase_ : Tuple=1.0 ): if isinstance(_lowerCamelCase , _lowerCamelCase ): SCREAMING_SNAKE_CASE : List[Any] = [] for i in range(len(_lowerCamelCase ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=_lowerCamelCase ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: SCREAMING_SNAKE_CASE : str = self.token_map[placeholder_token] SCREAMING_SNAKE_CASE : int = tokens[: 1 + int(len(_lowerCamelCase ) * prop_tokens_to_load )] if vector_shuffle: SCREAMING_SNAKE_CASE : Tuple = copy.copy(_lowerCamelCase ) random.shuffle(_lowerCamelCase ) SCREAMING_SNAKE_CASE : List[Any] = text.replace(_lowerCamelCase , " ".join(_lowerCamelCase ) ) return text def __call__( self : int , UpperCAmelCase_ : Optional[int] , *UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : str=False , UpperCAmelCase_ : Tuple=1.0 , **UpperCAmelCase_ : int ): return super().__call__( self.replace_placeholder_tokens_in_text( _lowerCamelCase , vector_shuffle=_lowerCamelCase , prop_tokens_to_load=_lowerCamelCase ) , *_lowerCamelCase , **_lowerCamelCase , ) def _A ( self : List[str] , UpperCAmelCase_ : Any , *UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Optional[Any]=False , UpperCAmelCase_ : Optional[Any]=1.0 , **UpperCAmelCase_ : Any ): return super().encode( self.replace_placeholder_tokens_in_text( _lowerCamelCase , vector_shuffle=_lowerCamelCase , prop_tokens_to_load=_lowerCamelCase ) , *_lowerCamelCase , **_lowerCamelCase , )
369
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) snake_case = { """configuration_encodec""": [ """ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP""", """EncodecConfig""", ], """feature_extraction_encodec""": ["""EncodecFeatureExtractor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case = [ """ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST""", """EncodecModel""", """EncodecPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_encodec import ( ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP, EncodecConfig, ) from .feature_extraction_encodec import EncodecFeatureExtractor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encodec import ( ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST, EncodecModel, EncodecPreTrainedModel, ) else: import sys snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
319
0
"""simple docstring""" import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class a__ ( a_, unittest.TestCase ): __lowerCAmelCase = GPTSanJapaneseTokenizer __lowerCAmelCase = False __lowerCAmelCase = {"""do_clean_text""": False, """add_prefix_space""": False} def __magic_name__ ( self ): super().setUp() # fmt: off lowercase : Any = ["こん", "こんに", "にちは", "ばんは", "世界,㔺界", "、", "。", "<BR>", "<SP>", "<TAB>", "<URL>", "<EMAIL>", "<TEL>", "<DATE>", "<PRICE>", "<BLOCK>", "<KIGOU>", "<U2000U2BFF>", "<|emoji1|>", "<unk>", "<|bagoftoken|>", "<|endoftext|>"] # fmt: on lowercase : str = {"emoji": {"\ud83d\ude00": "<|emoji1|>"}, "emoji_inv": {"<|emoji1|>": "\ud83d\ude00"}} # 😀 lowercase : List[Any] = {"unk_token": "<unk>"} lowercase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) lowercase : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["emoji_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) with open(self.emoji_file , "w" ) as emoji_writer: emoji_writer.write(json.dumps(_a ) ) def __magic_name__ ( self , **_a ): kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **_a ) def __magic_name__ ( self , _a ): lowercase : Dict = "こんにちは、世界。 \nこんばんは、㔺界。😀" lowercase : str = "こんにちは、世界。 \nこんばんは、世界。😀" return input_text, output_text def __magic_name__ ( self , _a ): lowercase , lowercase : str = self.get_input_output_texts(_a ) lowercase : List[str] = tokenizer.encode(_a , add_special_tokens=_a ) lowercase : Optional[Any] = tokenizer.decode(_a , clean_up_tokenization_spaces=_a ) return text, ids def __magic_name__ ( self ): pass # TODO add if relevant def __magic_name__ ( self ): pass # TODO add if relevant def __magic_name__ ( self ): pass # TODO add if relevant def __magic_name__ ( self ): lowercase : List[str] = self.get_tokenizer() # Testing tokenization lowercase : Optional[Any] = "こんにちは、世界。 こんばんは、㔺界。" lowercase : int = ["こん", "にちは", "、", "世界", "。", "<SP>", "こん", "ばんは", "、", "㔺界", "。"] lowercase : List[str] = tokenizer.tokenize(_a ) self.assertListEqual(_a , _a ) # Testing conversion to ids without special tokens lowercase : int = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] lowercase : Optional[int] = tokenizer.convert_tokens_to_ids(_a ) self.assertListEqual(_a , _a ) # Testing conversion to ids with special tokens lowercase : List[Any] = tokens + [tokenizer.unk_token] lowercase : str = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] lowercase : List[Any] = tokenizer.convert_tokens_to_ids(_a ) self.assertListEqual(_a , _a ) def __magic_name__ ( self ): lowercase : str = self.get_tokenizer() # Testing tokenization lowercase : Union[str, Any] = "こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。" lowercase : str = "こんにちは、、、、世界。こんばんは、、、、世界。" lowercase : List[str] = tokenizer.encode(_a ) lowercase : List[str] = tokenizer.decode(_a ) self.assertEqual(_a , _a ) @slow def __magic_name__ ( self ): lowercase : List[str] = self.tokenizer_class.from_pretrained("Tanrei/GPTSAN-japanese" ) # Testing tokenization lowercase : int = "こんにちは、世界。" lowercase : List[Any] = "こんばんは、㔺界。😀" lowercase : Dict = "こんにちは、世界。こんばんは、世界。😀" lowercase : str = tokenizer.encode(prefix_text + input_text ) lowercase : Optional[Any] = tokenizer.encode("" , prefix_text=prefix_text + input_text ) lowercase : Dict = tokenizer.encode(_a , prefix_text=_a ) lowercase : str = tokenizer.decode(_a ) lowercase : Union[str, Any] = tokenizer.decode(_a ) lowercase : int = tokenizer.decode(_a ) self.assertEqual(_a , _a ) self.assertEqual(_a , _a ) self.assertEqual(_a , _a ) @slow def __magic_name__ ( self ): lowercase : Dict = self.tokenizer_class.from_pretrained("Tanrei/GPTSAN-japanese" ) # Testing tokenization lowercase : Union[str, Any] = "こんにちは、世界。" lowercase : List[Any] = "こんばんは、㔺界。😀" lowercase : Tuple = len(tokenizer.encode(_a ) ) - 2 lowercase : Dict = len(tokenizer.encode(_a ) ) - 2 lowercase : Union[str, Any] = [1] + [0] * (len_prefix + len_text + 1) lowercase : List[str] = [1] * (len_prefix + len_text + 1) + [0] lowercase : List[str] = [1] + [1] * (len_prefix) + [0] * (len_text + 1) lowercase : Optional[Any] = tokenizer(prefix_text + input_text ).token_type_ids lowercase : Optional[Any] = tokenizer("" , prefix_text=prefix_text + input_text ).token_type_ids lowercase : Dict = tokenizer(_a , prefix_text=_a ).token_type_ids self.assertListEqual(_a , _a ) self.assertListEqual(_a , _a ) self.assertListEqual(_a , _a ) @slow def __magic_name__ ( self ): lowercase : Optional[int] = self.tokenizer_class.from_pretrained("Tanrei/GPTSAN-japanese" ) lowercase : str = tokenizer.encode("あンいワ" ) lowercase : Any = tokenizer.encode("" , prefix_text="あンいワ" ) lowercase : str = tokenizer.encode("いワ" , prefix_text="あン" ) self.assertEqual(tokenizer.decode(_a ) , tokenizer.decode(_a ) ) self.assertEqual(tokenizer.decode(_a ) , tokenizer.decode(_a ) ) self.assertNotEqual(_a , _a ) self.assertNotEqual(_a , _a ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def __magic_name__ ( self ): lowercase : Tuple = self.tokenizer_class.from_pretrained("Tanrei/GPTSAN-japanese" ) lowercase : Optional[Any] = [["武田信玄", "は、"], ["織田信長", "の配下の、"]] lowercase : Optional[int] = tokenizer(_a , padding=_a ) lowercase : List[str] = tokenizer.batch_encode_plus(_a , padding=_a ) # fmt: off lowercase : Any = [[35_993, 8_640, 25_948, 35_998, 30_647, 35_675, 35_999, 35_999], [35_993, 10_382, 9_868, 35_998, 30_646, 9_459, 30_646, 35_675]] lowercase : List[str] = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] lowercase : int = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , _a ) self.assertListEqual(x_token.token_type_ids , _a ) self.assertListEqual(x_token.attention_mask , _a ) self.assertListEqual(x_token_a.input_ids , _a ) self.assertListEqual(x_token_a.token_type_ids , _a ) self.assertListEqual(x_token_a.attention_mask , _a ) def __magic_name__ ( self ): # Intentionally convert some words to accommodate character fluctuations unique to Japanese pass def __magic_name__ ( self ): # tokenizer has no padding token pass
202
"""simple docstring""" import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin _A : int = """ Hugging Face was founded in 2016 by French entrepreneurs Clément Delangue, Julien Chaumond, and Thomas Wolf originally as a company that developed a chatbot app targeted at teenagers.[2] After open-sourcing the model behind the chatbot, the company pivoted to focus on being a platform for machine learning. In March 2021, Hugging Face raised $40 million in a Series B funding round.[3] On April 28, 2021, the company launched the BigScience Research Workshop in collaboration with several other research groups to release an open large language model.[4] In 2022, the workshop concluded with the announcement of BLOOM, a multilingual large language model with 176 billion parameters.[5] """ class a__ ( unittest.TestCase, a_ ): def __magic_name__ ( self ): lowercase : Tuple = load_tool("text-question-answering" ) self.tool.setup() lowercase : Dict = load_tool("text-question-answering" , remote=_a ) def __magic_name__ ( self ): lowercase : str = self.tool(_a , "What did Hugging Face do in April 2021?" ) self.assertEqual(_a , "launched the BigScience Research Workshop" ) def __magic_name__ ( self ): lowercase : Union[str, Any] = self.remote_tool(_a , "What did Hugging Face do in April 2021?" ) self.assertEqual(_a , "launched the BigScience Research Workshop" ) def __magic_name__ ( self ): lowercase : int = self.tool(text=_a , question="What did Hugging Face do in April 2021?" ) self.assertEqual(_a , "launched the BigScience Research Workshop" ) def __magic_name__ ( self ): lowercase : Optional[Any] = self.remote_tool(text=_a , question="What did Hugging Face do in April 2021?" ) self.assertEqual(_a , "launched the BigScience Research Workshop" )
202
1
import timeit import numpy as np import datasets from datasets.arrow_writer import ArrowWriter from datasets.features.features import _ArrayXD def _a ( SCREAMING_SNAKE_CASE : Dict ): """simple docstring""" def wrapper(*SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : Dict ): UpperCamelCase__ : Optional[Any] = timeit.default_timer() UpperCamelCase__ : Dict = func(*SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) UpperCamelCase__ : int = timeit.default_timer() - starttime return delta UpperCamelCase__ : Any = func.__name__ return wrapper def _a ( SCREAMING_SNAKE_CASE : dict , SCREAMING_SNAKE_CASE : int=100 , SCREAMING_SNAKE_CASE : List[Any]=None ): """simple docstring""" UpperCamelCase__ : Dict = [] UpperCamelCase__ : List[str] = seq_shapes or {} for i in range(SCREAMING_SNAKE_CASE ): UpperCamelCase__ : List[Any] = {} for col_id, (k, v) in enumerate(features.items() ): if isinstance(SCREAMING_SNAKE_CASE , _ArrayXD ): UpperCamelCase__ : Optional[int] = np.random.rand(*v.shape ).astype(v.dtype ) elif isinstance(SCREAMING_SNAKE_CASE , datasets.Value ): if v.dtype == "string": UpperCamelCase__ : Any = '''The small grey turtle was surprisingly fast when challenged.''' else: UpperCamelCase__ : str = np.random.randint(10 , size=1 ).astype(v.dtype ).item() elif isinstance(SCREAMING_SNAKE_CASE , datasets.Sequence ): while isinstance(SCREAMING_SNAKE_CASE , datasets.Sequence ): UpperCamelCase__ : Union[str, Any] = v.feature UpperCamelCase__ : Optional[Any] = seq_shapes[k] UpperCamelCase__ : int = np.random.rand(*SCREAMING_SNAKE_CASE ).astype(v.dtype ) UpperCamelCase__ : int = data dummy_data.append((i, example) ) return dummy_data def _a ( SCREAMING_SNAKE_CASE : List[str] , SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : List[Any]=100 , SCREAMING_SNAKE_CASE : List[Any]=None ): """simple docstring""" UpperCamelCase__ : Dict = generate_examples(SCREAMING_SNAKE_CASE , num_examples=SCREAMING_SNAKE_CASE , seq_shapes=SCREAMING_SNAKE_CASE ) with ArrowWriter(features=SCREAMING_SNAKE_CASE , path=SCREAMING_SNAKE_CASE ) as writer: for key, record in dummy_data: UpperCamelCase__ : Optional[Any] = features.encode_example(SCREAMING_SNAKE_CASE ) writer.write(SCREAMING_SNAKE_CASE ) UpperCamelCase__ : str = writer.finalize() if not num_final_examples == num_examples: raise ValueError( F"Error writing the dataset, wrote {num_final_examples} examples but should have written {num_examples}." ) UpperCamelCase__ : Optional[Any] = datasets.Dataset.from_file(filename=SCREAMING_SNAKE_CASE , info=datasets.DatasetInfo(features=SCREAMING_SNAKE_CASE ) ) return dataset
365
import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline __UpperCamelCase : Optional[int] = argparse.ArgumentParser("Stable Diffusion script with intel optimization", add_help=False) parser.add_argument("--dpm", action="store_true", help="Enable DPMSolver or not") parser.add_argument("--steps", default=None, type=int, help="Num inference steps") __UpperCamelCase : Optional[int] = parser.parse_args() __UpperCamelCase : Union[str, Any] = "cpu" __UpperCamelCase : Dict = "a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings" __UpperCamelCase : int = "path-to-your-trained-model" __UpperCamelCase : List[str] = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: __UpperCamelCase : Optional[Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) __UpperCamelCase : Optional[Any] = pipe.to(device) # to channels last __UpperCamelCase : Tuple = pipe.unet.to(memory_format=torch.channels_last) __UpperCamelCase : Optional[int] = pipe.vae.to(memory_format=torch.channels_last) __UpperCamelCase : int = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: __UpperCamelCase : Tuple = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex __UpperCamelCase : Tuple = torch.randn(2, 4, 64, 64) __UpperCamelCase : Any = torch.rand(1) * 999 __UpperCamelCase : Any = torch.randn(2, 77, 768) __UpperCamelCase : List[Any] = (sample, timestep, encoder_hidden_status) try: __UpperCamelCase : Union[str, Any] = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: __UpperCamelCase : str = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) __UpperCamelCase : Tuple = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) __UpperCamelCase : str = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: __UpperCamelCase : List[Any] = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute __UpperCamelCase : Optional[Any] = 666 __UpperCamelCase : int = torch.Generator(device).manual_seed(seed) __UpperCamelCase : int = {"generator": generator} if args.steps is not None: __UpperCamelCase : str = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): __UpperCamelCase : str = pipe(prompt, **generate_kwargs).images[0] # save image image.save("generated.png")
51
0
"""simple docstring""" from statistics import mean import numpy as np def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->list: a__: List[str] = 0 # Number of processes finished a__: Optional[int] = 0 # Displays the finished process. # If it is 0, the performance is completed if it is 1, before the performance. a__: Union[str, Any] = [0] * no_of_process # List to include calculation results a__: Union[str, Any] = [0] * no_of_process # Sort by arrival time. a__: List[Any] = [burst_time[i] for i in np.argsort(_SCREAMING_SNAKE_CASE )] a__: Optional[int] = [process_name[i] for i in np.argsort(_SCREAMING_SNAKE_CASE )] arrival_time.sort() while no_of_process > finished_process_count: a__: Optional[int] = 0 while finished_process[i] == 1: i += 1 if current_time < arrival_time[i]: a__: List[Any] = arrival_time[i] a__: str = 0 # Index showing the location of the process being performed a__: Dict = 0 # Saves the current response ratio. a__: Optional[Any] = 0 for i in range(0 , _SCREAMING_SNAKE_CASE ): if finished_process[i] == 0 and arrival_time[i] <= current_time: a__: Tuple = (burst_time[i] + (current_time - arrival_time[i])) / burst_time[ i ] if response_ratio < temp: a__: Union[str, Any] = temp a__: Any = i # Calculate the turn around time a__: Dict = current_time + burst_time[loc] - arrival_time[loc] current_time += burst_time[loc] # Indicates that the process has been performed. a__: Any = 1 # Increase finished_process_count by 1 finished_process_count += 1 return turn_around_time def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->list: a__: Dict = [0] * no_of_process for i in range(0 , _SCREAMING_SNAKE_CASE ): a__: Optional[int] = turn_around_time[i] - burst_time[i] return waiting_time if __name__ == "__main__": lowercase__ = 5 lowercase__ = ['A', 'B', 'C', 'D', 'E'] lowercase__ = [1, 2, 3, 4, 5] lowercase__ = [1, 2, 3, 4, 5] lowercase__ = calculate_turn_around_time( process_name, arrival_time, burst_time, no_of_process ) lowercase__ = calculate_waiting_time( process_name, turn_around_time, burst_time, no_of_process ) print('Process name \tArrival time \tBurst time \tTurn around time \tWaiting time') for i in range(0, no_of_process): print( f"{process_name[i]}\t\t{arrival_time[i]}\t\t{burst_time[i]}\t\t" f"{turn_around_time[i]}\t\t\t{waiting_time[i]}" ) print(f"average waiting time : {mean(waiting_time):.5f}") print(f"average turn around time : {mean(turn_around_time):.5f}")
290
"""simple docstring""" def __a ( _SCREAMING_SNAKE_CASE = 1000000 ) ->int: a__: int = limit + 1 a__: Optional[int] = [0] * limit for first_term in range(1 , _SCREAMING_SNAKE_CASE ): for n in range(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): a__: List[Any] = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a a__: Any = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(f"{solution() = }")
290
1
import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase_ = get_tests_dir('''fixtures/test_sentencepiece.model''') lowerCAmelCase_ = get_tests_dir('''fixtures/test_sentencepiece_bpe.model''') lowerCAmelCase_ = '''pt''' if is_torch_available() else '''tf''' @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( _a, unittest.TestCase ): lowerCamelCase_ : Tuple = CamembertTokenizer lowerCamelCase_ : int = CamembertTokenizerFast lowerCamelCase_ : Union[str, Any] = True lowerCamelCase_ : Optional[Any] = True def lowerCamelCase (self ) -> Optional[int]: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing snake_case_ : Optional[int] = CamembertTokenizer(__magic_name__ ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase (self ) -> str: '''simple docstring''' snake_case_ : Optional[int] = '<pad>' snake_case_ : Tuple = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__magic_name__ ) , __magic_name__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__magic_name__ ) , __magic_name__ ) def lowerCamelCase (self ) -> int: '''simple docstring''' snake_case_ : Tuple = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>NOTUSED''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(__magic_name__ ) , 1004 ) def lowerCamelCase (self ) -> List[str]: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1005 ) def lowerCamelCase (self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : List[Any] = CamembertTokenizer(__magic_name__ ) tokenizer.save_pretrained(self.tmpdirname ) snake_case_ : int = CamembertTokenizerFast.from_pretrained(self.tmpdirname ) snake_case_ : Any = 'I was born in 92000, and this is falsé.' snake_case_ : str = tokenizer.encode(__magic_name__ ) snake_case_ : List[str] = rust_tokenizer.encode(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) snake_case_ : List[str] = tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) snake_case_ : Optional[int] = rust_tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) # <unk> tokens are not the same for `rust` than for `slow`. # Because spm gives back raw token instead of `unk` in EncodeAsPieces # tokens = tokenizer.tokenize(sequence) snake_case_ : List[Any] = tokenizer.convert_ids_to_tokens(__magic_name__ ) snake_case_ : Optional[Any] = rust_tokenizer.tokenize(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) def lowerCamelCase (self ) -> str: '''simple docstring''' if not self.test_rust_tokenizer: return snake_case_ : List[Any] = self.get_tokenizer() snake_case_ : Optional[Any] = self.get_rust_tokenizer() snake_case_ : List[Any] = 'I was born in 92000, and this is falsé.' snake_case_ : Union[str, Any] = tokenizer.tokenize(__magic_name__ ) snake_case_ : List[Any] = rust_tokenizer.tokenize(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) snake_case_ : str = tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) snake_case_ : List[str] = rust_tokenizer.encode(__magic_name__ , add_special_tokens=__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) snake_case_ : Optional[int] = self.get_rust_tokenizer() snake_case_ : Any = tokenizer.encode(__magic_name__ ) snake_case_ : str = rust_tokenizer.encode(__magic_name__ ) self.assertListEqual(__magic_name__ , __magic_name__ ) @slow def lowerCamelCase (self ) -> str: '''simple docstring''' snake_case_ : Union[str, Any] = {'input_ids': [[5, 54, 7196, 297, 30, 23, 776, 18, 11, 3215, 3705, 8252, 22, 3164, 1181, 2116, 29, 16, 813, 25, 791, 3314, 20, 3446, 38, 2_7575, 120, 6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [5, 468, 17, 11, 9088, 20, 1517, 8, 2_2804, 1_8818, 10, 38, 629, 607, 607, 142, 19, 7196, 867, 56, 1_0326, 24, 2267, 20, 416, 5072, 1_5612, 233, 734, 7, 2399, 27, 16, 3015, 1649, 7, 24, 20, 4338, 2399, 27, 13, 3400, 14, 13, 6189, 8, 930, 9, 6]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # camembert is a french model. So we also use french texts. snake_case_ : List[Any] = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=__magic_name__ , model_name='''camembert-base''' , revision='''3a0641d9a1aeb7e848a74299e7e4c4bca216b4cf''' , sequences=__magic_name__ , )
362
import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase_ = get_tests_dir('''fixtures/test_sentencepiece_with_bytefallback.model''') @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( _a, unittest.TestCase ): lowerCamelCase_ : List[str] = GPTSwaTokenizer lowerCamelCase_ : str = False lowerCamelCase_ : str = True lowerCamelCase_ : List[Any] = False def lowerCamelCase (self ) -> List[str]: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing snake_case_ : Union[str, Any] = GPTSwaTokenizer(__magic_name__ , eos_token='''<unk>''' , bos_token='''<unk>''' , pad_token='''<unk>''' ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase (self , __magic_name__ ) -> int: '''simple docstring''' snake_case_ : Any = '''This is a test''' snake_case_ : str = '''This is a test''' return input_text, output_text def lowerCamelCase (self ) -> Any: '''simple docstring''' snake_case_ : int = '''<s>''' snake_case_ : Any = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__magic_name__ ) , __magic_name__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__magic_name__ ) , __magic_name__ ) def lowerCamelCase (self ) -> List[Any]: '''simple docstring''' snake_case_ : Union[str, Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(__magic_name__ ) , 2000 ) def lowerCamelCase (self ) -> List[str]: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 2000 ) def lowerCamelCase (self ) -> List[str]: '''simple docstring''' snake_case_ : str = GPTSwaTokenizer(__magic_name__ ) snake_case_ : List[str] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__magic_name__ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__magic_name__ ) , [465, 287, 265, 631, 842] ) snake_case_ : Optional[int] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) # fmt: off self.assertListEqual( __magic_name__ , ['''▁I''', '''▁was''', '''▁bor''', '''n''', '''▁in''', '''▁''', '''<0x39>''', '''2''', '''0''', '''0''', '''0''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁f''', '''al''', '''s''', '''<0xC3>''', '''<0xA9>''', '''.'''] , ) # fmt: on snake_case_ : int = tokenizer.convert_tokens_to_ids(__magic_name__ ) self.assertListEqual( __magic_name__ , [262, 272, 1525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260] , ) snake_case_ : Dict = tokenizer.convert_ids_to_tokens(__magic_name__ ) # fmt: off self.assertListEqual( __magic_name__ , ['''▁I''', '''▁was''', '''▁bor''', '''n''', '''▁in''', '''▁''', '''<0x39>''', '''2''', '''0''', '''0''', '''0''', ''',''', '''▁and''', '''▁this''', '''▁is''', '''▁f''', '''al''', '''s''', '''<0xC3>''', '''<0xA9>''', '''.'''] ) # fmt: on def lowerCamelCase (self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[Any] = GPTSwaTokenizer(__magic_name__ ) snake_case_ : Tuple = ['''This is a test''', '''I was born in 92000, and this is falsé.'''] snake_case_ : Optional[int] = [ [465, 287, 265, 631, 842], [262, 272, 1525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(__magic_name__ , __magic_name__ ): self.assertListEqual(tokenizer.encode_fast(__magic_name__ ) , __magic_name__ ) # Test that decode_fast returns the input text for text, token_ids in zip(__magic_name__ , __magic_name__ ): self.assertEqual(tokenizer.decode_fast(__magic_name__ ) , __magic_name__ ) @slow def lowerCamelCase (self ) -> Tuple: '''simple docstring''' snake_case_ : str = [ '''<|python|>def fibonacci(n)\n if n < 0:\n print(\'Incorrect input\')''', '''Hey there, how are you doing this fine day?''', '''This is a text with a trailing spaces followed by a dot .''', '''Häj sväjs lillebrör! =)''', '''Det är inget fel på Mr. Cool''', ] # fmt: off snake_case_ : str = {'''input_ids''': [[6_3423, 5, 6811, 1_4954, 282, 816, 3821, 6_3466, 6_3425, 6_3462, 18, 6_3978, 678, 301, 1320, 6_3423, 6_3455, 6_3458, 18, 6_3982, 4246, 3940, 1901, 4_7789, 5547, 1_8994], [1_9630, 1100, 6_3446, 1342, 633, 544, 4488, 593, 5102, 2416, 6_3495, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1652, 428, 268, 1936, 515, 268, 5_8593, 2_2413, 9106, 546, 268, 3_3213, 6_3979, 698, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [5_5130, 6_3450, 924, 6_3449, 2249, 4062, 1558, 318, 6_3504, 2_1498, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [509, 377, 2827, 2559, 332, 6575, 6_3443, 2_6801, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''token_type_ids''': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=__magic_name__ , model_name='''AI-Sweden/gpt-sw3-126m''' , sequences=__magic_name__ , )
279
0
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Optional[int] = ["""image_processor""", """tokenizer"""] a__ : Optional[int] = """LayoutLMv2ImageProcessor""" a__ : Any = ("""LayoutXLMTokenizer""", """LayoutXLMTokenizerFast""") def __init__( self , __lowercase=None , __lowercase=None , **__lowercase) -> Any: if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , __lowercase , ) __UpperCamelCase :int = kwargs.pop('''feature_extractor''') __UpperCamelCase :Tuple = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''') if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''') super().__init__(__lowercase , __lowercase) def __call__( self , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = True , __lowercase = False , __lowercase = None , __lowercase = None , __lowercase = 0 , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = False , __lowercase = False , __lowercase = False , __lowercase = False , __lowercase = True , __lowercase = None , **__lowercase , ) -> BatchEncoding: # verify input if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( '''You cannot provide bounding boxes ''' '''if you initialized the image processor with apply_ocr set to True.''') if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( '''You cannot provide word labels if you initialized the image processor with apply_ocr set to True.''') if return_overflowing_tokens is True and return_offsets_mapping is False: raise ValueError('''You cannot return overflowing tokens without returning the offsets mapping.''') # first, apply the image processor __UpperCamelCase :int = self.image_processor(images=__lowercase , return_tensors=__lowercase) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(__lowercase , __lowercase): __UpperCamelCase :Dict = [text] # add batch dimension (as the image processor always adds a batch dimension) __UpperCamelCase :Optional[int] = features['''words'''] __UpperCamelCase :List[Any] = self.tokenizer( text=text if text is not None else features['''words'''] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['''boxes'''] , word_labels=__lowercase , add_special_tokens=__lowercase , padding=__lowercase , truncation=__lowercase , max_length=__lowercase , stride=__lowercase , pad_to_multiple_of=__lowercase , return_token_type_ids=__lowercase , return_attention_mask=__lowercase , return_overflowing_tokens=__lowercase , return_special_tokens_mask=__lowercase , return_offsets_mapping=__lowercase , return_length=__lowercase , verbose=__lowercase , return_tensors=__lowercase , **__lowercase , ) # add pixel values __UpperCamelCase :Dict = features.pop('''pixel_values''') if return_overflowing_tokens is True: __UpperCamelCase :Optional[Any] = self.get_overflowing_images(__lowercase , encoded_inputs['''overflow_to_sample_mapping''']) __UpperCamelCase :int = images return encoded_inputs def UpperCamelCase__ ( self , __lowercase , __lowercase) -> Tuple: # in case there's an overflow, ensure each `input_ids` sample is mapped to its corresponding image __UpperCamelCase :int = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx]) if len(__lowercase) != len(__lowercase): raise ValueError( '''Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got''' f""" {len(__lowercase)} and {len(__lowercase)}""") return images_with_overflow def UpperCamelCase__ ( self , *__lowercase , **__lowercase) -> List[Any]: return self.tokenizer.batch_decode(*__lowercase , **__lowercase) def UpperCamelCase__ ( self , *__lowercase , **__lowercase) -> List[str]: return self.tokenizer.decode(*__lowercase , **__lowercase) @property def UpperCamelCase__ ( self) -> int: return ["input_ids", "bbox", "attention_mask", "image"] @property def UpperCamelCase__ ( self) -> List[Any]: warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , __lowercase , ) return self.image_processor_class @property def UpperCamelCase__ ( self) -> str: warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , __lowercase , ) return self.image_processor
43
import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler __lowercase = 16 __lowercase = 32 def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 16 , SCREAMING_SNAKE_CASE = "bert-base-cased" ): '''simple docstring''' __UpperCamelCase :List[str] = AutoTokenizer.from_pretrained(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(SCREAMING_SNAKE_CASE ): # max_length=None => use the model max length (it's actually the default) __UpperCamelCase :int = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset __UpperCamelCase :Tuple = datasets.map( SCREAMING_SNAKE_CASE , batched=SCREAMING_SNAKE_CASE , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , load_from_cache_file=SCREAMING_SNAKE_CASE ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __UpperCamelCase :List[str] = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(SCREAMING_SNAKE_CASE ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(SCREAMING_SNAKE_CASE , padding='''max_length''' , max_length=128 , return_tensors='''pt''' ) return tokenizer.pad(SCREAMING_SNAKE_CASE , padding='''longest''' , return_tensors='''pt''' ) # Instantiate dataloaders. __UpperCamelCase :Union[str, Any] = DataLoader( tokenized_datasets['''train'''] , shuffle=SCREAMING_SNAKE_CASE , collate_fn=SCREAMING_SNAKE_CASE , batch_size=SCREAMING_SNAKE_CASE ) __UpperCamelCase :Dict = DataLoader( tokenized_datasets['''validation'''] , shuffle=SCREAMING_SNAKE_CASE , collate_fn=SCREAMING_SNAKE_CASE , batch_size=SCREAMING_SNAKE_CASE ) return train_dataloader, eval_dataloader def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __UpperCamelCase :int = config['''lr'''] __UpperCamelCase :str = int(config['''num_epochs'''] ) __UpperCamelCase :Any = int(config['''seed'''] ) __UpperCamelCase :Dict = int(config['''batch_size'''] ) __UpperCamelCase :Optional[Any] = args.model_name_or_path set_seed(SCREAMING_SNAKE_CASE ) __UpperCamelCase , __UpperCamelCase :Dict = get_dataloaders(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __UpperCamelCase :Any = AutoModelForSequenceClassification.from_pretrained(SCREAMING_SNAKE_CASE , return_dict=SCREAMING_SNAKE_CASE ) # Instantiate optimizer __UpperCamelCase :List[str] = ( AdamW if accelerator.state.deepspeed_plugin is None or '''optimizer''' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) __UpperCamelCase :Optional[Any] = optimizer_cls(params=model.parameters() , lr=SCREAMING_SNAKE_CASE ) if accelerator.state.deepspeed_plugin is not None: __UpperCamelCase :Dict = accelerator.state.deepspeed_plugin.deepspeed_config[ '''gradient_accumulation_steps''' ] else: __UpperCamelCase :Dict = 1 __UpperCamelCase :Tuple = (len(SCREAMING_SNAKE_CASE ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): __UpperCamelCase :str = get_linear_schedule_with_warmup( optimizer=SCREAMING_SNAKE_CASE , num_warmup_steps=0 , num_training_steps=SCREAMING_SNAKE_CASE , ) else: __UpperCamelCase :Dict = DummyScheduler(SCREAMING_SNAKE_CASE , total_num_steps=SCREAMING_SNAKE_CASE , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :int = accelerator.prepare( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # We need to keep track of how many total steps we have iterated over __UpperCamelCase :List[Any] = 0 # We also need to keep track of the stating epoch so files are named properly __UpperCamelCase :Dict = 0 # Now we train the model __UpperCamelCase :Any = evaluate.load('''glue''' , '''mrpc''' ) __UpperCamelCase :Union[str, Any] = 0 __UpperCamelCase :Optional[int] = {} for epoch in range(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): model.train() for step, batch in enumerate(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = model(**SCREAMING_SNAKE_CASE ) __UpperCamelCase :Tuple = outputs.loss __UpperCamelCase :str = loss / gradient_accumulation_steps accelerator.backward(SCREAMING_SNAKE_CASE ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() __UpperCamelCase :Any = 0 for step, batch in enumerate(SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): __UpperCamelCase :Any = model(**SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times __UpperCamelCase , __UpperCamelCase :List[Any] = accelerator.gather( (predictions, batch['''labels''']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(SCREAMING_SNAKE_CASE ) - 1: __UpperCamelCase :List[str] = predictions[: len(eval_dataloader.dataset ) - samples_seen] __UpperCamelCase :Optional[int] = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=SCREAMING_SNAKE_CASE , references=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :Dict = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , SCREAMING_SNAKE_CASE ) __UpperCamelCase :str = eval_metric['''accuracy'''] if best_performance < eval_metric["accuracy"]: __UpperCamelCase :int = eval_metric['''accuracy'''] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f"""Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}""" accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , '''all_results.json''' ) , '''w''' ) as f: json.dump(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = argparse.ArgumentParser(description='''Simple example of training script tracking peak GPU memory usage.''' ) parser.add_argument( '''--model_name_or_path''' , type=SCREAMING_SNAKE_CASE , default='''bert-base-cased''' , help='''Path to pretrained model or model identifier from huggingface.co/models.''' , required=SCREAMING_SNAKE_CASE , ) parser.add_argument( '''--output_dir''' , type=SCREAMING_SNAKE_CASE , default='''.''' , help='''Optional save directory where all checkpoint folders will be stored. Default is the current working directory.''' , ) parser.add_argument( '''--performance_lower_bound''' , type=SCREAMING_SNAKE_CASE , default=SCREAMING_SNAKE_CASE , help='''Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.''' , ) parser.add_argument( '''--num_epochs''' , type=SCREAMING_SNAKE_CASE , default=3 , help='''Number of train epochs.''' , ) __UpperCamelCase :List[str] = parser.parse_args() __UpperCamelCase :Tuple = {'''lr''': 2e-5, '''num_epochs''': args.num_epochs, '''seed''': 42, '''batch_size''': 16} training_function(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
43
1
UpperCamelCase = '''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def __lowerCamelCase ( snake_case__ ) -> bytes: """simple docstring""" if not isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = F'a bytes-like object is required, not \'{data.__class__.__name__}\'' raise TypeError(snake_case__ ) _SCREAMING_SNAKE_CASE = """""".join(bin(snake_case__ )[2:].zfill(8 ) for byte in data ) _SCREAMING_SNAKE_CASE = len(snake_case__ ) % 6 != 0 if padding_needed: # The padding that will be added later _SCREAMING_SNAKE_CASE = b"""=""" * ((6 - len(snake_case__ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(snake_case__ ) % 6) else: _SCREAMING_SNAKE_CASE = b"""""" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] ,2 )] for index in range(0 ,len(snake_case__ ) ,6 ) ).encode() + padding ) def __lowerCamelCase ( snake_case__ ) -> bytes: """simple docstring""" if not isinstance(snake_case__ ,snake_case__ ) and not isinstance(snake_case__ ,snake_case__ ): _SCREAMING_SNAKE_CASE = ( """argument should be a bytes-like object or ASCII string, """ F'not \'{encoded_data.__class__.__name__}\'' ) raise TypeError(snake_case__ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(snake_case__ ,snake_case__ ): try: _SCREAMING_SNAKE_CASE = encoded_data.decode("""utf-8""" ) except UnicodeDecodeError: raise ValueError("""base64 encoded data should only contain ASCII characters""" ) _SCREAMING_SNAKE_CASE = encoded_data.count("""=""" ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(snake_case__ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one _SCREAMING_SNAKE_CASE = encoded_data[:-padding] _SCREAMING_SNAKE_CASE = """""".join( bin(B64_CHARSET.index(snake_case__ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: _SCREAMING_SNAKE_CASE = """""".join( bin(B64_CHARSET.index(snake_case__ ) )[2:].zfill(6 ) for char in encoded_data ) _SCREAMING_SNAKE_CASE = [ int(binary_stream[index : index + 8] ,2 ) for index in range(0 ,len(snake_case__ ) ,8 ) ] return bytes(snake_case__ ) if __name__ == "__main__": import doctest doctest.testmod()
125
import argparse from pathlib import Path import requests import torch from PIL import Image from transformers import ( RobertaTokenizer, TrOCRConfig, TrOCRForCausalLM, TrOCRProcessor, VisionEncoderDecoderModel, ViTConfig, ViTImageProcessor, ViTModel, ) from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase = logging.get_logger(__name__) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Any: """simple docstring""" _SCREAMING_SNAKE_CASE = [] for i in range(encoder_config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F'encoder.deit.blocks.{i}.norm1.weight', F'encoder.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((F'encoder.deit.blocks.{i}.norm1.bias', F'encoder.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append( (F'encoder.deit.blocks.{i}.attn.proj.weight', F'encoder.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append( (F'encoder.deit.blocks.{i}.attn.proj.bias', F'encoder.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append( (F'encoder.deit.blocks.{i}.norm2.weight', F'encoder.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((F'encoder.deit.blocks.{i}.norm2.bias', F'encoder.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append( (F'encoder.deit.blocks.{i}.mlp.fc1.weight', F'encoder.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append( (F'encoder.deit.blocks.{i}.mlp.fc1.bias', F'encoder.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append( (F'encoder.deit.blocks.{i}.mlp.fc2.weight', F'encoder.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((F'encoder.deit.blocks.{i}.mlp.fc2.bias', F'encoder.encoder.layer.{i}.output.dense.bias') ) # cls token, position embeddings and patch embeddings of encoder rename_keys.extend( [ ("""encoder.deit.cls_token""", """encoder.embeddings.cls_token"""), ("""encoder.deit.pos_embed""", """encoder.embeddings.position_embeddings"""), ("""encoder.deit.patch_embed.proj.weight""", """encoder.embeddings.patch_embeddings.projection.weight"""), ("""encoder.deit.patch_embed.proj.bias""", """encoder.embeddings.patch_embeddings.projection.bias"""), ("""encoder.deit.norm.weight""", """encoder.layernorm.weight"""), ("""encoder.deit.norm.bias""", """encoder.layernorm.bias"""), ] ) return rename_keys def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Dict: """simple docstring""" for i in range(encoder_config.num_hidden_layers ): # queries, keys and values (only weights, no biases) _SCREAMING_SNAKE_CASE = state_dict.pop(F'encoder.deit.blocks.{i}.attn.qkv.weight' ) _SCREAMING_SNAKE_CASE = in_proj_weight[ : encoder_config.hidden_size, : ] _SCREAMING_SNAKE_CASE = in_proj_weight[ encoder_config.hidden_size : encoder_config.hidden_size * 2, : ] _SCREAMING_SNAKE_CASE = in_proj_weight[ -encoder_config.hidden_size :, : ] def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = dct.pop(snake_case__ ) _SCREAMING_SNAKE_CASE = val def __lowerCamelCase ( snake_case__ ) -> Union[str, Any]: """simple docstring""" if "handwritten" in checkpoint_url: _SCREAMING_SNAKE_CASE = """https://fki.tic.heia-fr.ch/static/img/a01-122-02-00.jpg""" # industry # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02-12.jpg" # have # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02-10.jpg" # let # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02.jpg" # # url = "https://fki.tic.heia-fr.ch/static/img/a01-122.jpg" elif "printed" in checkpoint_url or "stage1" in checkpoint_url: _SCREAMING_SNAKE_CASE = """https://www.researchgate.net/profile/Dinh-Sang/publication/338099565/figure/fig8/AS:840413229350922@1577381536857/An-receipt-example-in-the-SROIE-2019-dataset_Q640.jpg""" _SCREAMING_SNAKE_CASE = Image.open(requests.get(snake_case__ ,stream=snake_case__ ).raw ).convert("""RGB""" ) return im @torch.no_grad() def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Tuple: """simple docstring""" _SCREAMING_SNAKE_CASE = ViTConfig(image_size=3_84 ,qkv_bias=snake_case__ ) _SCREAMING_SNAKE_CASE = TrOCRConfig() # size of the architecture if "base" in checkpoint_url: _SCREAMING_SNAKE_CASE = 7_68 elif "large" in checkpoint_url: # use ViT-large encoder _SCREAMING_SNAKE_CASE = 10_24 _SCREAMING_SNAKE_CASE = 40_96 _SCREAMING_SNAKE_CASE = 24 _SCREAMING_SNAKE_CASE = 16 _SCREAMING_SNAKE_CASE = 10_24 else: raise ValueError("""Should either find 'base' or 'large' in checkpoint URL""" ) # the large-printed + stage1 checkpoints uses sinusoidal position embeddings, no layernorm afterwards if "large-printed" in checkpoint_url or "stage1" in checkpoint_url: _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = """relu""" _SCREAMING_SNAKE_CASE = 10_24 _SCREAMING_SNAKE_CASE = True _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = False # load HuggingFace model _SCREAMING_SNAKE_CASE = ViTModel(snake_case__ ,add_pooling_layer=snake_case__ ) _SCREAMING_SNAKE_CASE = TrOCRForCausalLM(snake_case__ ) _SCREAMING_SNAKE_CASE = VisionEncoderDecoderModel(encoder=snake_case__ ,decoder=snake_case__ ) model.eval() # load state_dict of original model, rename some keys _SCREAMING_SNAKE_CASE = torch.hub.load_state_dict_from_url(snake_case__ ,map_location="""cpu""" ,check_hash=snake_case__ )["""model"""] _SCREAMING_SNAKE_CASE = create_rename_keys(snake_case__ ,snake_case__ ) for src, dest in rename_keys: rename_key(snake_case__ ,snake_case__ ,snake_case__ ) read_in_q_k_v(snake_case__ ,snake_case__ ) # remove parameters we don't need del state_dict["encoder.deit.head.weight"] del state_dict["encoder.deit.head.bias"] del state_dict["decoder.version"] # add prefix to decoder keys for key, val in state_dict.copy().items(): _SCREAMING_SNAKE_CASE = state_dict.pop(snake_case__ ) if key.startswith("""decoder""" ) and "output_projection" not in key: _SCREAMING_SNAKE_CASE = val else: _SCREAMING_SNAKE_CASE = val # load state dict model.load_state_dict(snake_case__ ) # Check outputs on an image _SCREAMING_SNAKE_CASE = ViTImageProcessor(size=encoder_config.image_size ) _SCREAMING_SNAKE_CASE = RobertaTokenizer.from_pretrained("""roberta-large""" ) _SCREAMING_SNAKE_CASE = TrOCRProcessor(snake_case__ ,snake_case__ ) _SCREAMING_SNAKE_CASE = processor(images=prepare_img(snake_case__ ) ,return_tensors="""pt""" ).pixel_values # verify logits _SCREAMING_SNAKE_CASE = torch.tensor([[model.config.decoder.decoder_start_token_id]] ) _SCREAMING_SNAKE_CASE = model(pixel_values=snake_case__ ,decoder_input_ids=snake_case__ ) _SCREAMING_SNAKE_CASE = outputs.logits _SCREAMING_SNAKE_CASE = torch.Size([1, 1, 5_02_65] ) if "trocr-base-handwritten" in checkpoint_url: _SCREAMING_SNAKE_CASE = torch.tensor( [-1.4_502, -4.6_683, -0.5_347, -2.9_291, 9.1_435, -3.0_571, 8.9_764, 1.7_560, 8.7_358, -1.5_311] ) elif "trocr-large-handwritten" in checkpoint_url: _SCREAMING_SNAKE_CASE = torch.tensor( [-2.6_437, -1.3_129, -2.2_596, -5.3_455, 6.3_539, 1.7_604, 5.4_991, 1.4_702, 5.6_113, 2.0_170] ) elif "trocr-base-printed" in checkpoint_url: _SCREAMING_SNAKE_CASE = torch.tensor( [-5.6_816, -5.8_388, 1.1_398, -6.9_034, 6.8_505, -2.4_393, 1.2_284, -1.0_232, -1.9_661, -3.9_210] ) elif "trocr-large-printed" in checkpoint_url: _SCREAMING_SNAKE_CASE = torch.tensor( [-6.0_162, -7.0_959, 4.4_155, -5.1_063, 7.0_468, -3.1_631, 2.6_466, -0.3_081, -0.8_106, -1.7_535] ) if "stage1" not in checkpoint_url: assert logits.shape == expected_shape, "Shape of logits not as expected" assert torch.allclose(logits[0, 0, :10] ,snake_case__ ,atol=1e-3 ), "First elements of logits not as expected" Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(F'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(snake_case__ ) print(F'Saving processor to {pytorch_dump_folder_path}' ) processor.save_pretrained(snake_case__ ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() parser.add_argument( '''--checkpoint_url''', default='''https://layoutlm.blob.core.windows.net/trocr/model_zoo/fairseq/trocr-base-handwritten.pt''', type=str, help='''URL to the original PyTorch checkpoint (.pth file).''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) UpperCamelCase = parser.parse_args() convert_tr_ocr_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
125
1
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class a__ ( lowerCamelCase_ ): def __init__( self , _UpperCamelCase , _UpperCamelCase=None , _UpperCamelCase=True , _UpperCamelCase=None , **_UpperCamelCase ): """simple docstring""" _lowercase : Optional[Any] = parent _lowercase : Dict = config_class _lowercase : str = has_text_modality _lowercase : Dict = kwargs _lowercase : Tuple = common_properties def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Union[str, Any] = self.config_class(**self.inputs_dict ) _lowercase : List[Any] = ( ["hidden_size", "num_attention_heads", "num_hidden_layers"] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(["vocab_size"] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(_UpperCamelCase , _UpperCamelCase ) , msg=f'''`{prop}` does not exist''' ) # Test that config has the common properties as setter for idx, name in enumerate(_UpperCamelCase ): try: setattr(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) self.parent.assertEqual( getattr(_UpperCamelCase , _UpperCamelCase ) , _UpperCamelCase , msg=f'''`{name} value {idx} expected, but was {getattr(_UpperCamelCase , _UpperCamelCase )}''' ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(_UpperCamelCase ): try: _lowercase : str = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(_UpperCamelCase , _UpperCamelCase ) , _UpperCamelCase , msg=f'''`{name} value {idx} expected, but was {getattr(_UpperCamelCase , _UpperCamelCase )}''' ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Tuple = self.config_class(**self.inputs_dict ) _lowercase : Dict = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] , _UpperCamelCase ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Tuple = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _lowercase : Optional[int] = os.path.join(_UpperCamelCase , "config.json" ) config_first.to_json_file(_UpperCamelCase ) _lowercase : Tuple = self.config_class.from_json_file(_UpperCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase : List[str] = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(_UpperCamelCase ) _lowercase : Tuple = self.config_class.from_pretrained(_UpperCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Optional[Any] = self.config_class(**self.inputs_dict ) _lowercase : Tuple = "test" with tempfile.TemporaryDirectory() as tmpdirname: _lowercase : Optional[int] = os.path.join(_UpperCamelCase , _UpperCamelCase ) config_first.save_pretrained(_UpperCamelCase ) _lowercase : Union[str, Any] = self.config_class.from_pretrained(_UpperCamelCase , subfolder=_UpperCamelCase ) self.parent.assertEqual(config_second.to_dict() , config_first.to_dict() ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase : int = self.config_class(**self.inputs_dict , num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) , 5 ) self.parent.assertEqual(len(config.labelaid ) , 5 ) _lowercase : int = 3 self.parent.assertEqual(len(config.idalabel ) , 3 ) self.parent.assertEqual(len(config.labelaid ) , 3 ) def _lowerCamelCase ( self ): """simple docstring""" if self.config_class.is_composition: return _lowercase : Union[str, Any] = self.config_class() self.parent.assertIsNotNone(_UpperCamelCase ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase : str = copy.deepcopy(_UpperCamelCase ) _lowercase : int = self.config_class(**_UpperCamelCase ) _lowercase : Optional[int] = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(("torch_dtype", config.torch_dtype, torch.floataa) ) elif getattr(_UpperCamelCase , _UpperCamelCase ) != value: wrong_values.append((key, getattr(_UpperCamelCase , _UpperCamelCase ), value) ) if len(_UpperCamelCase ) > 0: _lowercase : Optional[Any] = "\n".join([f'''- {v[0]}: got {v[1]} instead of {v[2]}''' for v in wrong_values] ) raise ValueError(f'''The following keys were not properly set in the config:\n{errors}''' ) def _lowerCamelCase ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
250
'''simple docstring''' import doctest import glob import importlib import inspect import os import re from contextlib import contextmanager from functools import wraps from unittest.mock import patch import numpy as np import pytest from absl.testing import parameterized import datasets from datasets import load_metric from .utils import for_all_test_methods, local, slow # mark all tests as integration _snake_case = pytest.mark.integration _snake_case = {'comet'} _snake_case = importlib.util.find_spec('fairseq') is not None _snake_case = {'code_eval'} _snake_case = os.name == 'nt' _snake_case = {'bertscore', 'frugalscore', 'perplexity'} _snake_case = importlib.util.find_spec('transformers') is not None def _A ( snake_case ) -> Tuple: @wraps(snake_case ) def wrapper(self , snake_case ): if not _has_fairseq and metric_name in REQUIRE_FAIRSEQ: self.skipTest("\"test requires Fairseq\"" ) else: test_case(self , snake_case ) return wrapper def _A ( snake_case ) -> Optional[int]: @wraps(snake_case ) def wrapper(self , snake_case ): if not _has_transformers and metric_name in REQUIRE_TRANSFORMERS: self.skipTest("\"test requires transformers\"" ) else: test_case(self , snake_case ) return wrapper def _A ( snake_case ) -> List[Any]: @wraps(snake_case ) def wrapper(self , snake_case ): if _on_windows and metric_name in UNSUPPORTED_ON_WINDOWS: self.skipTest("\"test not supported on Windows\"" ) else: test_case(self , snake_case ) return wrapper def _A ( ) -> List[Any]: _lowercase : Any = [metric_dir.split(os.sep )[-2] for metric_dir in glob.glob("./metrics/*/" )] return [{"testcase_name": x, "metric_name": x} for x in metrics if x != "gleu"] # gleu is unfinished @parameterized.named_parameters(get_local_metric_names() ) @for_all_test_methods( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) @local class a__ ( parameterized.TestCase ): _SCREAMING_SNAKE_CASE : Any = {} _SCREAMING_SNAKE_CASE : Any = None @pytest.mark.filterwarnings("ignore:metric_module_factory is deprecated:FutureWarning" ) @pytest.mark.filterwarnings("ignore:load_metric is deprecated:FutureWarning" ) def _lowerCamelCase ( self , _UpperCamelCase ): """simple docstring""" _lowercase : Tuple = "[...]" _lowercase : Optional[Any] = importlib.import_module( datasets.load.metric_module_factory(os.path.join("metrics" , _UpperCamelCase ) ).module_path ) _lowercase : Union[str, Any] = datasets.load.import_main_class(metric_module.__name__ , dataset=_UpperCamelCase ) # check parameters _lowercase : str = inspect.signature(metric._compute ).parameters self.assertTrue(all(p.kind != p.VAR_KEYWORD for p in parameters.values() ) ) # no **kwargs # run doctest with self.patch_intensive_calls(_UpperCamelCase , metric_module.__name__ ): with self.use_local_metrics(): try: _lowercase : int = doctest.testmod(_UpperCamelCase , verbose=_UpperCamelCase , raise_on_error=_UpperCamelCase ) except doctest.UnexpectedException as e: raise e.exc_info[1] # raise the exception that doctest caught self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @slow def _lowerCamelCase ( self , _UpperCamelCase ): """simple docstring""" _lowercase : Any = "[...]" _lowercase : Dict = importlib.import_module( datasets.load.metric_module_factory(os.path.join("metrics" , _UpperCamelCase ) ).module_path ) # run doctest with self.use_local_metrics(): _lowercase : str = doctest.testmod(_UpperCamelCase , verbose=_UpperCamelCase , raise_on_error=_UpperCamelCase ) self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @contextmanager def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" if metric_name in self.INTENSIVE_CALLS_PATCHER: with self.INTENSIVE_CALLS_PATCHER[metric_name](_UpperCamelCase ): yield else: yield @contextmanager def _lowerCamelCase ( self ): """simple docstring""" def load_local_metric(_UpperCamelCase , *_UpperCamelCase , **_UpperCamelCase ): return load_metric(os.path.join("metrics" , _UpperCamelCase ) , *_UpperCamelCase , **_UpperCamelCase ) with patch("datasets.load_metric" ) as mock_load_metric: _lowercase : List[Any] = load_local_metric yield @classmethod def _lowerCamelCase ( cls , _UpperCamelCase ): """simple docstring""" def wrapper(_UpperCamelCase ): _lowercase : str = contextmanager(_UpperCamelCase ) _lowercase : Any = patcher return patcher return wrapper @LocalMetricTest.register_intensive_calls_patcher("bleurt" ) def _A ( snake_case ) -> List[Any]: import tensorflow.compat.va as tf from bleurt.score import Predictor tf.flags.DEFINE_string("sv" , "" , "" ) # handle pytest cli flags class a__ ( lowerCamelCase_ ): def _lowerCamelCase ( self , _UpperCamelCase ): """simple docstring""" assert len(input_dict["input_ids"] ) == 2 return np.array([1.0_3, 1.0_4] ) # mock predict_fn which is supposed to do a forward pass with a bleurt model with patch("bleurt.score._create_predictor" ) as mock_create_predictor: _lowercase : List[Any] = MockedPredictor() yield @LocalMetricTest.register_intensive_calls_patcher("bertscore" ) def _A ( snake_case ) -> Tuple: import torch def bert_cos_score_idf(snake_case , snake_case , *snake_case , **snake_case ): return torch.tensor([[1.0, 1.0, 1.0]] * len(snake_case ) ) # mock get_model which is supposed to do download a bert model # mock bert_cos_score_idf which is supposed to do a forward pass with a bert model with patch("bert_score.scorer.get_model" ), patch( "bert_score.scorer.bert_cos_score_idf" ) as mock_bert_cos_score_idf: _lowercase : List[str] = bert_cos_score_idf yield @LocalMetricTest.register_intensive_calls_patcher("comet" ) def _A ( snake_case ) -> Optional[int]: def load_from_checkpoint(snake_case ): class a__ : def _lowerCamelCase ( self , _UpperCamelCase , *_UpperCamelCase , **_UpperCamelCase ): """simple docstring""" assert len(_UpperCamelCase ) == 2 _lowercase : Tuple = [0.1_9, 0.9_2] return scores, sum(_UpperCamelCase ) / len(_UpperCamelCase ) return Model() # mock load_from_checkpoint which is supposed to do download a bert model # mock load_from_checkpoint which is supposed to do download a bert model with patch("comet.download_model" ) as mock_download_model: _lowercase : Union[str, Any] = None with patch("comet.load_from_checkpoint" ) as mock_load_from_checkpoint: _lowercase : str = load_from_checkpoint yield def _A ( ) -> Optional[Any]: _lowercase : str = load_metric(os.path.join("metrics" , "seqeval" ) ) _lowercase : Optional[int] = "ERROR" _lowercase : Any = F'''Scheme should be one of [IOB1, IOB2, IOE1, IOE2, IOBES, BILOU], got {wrong_scheme}''' with pytest.raises(snake_case , match=re.escape(snake_case ) ): metric.compute(predictions=[] , references=[] , scheme=snake_case )
250
1
from ... import PretrainedConfig lowerCAmelCase = { 'sijunhe/nezha-cn-base': 'https://huggingface.co/sijunhe/nezha-cn-base/resolve/main/config.json', } class _a ( UpperCamelCase__ ): """simple docstring""" _lowercase : List[str] = NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP _lowercase : Union[str, Any] = '''nezha''' def __init__( self: Optional[int] , UpperCamelCase_: Tuple=21_128 , UpperCamelCase_: Optional[Any]=768 , UpperCamelCase_: Optional[Any]=12 , UpperCamelCase_: List[Any]=12 , UpperCamelCase_: List[str]=3_072 , UpperCamelCase_: List[Any]="gelu" , UpperCamelCase_: int=0.1 , UpperCamelCase_: List[Any]=0.1 , UpperCamelCase_: int=512 , UpperCamelCase_: Any=64 , UpperCamelCase_: Optional[int]=2 , UpperCamelCase_: str=0.02 , UpperCamelCase_: Any=1E-1_2 , UpperCamelCase_: str=0.1 , UpperCamelCase_: Optional[Any]=0 , UpperCamelCase_: Dict=2 , UpperCamelCase_: List[Any]=3 , UpperCamelCase_: Any=True , **UpperCamelCase_: Union[str, Any] , ) -> List[str]: """simple docstring""" super().__init__(pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , **__lowerCamelCase ) lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = hidden_act lowercase__ = intermediate_size lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = max_relative_position lowercase__ = type_vocab_size lowercase__ = initializer_range lowercase__ = layer_norm_eps lowercase__ = classifier_dropout lowercase__ = use_cache
367
def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = len(SCREAMING_SNAKE_CASE ) lowercase__ = [] for i in range(len(SCREAMING_SNAKE_CASE ) - pat_len + 1 ): lowercase__ = True for j in range(SCREAMING_SNAKE_CASE ): if s[i + j] != pattern[j]: lowercase__ = False break if match_found: position.append(SCREAMING_SNAKE_CASE ) return position if __name__ == "__main__": assert naive_pattern_search('ABCDEFG', 'DE') == [3] print(naive_pattern_search('ABAAABCDBBABCDDEBCABC', 'ABC'))
93
0
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, BlipProcessor, PreTrainedTokenizerFast @require_vision class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self : Dict ): __A = tempfile.mkdtemp() __A = BlipImageProcessor() __A = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-BertModel" ) __A = BlipProcessor(A ,A ) processor.save_pretrained(self.tmpdirname ) def UpperCamelCase_ ( self : Union[str, Any] ,**A : int ): return AutoProcessor.from_pretrained(self.tmpdirname ,**A ).tokenizer def UpperCamelCase_ ( self : List[str] ,**A : Optional[Any] ): return AutoProcessor.from_pretrained(self.tmpdirname ,**A ).image_processor def UpperCamelCase_ ( self : Optional[int] ): shutil.rmtree(self.tmpdirname ) def UpperCamelCase_ ( self : int ): __A = [np.random.randint(2_55 ,size=(3, 30, 4_00) ,dtype=np.uinta )] __A = [Image.fromarray(np.moveaxis(A ,0 ,-1 ) ) for x in image_inputs] return image_inputs def UpperCamelCase_ ( self : List[Any] ): __A = BlipProcessor(tokenizer=self.get_tokenizer() ,image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __A = self.get_tokenizer(bos_token="(BOS)" ,eos_token="(EOS)" ) __A = self.get_image_processor(do_normalize=A ,padding_value=1.0 ) __A = BlipProcessor.from_pretrained( self.tmpdirname ,bos_token="(BOS)" ,eos_token="(EOS)" ,do_normalize=A ,padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() ,tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer ,A ) self.assertEqual(processor.image_processor.to_json_string() ,image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor ,A ) def UpperCamelCase_ ( self : List[Any] ): __A = self.get_image_processor() __A = self.get_tokenizer() __A = BlipProcessor(tokenizer=A ,image_processor=A ) __A = self.prepare_image_inputs() __A = image_processor(A ,return_tensors="np" ) __A = processor(images=A ,return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() ,input_processor[key].sum() ,delta=1E-2 ) def UpperCamelCase_ ( self : Optional[int] ): __A = self.get_image_processor() __A = self.get_tokenizer() __A = BlipProcessor(tokenizer=A ,image_processor=A ) __A = "lower newer" __A = processor(text=A ) __A = tokenizer(A ,return_token_type_ids=A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] ,encoded_processor[key] ) def UpperCamelCase_ ( self : List[Any] ): __A = self.get_image_processor() __A = self.get_tokenizer() __A = BlipProcessor(tokenizer=A ,image_processor=A ) __A = "lower newer" __A = self.prepare_image_inputs() __A = processor(text=A ,images=A ) self.assertListEqual(list(inputs.keys() ) ,["pixel_values", "input_ids", "attention_mask"] ) # test if it raises when no input is passed with pytest.raises(A ): processor() def UpperCamelCase_ ( self : Optional[int] ): __A = self.get_image_processor() __A = self.get_tokenizer() __A = BlipProcessor(tokenizer=A ,image_processor=A ) __A = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __A = processor.batch_decode(A ) __A = tokenizer.batch_decode(A ) self.assertListEqual(A ,A ) def UpperCamelCase_ ( self : Union[str, Any] ): __A = self.get_image_processor() __A = self.get_tokenizer() __A = BlipProcessor(tokenizer=A ,image_processor=A ) __A = "lower newer" __A = self.prepare_image_inputs() __A = processor(text=A ,images=A ) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys() ) ,["pixel_values", "input_ids", "attention_mask"] )
15
from heapq import heappop, heappush import numpy as np def __lowercase ( a__ , a__ , a__ , a__ , ) -> tuple[float | int, list[tuple[int, int]]]: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = grid.shape __SCREAMING_SNAKE_CASE = [-1, 1, 0, 0] __SCREAMING_SNAKE_CASE = [0, 0, -1, 1] if allow_diagonal: dx += [-1, -1, 1, 1] dy += [-1, 1, -1, 1] __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = [(0, source)], set() __SCREAMING_SNAKE_CASE = np.full((rows, cols) , np.inf ) __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = np.empty((rows, cols) , dtype=a__ ) __SCREAMING_SNAKE_CASE = None while queue: ((__SCREAMING_SNAKE_CASE) , (__SCREAMING_SNAKE_CASE)) = heappop(a__ ) if (x, y) in visited: continue visited.add((x, y) ) if (x, y) == destination: __SCREAMING_SNAKE_CASE = [] while (x, y) != source: path.append((x, y) ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = predecessors[x, y] path.append(a__ ) # add the source manually path.reverse() return matrix[destination], path for i in range(len(a__ ) ): __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = x + dx[i], y + dy[i] if 0 <= nx < rows and 0 <= ny < cols: __SCREAMING_SNAKE_CASE = grid[nx][ny] if next_node == 1 and matrix[nx, ny] > dist + 1: heappush(a__ , (dist + 1, (nx, ny)) ) __SCREAMING_SNAKE_CASE = dist + 1 __SCREAMING_SNAKE_CASE = (x, y) return np.inf, [] if __name__ == "__main__": import doctest doctest.testmod()
257
0
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record _UpperCAmelCase = '\\n@article{wang2019superglue,\n title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems},\n author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R},\n journal={arXiv preprint arXiv:1905.00537},\n year={2019}\n}\n' _UpperCAmelCase = '\\nSuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after\nGLUE with a new set of more difficult language understanding tasks, improved\nresources, and a new public leaderboard.\n' _UpperCAmelCase = '\nCompute SuperGLUE evaluation metric associated to each SuperGLUE dataset.\nArgs:\n predictions: list of predictions to score. Depending on the SuperGlUE subset:\n - for \'record\': list of question-answer dictionaries with the following keys:\n - \'idx\': index of the question as specified by the dataset\n - \'prediction_text\': the predicted answer text\n - for \'multirc\': list of question-answer dictionaries with the following keys:\n - \'idx\': index of the question-answer pair as specified by the dataset\n - \'prediction\': the predicted answer label\n - otherwise: list of predicted labels\n references: list of reference labels. Depending on the SuperGLUE subset:\n - for \'record\': list of question-answers dictionaries with the following keys:\n - \'idx\': index of the question as specified by the dataset\n - \'answers\': list of possible answers\n - otherwise: list of reference labels\nReturns: depending on the SuperGLUE subset:\n - for \'record\':\n - \'exact_match\': Exact match between answer and gold answer\n - \'f1\': F1 score\n - for \'multirc\':\n - \'exact_match\': Exact match between answer and gold answer\n - \'f1_m\': Per-question macro-F1 score\n - \'f1_a\': Average F1 score over all answers\n - for \'axb\':\n \'matthews_correlation\': Matthew Correlation\n - for \'cb\':\n - \'accuracy\': Accuracy\n - \'f1\': F1 score\n - for all others:\n - \'accuracy\': Accuracy\nExamples:\n\n >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'copa\') # any of ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n\n >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'cb\')\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0, \'f1\': 1.0}\n\n >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'record\')\n >>> predictions = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'prediction_text\': \'answer\'}]\n >>> references = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'answers\': [\'answer\', \'another_answer\']}]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'exact_match\': 1.0, \'f1\': 1.0}\n\n >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'multirc\')\n >>> predictions = [{\'idx\': {\'answer\': 0, \'paragraph\': 0, \'question\': 0}, \'prediction\': 0}, {\'idx\': {\'answer\': 1, \'paragraph\': 2, \'question\': 3}, \'prediction\': 1}]\n >>> references = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'exact_match\': 1.0, \'f1_m\': 1.0, \'f1_a\': 1.0}\n\n >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'axb\')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = super_glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'matthews_correlation\': 1.0}\n' def lowerCAmelCase_ ( UpperCamelCase_ , UpperCamelCase_ ) -> Optional[Any]: return float((preds == labels).mean() ) def lowerCAmelCase_ ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_="binary" ) -> Tuple: UpperCamelCase_ = simple_accuracy(UpperCamelCase_ , UpperCamelCase_ ) UpperCamelCase_ = float(fa_score(y_true=UpperCamelCase_ , y_pred=UpperCamelCase_ , average=UpperCamelCase_ ) ) return { "accuracy": acc, "f1": fa, } def lowerCAmelCase_ ( UpperCamelCase_ , UpperCamelCase_ ) -> List[Any]: UpperCamelCase_ = {} for id_pred, label in zip(UpperCamelCase_ , UpperCamelCase_ ): UpperCamelCase_ = F'''{id_pred["idx"]["paragraph"]}-{id_pred["idx"]["question"]}''' UpperCamelCase_ = id_pred["prediction"] if question_id in question_map: question_map[question_id].append((pred, label) ) else: UpperCamelCase_ = [(pred, label)] UpperCamelCase_ , UpperCamelCase_ = [], [] for question, preds_labels in question_map.items(): UpperCamelCase_ , UpperCamelCase_ = zip(*UpperCamelCase_ ) UpperCamelCase_ = fa_score(y_true=UpperCamelCase_ , y_pred=UpperCamelCase_ , average="macro" ) fas.append(UpperCamelCase_ ) UpperCamelCase_ = int(sum(pred == label for pred, label in preds_labels ) == len(UpperCamelCase_ ) ) ems.append(UpperCamelCase_ ) UpperCamelCase_ = float(sum(UpperCamelCase_ ) / len(UpperCamelCase_ ) ) UpperCamelCase_ = sum(UpperCamelCase_ ) / len(UpperCamelCase_ ) UpperCamelCase_ = float(fa_score(y_true=UpperCamelCase_ , y_pred=[id_pred["prediction"] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _UpperCamelCase ( datasets.Metric ): def lowercase ( self: Optional[int] ) -> Optional[int]: """simple docstring""" if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( "You should supply a configuration name selected in " "[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format="numpy" if not self.config_name == "record" and not self.config_name == "multirc" else None , ) def lowercase ( self: List[Any] ) -> int: """simple docstring""" if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value("int64" ), "query": datasets.Value("int64" ), }, "prediction_text": datasets.Value("string" ), }, "references": { "idx": { "passage": datasets.Value("int64" ), "query": datasets.Value("int64" ), }, "answers": datasets.Sequence(datasets.Value("string" ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value("int64" ), "paragraph": datasets.Value("int64" ), "question": datasets.Value("int64" ), }, "prediction": datasets.Value("int64" ), }, "references": datasets.Value("int64" ), } else: return { "predictions": datasets.Value("int64" ), "references": datasets.Value("int64" ), } def lowercase ( self: Tuple , _SCREAMING_SNAKE_CASE: Optional[Any] , _SCREAMING_SNAKE_CASE: List[str] ) -> Dict: """simple docstring""" if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )} elif self.config_name == "cb": return acc_and_fa(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , fa_avg="macro" ) elif self.config_name == "record": UpperCamelCase_ = [ { "qas": [ {"id": ref["idx"]["query"], "answers": [{"text": ans} for ans in ref["answers"]]} for ref in references ] } ] UpperCamelCase_ = {pred["idx"]["query"]: pred["prediction_text"] for pred in predictions} return evaluate_record(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )[0] elif self.config_name == "multirc": return evaluate_multirc(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE )} else: raise KeyError( "You should supply a configuration name selected in " "[\"boolq\", \"cb\", \"copa\", \"multirc\", \"record\", \"rte\", \"wic\", \"wsc\", \"wsc.fixed\", \"axb\", \"axg\",]" )
361
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _UpperCAmelCase = { 'configuration_tapas': ['TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TapasConfig'], 'tokenization_tapas': ['TapasTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase = [ 'TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TapasForMaskedLM', 'TapasForQuestionAnswering', 'TapasForSequenceClassification', 'TapasModel', 'TapasPreTrainedModel', 'load_tf_weights_in_tapas', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase = [ 'TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFTapasForMaskedLM', 'TFTapasForQuestionAnswering', 'TFTapasForSequenceClassification', 'TFTapasModel', 'TFTapasPreTrainedModel', ] if TYPE_CHECKING: from .configuration_tapas import TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP, TapasConfig from .tokenization_tapas import TapasTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tapas import ( TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasPreTrainedModel, load_tf_weights_in_tapas, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_tapas import ( TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TFTapasForMaskedLM, TFTapasForQuestionAnswering, TFTapasForSequenceClassification, TFTapasModel, TFTapasPreTrainedModel, ) else: import sys _UpperCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
328
0
'''simple docstring''' from __future__ import annotations import time from collections.abc import Sequence from random import randint from matplotlib import pyplot as plt def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> tuple[int | None, int | None, float]: if not arr: return None, None, 0 if low == high: return low, high, arr[low] _a : Any = (low + high) // 2 _a , _a , _a : Tuple = max_subarray(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _a , _a , _a : List[Any] = max_subarray(lowerCAmelCase_ , mid + 1 , lowerCAmelCase_ ) _a , _a , _a : Optional[int] = max_cross_sum(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) if left_sum >= right_sum and left_sum >= cross_sum: return left_low, left_high, left_sum elif right_sum >= left_sum and right_sum >= cross_sum: return right_low, right_high, right_sum return cross_left, cross_right, cross_sum def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> tuple[int, int, float]: _a , _a : List[Any] = float('-inf' ), -1 _a , _a : List[str] = float('-inf' ), -1 _a : int | float = 0 for i in range(lowerCAmelCase_ , low - 1 , -1 ): summ += arr[i] if summ > left_sum: _a : int = summ _a : List[str] = i _a : List[str] = 0 for i in range(mid + 1 , high + 1 ): summ += arr[i] if summ > right_sum: _a : List[Any] = summ _a : Optional[Any] = i return max_left, max_right, (left_sum + right_sum) def __lowerCamelCase ( lowerCAmelCase_ ) -> float: _a : List[Any] = [randint(1 , lowerCAmelCase_ ) for _ in range(lowerCAmelCase_ )] _a : str = time.time() max_subarray(lowerCAmelCase_ , 0 , input_size - 1 ) _a : Optional[Any] = time.time() return end - start def __lowerCamelCase ( ) -> None: _a : Tuple = [10, 100, 1000, 10000, 50000, 100000, 200000, 300000, 400000, 500000] _a : Union[str, Any] = [time_max_subarray(lowerCAmelCase_ ) for input_size in input_sizes] print('No of Inputs\t\tTime Taken' ) for input_size, runtime in zip(lowerCAmelCase_ , lowerCAmelCase_ ): print(lowerCAmelCase_ , '\t\t' , lowerCAmelCase_ ) plt.plot(lowerCAmelCase_ , lowerCAmelCase_ ) plt.xlabel('Number of Inputs' ) plt.ylabel('Time taken in seconds' ) plt.show() if __name__ == "__main__": from doctest import testmod testmod()
89
'''simple docstring''' import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class __magic_name__ ( _UpperCamelCase ): @require_torch def __lowercase ( self : Tuple ): # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _a : Optional[int] = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _a : List[str] = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _a : Tuple = '\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _a : List[Any] = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(_UpperCAmelCase ) BertModel.from_pretrained(_UpperCAmelCase ) BertTokenizer.from_pretrained(_UpperCAmelCase ) pipeline(task='fill-mask' ,model=_UpperCAmelCase ) # baseline - just load from_pretrained with normal network _a : Optional[int] = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _a : Tuple = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _a : int = '1' _a : List[Any] = subprocess.run(_UpperCAmelCase ,env=_UpperCAmelCase ,check=_UpperCAmelCase ,capture_output=_UpperCAmelCase ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def __lowercase ( self : Any ): # python one-liner segments # this must be loaded before socket.socket is monkey-patched _a : Dict = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _a : Optional[int] = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _a : Optional[Any] = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _a : int = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(_UpperCAmelCase ) BertModel.from_pretrained(_UpperCAmelCase ) BertTokenizer.from_pretrained(_UpperCAmelCase ) pipeline(task='fill-mask' ,model=_UpperCAmelCase ) # baseline - just load from_pretrained with normal network _a : Optional[int] = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _a : str = self.get_env() _a : Optional[Any] = subprocess.run(_UpperCAmelCase ,env=_UpperCAmelCase ,check=_UpperCAmelCase ,capture_output=_UpperCAmelCase ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def __lowercase ( self : List[str] ): # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched _a : Union[str, Any] = '\nfrom transformers import BertConfig, BertModel, BertTokenizer\n ' _a : Optional[Any] = '\nmname = "hf-internal-testing/tiny-random-bert-sharded"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint("success")\n ' _a : str = '\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled")\nsocket.socket = offline_socket\n ' # baseline - just load from_pretrained with normal network _a : Optional[Any] = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _a : Dict = self.get_env() _a : int = subprocess.run(_UpperCAmelCase ,env=_UpperCAmelCase ,check=_UpperCAmelCase ,capture_output=_UpperCAmelCase ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) # next emulate no network _a : List[Any] = [sys.executable, '-c', '\n'.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _a : int = '1' _a : Any = subprocess.run(_UpperCAmelCase ,env=_UpperCAmelCase ,check=_UpperCAmelCase ,capture_output=_UpperCAmelCase ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) @require_torch def __lowercase ( self : int ): _a : Optional[Any] = '\nfrom transformers import pipeline\n ' _a : str = '\nmname = "hf-internal-testing/tiny-random-bert"\npipe = pipeline(model=mname)\n ' _a : List[str] = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled")\nsocket.socket = offline_socket\n ' _a : List[Any] = self.get_env() _a : Dict = '1' _a : Dict = [sys.executable, '-c', '\n'.join([load, mock, run] )] _a : str = subprocess.run(_UpperCAmelCase ,env=_UpperCAmelCase ,check=_UpperCAmelCase ,capture_output=_UpperCAmelCase ) self.assertEqual(result.returncode ,1 ,result.stderr ) self.assertIn( 'You cannot infer task automatically within `pipeline` when using offline mode' ,result.stderr.decode().replace('\n' ,'' ) ,) @require_torch def __lowercase ( self : int ): _a : Optional[int] = '\nfrom transformers import AutoModel\n ' _a : List[Any] = '\nmname = "hf-internal-testing/test_dynamic_model"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint("success")\n ' # baseline - just load from_pretrained with normal network _a : Union[str, Any] = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _a : Tuple = self.get_env() _a : List[str] = subprocess.run(_UpperCAmelCase ,env=_UpperCAmelCase ,check=_UpperCAmelCase ,capture_output=_UpperCAmelCase ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _a : Optional[Any] = '1' _a : Any = subprocess.run(_UpperCAmelCase ,env=_UpperCAmelCase ,check=_UpperCAmelCase ,capture_output=_UpperCAmelCase ) self.assertEqual(result.returncode ,0 ,result.stderr ) self.assertIn('success' ,result.stdout.decode() )
89
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import tensorflow as tf from transformers import AutoTokenizer, TFAutoModelForSeqaSeqLM @require_tf @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE ( self : Any ) -> str: a_ : Tuple = TFAutoModelForSeqaSeqLM.from_pretrained('google/mt5-small' ) a_ : int = AutoTokenizer.from_pretrained('google/mt5-small' ) a_ : Dict = tokenizer('Hello there' , return_tensors='tf' ).input_ids a_ : str = tokenizer('Hi I am' , return_tensors='tf' ).input_ids a_ : int = model(SCREAMING_SNAKE_CASE__ , labels=SCREAMING_SNAKE_CASE__ ).loss a_ : List[str] = -tf.math.reduce_mean(SCREAMING_SNAKE_CASE__ ).numpy() a_ : Optional[int] = -21.228168 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 2E-4 )
350
def SCREAMING_SNAKE_CASE_ ( __A : list ) -> bool: """simple docstring""" if not isinstance(__A , __A ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(__A ) == 0: raise ValueError('Input list must be a non empty list' ) if len(__A ) == 1: return True a_ : Tuple = series[1] - series[0] for index in range(len(__A ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def SCREAMING_SNAKE_CASE_ ( __A : list ) -> float: """simple docstring""" if not isinstance(__A , __A ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(__A ) == 0: raise ValueError('Input list must be a non empty list' ) a_ : str = 0 for val in series: answer += val return answer / len(__A ) if __name__ == "__main__": import doctest doctest.testmod()
120
0
'''simple docstring''' import math class a_ : def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> int: """simple docstring""" UpperCamelCase = 0.0 UpperCamelCase = 0.0 for i in range(len(_SCREAMING_SNAKE_CASE ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def A__ ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> list[list[int | float]]: """simple docstring""" for i in range(len(_SCREAMING_SNAKE_CASE ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def lowercase__ ( )-> None: # Training Examples ( m, n ) UpperCamelCase = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) UpperCamelCase = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training UpperCamelCase = SelfOrganizingMap() UpperCamelCase = 3 UpperCamelCase = 0.5 for _ in range(__UpperCamelCase ): for j in range(len(__UpperCamelCase ) ): # training sample UpperCamelCase = training_samples[j] # Compute the winning vector UpperCamelCase = self_organizing_map.get_winner(__UpperCamelCase , __UpperCamelCase ) # Update the winning vector UpperCamelCase = self_organizing_map.update(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # classify test sample UpperCamelCase = [0, 0, 0, 1] UpperCamelCase = self_organizing_map.get_winner(__UpperCamelCase , __UpperCamelCase ) # results print(F"Clusters that the test sample belongs to : {winner}" ) print(F"Weights that have been trained : {weights}" ) # running the main() function if __name__ == "__main__": main()
321
'''simple docstring''' from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool from google.protobuf import symbol_database as _symbol_database from google.protobuf.internal import builder as _builder # @@protoc_insertion_point(imports) SCREAMING_SNAKE_CASE__ = _symbol_database.Default() SCREAMING_SNAKE_CASE__ = _descriptor_pool.Default().AddSerializedFile( b'\n\x19sentencepiece_model.proto\x12\rsentencepiece"\x80\x0c\n\x0bTrainerSpec\x12\r\n\x05input\x18\x01 \x03(\t\x12\x14\n\x0cinput_format\x18\x07 \x01(\t\x12\x14\n\x0cmodel_prefix\x18\x02 \x01(\t\x12\x41\n\nmodel_type\x18\x03 \x01(\x0e\x32$.sentencepiece.TrainerSpec.ModelType:\x07UNIGRAM\x12\x18\n\nvocab_size\x18\x04 \x01(\x05:\x04\x38\x30\x30\x30\x12\x17\n\x0f\x61\x63\x63\x65pt_language\x18\x05 \x03(\t\x12 \n\x15self_test_sample_size\x18\x06 \x01(\x05:\x01\x30\x12*\n\x1b\x65nable_differential_privacy\x18\x32 \x01(\x08:\x05\x66\x61lse\x12+\n differential_privacy_noise_level\x18\x33 \x01(\x02:\x01\x30\x12\x32\n\'differential_privacy_clipping_threshold\x18\x34 \x01(\x04:\x01\x30\x12"\n\x12\x63haracter_coverage\x18\n \x01(\x02:\x06\x30.9995\x12\x1e\n\x13input_sentence_size\x18\x0b \x01(\x04:\x01\x30\x12$\n\x16shuffle_input_sentence\x18\x13 \x01(\x08:\x04true\x12 \n\x14mining_sentence_size\x18\x0c \x01(\x05\x42\x02\x18\x01\x12"\n\x16training_sentence_size\x18\r \x01(\x05\x42\x02\x18\x01\x12(\n\x17seed_sentencepiece_size\x18\x0e \x01(\x05:\x07\x31\x30\x30\x30\x30\x30\x30\x12\x1e\n\x10shrinking_factor\x18\x0f \x01(\x02:\x04\x30.75\x12!\n\x13max_sentence_length\x18\x12 \x01(\x05:\x04\x34\x31\x39\x32\x12\x17\n\x0bnum_threads\x18\x10 \x01(\x05:\x02\x31\x36\x12\x1d\n\x12num_sub_iterations\x18\x11 \x01(\x05:\x01\x32\x12$\n\x18max_sentencepiece_length\x18\x14 \x01(\x05:\x02\x31\x36\x12%\n\x17split_by_unicode_script\x18\x15 \x01(\x08:\x04true\x12\x1d\n\x0fsplit_by_number\x18\x17 \x01(\x08:\x04true\x12!\n\x13split_by_whitespace\x18\x16 \x01(\x08:\x04true\x12)\n\x1atreat_whitespace_as_suffix\x18\x18 \x01(\x08:\x05\x66\x61lse\x12+\n\x1c\x61llow_whitespace_only_pieces\x18\x1a \x01(\x08:\x05\x66\x61lse\x12\x1b\n\x0csplit_digits\x18\x19 \x01(\x08:\x05\x66\x61lse\x12#\n\x19pretokenization_delimiter\x18\x35 \x01(\t:\x00\x12\x17\n\x0f\x63ontrol_symbols\x18\x1e \x03(\t\x12\x1c\n\x14user_defined_symbols\x18\x1f \x03(\t\x12\x16\n\x0erequired_chars\x18$ \x01(\t\x12\x1c\n\rbyte_fallback\x18# \x01(\x08:\x05\x66\x61lse\x12+\n\x1dvocabulary_output_piece_score\x18 \x01(\x08:\x04true\x12\x1e\n\x10hard_vocab_limit\x18! \x01(\x08:\x04true\x12\x1c\n\ruse_all_vocab\x18" \x01(\x08:\x05\x66\x61lse\x12\x11\n\x06unk_id\x18( \x01(\x05:\x01\x30\x12\x11\n\x06\x62os_id\x18) \x01(\x05:\x01\x31\x12\x11\n\x06\x65os_id\x18* \x01(\x05:\x01\x32\x12\x12\n\x06pad_id\x18+ \x01(\x05:\x02-1\x12\x18\n\tunk_piece\x18- \x01(\t:\x05<unk>\x12\x16\n\tbos_piece\x18. \x01(\t:\x03<s>\x12\x17\n\teos_piece\x18/ \x01(\t:\x04</s>\x12\x18\n\tpad_piece\x18\x30 \x01(\t:\x05<pad>\x12\x1a\n\x0bunk_surface\x18, \x01(\t:\x05 \xe2\x81\x87 \x12+\n\x1ctrain_extremely_large_corpus\x18\x31 \x01(\x08:\x05\x66\x61lse"5\n\tModelType\x12\x0b\n\x07UNIGRAM\x10\x01\x12\x07\n\x03\x42PE\x10\x02\x12\x08\n\x04WORD\x10\x03\x12\x08\n\x04\x43HAR\x10\x04*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02"\xd1\x01\n\x0eNormalizerSpec\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x1c\n\x14precompiled_charsmap\x18\x02 \x01(\x0c\x12\x1e\n\x10\x61\x64\x64_dummy_prefix\x18\x03 \x01(\x08:\x04true\x12&\n\x18remove_extra_whitespaces\x18\x04 \x01(\x08:\x04true\x12 \n\x12\x65scape_whitespaces\x18\x05 \x01(\x08:\x04true\x12\x1e\n\x16normalization_rule_tsv\x18\x06 \x01(\t*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02"y\n\x0cSelfTestData\x12\x33\n\x07samples\x18\x01 \x03(\x0b\x32".sentencepiece.SelfTestData.Sample\x1a)\n\x06Sample\x12\r\n\x05input\x18\x01 \x01(\t\x12\x10\n\x08\x65xpected\x18\x02 \x01(\t*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02"\xfe\x03\n\nModelProto\x12\x37\n\x06pieces\x18\x01 \x03(\x0b\x32\'.sentencepiece.ModelProto.SentencePiece\x12\x30\n\x0ctrainer_spec\x18\x02 \x01(\x0b\x32\x1a.sentencepiece.TrainerSpec\x12\x36\n\x0fnormalizer_spec\x18\x03 \x01(\x0b\x32\x1d.sentencepiece.NormalizerSpec\x12\x33\n\x0eself_test_data\x18\x04 \x01(\x0b\x32\x1b.sentencepiece.SelfTestData\x12\x38\n\x11\x64\x65normalizer_spec\x18\x05 \x01(\x0b\x32\x1d.sentencepiece.NormalizerSpec\x1a\xd2\x01\n\rSentencePiece\x12\r\n\x05piece\x18\x01 \x01(\t\x12\r\n\x05score\x18\x02 \x01(\x02\x12\x42\n\x04type\x18\x03 \x01(\x0e\x32,.sentencepiece.ModelProto.SentencePiece.Type:\x06NORMAL"T\n\x04Type\x12\n\n\x06NORMAL\x10\x01\x12\x0b\n\x07UNKNOWN\x10\x02\x12\x0b\n\x07\x43ONTROL\x10\x03\x12\x10\n\x0cUSER_DEFINED\x10\x04\x12\x08\n\x04\x42YTE\x10\x06\x12\n\n\x06UNUSED\x10\x05*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\x42\x02H\x03' ) SCREAMING_SNAKE_CASE__ = globals() _builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals) _builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, 'sentencepiece_model_pb2', _globals) if _descriptor._USE_C_DESCRIPTORS is False: SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = b'H\003' # (generated by protobuf compiler, but `_TRAINERSPEC` is not defined) # _TRAINERSPEC.fields_by_name["mining_sentence_size"]._options = None # _TRAINERSPEC.fields_by_name["mining_sentence_size"]._serialized_options = b"\030\001" # _TRAINERSPEC.fields_by_name["training_sentence_size"]._options = None # _TRAINERSPEC.fields_by_name["training_sentence_size"]._serialized_options = b"\030\001" SCREAMING_SNAKE_CASE__ = 4_5 SCREAMING_SNAKE_CASE__ = 1_5_8_1 SCREAMING_SNAKE_CASE__ = 1_5_1_7 SCREAMING_SNAKE_CASE__ = 1_5_7_0 SCREAMING_SNAKE_CASE__ = 1_5_8_4 SCREAMING_SNAKE_CASE__ = 1_7_9_3 SCREAMING_SNAKE_CASE__ = 1_7_9_5 SCREAMING_SNAKE_CASE__ = 1_9_1_6 SCREAMING_SNAKE_CASE__ = 1_8_6_4 SCREAMING_SNAKE_CASE__ = 1_9_0_5 SCREAMING_SNAKE_CASE__ = 1_9_1_9 SCREAMING_SNAKE_CASE__ = 2_4_2_9 SCREAMING_SNAKE_CASE__ = 2_2_0_8 SCREAMING_SNAKE_CASE__ = 2_4_1_8 SCREAMING_SNAKE_CASE__ = 2_3_2_3 SCREAMING_SNAKE_CASE__ = 2_4_0_7 # @@protoc_insertion_point(module_scope)
321
1
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor UpperCAmelCase : Optional[Any] = logging.get_logger(__name__) class __lowerCAmelCase ( UpperCamelCase__): def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> None: '''simple docstring''' warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCAmelCase__ , ) super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__ )
148
from __future__ import annotations from math import pi # Define the Reduced Planck Constant ℏ (H bar), speed of light C, value of # Pi and the function UpperCAmelCase : Union[str, Any] = 1.054571817E-34 # unit of ℏ : J * s UpperCAmelCase : Union[str, Any] = 3E8 # unit of c : m * s^-1 def _A ( SCREAMING_SNAKE_CASE : float , SCREAMING_SNAKE_CASE : float , SCREAMING_SNAKE_CASE : float ): """simple docstring""" if (force, area, distance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if force < 0: raise ValueError("Magnitude of force can not be negative" ) if distance < 0: raise ValueError("Distance can not be negative" ) if area < 0: raise ValueError("Area can not be negative" ) if force == 0: a__ : Tuple =(REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / ( 240 * (distance) ** 4 ) return {"force": force} elif area == 0: a__ : Any =(240 * force * (distance) ** 4) / ( REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 ) return {"area": area} elif distance == 0: a__ : List[str] =( (REDUCED_PLANCK_CONSTANT * SPEED_OF_LIGHT * pi**2 * area) / (240 * force) ) ** (1 / 4) return {"distance": distance} raise ValueError("One and only one argument must be 0" ) # Run doctest if __name__ == "__main__": import doctest doctest.testmod()
148
1
"""simple docstring""" from .configuration_bert_masked import MaskedBertConfig from .modeling_bert_masked import ( MaskedBertForMultipleChoice, MaskedBertForQuestionAnswering, MaskedBertForSequenceClassification, MaskedBertForTokenClassification, MaskedBertModel, ) from .modules import *
109
"""simple docstring""" import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING A: List[str] = logging.get_logger(__name__) A: Dict = { "microsoft/conditional-detr-resnet-50": ( "https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json" ), } class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase__ ): __lowerCAmelCase : Any = 'conditional_detr' __lowerCAmelCase : Union[str, Any] = ['past_key_values'] __lowerCAmelCase : int = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=300 , _SCREAMING_SNAKE_CASE=6 , _SCREAMING_SNAKE_CASE=2048 , _SCREAMING_SNAKE_CASE=8 , _SCREAMING_SNAKE_CASE=6 , _SCREAMING_SNAKE_CASE=2048 , _SCREAMING_SNAKE_CASE=8 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE="relu" , _SCREAMING_SNAKE_CASE=256 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=1.0 , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE="sine" , _SCREAMING_SNAKE_CASE="resnet50" , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=5 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=1 , _SCREAMING_SNAKE_CASE=1 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=5 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=0.25 , **_SCREAMING_SNAKE_CASE , ) -> Tuple: '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) UpperCAmelCase : Tuple = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): UpperCAmelCase : str = backbone_config.get("""model_type""" ) UpperCAmelCase : int = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase : Union[str, Any] = config_class.from_dict(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Union[str, Any] = use_timm_backbone UpperCAmelCase : Optional[int] = backbone_config UpperCAmelCase : List[str] = num_channels UpperCAmelCase : Any = num_queries UpperCAmelCase : Union[str, Any] = d_model UpperCAmelCase : List[str] = encoder_ffn_dim UpperCAmelCase : Optional[int] = encoder_layers UpperCAmelCase : Union[str, Any] = encoder_attention_heads UpperCAmelCase : Optional[Any] = decoder_ffn_dim UpperCAmelCase : Any = decoder_layers UpperCAmelCase : Optional[int] = decoder_attention_heads UpperCAmelCase : Optional[int] = dropout UpperCAmelCase : Dict = attention_dropout UpperCAmelCase : Dict = activation_dropout UpperCAmelCase : Any = activation_function UpperCAmelCase : Any = init_std UpperCAmelCase : Tuple = init_xavier_std UpperCAmelCase : Optional[int] = encoder_layerdrop UpperCAmelCase : Any = decoder_layerdrop UpperCAmelCase : Any = encoder_layers UpperCAmelCase : Optional[Any] = auxiliary_loss UpperCAmelCase : List[Any] = position_embedding_type UpperCAmelCase : Union[str, Any] = backbone UpperCAmelCase : List[Any] = use_pretrained_backbone UpperCAmelCase : Dict = dilation # Hungarian matcher UpperCAmelCase : Optional[int] = class_cost UpperCAmelCase : List[str] = bbox_cost UpperCAmelCase : List[str] = giou_cost # Loss coefficients UpperCAmelCase : List[Any] = mask_loss_coefficient UpperCAmelCase : List[str] = dice_loss_coefficient UpperCAmelCase : Optional[int] = cls_loss_coefficient UpperCAmelCase : Union[str, Any] = bbox_loss_coefficient UpperCAmelCase : Union[str, Any] = giou_loss_coefficient UpperCAmelCase : int = focal_alpha super().__init__(is_encoder_decoder=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) @property def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' return self.encoder_attention_heads @property def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' return self.d_model def SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' UpperCAmelCase : List[Any] = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: UpperCAmelCase : Union[str, Any] = self.backbone_config.to_dict() UpperCAmelCase : Dict = self.__class__.model_type return output class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase__ ): __lowerCAmelCase : Any = version.parse('1.11' ) @property def SCREAMING_SNAKE_CASE ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def SCREAMING_SNAKE_CASE ( self ) -> float: '''simple docstring''' return 1E-5 @property def SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' return 12
109
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class __lowercase (unittest.TestCase ): """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7 , lowerCAmelCase__=3 , lowerCAmelCase__=1_8 , lowerCAmelCase__=3_0 , lowerCAmelCase__=4_0_0 , lowerCAmelCase__=True , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=[0.5, 0.5, 0.5] , lowerCAmelCase__=[0.5, 0.5, 0.5] , ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = size if size is not None else {'shortest_edge': 1_8} SCREAMING_SNAKE_CASE_ : Optional[Any] = crop_size if crop_size is not None else {'height': 1_8, 'width': 1_8} SCREAMING_SNAKE_CASE_ : int = parent SCREAMING_SNAKE_CASE_ : str = batch_size SCREAMING_SNAKE_CASE_ : str = num_channels SCREAMING_SNAKE_CASE_ : List[Any] = image_size SCREAMING_SNAKE_CASE_ : str = min_resolution SCREAMING_SNAKE_CASE_ : Union[str, Any] = max_resolution SCREAMING_SNAKE_CASE_ : int = do_resize SCREAMING_SNAKE_CASE_ : List[Any] = size SCREAMING_SNAKE_CASE_ : Optional[int] = do_center_crop SCREAMING_SNAKE_CASE_ : Any = crop_size SCREAMING_SNAKE_CASE_ : List[Any] = do_normalize SCREAMING_SNAKE_CASE_ : List[str] = image_mean SCREAMING_SNAKE_CASE_ : Optional[int] = image_std def UpperCamelCase__ ( self ): """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "do_center_crop": self.do_center_crop, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class __lowercase (__SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" _UpperCAmelCase = LevitImageProcessor if is_vision_available() else None def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = LevitImageProcessingTester(self ) @property def UpperCamelCase__ ( self ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCAmelCase__ , 'image_mean' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , 'image_std' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , 'do_normalize' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , 'do_resize' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , 'do_center_crop' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , 'size' ) ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 1_8} ) self.assertEqual(image_processor.crop_size , {'height': 1_8, 'width': 1_8} ) SCREAMING_SNAKE_CASE_ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 , crop_size=8_4 ) self.assertEqual(image_processor.size , {'shortest_edge': 4_2} ) self.assertEqual(image_processor.crop_size , {'height': 8_4, 'width': 8_4} ) def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE_ : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , Image.Image ) # Test not batched input SCREAMING_SNAKE_CASE_ : str = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched SCREAMING_SNAKE_CASE_ : str = image_processing(lowerCAmelCase__ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors SCREAMING_SNAKE_CASE_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , numpify=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , np.ndarray ) # Test not batched input SCREAMING_SNAKE_CASE_ : List[Any] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched SCREAMING_SNAKE_CASE_ : Optional[int] = image_processing(lowerCAmelCase__ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE_ : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , torchify=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , torch.Tensor ) # Test not batched input SCREAMING_SNAKE_CASE_ : Dict = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched SCREAMING_SNAKE_CASE_ : List[Any] = image_processing(lowerCAmelCase__ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , )
162
import socket def a__ ( ): SCREAMING_SNAKE_CASE_ : Dict = socket.socket(socket.AF_INET, socket.SOCK_STREAM ) SCREAMING_SNAKE_CASE_ : Any = socket.gethostname() SCREAMING_SNAKE_CASE_ : List[str] = 1_2_3_1_2 sock.connect((host, port) ) sock.send(B'Hello server!' ) with open('Received_file', 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: SCREAMING_SNAKE_CASE_ : Tuple = sock.recv(1_0_2_4 ) if not data: break out_file.write(A__ ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
162
1
"""simple docstring""" def __A ( a_ :Any) -> Tuple: return [ txt[:a] + txt[a].upper() + txt[a + 1 :] for a in range(len(a__)) if txt[a].isalpha() ] if __name__ == "__main__": __import__('''doctest''').testmod()
160
import argparse import os import sys from unittest.mock import patch import pytorch_lightning as pl import timeout_decorator import torch from distillation import SummarizationDistiller, distill_main from finetune import SummarizationModule, main from transformers import MarianMTModel from transformers.file_utils import cached_path from transformers.testing_utils import TestCasePlus, require_torch_gpu, slow from utils import load_json __snake_case : Optional[int] = """sshleifer/mar_enro_6_3_student""" class A__(a_ ): """simple docstring""" def UpperCamelCase__ ( self ) -> Tuple: super().setUp() a_ : Union[str, Any] = cached_path( """https://cdn-datasets.huggingface.co/translation/wmt_en_ro-tr40k-va0.5k-te0.5k.tar.gz""" , extract_compressed_file=_lowercase , ) a_ : Union[str, Any] = F'''{data_cached}/wmt_en_ro-tr40k-va0.5k-te0.5k''' @slow @require_torch_gpu def UpperCamelCase__ ( self ) -> Tuple: MarianMTModel.from_pretrained(_lowercase ) @slow @require_torch_gpu def UpperCamelCase__ ( self ) -> int: a_ : Any = { """$MAX_LEN""": 64, """$BS""": 64, """$GAS""": 1, """$ENRO_DIR""": self.data_dir, """facebook/mbart-large-cc25""": MARIAN_MODEL, # "val_check_interval=0.25": "val_check_interval=1.0", """--learning_rate=3e-5""": """--learning_rate 3e-4""", """--num_train_epochs 6""": """--num_train_epochs 1""", } # Clean up bash script a_ : List[str] = (self.test_file_dir / """train_mbart_cc25_enro.sh""").open().read().split("""finetune.py""" )[1].strip() a_ : Dict = bash_script.replace("""\\\n""" , """""" ).strip().replace("""\"$@\"""" , """""" ) for k, v in env_vars_to_replace.items(): a_ : Optional[int] = bash_script.replace(_lowercase , str(_lowercase ) ) a_ : int = self.get_auto_remove_tmp_dir() # bash_script = bash_script.replace("--fp16 ", "") a_ : Dict = F''' --output_dir {output_dir} --tokenizer_name Helsinki-NLP/opus-mt-en-ro --sortish_sampler --do_predict --gpus 1 --freeze_encoder --n_train 40000 --n_val 500 --n_test 500 --fp16_opt_level O1 --num_sanity_val_steps 0 --eval_beams 2 '''.split() # XXX: args.gpus > 1 : handle multi_gpu in the future a_ : Union[str, Any] = ["""finetune.py"""] + bash_script.split() + args with patch.object(_lowercase , """argv""" , _lowercase ): a_ : Optional[Any] = argparse.ArgumentParser() a_ : Tuple = pl.Trainer.add_argparse_args(_lowercase ) a_ : Any = SummarizationModule.add_model_specific_args(_lowercase , os.getcwd() ) a_ : str = parser.parse_args() a_ : Union[str, Any] = main(_lowercase ) # Check metrics a_ : Any = load_json(model.metrics_save_path ) a_ : List[Any] = metrics["""val"""][0] a_ : Union[str, Any] = metrics["""val"""][-1] self.assertEqual(len(metrics["""val"""] ) , (args.max_epochs / args.val_check_interval) ) assert isinstance(last_step_stats[F'''val_avg_{model.val_metric}'''] , _lowercase ) self.assertGreater(last_step_stats["""val_avg_gen_time"""] , 0.0_1 ) # model hanging on generate. Maybe bad config was saved. (XXX: old comment/assert?) self.assertLessEqual(last_step_stats["""val_avg_gen_time"""] , 1.0 ) # test learning requirements: # 1. BLEU improves over the course of training by more than 2 pts self.assertGreater(last_step_stats["""val_avg_bleu"""] - first_step_stats["""val_avg_bleu"""] , 2 ) # 2. BLEU finishes above 17 self.assertGreater(last_step_stats["""val_avg_bleu"""] , 17 ) # 3. test BLEU and val BLEU within ~1.1 pt. self.assertLess(abs(metrics["""val"""][-1]["""val_avg_bleu"""] - metrics["""test"""][-1]["""test_avg_bleu"""] ) , 1.1 ) # check lightning ckpt can be loaded and has a reasonable statedict a_ : Optional[Any] = os.listdir(_lowercase ) a_ : Dict = [x for x in contents if x.endswith(""".ckpt""" )][0] a_ : str = os.path.join(args.output_dir , _lowercase ) a_ : Any = torch.load(_lowercase , map_location="""cpu""" ) a_ : Union[str, Any] = """model.model.decoder.layers.0.encoder_attn_layer_norm.weight""" assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: a_ : List[Any] = {os.path.basename(_lowercase ) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics["""test"""] ) == 1 class A__(a_ ): """simple docstring""" @timeout_decorator.timeout(600 ) @slow @require_torch_gpu def UpperCamelCase__ ( self ) -> Optional[Any]: a_ : Tuple = F'''{self.test_file_dir_str}/test_data/wmt_en_ro''' a_ : str = { """--fp16_opt_level=O1""": """""", """$MAX_LEN""": 128, """$BS""": 16, """$GAS""": 1, """$ENRO_DIR""": data_dir, """$m""": """sshleifer/student_marian_en_ro_6_1""", """val_check_interval=0.25""": """val_check_interval=1.0""", } # Clean up bash script a_ : Union[str, Any] = ( (self.test_file_dir / """distil_marian_no_teacher.sh""").open().read().split("""distillation.py""" )[1].strip() ) a_ : Union[str, Any] = bash_script.replace("""\\\n""" , """""" ).strip().replace("""\"$@\"""" , """""" ) a_ : Any = bash_script.replace("""--fp16 """ , """ """ ) for k, v in env_vars_to_replace.items(): a_ : Dict = bash_script.replace(_lowercase , str(_lowercase ) ) a_ : int = self.get_auto_remove_tmp_dir() a_ : Optional[Any] = bash_script.replace("""--fp16""" , """""" ) a_ : List[str] = 6 a_ : str = ( ["""distillation.py"""] + bash_script.split() + [ F'''--output_dir={output_dir}''', """--gpus=1""", """--learning_rate=1e-3""", F'''--num_train_epochs={epochs}''', """--warmup_steps=10""", """--val_check_interval=1.0""", """--do_predict""", ] ) with patch.object(_lowercase , """argv""" , _lowercase ): a_ : int = argparse.ArgumentParser() a_ : Any = pl.Trainer.add_argparse_args(_lowercase ) a_ : str = SummarizationDistiller.add_model_specific_args(_lowercase , os.getcwd() ) a_ : Any = parser.parse_args() # assert args.gpus == gpus THIS BREAKS for multi_gpu a_ : Dict = distill_main(_lowercase ) # Check metrics a_ : Any = load_json(model.metrics_save_path ) a_ : int = metrics["""val"""][0] a_ : Union[str, Any] = metrics["""val"""][-1] assert len(metrics["""val"""] ) >= (args.max_epochs / args.val_check_interval) # +1 accounts for val_sanity_check assert last_step_stats["val_avg_gen_time"] >= 0.0_1 assert first_step_stats["val_avg_bleu"] < last_step_stats["val_avg_bleu"] # model learned nothing assert 1.0 >= last_step_stats["val_avg_gen_time"] # model hanging on generate. Maybe bad config was saved. assert isinstance(last_step_stats[F'''val_avg_{model.val_metric}'''] , _lowercase ) # check lightning ckpt can be loaded and has a reasonable statedict a_ : Dict = os.listdir(_lowercase ) a_ : List[Any] = [x for x in contents if x.endswith(""".ckpt""" )][0] a_ : int = os.path.join(args.output_dir , _lowercase ) a_ : Union[str, Any] = torch.load(_lowercase , map_location="""cpu""" ) a_ : List[str] = """model.model.decoder.layers.0.encoder_attn_layer_norm.weight""" assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: a_ : List[str] = {os.path.basename(_lowercase ) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics["""test"""] ) == 1
248
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class _lowerCamelCase ( unittest.TestCase ): @slow def snake_case_ (self ) -> Optional[Any]: UpperCamelCase = TFCamembertModel.from_pretrained("jplu/tf-camembert-base" ) UpperCamelCase = tf.convert_to_tensor( [[5, 1_21, 11, 6_60, 16, 7_30, 2_55_43, 1_10, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" UpperCamelCase = model(__a )["last_hidden_state"] UpperCamelCase = tf.TensorShape((1, 10, 7_68) ) self.assertEqual(output.shape , __a ) # compare the actual values for a slice. UpperCamelCase = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1e-4 ) )
351
"""simple docstring""" import math import random from typing import Any from .hill_climbing import SearchProblem def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE = math.inf , _SCREAMING_SNAKE_CASE = -math.inf , _SCREAMING_SNAKE_CASE = math.inf , _SCREAMING_SNAKE_CASE = -math.inf , _SCREAMING_SNAKE_CASE = False , _SCREAMING_SNAKE_CASE = 100 , _SCREAMING_SNAKE_CASE = 0.01 , _SCREAMING_SNAKE_CASE = 1 , ): """simple docstring""" UpperCamelCase = False UpperCamelCase = search_prob UpperCamelCase = start_temperate UpperCamelCase = [] UpperCamelCase = 0 UpperCamelCase = None while not search_end: UpperCamelCase = current_state.score() if best_state is None or current_score > best_state.score(): UpperCamelCase = current_state scores.append(_SCREAMING_SNAKE_CASE ) iterations += 1 UpperCamelCase = None UpperCamelCase = current_state.get_neighbors() while ( next_state is None and neighbors ): # till we do not find a neighbor that we can move to UpperCamelCase = random.randint(0 , len(_SCREAMING_SNAKE_CASE ) - 1 ) # picking a random neighbor UpperCamelCase = neighbors.pop(_SCREAMING_SNAKE_CASE ) UpperCamelCase = picked_neighbor.score() - current_score if ( picked_neighbor.x > max_x or picked_neighbor.x < min_x or picked_neighbor.y > max_y or picked_neighbor.y < min_y ): continue # neighbor outside our bounds if not find_max: UpperCamelCase = change * -1 # in case we are finding minimum if change > 0: # improves the solution UpperCamelCase = picked_neighbor else: UpperCamelCase = (math.e) ** ( change / current_temp ) # probability generation function if random.random() < probability: # random number within probability UpperCamelCase = picked_neighbor UpperCamelCase = current_temp - (current_temp * rate_of_decrease) if current_temp < threshold_temp or next_state is None: # temperature below threshold, or could not find a suitable neighbor UpperCamelCase = True else: UpperCamelCase = next_state if visualization: from matplotlib import pyplot as plt plt.plot(range(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) plt.xlabel("Iterations" ) plt.ylabel("Function values" ) plt.show() return best_state if __name__ == "__main__": def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" return (x**2) + (y**2) # starting the problem with initial coordinates (12, 47) lowerCAmelCase__ = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing( prob, find_max=False, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( '''The minimum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' f'''and 50 > y > - 5 found via hill climbing: {local_min.score()}''' ) # starting the problem with initial coordinates (12, 47) lowerCAmelCase__ = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing( prob, find_max=True, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( '''The maximum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' f'''and 50 > y > - 5 found via hill climbing: {local_min.score()}''' ) def a__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): """simple docstring""" return (3 * x**2) - (6 * y) lowerCAmelCase__ = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing(prob, find_max=False, visualization=True) print( '''The minimum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' f'''{local_min.score()}''' ) lowerCAmelCase__ = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing(prob, find_max=True, visualization=True) print( '''The maximum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' f'''{local_min.score()}''' )
244
0
from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def a_ ( __lowercase : Tuple = "isbn/0140328726" ) -> dict: _snake_case = olid.strip().strip('/' ) # Remove leading/trailing whitespace & slashes if new_olid.count('/' ) != 1: _snake_case = f'''{olid} is not a valid Open Library olid''' raise ValueError(__a ) return requests.get(f'''https://openlibrary.org/{new_olid}.json''' ).json() def a_ ( __lowercase : Any ) -> dict: _snake_case = { '''title''': '''Title''', '''publish_date''': '''Publish date''', '''authors''': '''Authors''', '''number_of_pages''': '''Number of pages:''', '''first_sentence''': '''First sentence''', '''isbn_10''': '''ISBN (10)''', '''isbn_13''': '''ISBN (13)''', } _snake_case = {better_key: ol_book_data[key] for key, better_key in desired_keys.items()} _snake_case = [ get_openlibrary_data(author['key'] )['''name'''] for author in data['''Authors'''] ] _snake_case = data['''First sentence''']['''value'''] for key, value in data.items(): if isinstance(__a , __a ): _snake_case = ''', '''.join(__a ) return data if __name__ == "__main__": import doctest doctest.testmod() while True: _lowerCamelCase : Optional[Any] = input('''\nEnter the ISBN code to search (or \'quit\' to stop): ''').strip() if isbn.lower() in ("", "q", "quit", "exit", "stop"): break if len(isbn) not in (10, 13) or not isbn.isdigit(): print(F'Sorry, {isbn} is not a valid ISBN. Please, input a valid ISBN.') continue print(F'\nSearching Open Library for ISBN: {isbn}...\n') try: _lowerCamelCase : List[str] = summarize_book(get_openlibrary_data(F'isbn/{isbn}')) print('''\n'''.join(F'{key}: {value}' for key, value in book_summary.items())) except JSONDecodeError: # Workaround for requests.exceptions.RequestException: print(F'Sorry, there are no results for ISBN: {isbn}.')
282
"""simple docstring""" def _A (__a = 50 ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(f'''{solution() = }''')
91
0
from __future__ import annotations from random import random class __magic_name__ : """simple docstring""" def __init__( self :Any , snake_case :int | None = None ): '''simple docstring''' A_ : Tuple = value A_ : Any = random() A_ : Node | None = None A_ : Node | None = None def __repr__( self :Optional[int] ): '''simple docstring''' from pprint import pformat if self.left is None and self.right is None: return f"'{self.value}: {self.prior:.5}'" else: return pformat( {f"{self.value}: {self.prior:.5}": (self.left, self.right)} , indent=1 ) def __str__( self :Any ): '''simple docstring''' A_ : Tuple = str(self.value ) + " " A_ : Optional[int] = str(self.left or "" ) A_ : Union[str, Any] = str(self.right or "" ) return value + left + right def __snake_case ( _lowerCAmelCase : Node | None , _lowerCAmelCase : int ) -> tuple[Node | None, Node | None]: if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: A_ : List[Any] = split(root.left , _lowerCAmelCase ) return left, root else: A_ : Union[str, Any] = split(root.right , _lowerCAmelCase ) return root, right def __snake_case ( _lowerCAmelCase : Node | None , _lowerCAmelCase : Node | None ) -> Node | None: if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: A_ : Optional[Any] = merge(left.right , _lowerCAmelCase ) return left else: A_ : int = merge(_lowerCAmelCase , right.left ) return right def __snake_case ( _lowerCAmelCase : Node | None , _lowerCAmelCase : int ) -> Node | None: A_ : Union[str, Any] = Node(_lowerCAmelCase ) A_ : str = split(_lowerCAmelCase , _lowerCAmelCase ) return merge(merge(_lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase ) def __snake_case ( _lowerCAmelCase : Node | None , _lowerCAmelCase : int ) -> Node | None: A_ : Tuple = split(_lowerCAmelCase , value - 1 ) A_ : Optional[int] = split(_lowerCAmelCase , _lowerCAmelCase ) return merge(_lowerCAmelCase , _lowerCAmelCase ) def __snake_case ( _lowerCAmelCase : Node | None ) -> None: if not root: # None return else: inorder(root.left ) print(root.value , end="," ) inorder(root.right ) def __snake_case ( _lowerCAmelCase : Node | None , _lowerCAmelCase : str ) -> Node | None: for arg in args.split(): if arg[0] == "+": A_ : str = insert(_lowerCAmelCase , int(arg[1:] ) ) elif arg[0] == "-": A_ : Tuple = erase(_lowerCAmelCase , int(arg[1:] ) ) else: print("Unknown command" ) return root def __snake_case ( ) -> None: A_ : str = None print( "enter numbers to create a tree, + value to add value into treap, " "- value to erase all nodes with value. 'q' to quit. " ) A_ : Any = input() while args != "q": A_ : Tuple = interact_treap(_lowerCAmelCase , _lowerCAmelCase ) print(_lowerCAmelCase ) A_ : Optional[Any] = input() print("good by!" ) if __name__ == "__main__": import doctest doctest.testmod() main()
365
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCAmelCase : List[Any] = { '''configuration_funnel''': ['''FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''FunnelConfig'''], '''convert_funnel_original_tf_checkpoint_to_pytorch''': [], '''tokenization_funnel''': ['''FunnelTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Any = ['''FunnelTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Tuple = [ '''FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FunnelBaseModel''', '''FunnelForMaskedLM''', '''FunnelForMultipleChoice''', '''FunnelForPreTraining''', '''FunnelForQuestionAnswering''', '''FunnelForSequenceClassification''', '''FunnelForTokenClassification''', '''FunnelModel''', '''FunnelPreTrainedModel''', '''load_tf_weights_in_funnel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : int = [ '''TF_FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFFunnelBaseModel''', '''TFFunnelForMaskedLM''', '''TFFunnelForMultipleChoice''', '''TFFunnelForPreTraining''', '''TFFunnelForQuestionAnswering''', '''TFFunnelForSequenceClassification''', '''TFFunnelForTokenClassification''', '''TFFunnelModel''', '''TFFunnelPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_funnel import FUNNEL_PRETRAINED_CONFIG_ARCHIVE_MAP, FunnelConfig from .tokenization_funnel import FunnelTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_funnel_fast import FunnelTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_funnel import ( FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST, FunnelBaseModel, FunnelForMaskedLM, FunnelForMultipleChoice, FunnelForPreTraining, FunnelForQuestionAnswering, FunnelForSequenceClassification, FunnelForTokenClassification, FunnelModel, FunnelPreTrainedModel, load_tf_weights_in_funnel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_funnel import ( TF_FUNNEL_PRETRAINED_MODEL_ARCHIVE_LIST, TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, TFFunnelPreTrainedModel, ) else: import sys _lowerCAmelCase : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
70
0
"""simple docstring""" import numpy as np import datasets _a = '\nCompute the Mahalanobis Distance\n\nMahalonobis distance is the distance between a point and a distribution.\nAnd not between two distinct points. It is effectively a multivariate equivalent of the Euclidean distance.\nIt was introduced by Prof. P. C. Mahalanobis in 1936\nand has been used in various statistical applications ever since\n[source: https://www.machinelearningplus.com/statistics/mahalanobis-distance/]\n' _a = '\\n@article{de2000mahalanobis,\n title={The mahalanobis distance},\n author={De Maesschalck, Roy and Jouan-Rimbaud, Delphine and Massart, D{\'e}sir{\'e} L},\n journal={Chemometrics and intelligent laboratory systems},\n volume={50},\n number={1},\n pages={1--18},\n year={2000},\n publisher={Elsevier}\n}\n' _a = '\nArgs:\n X: List of datapoints to be compared with the `reference_distribution`.\n reference_distribution: List of datapoints from the reference distribution we want to compare to.\nReturns:\n mahalanobis: The Mahalonobis distance for each datapoint in `X`.\nExamples:\n\n >>> mahalanobis_metric = datasets.load_metric("mahalanobis")\n >>> results = mahalanobis_metric.compute(reference_distribution=[[0, 1], [1, 0]], X=[[0, 1]])\n >>> print(results)\n {\'mahalanobis\': array([0.5])}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): """simple docstring""" def _lowercase ( self : Tuple ): return datasets.MetricInfo( description=_DESCRIPTION, citation=_CITATION, inputs_description=_KWARGS_DESCRIPTION, features=datasets.Features( { "X": datasets.Sequence(datasets.Value("float", id="sequence" ), id="X" ), } ), ) def _lowercase ( self : int, UpperCAmelCase__ : List[str], UpperCAmelCase__ : List[Any] ): # convert to numpy arrays __lowercase = np.array(snake_case_ ) __lowercase = np.array(snake_case_ ) # Assert that arrays are 2D if len(X.shape ) != 2: raise ValueError("Expected `X` to be a 2D vector" ) if len(reference_distribution.shape ) != 2: raise ValueError("Expected `reference_distribution` to be a 2D vector" ) if reference_distribution.shape[0] < 2: raise ValueError( "Expected `reference_distribution` to be a 2D vector with more than one element in the first dimension" ) # Get mahalanobis distance for each prediction __lowercase = X - np.mean(snake_case_ ) __lowercase = np.cov(reference_distribution.T ) try: __lowercase = np.linalg.inv(snake_case_ ) except np.linalg.LinAlgError: __lowercase = np.linalg.pinv(snake_case_ ) __lowercase = np.dot(snake_case_, snake_case_ ) __lowercase = np.dot(snake_case_, X_minus_mu.T ).diagonal() return {"mahalanobis": mahal_dist}
17
from __future__ import annotations def __UpperCamelCase ( lowerCAmelCase__ : list[float] , lowerCAmelCase__ : list[float] ): __a : Dict = sorted(numsa + numsa ) __a , __a : Optional[Any] = divmod(len(lowerCAmelCase__ ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() lowercase__ =[float(x) for x in input('Enter the elements of first array: ').split()] lowercase__ =[float(x) for x in input('Enter the elements of second array: ').split()] print(F"""The median of two arrays is: {median_of_two_arrays(array_a, array_a)}""")
216
0
import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 _UpperCAmelCase = { 'return_dict': False, 'output_hidden_states': True, 'output_attentions': True, 'torchscript': True, 'torch_dtype': 'float16', 'use_bfloat16': True, 'tf_legacy_loss': True, 'pruned_heads': {'a': 1}, 'tie_word_embeddings': False, 'is_decoder': True, 'cross_attention_hidden_size': 128, 'add_cross_attention': True, 'tie_encoder_decoder': True, 'max_length': 50, 'min_length': 3, 'do_sample': True, 'early_stopping': True, 'num_beams': 3, 'num_beam_groups': 3, 'diversity_penalty': 0.5, 'temperature': 2.0, 'top_k': 10, 'top_p': 0.7, 'typical_p': 0.2, 'repetition_penalty': 0.8, 'length_penalty': 0.8, 'no_repeat_ngram_size': 5, 'encoder_no_repeat_ngram_size': 5, 'bad_words_ids': [1, 2, 3], 'num_return_sequences': 3, 'chunk_size_feed_forward': 5, 'output_scores': True, 'return_dict_in_generate': True, 'forced_bos_token_id': 2, 'forced_eos_token_id': 3, 'remove_invalid_values': True, 'architectures': ['BertModel'], 'finetuning_task': 'translation', 'id2label': {0: 'label'}, 'label2id': {'label': '0'}, 'tokenizer_class': 'BertTokenizerFast', 'prefix': 'prefix', 'bos_token_id': 6, 'pad_token_id': 7, 'eos_token_id': 8, 'sep_token_id': 9, 'decoder_start_token_id': 10, 'exponential_decay_length_penalty': (5, 1.01), 'suppress_tokens': [0, 1], 'begin_suppress_tokens': 2, 'task_specific_params': {'translation': 'some_params'}, 'problem_type': 'regression', } @is_staging_test class snake_case_ ( unittest.TestCase ): @classmethod def UpperCAmelCase__ ( cls : Optional[int] )->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Any = TOKEN HfFolder.save_token(_snake_case ) @classmethod def UpperCAmelCase__ ( cls : int )->Dict: '''simple docstring''' try: delete_repo(token=cls._token , repo_id="""test-config""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-config-org""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""test-dynamic-config""" ) except HTTPError: pass def UpperCAmelCase__ ( self : Optional[Any] )->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Any = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("""test-config""" , use_auth_token=self._token ) __lowerCAmelCase : int = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(_snake_case , getattr(_snake_case , _snake_case ) ) # Reset repo delete_repo(token=self._token , repo_id="""test-config""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(_snake_case , repo_id="""test-config""" , push_to_hub=_snake_case , use_auth_token=self._token ) __lowerCAmelCase : Optional[int] = BertConfig.from_pretrained(F'''{USER}/test-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(_snake_case , getattr(_snake_case , _snake_case ) ) def UpperCAmelCase__ ( self : Optional[int] )->List[str]: '''simple docstring''' __lowerCAmelCase : List[str] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("""valid_org/test-config-org""" , use_auth_token=self._token ) __lowerCAmelCase : List[Any] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(_snake_case , getattr(_snake_case , _snake_case ) ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-config-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( _snake_case , repo_id="""valid_org/test-config-org""" , push_to_hub=_snake_case , use_auth_token=self._token ) __lowerCAmelCase : Optional[Any] = BertConfig.from_pretrained("""valid_org/test-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(_snake_case , getattr(_snake_case , _snake_case ) ) def UpperCAmelCase__ ( self : Dict )->Dict: '''simple docstring''' CustomConfig.register_for_auto_class() __lowerCAmelCase : Dict = CustomConfig(attribute=42 ) config.push_to_hub("""test-dynamic-config""" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"""AutoConfig""": """custom_configuration.CustomConfig"""} ) __lowerCAmelCase : str = AutoConfig.from_pretrained(F'''{USER}/test-dynamic-config''' , trust_remote_code=_snake_case ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , """CustomConfig""" ) self.assertEqual(new_config.attribute , 42 ) class snake_case_ ( unittest.TestCase ): def UpperCAmelCase__ ( self : List[Any] )->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Any = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated __lowerCAmelCase : List[str] = c.n_embd + 1 # int __lowerCAmelCase : Union[str, Any] = c.resid_pdrop + 1.0 # float __lowerCAmelCase : Dict = not c.scale_attn_weights # bool __lowerCAmelCase : Optional[Any] = c.summary_type + """foo""" # str c.update_from_string( F'''n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}''' ) self.assertEqual(_snake_case , c.n_embd , """mismatch for key: n_embd""" ) self.assertEqual(_snake_case , c.resid_pdrop , """mismatch for key: resid_pdrop""" ) self.assertEqual(_snake_case , c.scale_attn_weights , """mismatch for key: scale_attn_weights""" ) self.assertEqual(_snake_case , c.summary_type , """mismatch for key: summary_type""" ) def UpperCAmelCase__ ( self : List[Any] )->Optional[Any]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = PretrainedConfig() __lowerCAmelCase : List[Any] = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( _snake_case , ["""is_encoder_decoder""", """_name_or_path""", """_commit_hash""", """transformers_version"""] ) __lowerCAmelCase : Tuple = [key for key, value in config_common_kwargs.items() if value == getattr(_snake_case , _snake_case )] if len(_snake_case ) > 0: raise ValueError( """The following keys are set with the default values in""" """ `test_configuration_common.config_common_kwargs` pick another value for them:""" F''' {", ".join(_snake_case )}.''' ) def UpperCAmelCase__ ( self : Optional[Any] )->List[Any]: '''simple docstring''' with self.assertRaises(_snake_case ): # config is in subfolder, the following should not work without specifying the subfolder __lowerCAmelCase : Optional[int] = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" ) __lowerCAmelCase : int = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert-subfolder""" , subfolder="""bert""" ) self.assertIsNotNone(_snake_case ) def UpperCAmelCase__ ( self : Optional[int] )->Optional[int]: '''simple docstring''' __lowerCAmelCase : int = mock.Mock() __lowerCAmelCase : List[Any] = 500 __lowerCAmelCase : List[str] = {} __lowerCAmelCase : Tuple = HTTPError __lowerCAmelCase : Optional[int] = {} # Download this model to make sure it's in the cache. __lowerCAmelCase : int = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("""requests.Session.request""" , return_value=_snake_case ) as mock_head: __lowerCAmelCase : Optional[Any] = BertConfig.from_pretrained("""hf-internal-testing/tiny-random-bert""" ) # This check we did call the fake head request mock_head.assert_called() def UpperCAmelCase__ ( self : str )->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : List[str] = BertConfig.from_pretrained( """https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json""" ) def UpperCAmelCase__ ( self : List[Any] )->Dict: '''simple docstring''' __lowerCAmelCase : Union[str, Any] = AutoConfig.from_pretrained("""bert-base-cased""" ) __lowerCAmelCase : str = ["""config.4.0.0.json"""] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(_snake_case ) __lowerCAmelCase : Optional[Any] = 2 json.dump(configuration.to_dict() , open(os.path.join(_snake_case , """config.4.0.0.json""" ) , """w""" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 __lowerCAmelCase : Union[str, Any] = AutoConfig.from_pretrained(_snake_case ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 __lowerCAmelCase : str = ["""config.42.0.0.json"""] __lowerCAmelCase : Tuple = 768 configuration.save_pretrained(_snake_case ) shutil.move(os.path.join(_snake_case , """config.4.0.0.json""" ) , os.path.join(_snake_case , """config.42.0.0.json""" ) ) __lowerCAmelCase : Any = AutoConfig.from_pretrained(_snake_case ) self.assertEqual(new_configuration.hidden_size , 768 ) def UpperCAmelCase__ ( self : Any )->Any: '''simple docstring''' __lowerCAmelCase : Optional[int] = """hf-internal-testing/test-two-configs""" import transformers as new_transformers __lowerCAmelCase : Optional[int] = """v4.0.0""" __lowerCAmelCase , __lowerCAmelCase : Dict = new_transformers.models.auto.AutoConfig.from_pretrained( _snake_case , return_unused_kwargs=_snake_case ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(_snake_case , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers __lowerCAmelCase : Union[str, Any] = """v3.0.0""" __lowerCAmelCase : Union[str, Any] = old_transformers.models.auto.AutoConfig.from_pretrained(_snake_case ) self.assertEqual(old_configuration.hidden_size , 768 )
232
def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :list[int] ) -> int: if not numbers: return 0 if not isinstance(SCREAMING_SNAKE_CASE , (list, tuple) ) or not all( isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for number in numbers ): raise ValueError("""numbers must be an iterable of integers""" ) __lowerCAmelCase : int = numbers[0] for i in range(1 , len(SCREAMING_SNAKE_CASE ) ): # update the maximum and minimum subarray products __lowerCAmelCase : List[str] = numbers[i] if number < 0: __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = min_till_now, max_till_now __lowerCAmelCase : Optional[int] = max(SCREAMING_SNAKE_CASE , max_till_now * number ) __lowerCAmelCase : List[Any] = min(SCREAMING_SNAKE_CASE , min_till_now * number ) # update the maximum product found till now __lowerCAmelCase : List[str] = max(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return max_prod
232
1
"""simple docstring""" import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder _A = """__DUMMY_TRANSFORMERS_USER__""" _A = """Dummy User""" _A = """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""" _A = """https://hub-ci.huggingface.co""" _A = CI_HUB_ENDPOINT + """/datasets/{repo_id}/resolve/{revision}/{path}""" _A = CI_HUB_ENDPOINT + """/{repo_id}/resolve/{revision}/{filename}""" _A = Path("""~/.huggingface/hub_ci_token""").expanduser() @pytest.fixture def a__ ( lowerCAmelCase ) -> List[str]: monkeypatch.setattr( """huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE""" , __a ) @pytest.fixture def a__ ( lowerCAmelCase ) -> Any: monkeypatch.setattr("""datasets.config.HF_ENDPOINT""" , __a ) monkeypatch.setattr("""datasets.config.HUB_DATASETS_URL""" , __a ) @pytest.fixture def a__ ( lowerCAmelCase ) -> Any: monkeypatch.setattr("""huggingface_hub.hf_api.HfFolder.path_token""" , __a ) @pytest.fixture def a__ ( lowerCAmelCase , lowerCAmelCase ) -> Dict: HfFolder.save_token(__a ) yield HfFolder.delete_token() @pytest.fixture(scope="""session""" ) def a__ ( ) -> int: return HfApi(endpoint=__a ) @pytest.fixture(scope="""session""" ) def a__ ( lowerCAmelCase ) -> Any: UpperCAmelCase__ : List[str] = HfFolder.get_token() HfFolder.save_token(__a ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(__a ) @pytest.fixture def a__ ( lowerCAmelCase ) -> Dict: def _cleanup_repo(lowerCAmelCase ): hf_api.delete_repo(__a , token=__a , repo_type="""dataset""" ) return _cleanup_repo @pytest.fixture def a__ ( lowerCAmelCase ) -> Optional[int]: @contextmanager def _temporary_repo(lowerCAmelCase ): try: yield repo_id finally: cleanup_repo(__a ) return _temporary_repo @pytest.fixture(scope="""session""" ) def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Dict: UpperCAmelCase__ : List[str] = F"""repo_txt_data-{int(time.time() * 10E3 )}""" UpperCAmelCase__ : Any = F"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(__a , token=__a , repo_type="""dataset""" , private=__a ) hf_api.upload_file( token=__a , path_or_fileobj=str(__a ) , path_in_repo="""data/text_data.txt""" , repo_id=__a , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(__a , token=__a , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Optional[Any]: return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope="""session""" ) def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Any: UpperCAmelCase__ : Union[str, Any] = F"""repo_zipped_txt_data-{int(time.time() * 10E3 )}""" UpperCAmelCase__ : Optional[int] = F"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(__a , token=__a , repo_type="""dataset""" , private=__a ) hf_api.upload_file( token=__a , path_or_fileobj=str(__a ) , path_in_repo="""data.zip""" , repo_id=__a , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(__a , token=__a , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> int: return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope="""session""" ) def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Union[str, Any]: UpperCAmelCase__ : str = F"""repo_zipped_img_data-{int(time.time() * 10E3 )}""" UpperCAmelCase__ : Union[str, Any] = F"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(__a , token=__a , repo_type="""dataset""" , private=__a ) hf_api.upload_file( token=__a , path_or_fileobj=str(__a ) , path_in_repo="""data.zip""" , repo_id=__a , repo_type="""dataset""" , ) yield repo_id try: hf_api.delete_repo(__a , token=__a , repo_type="""dataset""" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Tuple: return hf_private_dataset_repo_zipped_img_data_
171
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class UpperCAmelCase_ ( __lowercase ): """simple docstring""" UpperCAmelCase__ : str = ["image_processor", "tokenizer"] UpperCAmelCase__ : Optional[int] = "CLIPImageProcessor" UpperCAmelCase__ : Tuple = ("XLMRobertaTokenizer", "XLMRobertaTokenizerFast") def __init__( self , _a=None , _a=None , **_a ) -> Any: _a : int = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _a , ) _a : int = kwargs.pop('''feature_extractor''' ) _a : int = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_a , _a ) def __call__( self , _a=None , _a=None , _a=None , **_a ) -> List[Any]: if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: _a : str = self.tokenizer(_a , return_tensors=_a , **_a ) if images is not None: _a : Optional[Any] = self.image_processor(_a , return_tensors=_a , **_a ) if text is not None and images is not None: _a : List[Any] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_a ) , tensor_type=_a ) def __lowercase ( self , *_a , **_a ) -> Optional[Any]: return self.tokenizer.batch_decode(*_a , **_a ) def __lowercase ( self , *_a , **_a ) -> Optional[int]: return self.tokenizer.decode(*_a , **_a ) @property def __lowercase ( self ) -> Optional[int]: _a : Dict = self.tokenizer.model_input_names _a : Dict = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
235
0
"""simple docstring""" import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class lowercase( __a ): '''simple docstring''' lowercase__ = ["image_processor", "tokenizer"] lowercase__ = "AutoImageProcessor" lowercase__ = "AutoTokenizer" def __init__( self: Dict, a_: Union[str, Any]=None, a_: Union[str, Any]=None, **a_: str ): '''simple docstring''' _snake_case : Any = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""", a_, ) _snake_case : List[str] = kwargs.pop("""feature_extractor""" ) _snake_case : List[Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(a_, a_ ) _snake_case : Dict = self.image_processor _snake_case : Dict = False def __call__( self: Dict, *a_: Tuple, **a_: Optional[int] ): '''simple docstring''' if self._in_target_context_manager: return self.current_processor(*a_, **a_ ) _snake_case : List[Any] = kwargs.pop("""images""", a_ ) _snake_case : Optional[Any] = kwargs.pop("""text""", a_ ) if len(a_ ) > 0: _snake_case : Tuple = args[0] _snake_case : Optional[int] = args[1:] if images is None and text is None: raise ValueError("""You need to specify either an `images` or `text` input to process.""" ) if images is not None: _snake_case : List[str] = self.image_processor(a_, *a_, **a_ ) if text is not None: _snake_case : str = self.tokenizer(a_, **a_ ) if text is None: return inputs elif images is None: return encodings else: _snake_case : Optional[Any] = encodings["""input_ids"""] return inputs def UpperCamelCase_ ( self: List[str], *a_: str, **a_: Tuple ): '''simple docstring''' return self.tokenizer.batch_decode(*a_, **a_ ) def UpperCamelCase_ ( self: str, *a_: Dict, **a_: str ): '''simple docstring''' return self.tokenizer.decode(*a_, **a_ ) @contextmanager def UpperCamelCase_ ( self: Dict ): '''simple docstring''' warnings.warn( """`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your """ """labels by using the argument `text` of the regular `__call__` method (either in the same call as """ """your images inputs, or in a separate call.""" ) _snake_case : List[Any] = True _snake_case : Tuple = self.tokenizer yield _snake_case : List[str] = self.image_processor _snake_case : int = False def UpperCamelCase_ ( self: Tuple, a_: Optional[Any], a_: int=False, a_: List[Any]=None ): '''simple docstring''' if added_vocab is None: _snake_case : List[str] = self.tokenizer.get_added_vocab() _snake_case : List[Any] = {} while tokens: _snake_case : List[Any] = re.search(r"""<s_(.*?)>""", a_, re.IGNORECASE ) if start_token is None: break _snake_case : str = start_token.group(1 ) _snake_case : int = re.search(rf"</s_{key}>", a_, re.IGNORECASE ) _snake_case : Optional[Any] = start_token.group() if end_token is None: _snake_case : Dict = tokens.replace(a_, """""" ) else: _snake_case : Tuple = end_token.group() _snake_case : List[str] = re.escape(a_ ) _snake_case : List[Any] = re.escape(a_ ) _snake_case : Tuple = re.search(f"{start_token_escaped}(.*?){end_token_escaped}", a_, re.IGNORECASE ) if content is not None: _snake_case : Tuple = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node _snake_case : Dict = self.tokenajson(a_, is_inner_value=a_, added_vocab=a_ ) if value: if len(a_ ) == 1: _snake_case : Optional[int] = value[0] _snake_case : int = value else: # leaf nodes _snake_case : List[str] = [] for leaf in content.split(r"""<sep/>""" ): _snake_case : List[Any] = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": _snake_case : List[str] = leaf[1:-2] # for categorical special tokens output[key].append(a_ ) if len(output[key] ) == 1: _snake_case : int = output[key][0] _snake_case : Optional[int] = tokens[tokens.find(a_ ) + len(a_ ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:], is_inner_value=a_, added_vocab=a_ ) if len(a_ ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def UpperCamelCase_ ( self: int ): '''simple docstring''' warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""", a_, ) return self.image_processor_class @property def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""", a_, ) return self.image_processor
132
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : int = 10_00 ): """simple docstring""" _snake_case , _snake_case : List[Any] = 1, 1 _snake_case : str = [] for i in range(1 , n + 1 ): _snake_case : Any = prev_numerator + 2 * prev_denominator _snake_case : Optional[Any] = prev_numerator + prev_denominator if len(str(snake_case__ ) ) > len(str(snake_case__ ) ): result.append(snake_case__ ) _snake_case : int = numerator _snake_case : Any = denominator return len(snake_case__ ) if __name__ == "__main__": print(F'''{solution() = }''')
132
1
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class __lowerCAmelCase : '''simple docstring''' def __init__( self , _a , _a=13 , _a=2 , _a=24 , _a=16 , _a=True , _a=True , _a=32 , _a=5 , _a=4 , _a=37 , _a="gelu" , _a=0.1 , _a=0.1 , _a=10 , _a=0.02 , _a=None , _a=2 , _a=2 , ): __a = parent __a = batch_size __a = patch_size __a = max_length __a = num_mel_bins __a = is_training __a = use_labels __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = type_sequence_label_size __a = initializer_range __a = scope __a = frequency_stride __a = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) __a = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 __a = (self.max_length - self.patch_size) // self.time_stride + 1 __a = frequency_out_dimension * time_out_dimension __a = num_patches + 2 def __UpperCAmelCase ( self ): __a = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a = self.get_config() return config, input_values, labels def __UpperCAmelCase ( self ): return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_a , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def __UpperCAmelCase ( self , _a , _a , _a ): __a = ASTModel(config=_a ) model.to(_a ) model.eval() __a = model(_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCAmelCase ( self ): __a = self.prepare_config_and_inputs() ( ( __a ) , ( __a ) , ( __a ) , ) = config_and_inputs __a = {'''input_values''': input_values} return config, inputs_dict @require_torch class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' __UpperCAmelCase : Optional[int] = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) __UpperCAmelCase : Union[str, Any] = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : str = False __UpperCAmelCase : Any = False __UpperCAmelCase : List[str] = False def __UpperCAmelCase ( self , _a , _a , _a , _a , _a ): if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def __UpperCAmelCase ( self ): __a = ASTModelTester(self ) __a = ConfigTester(self , config_class=_a , has_text_modality=_a , hidden_size=37 ) def __UpperCAmelCase ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def __UpperCAmelCase ( self ): pass def __UpperCAmelCase ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(_a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __a = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_a , nn.Linear ) ) def __UpperCAmelCase ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(_a ) __a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a = [*signature.parameters.keys()] __a = ['''input_values'''] self.assertListEqual(arg_names[:1] , _a ) def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) @slow def __UpperCAmelCase ( self ): for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = ASTModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def lowercase ( ) -> int: __a = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' ) __a , __a = torchaudio.load(lowerCAmelCase__ ) return audio, sampling_rate @require_torch @require_torchaudio class __lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def __UpperCAmelCase ( self ): return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def __UpperCAmelCase ( self ): __a = self.default_feature_extractor __a = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(_a ) __a = self.default_feature_extractor __a , __a = prepare_audio() __a = audio.squeeze().numpy() __a = feature_extractor(_a , sampling_rate=_a , return_tensors='''pt''' ).to(_a ) # forward pass with torch.no_grad(): __a = model(**_a ) # verify the logits __a = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , _a ) __a = torch.tensor([-0.8760, -7.0042, -8.6602] ).to(_a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _a , atol=1E-4 ) )
45
import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency a ={ """E""": 12.70, """T""": 9.06, """A""": 8.17, """O""": 7.51, """I""": 6.97, """N""": 6.75, """S""": 6.33, """H""": 6.09, """R""": 5.99, """D""": 4.25, """L""": 4.03, """C""": 2.78, """U""": 2.76, """M""": 2.41, """W""": 2.36, """F""": 2.23, """G""": 2.02, """Y""": 1.97, """P""": 1.93, """B""": 1.29, """V""": 0.98, """K""": 0.77, """J""": 0.15, """X""": 0.15, """Q""": 0.10, """Z""": 0.07, } a ="""ETAOINSHRDLCUMWFGYPBVKJXQZ""" a ="""ABCDEFGHIJKLMNOPQRSTUVWXYZ""" def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> dict[str, int]: __lowerCamelCase : Tuple = {letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> str: return x[0] def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> str: __lowerCamelCase : List[str] = get_letter_count(lowerCamelCase__ ) __lowerCamelCase : dict[int, list[str]] = { freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(lowerCamelCase__ ) __lowerCamelCase : dict[int, str] = {} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find , reverse=lowerCamelCase__ ) __lowerCamelCase : Optional[Any] = ''.join(freq_to_letter[freq] ) __lowerCamelCase : int = list(freq_to_letter_str.items() ) freq_pairs.sort(key=lowerCamelCase__ , reverse=lowerCamelCase__ ) __lowerCamelCase : list[str] = [freq_pair[1] for freq_pair in freq_pairs] return "".join(lowerCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> int: __lowerCamelCase : str = get_frequency_order(lowerCamelCase__ ) __lowerCamelCase : Optional[Any] = 0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
73
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging snake_case_ : Any = logging.get_logger(__name__) snake_case_ : List[Any] = "▁" snake_case_ : Any = {"vocab_file": "sentencepiece.bpe.model"} snake_case_ : List[Any] = { "vocab_file": { "facebook/mbart-large-50-one-to-many-mmt": ( "https://huggingface.co/facebook/mbart-large-50-one-to-many-mmt/resolve/main/sentencepiece.bpe.model" ), } } snake_case_ : List[str] = { "facebook/mbart-large-50-one-to-many-mmt": 1024, } # fmt: off snake_case_ : int = ["ar_AR", "cs_CZ", "de_DE", "en_XX", "es_XX", "et_EE", "fi_FI", "fr_XX", "gu_IN", "hi_IN", "it_IT", "ja_XX", "kk_KZ", "ko_KR", "lt_LT", "lv_LV", "my_MM", "ne_NP", "nl_XX", "ro_RO", "ru_RU", "si_LK", "tr_TR", "vi_VN", "zh_CN", "af_ZA", "az_AZ", "bn_IN", "fa_IR", "he_IL", "hr_HR", "id_ID", "ka_GE", "km_KH", "mk_MK", "ml_IN", "mn_MN", "mr_IN", "pl_PL", "ps_AF", "pt_XX", "sv_SE", "sw_KE", "ta_IN", "te_IN", "th_TH", "tl_XX", "uk_UA", "ur_PK", "xh_ZA", "gl_ES", "sl_SI"] class __snake_case ( a ): UpperCAmelCase__ : int = VOCAB_FILES_NAMES UpperCAmelCase__ : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ : Tuple = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : List[str] = ['''input_ids''', '''attention_mask'''] UpperCAmelCase__ : List[int] = [] UpperCAmelCase__ : List[int] = [] def __init__( self : List[Any] , _snake_case : Tuple , _snake_case : Optional[Any]=None , _snake_case : Any=None , _snake_case : Dict="</s>" , _snake_case : int="</s>" , _snake_case : Optional[Any]="<s>" , _snake_case : List[Any]="<unk>" , _snake_case : List[str]="<pad>" , _snake_case : List[str]="<mask>" , _snake_case : Optional[Dict[str, Any]] = None , **_snake_case : str , ): """simple docstring""" UpperCAmelCase_ = AddedToken(_snake_case , lstrip=_snake_case , rstrip=_snake_case) if isinstance(_snake_case , _snake_case) else mask_token UpperCAmelCase_ = {} if sp_model_kwargs is None else sp_model_kwargs UpperCAmelCase_ = kwargs.get('''additional_special_tokens''' , []) kwargs["additional_special_tokens"] += [ code for code in FAIRSEQ_LANGUAGE_CODES if code not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=_snake_case , tgt_lang=_snake_case , eos_token=_snake_case , unk_token=_snake_case , sep_token=_snake_case , cls_token=_snake_case , pad_token=_snake_case , mask_token=_snake_case , sp_model_kwargs=self.sp_model_kwargs , **_snake_case , ) UpperCAmelCase_ = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(str(_snake_case)) UpperCAmelCase_ = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token UpperCAmelCase_ = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab UpperCAmelCase_ = 1 UpperCAmelCase_ = len(self.sp_model) UpperCAmelCase_ = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(_snake_case) } UpperCAmelCase_ = {v: k for k, v in self.lang_code_to_id.items()} UpperCAmelCase_ = len(self.sp_model) + len(self.lang_code_to_id) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id) UpperCAmelCase_ = {v: k for k, v in self.fairseq_tokens_to_ids.items()} UpperCAmelCase_ = src_lang if src_lang is not None else '''en_XX''' UpperCAmelCase_ = self.lang_code_to_id[self._src_lang] UpperCAmelCase_ = tgt_lang self.set_src_lang_special_tokens(self._src_lang) @property def lowerCamelCase ( self : Optional[Any]): """simple docstring""" return len(self.sp_model) + len(self.lang_code_to_id) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def lowerCamelCase ( self : int): """simple docstring""" return self._src_lang @src_lang.setter def lowerCamelCase ( self : Optional[Any] , _snake_case : str): """simple docstring""" UpperCAmelCase_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def __getstate__( self : int): """simple docstring""" UpperCAmelCase_ = self.__dict__.copy() UpperCAmelCase_ = None return state def __setstate__( self : str , _snake_case : Dict): """simple docstring""" UpperCAmelCase_ = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs'''): UpperCAmelCase_ = {} UpperCAmelCase_ = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def lowerCamelCase ( self : Dict): """simple docstring""" UpperCAmelCase_ = {self.convert_ids_to_tokens(_snake_case): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def lowerCamelCase ( self : int , _snake_case : str): """simple docstring""" return self.sp_model.encode(_snake_case , out_type=_snake_case) def lowerCamelCase ( self : Dict , _snake_case : str): """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCAmelCase_ = self.sp_model.PieceToId(_snake_case) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def lowerCamelCase ( self : Dict , _snake_case : int): """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset) def lowerCamelCase ( self : str , _snake_case : Any): """simple docstring""" UpperCAmelCase_ = [] UpperCAmelCase_ = '''''' UpperCAmelCase_ = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(_snake_case) + token UpperCAmelCase_ = True UpperCAmelCase_ = [] else: current_sub_tokens.append(_snake_case) UpperCAmelCase_ = False out_string += self.sp_model.decode(_snake_case) return out_string.strip() def lowerCamelCase ( self : Tuple , _snake_case : str , _snake_case : Optional[str] = None): """simple docstring""" if not os.path.isdir(_snake_case): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""") return UpperCAmelCase_ = os.path.join( _snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) if os.path.abspath(self.vocab_file) != os.path.abspath(_snake_case) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , _snake_case) elif not os.path.isfile(self.vocab_file): with open(_snake_case , '''wb''') as fi: UpperCAmelCase_ = self.sp_model.serialized_model_proto() fi.write(_snake_case) return (out_vocab_file,) def lowerCamelCase ( self : Optional[Any] , _snake_case : List[int] , _snake_case : Optional[List[int]] = None , _snake_case : bool = False): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_snake_case , token_ids_a=_snake_case , already_has_special_tokens=_snake_case) UpperCAmelCase_ = [1] * len(self.prefix_tokens) UpperCAmelCase_ = [1] * len(self.suffix_tokens) if token_ids_a is None: return prefix_ones + ([0] * len(_snake_case)) + suffix_ones return prefix_ones + ([0] * len(_snake_case)) + ([0] * len(_snake_case)) + suffix_ones def lowerCamelCase ( self : int , _snake_case : List[int] , _snake_case : Optional[List[int]] = None): """simple docstring""" if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowerCamelCase ( self : List[str] , _snake_case : List[str] , _snake_case : str , _snake_case : Optional[str] , _snake_case : Optional[str] , **_snake_case : Any): """simple docstring""" if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''') UpperCAmelCase_ = src_lang UpperCAmelCase_ = self(_snake_case , add_special_tokens=_snake_case , return_tensors=_snake_case , **_snake_case) UpperCAmelCase_ = self.convert_tokens_to_ids(_snake_case) UpperCAmelCase_ = tgt_lang_id return inputs def lowerCamelCase ( self : str , _snake_case : List[str] , _snake_case : str = "en_XX" , _snake_case : Optional[List[str]] = None , _snake_case : str = "ro_RO" , **_snake_case : Optional[int] , ): """simple docstring""" UpperCAmelCase_ = src_lang UpperCAmelCase_ = tgt_lang return super().prepare_seqaseq_batch(_snake_case , _snake_case , **_snake_case) def lowerCamelCase ( self : str): """simple docstring""" return self.set_src_lang_special_tokens(self.src_lang) def lowerCamelCase ( self : Union[str, Any]): """simple docstring""" return self.set_tgt_lang_special_tokens(self.tgt_lang) def lowerCamelCase ( self : Tuple , _snake_case : str): """simple docstring""" UpperCAmelCase_ = self.lang_code_to_id[src_lang] UpperCAmelCase_ = [self.cur_lang_code_id] UpperCAmelCase_ = [self.eos_token_id] def lowerCamelCase ( self : Any , _snake_case : str): """simple docstring""" UpperCAmelCase_ = self.lang_code_to_id[tgt_lang] UpperCAmelCase_ = [self.cur_lang_code_id] UpperCAmelCase_ = [self.eos_token_id]
7
from timeit import timeit def A (__A : int ) -> int: """simple docstring""" if number < 0: raise ValueError('''the value of input must not be negative''' ) UpperCAmelCase_ = 0 while number: number &= number - 1 result += 1 return result def A (__A : int ) -> int: """simple docstring""" if number < 0: raise ValueError('''the value of input must not be negative''' ) UpperCAmelCase_ = 0 while number: if number % 2 == 1: result += 1 number >>= 1 return result def A () -> None: """simple docstring""" def do_benchmark(__A : int ) -> None: UpperCAmelCase_ = '''import __main__ as z''' print(F"""Benchmark when {number = }:""" ) print(F"""{get_set_bits_count_using_modulo_operator(__A ) = }""" ) UpperCAmelCase_ = timeit('''z.get_set_bits_count_using_modulo_operator(25)''' , setup=__A ) print(F"""timeit() runs in {timing} seconds""" ) print(F"""{get_set_bits_count_using_brian_kernighans_algorithm(__A ) = }""" ) UpperCAmelCase_ = timeit( '''z.get_set_bits_count_using_brian_kernighans_algorithm(25)''' , setup=__A , ) print(F"""timeit() runs in {timing} seconds""" ) for number in (25, 37, 58, 0): do_benchmark(__A ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
7
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_electra import ElectraTokenizer A__ = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} A__ = { """vocab_file""": { """google/electra-small-generator""": ( """https://huggingface.co/google/electra-small-generator/resolve/main/vocab.txt""" ), """google/electra-base-generator""": """https://huggingface.co/google/electra-base-generator/resolve/main/vocab.txt""", """google/electra-large-generator""": ( """https://huggingface.co/google/electra-large-generator/resolve/main/vocab.txt""" ), """google/electra-small-discriminator""": ( """https://huggingface.co/google/electra-small-discriminator/resolve/main/vocab.txt""" ), """google/electra-base-discriminator""": ( """https://huggingface.co/google/electra-base-discriminator/resolve/main/vocab.txt""" ), """google/electra-large-discriminator""": ( """https://huggingface.co/google/electra-large-discriminator/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """google/electra-small-generator""": ( """https://huggingface.co/google/electra-small-generator/resolve/main/tokenizer.json""" ), """google/electra-base-generator""": ( """https://huggingface.co/google/electra-base-generator/resolve/main/tokenizer.json""" ), """google/electra-large-generator""": ( """https://huggingface.co/google/electra-large-generator/resolve/main/tokenizer.json""" ), """google/electra-small-discriminator""": ( """https://huggingface.co/google/electra-small-discriminator/resolve/main/tokenizer.json""" ), """google/electra-base-discriminator""": ( """https://huggingface.co/google/electra-base-discriminator/resolve/main/tokenizer.json""" ), """google/electra-large-discriminator""": ( """https://huggingface.co/google/electra-large-discriminator/resolve/main/tokenizer.json""" ), }, } A__ = { """google/electra-small-generator""": 5_12, """google/electra-base-generator""": 5_12, """google/electra-large-generator""": 5_12, """google/electra-small-discriminator""": 5_12, """google/electra-base-discriminator""": 5_12, """google/electra-large-discriminator""": 5_12, } A__ = { """google/electra-small-generator""": {"""do_lower_case""": True}, """google/electra-base-generator""": {"""do_lower_case""": True}, """google/electra-large-generator""": {"""do_lower_case""": True}, """google/electra-small-discriminator""": {"""do_lower_case""": True}, """google/electra-base-discriminator""": {"""do_lower_case""": True}, """google/electra-large-discriminator""": {"""do_lower_case""": True}, } class __lowerCAmelCase ( lowerCamelCase__ ): __lowerCamelCase = VOCAB_FILES_NAMES __lowerCamelCase = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase = PRETRAINED_INIT_CONFIGURATION __lowerCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase = ElectraTokenizer def __init__( self , _snake_case=None , _snake_case=None , _snake_case=True , _snake_case="[UNK]" , _snake_case="[SEP]" , _snake_case="[PAD]" , _snake_case="[CLS]" , _snake_case="[MASK]" , _snake_case=True , _snake_case=None , **_snake_case , ): """simple docstring""" super().__init__( _snake_case , tokenizer_file=_snake_case , do_lower_case=_snake_case , unk_token=_snake_case , sep_token=_snake_case , pad_token=_snake_case , cls_token=_snake_case , mask_token=_snake_case , tokenize_chinese_chars=_snake_case , strip_accents=_snake_case , **_snake_case , ) _lowerCAmelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("""lowercase""" , _snake_case ) != do_lower_case or normalizer_state.get("""strip_accents""" , _snake_case ) != strip_accents or normalizer_state.get("""handle_chinese_chars""" , _snake_case ) != tokenize_chinese_chars ): _lowerCAmelCase = getattr(_snake_case , normalizer_state.pop("""type""" ) ) _lowerCAmelCase = do_lower_case _lowerCAmelCase = strip_accents _lowerCAmelCase = tokenize_chinese_chars _lowerCAmelCase = normalizer_class(**_snake_case ) _lowerCAmelCase = do_lower_case def snake_case ( self , _snake_case , _snake_case=None ): """simple docstring""" _lowerCAmelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def snake_case ( self , _snake_case , _snake_case = None ): """simple docstring""" _lowerCAmelCase = [self.sep_token_id] _lowerCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def snake_case ( self , _snake_case , _snake_case = None ): """simple docstring""" _lowerCAmelCase = self._tokenizer.model.save(_snake_case , name=_snake_case ) return tuple(_snake_case )
82
A__ = [0, 2, 4, 6, 8] A__ = [1, 3, 5, 7, 9] def _UpperCAmelCase ( snake_case , snake_case , snake_case , snake_case ): """simple docstring""" if remaining_length == 0: if digits[0] == 0 or digits[-1] == 0: return 0 for i in range(length // 2 - 1 , -1 , -1 ): remainder += digits[i] + digits[length - i - 1] if remainder % 2 == 0: return 0 remainder //= 10 return 1 if remaining_length == 1: if remainder % 2 == 0: return 0 _lowerCAmelCase = 0 for digit in range(10 ): _lowerCAmelCase = digit result += reversible_numbers( 0 , (remainder + 2 * digit) // 10 , snake_case , snake_case ) return result _lowerCAmelCase = 0 for digita in range(10 ): _lowerCAmelCase = digita if (remainder + digita) % 2 == 0: _lowerCAmelCase = ODD_DIGITS else: _lowerCAmelCase = EVEN_DIGITS for digita in other_parity_digits: _lowerCAmelCase = digita result += reversible_numbers( remaining_length - 2 , (remainder + digita + digita) // 10 , snake_case , snake_case , ) return result def _UpperCAmelCase ( snake_case = 9 ): """simple docstring""" _lowerCAmelCase = 0 for length in range(1 , max_power + 1 ): result += reversible_numbers(snake_case , 0 , [0] * length , snake_case ) return result if __name__ == "__main__": print(f"{solution() = }")
82
1
"""simple docstring""" from statistics import mean, stdev def __lowerCAmelCase ( lowercase : list , lowercase : int = 3 ) -> list: """simple docstring""" snake_case : Dict = min(lowercase ) snake_case : List[Any] = max(lowercase ) # normalize data return [round((x - x_min) / (x_max - x_min) , lowercase ) for x in data] def __lowerCAmelCase ( lowercase : list , lowercase : int = 3 ) -> list: """simple docstring""" snake_case : List[str] = mean(lowercase ) snake_case : Dict = stdev(lowercase ) # standardize data return [round((x - mu) / (sigma) , lowercase ) for x in data]
112
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __snake_case = { """configuration_conditional_detr""": [ """CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ConditionalDetrConfig""", """ConditionalDetrOnnxConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = ["""ConditionalDetrFeatureExtractor"""] __snake_case = ["""ConditionalDetrImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = [ """CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST""", """ConditionalDetrForObjectDetection""", """ConditionalDetrForSegmentation""", """ConditionalDetrModel""", """ConditionalDetrPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP, ConditionalDetrConfig, ConditionalDetrOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_conditional_detr import ConditionalDetrFeatureExtractor from .image_processing_conditional_detr import ConditionalDetrImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_conditional_detr import ( CONDITIONAL_DETR_PRETRAINED_MODEL_ARCHIVE_LIST, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrModel, ConditionalDetrPreTrainedModel, ) else: import sys __snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
112
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A : Dict = { "configuration_x_clip": [ "XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "XCLIPConfig", "XCLIPTextConfig", "XCLIPVisionConfig", ], "processing_x_clip": ["XCLIPProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Dict = [ "XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "XCLIPModel", "XCLIPPreTrainedModel", "XCLIPTextModel", "XCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys A : int = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
184
class _lowercase : """simple docstring""" def __init__( self : Any , __lowerCamelCase : int ): '''simple docstring''' lowerCamelCase__ : List[str] = n lowerCamelCase__ : Union[str, Any] = [None] * self.n lowerCamelCase__ : List[str] = 0 # index of the first element lowerCamelCase__ : Any = 0 lowerCamelCase__ : Any = 0 def __len__( self : Tuple ): '''simple docstring''' return self.size def lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' return self.size == 0 def lowerCAmelCase ( self : Union[str, Any] ): '''simple docstring''' return False if self.is_empty() else self.array[self.front] def lowerCAmelCase ( self : str , __lowerCamelCase : List[str] ): '''simple docstring''' if self.size >= self.n: raise Exception("QUEUE IS FULL" ) lowerCamelCase__ : Optional[Any] = data lowerCamelCase__ : Tuple = (self.rear + 1) % self.n self.size += 1 return self def lowerCAmelCase ( self : Tuple ): '''simple docstring''' if self.size == 0: raise Exception("UNDERFLOW" ) lowerCamelCase__ : Any = self.array[self.front] lowerCamelCase__ : List[Any] = None lowerCamelCase__ : str = (self.front + 1) % self.n self.size -= 1 return temp
184
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class A_ ( metaclass=SCREAMING_SNAKE_CASE_ ): """simple docstring""" __UpperCamelCase = ["""speech"""] def __init__( self :str , *lowercase_ :Union[str, Any] , **lowercase_ :Union[str, Any] ) -> int: requires_backends(self , ['speech'] ) class A_ ( metaclass=SCREAMING_SNAKE_CASE_ ): """simple docstring""" __UpperCamelCase = ["""speech"""] def __init__( self :Tuple , *lowercase_ :Optional[int] , **lowercase_ :Union[str, Any] ) -> Optional[int]: requires_backends(self , ['speech'] )
181
"""simple docstring""" from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging snake_case_ = logging.get_logger(__name__) # pylint: disable=invalid-name class A_ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" def __init__( self :Optional[Any] , lowercase_ :CLIPSegForImageSegmentation , lowercase_ :CLIPSegProcessor , lowercase_ :AutoencoderKL , lowercase_ :CLIPTextModel , lowercase_ :CLIPTokenizer , lowercase_ :UNetaDConditionModel , lowercase_ :Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowercase_ :StableDiffusionSafetyChecker , lowercase_ :CLIPImageProcessor , ) -> List[str]: super().__init__() if hasattr(scheduler.config , 'steps_offset' ) and scheduler.config.steps_offset != 1: UpperCAmelCase = ( f"""The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`""" f""" should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure """ 'to update the config accordingly as leaving `steps_offset` might led to incorrect results' ' in future versions. If you have downloaded this checkpoint from the Hugging Face Hub,' ' it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`' ' file' ) deprecate('steps_offset!=1' , '1.0.0' , lowercase_ , standard_warn=lowercase_ ) UpperCAmelCase = dict(scheduler.config ) UpperCAmelCase = 1 UpperCAmelCase = FrozenDict(lowercase_ ) if hasattr(scheduler.config , 'skip_prk_steps' ) and scheduler.config.skip_prk_steps is False: UpperCAmelCase = ( f"""The configuration file of this scheduler: {scheduler} has not set the configuration""" ' `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make' ' sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to' ' incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face' ' Hub, it would be very nice if you could open a Pull request for the' ' `scheduler/scheduler_config.json` file' ) deprecate('skip_prk_steps not set' , '1.0.0' , lowercase_ , standard_warn=lowercase_ ) UpperCAmelCase = dict(scheduler.config ) UpperCAmelCase = True UpperCAmelCase = FrozenDict(lowercase_ ) if safety_checker is None: logger.warning( f"""You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure""" ' that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered' ' results in services or applications open to the public. Both the diffusers team and Hugging Face' ' strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling' ' it only for use-cases that involve analyzing network behavior or auditing its results. For more' ' information, please have a look at https://github.com/huggingface/diffusers/pull/254 .' ) self.register_modules( segmentation_model=lowercase_ , segmentation_processor=lowercase_ , vae=lowercase_ , text_encoder=lowercase_ , tokenizer=lowercase_ , unet=lowercase_ , scheduler=lowercase_ , safety_checker=lowercase_ , feature_extractor=lowercase_ , ) def UpperCAmelCase__ ( self :Optional[int] , lowercase_ :Optional[Union[str, int]] = "auto" ) -> int: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory UpperCAmelCase = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowercase_ ) def UpperCAmelCase__ ( self :Optional[int] ) -> List[str]: self.enable_attention_slicing(lowercase_ ) def UpperCAmelCase__ ( self :int ) -> int: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) UpperCAmelCase = torch.device('cuda' ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowercase_ , lowercase_ ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCAmelCase__ ( self :Optional[Any] ) -> List[str]: if self.device != torch.device('meta' ) or not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(lowercase_ , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self :Optional[Any] , lowercase_ :Union[str, List[str]] , lowercase_ :Union[torch.FloatTensor, PIL.Image.Image] , lowercase_ :str , lowercase_ :int = 5_12 , lowercase_ :int = 5_12 , lowercase_ :int = 50 , lowercase_ :float = 7.5 , lowercase_ :Optional[Union[str, List[str]]] = None , lowercase_ :Optional[int] = 1 , lowercase_ :float = 0.0 , lowercase_ :Optional[torch.Generator] = None , lowercase_ :Optional[torch.FloatTensor] = None , lowercase_ :Optional[str] = "pil" , lowercase_ :bool = True , lowercase_ :Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowercase_ :int = 1 , **lowercase_ :int , ) -> int: UpperCAmelCase = self.segmentation_processor( text=[text] , images=[image] , padding='max_length' , return_tensors='pt' ).to(self.device ) UpperCAmelCase = self.segmentation_model(**lowercase_ ) UpperCAmelCase = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() UpperCAmelCase = self.numpy_to_pil(lowercase_ )[0].resize(image.size ) # Run inpainting pipeline with the generated mask UpperCAmelCase = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowercase_ , image=lowercase_ , mask_image=lowercase_ , height=lowercase_ , width=lowercase_ , num_inference_steps=lowercase_ , guidance_scale=lowercase_ , negative_prompt=lowercase_ , num_images_per_prompt=lowercase_ , eta=lowercase_ , generator=lowercase_ , latents=lowercase_ , output_type=lowercase_ , return_dict=lowercase_ , callback=lowercase_ , callback_steps=lowercase_ , )
181
1
"""simple docstring""" from typing import Dict from .base import GenericTensor, Pipeline class a ( A__ ): def UpperCamelCase_ ( self , _lowerCamelCase=None , _lowerCamelCase=None , _lowerCamelCase=None , **_lowerCamelCase ): if tokenize_kwargs is None: lowercase = {} if truncation is not None: if "truncation" in tokenize_kwargs: raise ValueError( 'truncation parameter defined twice (given as keyword argument as well as in tokenize_kwargs)' ) lowercase = truncation lowercase = tokenize_kwargs lowercase = {} if return_tensors is not None: lowercase = return_tensors return preprocess_params, {}, postprocess_params def UpperCamelCase_ ( self , _lowerCamelCase , **_lowerCamelCase ): lowercase = self.framework lowercase = self.tokenizer(__lowerCamelCase , return_tensors=__lowerCamelCase , **__lowerCamelCase ) return model_inputs def UpperCamelCase_ ( self , _lowerCamelCase ): lowercase = self.model(**__lowerCamelCase ) return model_outputs def UpperCamelCase_ ( self , _lowerCamelCase , _lowerCamelCase=False ): # [0] is the first available tensor, logits or last_hidden_state. if return_tensors: return model_outputs[0] if self.framework == "pt": return model_outputs[0].tolist() elif self.framework == "tf": return model_outputs[0].numpy().tolist() def __call__( self , *_lowerCamelCase , **_lowerCamelCase ): return super().__call__(*__lowerCamelCase , **__lowerCamelCase )
220
import requests from bsa import BeautifulSoup def UpperCAmelCase_ ( _A = "AAPL" ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = F'''https://in.finance.yahoo.com/quote/{symbol}?s={symbol}''' SCREAMING_SNAKE_CASE__ = BeautifulSoup(requests.get(_A ).text , '''html.parser''' ) SCREAMING_SNAKE_CASE__ = '''My(6px) Pos(r) smartphone_Mt(6px)''' return soup.find('''div''' , class_=class_ ).find('''span''' ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(F"Current {symbol:<4} stock price is {stock_price(symbol):>8}")
314
0
import json import re from typing import TYPE_CHECKING, List, Optional, Tuple, Union import numpy as np from ...utils import is_tf_available, is_torch_available, logging if TYPE_CHECKING: if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_codegen import CodeGenTokenizer __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} __UpperCamelCase = { "vocab_file": { "Salesforce/codegen-350M-mono": "https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/vocab.json", }, "merges_file": { "Salesforce/codegen-350M-mono": "https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/merges.txt", }, "tokenizer_file": { "Salesforce/codegen-350M-mono": ( "https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/tokenizer.json" ), }, } __UpperCamelCase = { "Salesforce/codegen-350M-mono": 2048, } class __magic_name__ ( __lowerCAmelCase): A: Dict = VOCAB_FILES_NAMES A: Optional[Any] = PRETRAINED_VOCAB_FILES_MAP A: Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A: Any = ["input_ids", "attention_mask"] A: Dict = CodeGenTokenizer def __init__( self : int , lowerCamelCase__ : Dict=None , lowerCamelCase__ : List[str]=None , lowerCamelCase__ : int=None , lowerCamelCase__ : Optional[Any]="<|endoftext|>" , lowerCamelCase__ : List[str]="<|endoftext|>" , lowerCamelCase__ : List[Any]="<|endoftext|>" , lowerCamelCase__ : Optional[Any]=False , **lowerCamelCase__ : Tuple , ) -> Any: '''simple docstring''' super().__init__( lowerCamelCase__ , lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , unk_token=lowerCamelCase__ , bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , add_prefix_space=lowerCamelCase__ , **lowerCamelCase__ , ) if kwargs.pop('''add_bos_token''' , lowerCamelCase__ ): UpperCamelCase__ : str = kwargs.pop('''name_or_path''' , '''''' ) raise ValueError( '''Currenty GPT2\'s fast tokenizer does NOT support adding a BOS token.''' '''Instead you should use GPT2\'s slow tokenizer class `CodeGenTokenizer` as follows: \n''' F"`CodeGenTokenizer.from_pretrained('{model_id}')`\nor\n" F"`AutoTokenizer.from_pretrained('{model_id}', use_fast=False)`\n" '''This issue will be fixed soon, see: https://github.com/huggingface/tokenizers/pull/1005.''' ''' so that the fast tokenizer works correctly.''' ) UpperCamelCase__ : List[str] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , lowerCamelCase__ ) != add_prefix_space: UpperCamelCase__ : Dict = getattr(lowerCamelCase__ , pre_tok_state.pop('''type''' ) ) UpperCamelCase__ : Union[str, Any] = add_prefix_space UpperCamelCase__ : List[str] = pre_tok_class(**lowerCamelCase__ ) UpperCamelCase__ : Optional[Any] = add_prefix_space def UpperCAmelCase__ ( self : List[Any] , *lowerCamelCase__ : Dict , **lowerCamelCase__ : List[Any] ) -> BatchEncoding: '''simple docstring''' UpperCamelCase__ : Optional[Any] = kwargs.get('''is_split_into_words''' , lowerCamelCase__ ) assert self.add_prefix_space or not is_split_into_words, ( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*lowerCamelCase__ , **lowerCamelCase__ ) def UpperCAmelCase__ ( self : Tuple , *lowerCamelCase__ : int , **lowerCamelCase__ : Optional[Any] ) -> BatchEncoding: '''simple docstring''' UpperCamelCase__ : Optional[Any] = kwargs.get('''is_split_into_words''' , lowerCamelCase__ ) assert self.add_prefix_space or not is_split_into_words, ( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*lowerCamelCase__ , **lowerCamelCase__ ) def UpperCAmelCase__ ( self : List[Any] , lowerCamelCase__ : str , lowerCamelCase__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' UpperCamelCase__ : Optional[int] = self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ ) def UpperCAmelCase__ ( self : Optional[int] , lowerCamelCase__ : Union[int, List[int], "np.ndarray", "torch.Tensor", "tf.Tensor"] , lowerCamelCase__ : bool = False , lowerCamelCase__ : bool = None , lowerCamelCase__ : Optional[List[str]] = None , **lowerCamelCase__ : Tuple , ) -> str: '''simple docstring''' UpperCamelCase__ : int = super().decode( token_ids=lowerCamelCase__ , skip_special_tokens=lowerCamelCase__ , clean_up_tokenization_spaces=lowerCamelCase__ , **lowerCamelCase__ , ) if truncate_before_pattern is not None and len(lowerCamelCase__ ) > 0: UpperCamelCase__ : List[Any] = self.truncate(lowerCamelCase__ , lowerCamelCase__ ) return decoded_text def UpperCAmelCase__ ( self : Union[str, Any] , lowerCamelCase__ : Any , lowerCamelCase__ : Optional[int] ) -> Dict: '''simple docstring''' def find_re(lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Tuple , lowerCamelCase__ : Tuple ): UpperCamelCase__ : List[Any] = pattern.search(lowerCamelCase__ , lowerCamelCase__ ) return m.start() if m else -1 UpperCamelCase__ : int = [re.compile(lowerCamelCase__ , re.MULTILINE ) for pattern in truncate_before_pattern] UpperCamelCase__ : Any = list(re.finditer('''^print''' , lowerCamelCase__ , re.MULTILINE ) ) if len(lowerCamelCase__ ) > 1: UpperCamelCase__ : Any = completion[: prints[1].start()] UpperCamelCase__ : Any = list(re.finditer('''^def''' , lowerCamelCase__ , re.MULTILINE ) ) if len(lowerCamelCase__ ) > 1: UpperCamelCase__ : Dict = completion[: defs[1].start()] UpperCamelCase__ : Any = 0 UpperCamelCase__ : List[str] = [ pos for pos in [find_re(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) for terminal in terminals] if pos != -1 ] if len(lowerCamelCase__ ) > 0: return completion[: min(lowerCamelCase__ )] else: return completion
360
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _interleave_iterable_datasets from .splits import NamedSplit from .utils import logging from .utils.py_utils import Literal __UpperCamelCase : Tuple = logging.get_logger(__name__) __UpperCamelCase : Tuple = TypeVar("DatasetType", Dataset, IterableDataset) def _a ( SCREAMING_SNAKE_CASE : List[DatasetType] , SCREAMING_SNAKE_CASE : Optional[List[float]] = None , SCREAMING_SNAKE_CASE : Optional[int] = None , SCREAMING_SNAKE_CASE : Optional[DatasetInfo] = None , SCREAMING_SNAKE_CASE : Optional[NamedSplit] = None , SCREAMING_SNAKE_CASE : Literal["first_exhausted", "all_exhausted"] = "first_exhausted" , ): """simple docstring""" from .arrow_dataset import Dataset from .iterable_dataset import IterableDataset if not datasets: raise ValueError('''Unable to interleave an empty list of datasets.''' ) for i, dataset in enumerate(SCREAMING_SNAKE_CASE ): if not isinstance(SCREAMING_SNAKE_CASE , (Dataset, IterableDataset) ): if isinstance(SCREAMING_SNAKE_CASE , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( F"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} " '''is an empty dataset dictionary.''' ) raise ValueError( F"Dataset at position {i} has at least one split: {list(SCREAMING_SNAKE_CASE )}\n" F"Please pick one to interleave with the other datasets, for example: dataset['{next(iter(SCREAMING_SNAKE_CASE ) )}']" ) raise ValueError( F"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(SCREAMING_SNAKE_CASE ).__name__}." ) if i == 0: UpperCamelCase__ , UpperCamelCase__ : List[Any] = ( (Dataset, IterableDataset) if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) else (IterableDataset, Dataset) ) elif not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): raise ValueError( F"Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects." ) if stopping_strategy not in ["first_exhausted", "all_exhausted"]: raise ValueError(F"{stopping_strategy} is not supported. Please enter a valid stopping_strategy." ) if dataset_type is Dataset: return _interleave_map_style_datasets( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , info=SCREAMING_SNAKE_CASE , split=SCREAMING_SNAKE_CASE , stopping_strategy=SCREAMING_SNAKE_CASE ) else: return _interleave_iterable_datasets( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , info=SCREAMING_SNAKE_CASE , split=SCREAMING_SNAKE_CASE , stopping_strategy=SCREAMING_SNAKE_CASE ) def _a ( SCREAMING_SNAKE_CASE : List[DatasetType] , SCREAMING_SNAKE_CASE : Optional[DatasetInfo] = None , SCREAMING_SNAKE_CASE : Optional[NamedSplit] = None , SCREAMING_SNAKE_CASE : int = 0 , ): """simple docstring""" if not dsets: raise ValueError('''Unable to concatenate an empty list of datasets.''' ) for i, dataset in enumerate(SCREAMING_SNAKE_CASE ): if not isinstance(SCREAMING_SNAKE_CASE , (Dataset, IterableDataset) ): if isinstance(SCREAMING_SNAKE_CASE , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( F"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} " '''is an empty dataset dictionary.''' ) raise ValueError( F"Dataset at position {i} has at least one split: {list(SCREAMING_SNAKE_CASE )}\n" F"Please pick one to interleave with the other datasets, for example: dataset['{next(iter(SCREAMING_SNAKE_CASE ) )}']" ) raise ValueError( F"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(SCREAMING_SNAKE_CASE ).__name__}." ) if i == 0: UpperCamelCase__ , UpperCamelCase__ : List[str] = ( (Dataset, IterableDataset) if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) else (IterableDataset, Dataset) ) elif not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): raise ValueError( F"Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects." ) if dataset_type is Dataset: return _concatenate_map_style_datasets(SCREAMING_SNAKE_CASE , info=SCREAMING_SNAKE_CASE , split=SCREAMING_SNAKE_CASE , axis=SCREAMING_SNAKE_CASE ) else: return _concatenate_iterable_datasets(SCREAMING_SNAKE_CASE , info=SCREAMING_SNAKE_CASE , split=SCREAMING_SNAKE_CASE , axis=SCREAMING_SNAKE_CASE )
51
0
# Lint as: python3 import dataclasses import re from dataclasses import dataclass from functools import total_ordering from typing import Optional, Union lowercase_ = re.compile(R'^(?P<major>\d+)' R'\.(?P<minor>\d+)' R'\.(?P<patch>\d+)$') @total_ordering @dataclass class __lowerCAmelCase : _a = 42 _a = None _a = None _a = None _a = None def A__ ( self ) -> str: '''simple docstring''' _lowercase , _lowercase , _lowercase =_str_to_version_tuple(self.version_str ) def __repr__( self ) -> Dict: '''simple docstring''' return F'''{self.tuple[0]}.{self.tuple[1]}.{self.tuple[2]}''' @property def A__ ( self ) -> Dict: '''simple docstring''' return self.major, self.minor, self.patch def A__ ( self , lowerCAmelCase ) -> List[Any]: '''simple docstring''' if isinstance(lowerCAmelCase , lowerCAmelCase ): return Version(lowerCAmelCase ) elif isinstance(lowerCAmelCase , lowerCAmelCase ): return other raise TypeError(F'''{other} (type {type(lowerCAmelCase )}) cannot be compared to version.''' ) def __eq__( self , lowerCAmelCase ) -> List[Any]: '''simple docstring''' try: _lowercase =self._validate_operand(lowerCAmelCase ) except (TypeError, ValueError): return False else: return self.tuple == other.tuple def __lt__( self , lowerCAmelCase ) -> Union[str, Any]: '''simple docstring''' _lowercase =self._validate_operand(lowerCAmelCase ) return self.tuple < other.tuple def __hash__( self ) -> Any: '''simple docstring''' return hash(_version_tuple_to_str(self.tuple ) ) @classmethod def A__ ( cls , lowerCAmelCase ) -> Tuple: '''simple docstring''' _lowercase ={f.name for f in dataclasses.fields(cls )} return cls(**{k: v for k, v in dic.items() if k in field_names} ) def A__ ( self ) -> str: '''simple docstring''' return self.version_str def a ( A__ : List[str] ) -> int: """simple docstring""" _lowercase =_VERSION_REG.match(A__ ) if not res: raise ValueError(F'''Invalid version \'{version_str}\'. Format should be x.y.z with {{x,y,z}} being digits.''' ) return tuple(int(A__ ) for v in [res.group('major' ), res.group('minor' ), res.group('patch' )] ) def a ( A__ : Dict ) -> List[str]: """simple docstring""" return ".".join(str(A__ ) for v in version_tuple )
205
import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler lowercase_ = 1_6 lowercase_ = 3_2 def a ( A__ : Accelerator , A__ : int = 16 , A__ : str = "bert-base-cased" ) -> Optional[int]: """simple docstring""" _lowercase =AutoTokenizer.from_pretrained(A__ ) _lowercase =load_dataset('glue' , 'mrpc' ) def tokenize_function(A__ : Optional[int] ): # max_length=None => use the model max length (it's actually the default) _lowercase =tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=A__ , max_length=A__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset _lowercase =datasets.map( A__ , batched=A__ , remove_columns=['idx', 'sentence1', 'sentence2'] , load_from_cache_file=A__ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _lowercase =tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(A__ : List[str] ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(A__ , padding='max_length' , max_length=128 , return_tensors='pt' ) return tokenizer.pad(A__ , padding='longest' , return_tensors='pt' ) # Instantiate dataloaders. _lowercase =DataLoader( tokenized_datasets['train'] , shuffle=A__ , collate_fn=A__ , batch_size=A__ ) _lowercase =DataLoader( tokenized_datasets['validation'] , shuffle=A__ , collate_fn=A__ , batch_size=A__ ) return train_dataloader, eval_dataloader def a ( A__ : Optional[Any] , A__ : Optional[int] , A__ : List[str] , A__ : Dict ) -> Dict: """simple docstring""" model.eval() _lowercase =0 for step, batch in enumerate(A__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _lowercase =model(**A__ ) _lowercase =outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times _lowercase , _lowercase =accelerator.gather( (predictions, batch['labels']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(A__ ) - 1: _lowercase =predictions[: len(eval_dataloader.dataset ) - samples_seen] _lowercase =references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=A__ , references=A__ , ) _lowercase =metric.compute() return eval_metric["accuracy"] def a ( A__ : str , A__ : Optional[Any] ) -> Optional[Any]: """simple docstring""" _lowercase =Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _lowercase =config['lr'] _lowercase =int(config['num_epochs'] ) _lowercase =int(config['seed'] ) _lowercase =int(config['batch_size'] ) _lowercase =args.model_name_or_path set_seed(A__ ) _lowercase , _lowercase =get_dataloaders(A__ , A__ , A__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _lowercase =AutoModelForSequenceClassification.from_pretrained(A__ , return_dict=A__ ) # Instantiate optimizer _lowercase =( AdamW if accelerator.state.deepspeed_plugin is None or 'optimizer' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) _lowercase =optimizer_cls(params=model.parameters() , lr=A__ ) if accelerator.state.deepspeed_plugin is not None: _lowercase =accelerator.state.deepspeed_plugin.deepspeed_config[ 'gradient_accumulation_steps' ] else: _lowercase =1 _lowercase =(len(A__ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): _lowercase =get_linear_schedule_with_warmup( optimizer=A__ , num_warmup_steps=0 , num_training_steps=A__ , ) else: _lowercase =DummyScheduler(A__ , total_num_steps=A__ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _lowercase , _lowercase , _lowercase , _lowercase , _lowercase =accelerator.prepare( A__ , A__ , A__ , A__ , A__ ) # We need to keep track of how many total steps we have iterated over _lowercase =0 # We also need to keep track of the stating epoch so files are named properly _lowercase =0 _lowercase =evaluate.load('glue' , 'mrpc' ) _lowercase =num_epochs if args.partial_train_epoch is not None: _lowercase =args.partial_train_epoch if args.resume_from_checkpoint: accelerator.load_state(args.resume_from_checkpoint ) _lowercase =args.resume_from_checkpoint.split('epoch_' )[1] _lowercase ='' for char in epoch_string: if char.isdigit(): state_epoch_num += char else: break _lowercase =int(A__ ) + 1 _lowercase =evaluation_loop(A__ , A__ , A__ , A__ ) accelerator.print('resumed checkpoint performance:' , A__ ) accelerator.print('resumed checkpoint\'s scheduler\'s lr:' , lr_scheduler.get_lr()[0] ) accelerator.print('resumed optimizers\'s lr:' , optimizer.param_groups[0]['lr'] ) with open(os.path.join(args.output_dir , F'''state_{starting_epoch-1}.json''' ) , 'r' ) as f: _lowercase =json.load(A__ ) assert resumed_state["accuracy"] == accuracy, "Accuracy mismatch, loading from checkpoint failed" assert ( resumed_state["lr"] == lr_scheduler.get_lr()[0] ), "Scheduler learning rate mismatch, loading from checkpoint failed" assert ( resumed_state["optimizer_lr"] == optimizer.param_groups[0]["lr"] ), "Optimizer learning rate mismatch, loading from checkpoint failed" assert resumed_state["epoch"] == starting_epoch - 1, "Epoch mismatch, loading from checkpoint failed" return # Now we train the model _lowercase ={} for epoch in range(A__ , A__ ): model.train() for step, batch in enumerate(A__ ): _lowercase =model(**A__ ) _lowercase =outputs.loss _lowercase =loss / gradient_accumulation_steps accelerator.backward(A__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 _lowercase =F'''epoch_{epoch}''' _lowercase =os.path.join(args.output_dir , A__ ) accelerator.save_state(A__ ) _lowercase =evaluation_loop(A__ , A__ , A__ , A__ ) _lowercase =accuracy _lowercase =lr_scheduler.get_lr()[0] _lowercase =optimizer.param_groups[0]['lr'] _lowercase =epoch _lowercase =overall_step accelerator.print(F'''epoch {epoch}:''' , A__ ) accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , F'''state_{epoch}.json''' ) , 'w' ) as f: json.dump(A__ , A__ ) def a ( ) -> Tuple: """simple docstring""" _lowercase =argparse.ArgumentParser(description='Simple example of training script tracking peak GPU memory usage.' ) parser.add_argument( '--model_name_or_path' , type=A__ , default='bert-base-cased' , help='Path to pretrained model or model identifier from huggingface.co/models.' , required=A__ , ) parser.add_argument( '--output_dir' , type=A__ , default='.' , help='Optional save directory where all checkpoint folders will be stored. Default is the current working directory.' , ) parser.add_argument( '--resume_from_checkpoint' , type=A__ , default=A__ , help='If the training should continue from a checkpoint folder.' , ) parser.add_argument( '--partial_train_epoch' , type=A__ , default=A__ , help='If passed, the training will stop after this number of epochs.' , ) parser.add_argument( '--num_epochs' , type=A__ , default=2 , help='Number of train epochs.' , ) _lowercase =parser.parse_args() _lowercase ={'lr': 2e-5, 'num_epochs': args.num_epochs, 'seed': 42, 'batch_size': 16} training_function(A__ , A__ ) if __name__ == "__main__": main()
205
1
'''simple docstring''' import argparse import os from accelerate.test_utils import execute_subprocess_async def SCREAMING_SNAKE_CASE_ ( __A : Any=None ) -> Any: if subparsers is not None: _SCREAMING_SNAKE_CASE = subparsers.add_parser("test" ) else: _SCREAMING_SNAKE_CASE = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" , default=__A , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have " "such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed " "with \'huggingface\'." ) , ) if subparsers is not None: parser.set_defaults(func=__A ) return parser def SCREAMING_SNAKE_CASE_ ( __A : str ) -> List[str]: _SCREAMING_SNAKE_CASE = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: _SCREAMING_SNAKE_CASE = script_name else: _SCREAMING_SNAKE_CASE = f"""--config_file={args.config_file} {script_name}""" _SCREAMING_SNAKE_CASE = ["accelerate-launch"] + test_args.split() _SCREAMING_SNAKE_CASE = execute_subprocess_async(__A , env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def SCREAMING_SNAKE_CASE_ ( ) -> int: _SCREAMING_SNAKE_CASE = test_command_parser() _SCREAMING_SNAKE_CASE = parser.parse_args() test_command(__A ) if __name__ == "__main__": main()
361
'''simple docstring''' def SCREAMING_SNAKE_CASE_ ( __A : int ) -> int: if n == 1 or not isinstance(__A , __A ): return 0 elif n == 2: return 1 else: _SCREAMING_SNAKE_CASE = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def SCREAMING_SNAKE_CASE_ ( __A : int ) -> int: _SCREAMING_SNAKE_CASE = 0 _SCREAMING_SNAKE_CASE = 2 while digits < n: index += 1 _SCREAMING_SNAKE_CASE = len(str(fibonacci(__A ) ) ) return index def SCREAMING_SNAKE_CASE_ ( __A : int = 10_00 ) -> int: return fibonacci_digits_index(__A ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
111
0
from __future__ import annotations from fractions import Fraction def UpperCAmelCase ( a_ , a_ ) -> bool: """simple docstring""" return ( num != den and num % 1_0 == den // 1_0 and (num // 1_0) / (den % 1_0) == num / den ) def UpperCAmelCase ( a_ ) -> list[str]: """simple docstring""" __A = [] __A = 1_1 __A = int("1" + "0" * digit_len ) for num in range(a_ , a_ ): while den <= 9_9: if (num != den) and (num % 1_0 == den // 1_0) and (den % 1_0 != 0): if is_digit_cancelling(a_ , a_ ): solutions.append(F'''{num}/{den}''' ) den += 1 num += 1 __A = 1_0 return solutions def UpperCAmelCase ( a_ = 2 ) -> int: """simple docstring""" __A = 1.0 for fraction in fraction_list(a_ ): __A = Fraction(a_ ) result *= frac.denominator / frac.numerator return int(a_ ) if __name__ == "__main__": print(solution())
15
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __A =logging.get_logger(__name__) def a ( _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Dict ): '''simple docstring''' __UpperCAmelCase : List[str] = b.T __UpperCAmelCase : Any = np.sum(np.square(_UpperCAmelCase ) , axis=1 ) __UpperCAmelCase : int = np.sum(np.square(_UpperCAmelCase ) , axis=0 ) __UpperCAmelCase : Optional[int] = np.matmul(_UpperCAmelCase , _UpperCAmelCase ) __UpperCAmelCase : List[str] = aa[:, None] - 2 * ab + ba[None, :] return d def a ( _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Optional[Any] ): '''simple docstring''' __UpperCAmelCase : str = x.reshape(-1 , 3 ) __UpperCAmelCase : Optional[int] = squared_euclidean_distance(_UpperCAmelCase , _UpperCAmelCase ) return np.argmin(_UpperCAmelCase , axis=1 ) class UpperCAmelCase__ ( __UpperCamelCase ): '''simple docstring''' UpperCamelCase = ["""pixel_values"""] def __init__( self : str , a_ : Optional[Union[List[List[int]], np.ndarray]] = None , a_ : bool = True , a_ : Dict[str, int] = None , a_ : PILImageResampling = PILImageResampling.BILINEAR , a_ : bool = True , a_ : bool = True , **a_ : List[str] , ): '''simple docstring''' super().__init__(**a_ ) __UpperCAmelCase : Optional[int] = size if size is not None else {'''height''': 2_56, '''width''': 2_56} __UpperCAmelCase : List[str] = get_size_dict(a_ ) __UpperCAmelCase : str = np.array(a_ ) if clusters is not None else None __UpperCAmelCase : Dict = do_resize __UpperCAmelCase : Tuple = size __UpperCAmelCase : Union[str, Any] = resample __UpperCAmelCase : Tuple = do_normalize __UpperCAmelCase : Optional[int] = do_color_quantize def snake_case__ ( self : Optional[Any] , a_ : np.ndarray , a_ : Dict[str, int] , a_ : PILImageResampling = PILImageResampling.BILINEAR , a_ : Optional[Union[str, ChannelDimension]] = None , **a_ : Dict , ): '''simple docstring''' __UpperCAmelCase : Tuple = get_size_dict(a_ ) if "height" not in size or "width" not in size: raise ValueError(F'Size dictionary must contain both height and width keys. Got {size.keys()}' ) return resize( a_ , size=(size['''height'''], size['''width''']) , resample=a_ , data_format=a_ , **a_ ) def snake_case__ ( self : Tuple , a_ : np.ndarray , a_ : Optional[Union[str, ChannelDimension]] = None , ): '''simple docstring''' __UpperCAmelCase : Dict = rescale(image=a_ , scale=1 / 1_2_7.5 , data_format=a_ ) __UpperCAmelCase : Union[str, Any] = image - 1 return image def snake_case__ ( self : int , a_ : ImageInput , a_ : bool = None , a_ : Dict[str, int] = None , a_ : PILImageResampling = None , a_ : bool = None , a_ : Optional[bool] = None , a_ : Optional[Union[List[List[int]], np.ndarray]] = None , a_ : Optional[Union[str, TensorType]] = None , a_ : Optional[Union[str, ChannelDimension]] = ChannelDimension.FIRST , **a_ : Any , ): '''simple docstring''' __UpperCAmelCase : Any = do_resize if do_resize is not None else self.do_resize __UpperCAmelCase : List[str] = size if size is not None else self.size __UpperCAmelCase : Any = get_size_dict(a_ ) __UpperCAmelCase : Optional[int] = resample if resample is not None else self.resample __UpperCAmelCase : List[str] = do_normalize if do_normalize is not None else self.do_normalize __UpperCAmelCase : int = do_color_quantize if do_color_quantize is not None else self.do_color_quantize __UpperCAmelCase : Optional[int] = clusters if clusters is not None else self.clusters __UpperCAmelCase : Any = np.array(a_ ) __UpperCAmelCase : Optional[int] = make_list_of_images(a_ ) if not valid_images(a_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_color_quantize and clusters is None: raise ValueError('''Clusters must be specified if do_color_quantize is True.''' ) # All transformations expect numpy arrays. __UpperCAmelCase : List[Any] = [to_numpy_array(a_ ) for image in images] if do_resize: __UpperCAmelCase : List[str] = [self.resize(image=a_ , size=a_ , resample=a_ ) for image in images] if do_normalize: __UpperCAmelCase : Dict = [self.normalize(image=a_ ) for image in images] if do_color_quantize: __UpperCAmelCase : int = [to_channel_dimension_format(a_ , ChannelDimension.LAST ) for image in images] # color quantize from (batch_size, height, width, 3) to (batch_size, height, width) __UpperCAmelCase : List[str] = np.array(a_ ) __UpperCAmelCase : Dict = color_quantize(a_ , a_ ).reshape(images.shape[:-1] ) # flatten to (batch_size, height*width) __UpperCAmelCase : Any = images.shape[0] __UpperCAmelCase : Any = images.reshape(a_ , -1 ) # We need to convert back to a list of images to keep consistent behaviour across processors. __UpperCAmelCase : List[Any] = list(a_ ) else: __UpperCAmelCase : int = [to_channel_dimension_format(a_ , a_ ) for image in images] __UpperCAmelCase : int = {'''input_ids''': images} return BatchFeature(data=a_ , tensor_type=a_ )
226
0
'''simple docstring''' import argparse import os import torch from transformers import FlavaConfig, FlavaForPreTraining from transformers.models.flava.convert_dalle_to_flava_codebook import convert_dalle_checkpoint def _snake_case ( _SCREAMING_SNAKE_CASE : Tuple ) -> Dict: """simple docstring""" return sum(param.float().sum() if """encoder.embeddings""" not in key else 0 for key, param in state_dict.items() ) def _snake_case ( _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : int ) -> str: """simple docstring""" lowerCAmelCase = {} for key, value in state_dict.items(): if "text_encoder.embeddings" in key or "image_encoder.embeddings" in key: continue lowerCAmelCase = key.replace("""heads.cmd.mim_head.cls.predictions""" , """mmm_image_head""" ) lowerCAmelCase = key.replace("""heads.cmd.mlm_head.cls.predictions""" , """mmm_text_head""" ) lowerCAmelCase = key.replace("""heads.cmd.itm_head.cls""" , """itm_head""" ) lowerCAmelCase = key.replace("""heads.cmd.itm_head.pooler""" , """itm_head.pooler""" ) lowerCAmelCase = key.replace("""heads.cmd.clip_head.logit_scale""" , """flava.logit_scale""" ) lowerCAmelCase = key.replace("""heads.fairseq_mlm.cls.predictions""" , """mlm_head""" ) lowerCAmelCase = key.replace("""heads.imagenet.mim_head.cls.predictions""" , """mim_head""" ) lowerCAmelCase = key.replace("""mm_text_projection""" , """flava.text_to_mm_projection""" ) lowerCAmelCase = key.replace("""mm_image_projection""" , """flava.image_to_mm_projection""" ) lowerCAmelCase = key.replace("""image_encoder.module""" , """flava.image_model""" ) lowerCAmelCase = key.replace("""text_encoder.module""" , """flava.text_model""" ) lowerCAmelCase = key.replace("""mm_encoder.module.encoder.cls_token""" , """flava.multimodal_model.cls_token""" ) lowerCAmelCase = key.replace("""mm_encoder.module""" , """flava.multimodal_model""" ) lowerCAmelCase = key.replace("""text_projection""" , """flava.text_projection""" ) lowerCAmelCase = key.replace("""image_projection""" , """flava.image_projection""" ) lowerCAmelCase = value.float() for key, value in codebook_state_dict.items(): lowerCAmelCase = value return upgrade @torch.no_grad() def _snake_case ( _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : List[Any] , _SCREAMING_SNAKE_CASE : Dict=None ) -> Dict: """simple docstring""" if config_path is not None: lowerCAmelCase = FlavaConfig.from_pretrained(lowerCamelCase__ ) else: lowerCAmelCase = FlavaConfig() lowerCAmelCase = FlavaForPreTraining(lowerCamelCase__ ).eval() lowerCAmelCase = convert_dalle_checkpoint(lowerCamelCase__ , lowerCamelCase__ , save_checkpoint=lowerCamelCase__ ) if os.path.exists(lowerCamelCase__ ): lowerCAmelCase = torch.load(lowerCamelCase__ , map_location="""cpu""" ) else: lowerCAmelCase = torch.hub.load_state_dict_from_url(lowerCamelCase__ , map_location="""cpu""" ) lowerCAmelCase = upgrade_state_dict(lowerCamelCase__ , lowerCamelCase__ ) hf_model.load_state_dict(lowerCamelCase__ ) lowerCAmelCase = hf_model.state_dict() lowerCAmelCase = count_parameters(lowerCamelCase__ ) lowerCAmelCase = count_parameters(lowerCamelCase__ ) + count_parameters(lowerCamelCase__ ) assert torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1E-3 ) hf_model.save_pretrained(lowerCamelCase__ ) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to flava checkpoint') parser.add_argument('--codebook_path', default=None, type=str, help='Path to flava codebook checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') UpperCAmelCase = parser.parse_args() convert_flava_checkpoint(args.checkpoint_path, args.codebook_path, args.pytorch_dump_folder_path, args.config_path)
350
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) class __snake_case( _lowerCAmelCase ): '''simple docstring''' UpperCAmelCase : Optional[int] = "timm_backbone" def __init__( self , A_=None , A_=3 , A_=True , A_=True , A_=None , **A_ , ) -> int: super().__init__(**A_ ) lowerCAmelCase = backbone lowerCAmelCase = num_channels lowerCAmelCase = features_only lowerCAmelCase = use_pretrained_backbone lowerCAmelCase = True lowerCAmelCase = out_indices if out_indices is not None else (-1,)
187
0
def UpperCamelCase( __UpperCamelCase : int = 1000 ): return sum(e for e in range(3 ,__UpperCamelCase ) if e % 3 == 0 or e % 5 == 0 ) if __name__ == "__main__": print(F'''{solution() = }''')
103
"""simple docstring""" import os from typing import Dict, List, Tuple, TypeVar, Union A_ = TypeVar('''T''') A_ = Union[List[T], Tuple[T, ...]] A_ = Union[T, List[T], Dict[str, T]] A_ = Union[str, bytes, os.PathLike]
64
0
from __future__ import annotations from dataclasses import dataclass @dataclass class a : _lowercase = 42 _lowercase = None _lowercase = None def __SCREAMING_SNAKE_CASE ( lowerCAmelCase: TreeNode | None ) -> bool: # Validation def is_valid_tree(lowerCAmelCase: TreeNode | None ) -> bool: if node is None: return True if not isinstance(lowerCAmelCase , lowerCAmelCase ): return False try: float(node.data ) except (TypeError, ValueError): return False return is_valid_tree(node.left ) and is_valid_tree(node.right ) if not is_valid_tree(lowerCAmelCase ): raise ValueError( "Each node should be type of TreeNode and data should be float." ) def is_binary_search_tree_recursive_check( lowerCAmelCase: TreeNode | None , lowerCAmelCase: float , lowerCAmelCase: float ) -> bool: if node is None: return True return ( left_bound < node.data < right_bound and is_binary_search_tree_recursive_check(node.left , lowerCAmelCase , node.data ) and is_binary_search_tree_recursive_check( node.right , node.data , lowerCAmelCase ) ) return is_binary_search_tree_recursive_check(lowerCAmelCase , -float("inf" ) , float("inf" ) ) if __name__ == "__main__": import doctest doctest.testmod()
189
from typing import Any class a : def __init__( self , A_ ): '''simple docstring''' _UpperCAmelCase : List[Any] = data _UpperCAmelCase : Any = None class a : def __init__( self ): '''simple docstring''' _UpperCAmelCase : Union[str, Any] = None def _UpperCAmelCase ( self ): '''simple docstring''' _UpperCAmelCase : str = self.head while temp is not None: print(temp.data , end=" " ) _UpperCAmelCase : str = temp.next print() def _UpperCAmelCase ( self , A_ ): '''simple docstring''' _UpperCAmelCase : Optional[int] = Node(A_ ) _UpperCAmelCase : Tuple = self.head _UpperCAmelCase : Tuple = new_node def _UpperCAmelCase ( self , A_ , A_ ): '''simple docstring''' if node_data_a == node_data_a: return else: _UpperCAmelCase : int = self.head while node_a is not None and node_a.data != node_data_a: _UpperCAmelCase : Tuple = node_a.next _UpperCAmelCase : Dict = self.head while node_a is not None and node_a.data != node_data_a: _UpperCAmelCase : List[Any] = node_a.next if node_a is None or node_a is None: return _UpperCAmelCase , _UpperCAmelCase : Optional[int] = node_a.data, node_a.data if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print('After swapping') ll.print_list()
189
1
'''simple docstring''' import argparse import re from flax.traverse_util import flatten_dict, unflatten_dict from tax import checkpoints from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model from transformers.utils import logging logging.set_verbosity_info() # should not include what is already done by the `from_pt` argument lowerCAmelCase: int = { '/attention/': '/0/SelfAttention/', '/self_attention/': '/0/SelfAttention/', '/encoder_decoder_attention/': '/1/EncDecAttention/', 'value': 'v', 'query': 'q', 'key': 'k', 'out': 'o', 'pre_self_attention_layer_norm': '0/layer_norm', 'pre_cross_attention_layer_norm': '1/layer_norm', 'pre_attention_layer_norm': '0/layer_norm', # previously 1, but seems wrong 'token_embedder': 'shared', 'encoder_norm': 'final_layer_norm', 'decoder_norm': 'final_layer_norm', 'relpos_bias/rel_embedding': 'block/0/layer/0/SelfAttention/relative_attention_bias/weight', 'router/router_weights/w/': 'router/classifier/', 'roer/roer_weights/w/': 'router/classifier/', 'logits_dense': 'lm_head', } def lowerCamelCase__ ( _A ): # 1. in HF T5, we have block.{x}.layer.{y}. which corresponds to layer.{x} in # the original model a : Union[str, Any] = list(s_dict.keys() ) for key in keys: a : Optional[int] = R'''.*/layers_(\d+)''' a : List[Any] = key if re.match(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): a : Optional[Any] = re.sub(r'layers_(\d+)' , r'block/\1/layer' , SCREAMING_SNAKE_CASE__ ) a : List[str] = R'''(encoder|decoder)\/''' if re.match(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): a : Optional[Any] = re.match(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ).groups() if groups[0] == "encoder": a : Optional[Any] = re.sub(r'/mlp/' , r'/1/mlp/' , SCREAMING_SNAKE_CASE__ ) a : Any = re.sub(r'/pre_mlp_layer_norm/' , r'/1/layer_norm/' , SCREAMING_SNAKE_CASE__ ) elif groups[0] == "decoder": a : Dict = re.sub(r'/mlp/' , r'/2/mlp/' , SCREAMING_SNAKE_CASE__ ) a : Union[str, Any] = re.sub(r'/pre_mlp_layer_norm/' , r'/2/layer_norm/' , SCREAMING_SNAKE_CASE__ ) # 2. Convert other classic mappings for old_key, temp_key in MOE_LAYER_NAME_MAPPING.items(): if old_key in new_key: a : int = new_key.replace(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) print(f"""{key} -> {new_key}""" ) a : Any = s_dict.pop(SCREAMING_SNAKE_CASE__ ) if "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: a : Any = s_dict[ '''encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight''' ].T if "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: a : List[str] = s_dict[ '''decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight''' ].T # 3. Take extra care of the EXPERTS layer for key in list(s_dict.keys() ): if "expert" in key: a : Dict = s_dict[key].shape[0] a : int = s_dict[key] for idx in range(SCREAMING_SNAKE_CASE__ ): a : int = expert_weihts[idx] print(f"""{key} -> {key.replace("expert/" , "nested fstring" )}""" ) s_dict.pop(SCREAMING_SNAKE_CASE__ ) return s_dict lowerCAmelCase: int = { 'NUM_ENCODER_LAYERS': 'num_layers', 'NUM_DECODER_LAYERS': 'num_decoder_layers', 'NUM_HEADS': 'num_heads', 'HEAD_DIM': 'd_kv', 'EMBED_DIM': 'd_model', 'MLP_DIM': 'd_ff', 'NUM_SELECTED_EXPERTS': 'num_selected_experts', 'NUM_ENCODER_SPARSE_LAYERS': 'num_sparse_encoder_layers', 'NUM_DECODER_SPARSE_LAYERS': 'num_sparse_decoder_layers', 'dense.MlpBlock.activations': 'feed_forward_proj', } def lowerCamelCase__ ( _A , _A ): # Convert a google style config to the hugging face fromat import regex as re with open(SCREAMING_SNAKE_CASE__ , 'r' ) as f: a : str = f.read() a : Dict = re.findall(r'(.*) = ([0-9.]*)' , SCREAMING_SNAKE_CASE__ ) a : List[str] = {} for param, value in regex_match: if param in GIN_TO_CONFIG_MAPPING and value != "": a : List[Any] = float(SCREAMING_SNAKE_CASE__ ) if '''.''' in value else int(SCREAMING_SNAKE_CASE__ ) a : List[str] = re.findall(r'(.*activations) = \(\'(.*)\',\)' , SCREAMING_SNAKE_CASE__ )[0] a : Optional[int] = str(activation[1] ) a : int = num_experts a : Union[str, Any] = SwitchTransformersConfig(**SCREAMING_SNAKE_CASE__ ) return config def lowerCamelCase__ ( _A , _A , _A=None , _A="./" , _A=8 ): # Initialise PyTorch model print(f"""Loading flax weights from : {flax_checkpoint_path}""" ) a : Optional[int] = checkpoints.load_tax_checkpoint(SCREAMING_SNAKE_CASE__ ) if gin_file is not None: a : List[Any] = convert_gin_to_config(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else: a : List[str] = SwitchTransformersConfig.from_pretrained(SCREAMING_SNAKE_CASE__ ) a : Any = SwitchTransformersForConditionalGeneration(SCREAMING_SNAKE_CASE__ ) a : Optional[int] = flax_params['''target'''] a : Optional[int] = flatten_dict(SCREAMING_SNAKE_CASE__ , sep='/' ) a : Tuple = rename_keys(SCREAMING_SNAKE_CASE__ ) a : int = unflatten_dict(SCREAMING_SNAKE_CASE__ , sep='/' ) # Load the flax params in the PT model load_flax_weights_in_pytorch_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) print(f"""Save PyTorch model to {pytorch_dump_path}""" ) pt_model.save_pretrained(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": lowerCAmelCase: Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--switch_t5x_checkpoint_path', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained SwitchTransformers model. \nThis specifies the' ' model architecture. If not provided, a `gin_file` has to be provided.' ), ) parser.add_argument( '--gin_file', default=None, type=str, required=False, help='Path to the gin config file. If not provided, a `config_file` has to be passed ', ) parser.add_argument( '--config_name', default=None, type=str, required=False, help='Config name of SwitchTransformers model.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output pytorch model.' ) parser.add_argument('--num_experts', default=8, type=int, required=False, help='Number of experts') lowerCAmelCase: Any = parser.parse_args() convert_flax_checkpoint_to_pytorch( args.switch_tax_checkpoint_path, args.config_name, args.gin_file, args.pytorch_dump_folder_path, args.num_experts, )
297
import inspect import unittest from transformers import DPTConfig from transformers.file_utils import is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MODEL_MAPPING, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel from transformers.models.dpt.modeling_dpt import DPT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class UpperCAmelCase_ : """simple docstring""" def __init__( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=2 , SCREAMING_SNAKE_CASE_=32 , SCREAMING_SNAKE_CASE_=16 , SCREAMING_SNAKE_CASE_=3 , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=32 , SCREAMING_SNAKE_CASE_=4 , SCREAMING_SNAKE_CASE_=[0, 1, 2, 3] , SCREAMING_SNAKE_CASE_=4 , SCREAMING_SNAKE_CASE_=37 , SCREAMING_SNAKE_CASE_="gelu" , SCREAMING_SNAKE_CASE_=0.1 , SCREAMING_SNAKE_CASE_=0.1 , SCREAMING_SNAKE_CASE_=0.02 , SCREAMING_SNAKE_CASE_=3 , SCREAMING_SNAKE_CASE_=[1, 384, 24, 24] , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=None , ) -> int: UpperCamelCase :List[Any] = parent UpperCamelCase :List[str] = batch_size UpperCamelCase :Optional[Any] = image_size UpperCamelCase :Optional[Any] = patch_size UpperCamelCase :Optional[Any] = num_channels UpperCamelCase :Union[str, Any] = is_training UpperCamelCase :Dict = use_labels UpperCamelCase :List[Any] = hidden_size UpperCamelCase :Optional[int] = num_hidden_layers UpperCamelCase :Any = backbone_out_indices UpperCamelCase :int = num_attention_heads UpperCamelCase :Union[str, Any] = intermediate_size UpperCamelCase :List[str] = hidden_act UpperCamelCase :Optional[int] = hidden_dropout_prob UpperCamelCase :int = attention_probs_dropout_prob UpperCamelCase :Optional[Any] = initializer_range UpperCamelCase :List[Any] = num_labels UpperCamelCase :Any = backbone_featmap_shape UpperCamelCase :Optional[int] = scope UpperCamelCase :Optional[int] = is_hybrid # sequence length of DPT = num_patches + 1 (we add 1 for the [CLS] token) UpperCamelCase :Tuple = (image_size // patch_size) ** 2 UpperCamelCase :int = num_patches + 1 def UpperCAmelCase ( self ) -> str: UpperCamelCase :Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase :int = None if self.use_labels: UpperCamelCase :str = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) UpperCamelCase :Any = self.get_config() return config, pixel_values, labels def UpperCAmelCase ( self ) -> Tuple: UpperCamelCase :Tuple = { '''global_padding''': '''same''', '''layer_type''': '''bottleneck''', '''depths''': [3, 4, 9], '''out_features''': ['''stage1''', '''stage2''', '''stage3'''], '''embedding_dynamic_padding''': True, '''hidden_sizes''': [96, 192, 384, 768], '''num_groups''': 2, } return DPTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , backbone_out_indices=self.backbone_out_indices , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=SCREAMING_SNAKE_CASE_ , initializer_range=self.initializer_range , is_hybrid=self.is_hybrid , backbone_config=SCREAMING_SNAKE_CASE_ , backbone_featmap_shape=self.backbone_featmap_shape , ) def UpperCAmelCase ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Optional[int]: UpperCamelCase :Optional[int] = DPTModel(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() UpperCamelCase :Optional[int] = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: UpperCamelCase :Tuple = self.num_labels UpperCamelCase :Any = DPTForDepthEstimation(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() UpperCamelCase :Union[str, Any] = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.predicted_depth.shape , (self.batch_size, self.image_size, self.image_size) ) def UpperCAmelCase ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Tuple: UpperCamelCase :int = self.num_labels UpperCamelCase :str = DPTForSemanticSegmentation(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() UpperCamelCase :List[str] = model(SCREAMING_SNAKE_CASE_ , labels=SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def UpperCAmelCase ( self ) -> Dict: UpperCamelCase :List[Any] = self.prepare_config_and_inputs() UpperCamelCase , UpperCamelCase , UpperCamelCase :Optional[Any] = config_and_inputs UpperCamelCase :List[Any] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class UpperCAmelCase_ ( lowercase, lowercase, unittest.TestCase ): """simple docstring""" UpperCamelCase_ : Tuple =(DPTModel, DPTForDepthEstimation, DPTForSemanticSegmentation) if is_torch_available() else () UpperCamelCase_ : Optional[Any] =( { 'depth-estimation': DPTForDepthEstimation, 'feature-extraction': DPTModel, 'image-segmentation': DPTForSemanticSegmentation, } if is_torch_available() else {} ) UpperCamelCase_ : List[Any] =False UpperCamelCase_ : Optional[int] =False UpperCamelCase_ : Union[str, Any] =False def UpperCAmelCase ( self ) -> int: UpperCamelCase :Optional[Any] = DPTModelTester(self ) UpperCamelCase :List[Any] = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE_ , has_text_modality=SCREAMING_SNAKE_CASE_ , hidden_size=37 ) def UpperCAmelCase ( self ) -> Union[str, Any]: self.config_tester.run_common_tests() @unittest.skip(reason='''DPT does not use inputs_embeds''' ) def UpperCAmelCase ( self ) -> int: pass def UpperCAmelCase ( self ) -> Optional[int]: UpperCamelCase , UpperCamelCase :int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase :Union[str, Any] = model_class(SCREAMING_SNAKE_CASE_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCamelCase :Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(SCREAMING_SNAKE_CASE_ , nn.Linear ) ) def UpperCAmelCase ( self ) -> int: UpperCamelCase , UpperCamelCase :Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase :Optional[Any] = model_class(SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase :Tuple = [*signature.parameters.keys()] UpperCamelCase :Any = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase ( self ) -> Optional[Any]: UpperCamelCase :Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase ( self ) -> List[Any]: UpperCamelCase :Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_depth_estimation(*SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase ( self ) -> Tuple: UpperCamelCase :int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase ( self ) -> Any: for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue UpperCamelCase , UpperCamelCase :Dict = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase :int = True if model_class in get_values(SCREAMING_SNAKE_CASE_ ): continue UpperCamelCase :Union[str, Any] = model_class(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.train() UpperCamelCase :Union[str, Any] = self._prepare_for_class(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , return_labels=SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Dict = model(**SCREAMING_SNAKE_CASE_ ).loss loss.backward() def UpperCAmelCase ( self ) -> Optional[int]: for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue UpperCamelCase , UpperCamelCase :List[str] = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase :Union[str, Any] = False UpperCamelCase :Dict = True if model_class in get_values(SCREAMING_SNAKE_CASE_ ) or not model_class.supports_gradient_checkpointing: continue UpperCamelCase :Tuple = model_class(SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.gradient_checkpointing_enable() model.train() UpperCamelCase :List[Any] = self._prepare_for_class(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , return_labels=SCREAMING_SNAKE_CASE_ ) UpperCamelCase :List[str] = model(**SCREAMING_SNAKE_CASE_ ).loss loss.backward() def UpperCAmelCase ( self ) -> Dict: UpperCamelCase , UpperCamelCase :int = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase :Dict = _config_zero_init(SCREAMING_SNAKE_CASE_ ) for model_class in self.all_model_classes: UpperCamelCase :Tuple = model_class(config=SCREAMING_SNAKE_CASE_ ) # Skip the check for the backbone UpperCamelCase :List[str] = [] for name, module in model.named_modules(): if module.__class__.__name__ == "DPTViTHybridEmbeddings": UpperCamelCase :Tuple = [F'''{name}.{key}''' for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def UpperCAmelCase ( self ) -> Tuple: pass @slow def UpperCAmelCase ( self ) -> Any: for model_name in DPT_PRETRAINED_MODEL_ARCHIVE_LIST[1:]: UpperCamelCase :int = DPTModel.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase ( self ) -> List[Any]: # We do this test only for DPTForDepthEstimation since it is the only model that uses readout_type UpperCamelCase , UpperCamelCase :int = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase :Optional[Any] = '''add''' with self.assertRaises(SCREAMING_SNAKE_CASE_ ): UpperCamelCase :int = DPTForDepthEstimation(SCREAMING_SNAKE_CASE_ ) def _A ( ): UpperCamelCase :List[Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision @slow class UpperCAmelCase_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase ( self ) -> str: UpperCamelCase :Any = DPTImageProcessor.from_pretrained('''Intel/dpt-hybrid-midas''' ) UpperCamelCase :int = DPTForDepthEstimation.from_pretrained('''Intel/dpt-hybrid-midas''' ).to(SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Any = prepare_img() UpperCamelCase :Union[str, Any] = image_processor(images=SCREAMING_SNAKE_CASE_ , return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE_ ) # forward pass with torch.no_grad(): UpperCamelCase :Union[str, Any] = model(**SCREAMING_SNAKE_CASE_ ) UpperCamelCase :Optional[int] = outputs.predicted_depth # verify the predicted depth UpperCamelCase :List[str] = torch.Size((1, 384, 384) ) self.assertEqual(predicted_depth.shape , SCREAMING_SNAKE_CASE_ ) UpperCamelCase :str = torch.tensor( [[[5.6437, 5.6146, 5.6511], [5.4371, 5.5649, 5.5958], [5.5215, 5.5184, 5.5293]]] ).to(SCREAMING_SNAKE_CASE_ ) self.assertTrue(torch.allclose(outputs.predicted_depth[:3, :3, :3] / 100 , SCREAMING_SNAKE_CASE_ , atol=1e-4 ) )
259
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __UpperCamelCase = { '''configuration_roformer''': ['''ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''RoFormerConfig''', '''RoFormerOnnxConfig'''], '''tokenization_roformer''': ['''RoFormerTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = ['''RoFormerTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''RoFormerForCausalLM''', '''RoFormerForMaskedLM''', '''RoFormerForMultipleChoice''', '''RoFormerForQuestionAnswering''', '''RoFormerForSequenceClassification''', '''RoFormerForTokenClassification''', '''RoFormerLayer''', '''RoFormerModel''', '''RoFormerPreTrainedModel''', '''load_tf_weights_in_roformer''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFRoFormerForCausalLM''', '''TFRoFormerForMaskedLM''', '''TFRoFormerForMultipleChoice''', '''TFRoFormerForQuestionAnswering''', '''TFRoFormerForSequenceClassification''', '''TFRoFormerForTokenClassification''', '''TFRoFormerLayer''', '''TFRoFormerModel''', '''TFRoFormerPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase = [ '''FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FlaxRoFormerForMaskedLM''', '''FlaxRoFormerForMultipleChoice''', '''FlaxRoFormerForQuestionAnswering''', '''FlaxRoFormerForSequenceClassification''', '''FlaxRoFormerForTokenClassification''', '''FlaxRoFormerModel''', '''FlaxRoFormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_roformer import ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, RoFormerConfig, RoFormerOnnxConfig from .tokenization_roformer import RoFormerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roformer_fast import RoFormerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roformer import ( ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, RoFormerForCausalLM, RoFormerForMaskedLM, RoFormerForMultipleChoice, RoFormerForQuestionAnswering, RoFormerForSequenceClassification, RoFormerForTokenClassification, RoFormerLayer, RoFormerModel, RoFormerPreTrainedModel, load_tf_weights_in_roformer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roformer import ( TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerLayer, TFRoFormerModel, TFRoFormerPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roformer import ( FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, FlaxRoFormerPreTrainedModel, ) else: import sys __UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
362
"""simple docstring""" from __future__ import annotations def lowercase (SCREAMING_SNAKE_CASE_ : list[int] ) -> bool: return len(set(SCREAMING_SNAKE_CASE_ ) ) == len(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": import doctest doctest.testmod()
38
0
'''simple docstring''' from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers __a = [ "python", "tqdm", "regex", "requests", "packaging", "filelock", "numpy", "tokenizers", "huggingface-hub", "safetensors", "accelerate", "pyyaml", ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(F"can't find {pkg} in {deps.keys()}, check dependency_versions_table.py") def __snake_case( _lowerCAmelCase , _lowerCAmelCase=None ) -> int: require_version(deps[pkg] , _lowerCAmelCase )
35
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { "vinvino02/glpn-kitti": "https://huggingface.co/vinvino02/glpn-kitti/resolve/main/config.json", # See all GLPN models at https://huggingface.co/models?filter=glpn } class UpperCAmelCase_ ( _a ): """simple docstring""" lowercase = "glpn" def __init__( self : Optional[Any] , snake_case_ : List[str]=3 , snake_case_ : Dict=4 , snake_case_ : List[Any]=[2, 2, 2, 2] , snake_case_ : int=[8, 4, 2, 1] , snake_case_ : List[str]=[32, 64, 160, 256] , snake_case_ : Tuple=[7, 3, 3, 3] , snake_case_ : List[Any]=[4, 2, 2, 2] , snake_case_ : Tuple=[1, 2, 5, 8] , snake_case_ : List[str]=[4, 4, 4, 4] , snake_case_ : Optional[int]="gelu" , snake_case_ : Dict=0.0 , snake_case_ : Union[str, Any]=0.0 , snake_case_ : List[Any]=0.02 , snake_case_ : Tuple=0.1 , snake_case_ : Any=1E-6 , snake_case_ : Dict=64 , snake_case_ : Tuple=10 , snake_case_ : List[Any]=-1 , **snake_case_ : Optional[Any] , ): super().__init__(**snake_case_ ) snake_case__ : Optional[Any] = num_channels snake_case__ : Dict = num_encoder_blocks snake_case__ : Tuple = depths snake_case__ : Union[str, Any] = sr_ratios snake_case__ : Tuple = hidden_sizes snake_case__ : Optional[Any] = patch_sizes snake_case__ : int = strides snake_case__ : List[Any] = mlp_ratios snake_case__ : Optional[int] = num_attention_heads snake_case__ : Dict = hidden_act snake_case__ : int = hidden_dropout_prob snake_case__ : Optional[Any] = attention_probs_dropout_prob snake_case__ : str = initializer_range snake_case__ : List[str] = drop_path_rate snake_case__ : int = layer_norm_eps snake_case__ : Tuple = decoder_hidden_size snake_case__ : List[Any] = max_depth snake_case__ : Dict = head_in_index
35
1
def UpperCamelCase_( lowerCamelCase_ ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError('The given input must be positive' ) # get the generated string sequence _lowercase : Dict = gray_code_sequence_string(lowerCamelCase_ ) # # convert them to integers for i in range(len(lowerCamelCase_ ) ): _lowercase : Dict = int(sequence[i] , 2 ) return sequence def UpperCamelCase_( lowerCamelCase_ ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] _lowercase : List[str] = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits _lowercase : List[Any] = gray_code_sequence_string(bit_count - 1 ) _lowercase : Any = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): _lowercase : Optional[Any] = '0' + smaller_sequence[i] sequence.append(lowerCamelCase_ ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): _lowercase : Tuple = '1' + smaller_sequence[i] sequence.append(lowerCamelCase_ ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
84
import argparse import os import transformers from .convert_slow_tokenizer import SLOW_TO_FAST_CONVERTERS from .utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Dict = {name: getattr(transformers, name + "Fast") for name in SLOW_TO_FAST_CONVERTERS} def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> List[str]: if tokenizer_name is not None and tokenizer_name not in TOKENIZER_CLASSES: raise ValueError(F'''Unrecognized tokenizer name, should be one of {list(TOKENIZER_CLASSES.keys() )}.''' ) if tokenizer_name is None: _lowercase : Any = TOKENIZER_CLASSES else: _lowercase : Tuple = {tokenizer_name: getattr(lowerCamelCase_ , tokenizer_name + 'Fast' )} logger.info(F'''Loading tokenizer classes: {tokenizer_names}''' ) for tokenizer_name in tokenizer_names: _lowercase : Union[str, Any] = TOKENIZER_CLASSES[tokenizer_name] _lowercase : Any = True if checkpoint_name is None: _lowercase : int = list(tokenizer_class.max_model_input_sizes.keys() ) else: _lowercase : List[Any] = [checkpoint_name] logger.info(F'''For tokenizer {tokenizer_class.__class__.__name__} loading checkpoints: {checkpoint_names}''' ) for checkpoint in checkpoint_names: logger.info(F'''Loading {tokenizer_class.__class__.__name__} {checkpoint}''' ) # Load tokenizer _lowercase : Union[str, Any] = tokenizer_class.from_pretrained(lowerCamelCase_ , force_download=lowerCamelCase_ ) # Save fast tokenizer logger.info(F'''Save fast tokenizer to {dump_path} with prefix {checkpoint} add_prefix {add_prefix}''' ) # For organization names we create sub-directories if "/" in checkpoint: _lowercase , _lowercase : str = checkpoint.split('/' ) _lowercase : Any = os.path.join(lowerCamelCase_ , lowerCamelCase_ ) elif add_prefix: _lowercase : Union[str, Any] = checkpoint _lowercase : List[str] = dump_path else: _lowercase : str = None _lowercase : Any = dump_path logger.info(F'''=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}''' ) if checkpoint in list(tokenizer.pretrained_vocab_files_map.values() )[0]: _lowercase : Tuple = list(tokenizer.pretrained_vocab_files_map.values() )[0][checkpoint] _lowercase : List[Any] = file_path.split(lowerCamelCase_ )[-1][0] if next_char == "/": _lowercase : Any = os.path.join(lowerCamelCase_ , lowerCamelCase_ ) _lowercase : List[Any] = None logger.info(F'''=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}''' ) _lowercase : Optional[Any] = tokenizer.save_pretrained( lowerCamelCase_ , legacy_format=lowerCamelCase_ , filename_prefix=lowerCamelCase_ ) logger.info(F'''=> File names {file_names}''' ) for file_name in file_names: if not file_name.endswith('tokenizer.json' ): os.remove(lowerCamelCase_ ) logger.info(F'''=> removing {file_name}''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--dump_path", default=None, type=str, required=True, help="Path to output generated fast tokenizer files." ) parser.add_argument( "--tokenizer_name", default=None, type=str, help=( F"Optional tokenizer type selected in the list of {list(TOKENIZER_CLASSES.keys())}. If not given, will " "download and convert all the checkpoints from AWS." ), ) parser.add_argument( "--checkpoint_name", default=None, type=str, help="Optional checkpoint name. If not given, will download and convert the canonical checkpoints from AWS.", ) parser.add_argument( "--force_download", action="store_true", help="Re-download checkpoints.", ) SCREAMING_SNAKE_CASE : str = parser.parse_args() convert_slow_checkpoint_to_fast(args.tokenizer_name, args.checkpoint_name, args.dump_path, args.force_download)
84
1
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor from torchvision.transforms.functional import InterpolationMode import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, ViTImageProcessor, ViTMAEConfig, ViTMAEForPreTraining, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version _UpperCAmelCase : Union[str, Any] =logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") require_version("""datasets>=1.8.0""", """To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt""") @dataclass class snake_case__: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[str] = field( default="""cifar10""", metadata={"""help""": """Name of a dataset from the datasets package"""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """The column name of the images in the files."""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field(default=UpperCAmelCase__, metadata={"""help""": """A folder containing the training data."""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field(default=UpperCAmelCase__, metadata={"""help""": """A folder containing the validation data."""} ) SCREAMING_SNAKE_CASE__ : Optional[float] = field( default=0.15, metadata={"""help""": """Percent to split off of train for validation."""} ) SCREAMING_SNAKE_CASE__ : Optional[int] = field( default=UpperCAmelCase__, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) }, ) SCREAMING_SNAKE_CASE__ : Optional[int] = field( default=UpperCAmelCase__, metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) }, ) def lowercase_ ( self ) -> Tuple: lowerCAmelCase_ : Union[str, Any] = {} if self.train_dir is not None: lowerCAmelCase_ : int = self.train_dir if self.validation_dir is not None: lowerCAmelCase_ : int = self.validation_dir lowerCAmelCase_ : Optional[Any] = data_files if data_files else None @dataclass class snake_case__: '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = field( default=UpperCAmelCase__, metadata={ """help""": ( """The model checkpoint for weights initialization.Don't set if you want to train a model from scratch.""" ) }, ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """Pretrained config name or path if not the same as model_name_or_path"""} ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={ """help""": ( """Override some existing default config settings when a model is trained from scratch. Example: """ """n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index""" ) }, ) SCREAMING_SNAKE_CASE__ : Optional[str] = field( default=UpperCAmelCase__, metadata={"""help""": """Where do you want to store the pretrained models downloaded from s3"""} ) SCREAMING_SNAKE_CASE__ : str = field( default="""main""", metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""}, ) SCREAMING_SNAKE_CASE__ : str = field(default=UpperCAmelCase__, metadata={"""help""": """Name or path of preprocessor config."""} ) SCREAMING_SNAKE_CASE__ : bool = field( default=UpperCAmelCase__, metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) }, ) SCREAMING_SNAKE_CASE__ : float = field( default=0.75, metadata={"""help""": """The ratio of the number of masked tokens in the input sequence."""} ) SCREAMING_SNAKE_CASE__ : bool = field( default=UpperCAmelCase__, metadata={"""help""": """Whether or not to train with normalized pixel values as target."""} ) @dataclass class snake_case__( UpperCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : float = field( default=1e-3, metadata={"""help""": """Base learning rate: absolute_lr = base_lr * total_batch_size / 256."""} ) def lowerCAmelCase ( lowerCAmelCase_ )-> Dict: lowerCAmelCase_ : Union[str, Any] = torch.stack([example['''pixel_values'''] for example in examples] ) return {"pixel_values": pixel_values} def lowerCAmelCase ( )-> Tuple: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCAmelCase_ : List[Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, CustomTrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[str] = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_mae''' , lowerCAmelCase_ , lowerCAmelCase_ ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() lowerCAmelCase_ : str = training_args.get_process_log_level() logger.setLevel(lowerCAmelCase_ ) transformers.utils.logging.set_verbosity(lowerCAmelCase_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + f"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(f"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. lowerCAmelCase_ : List[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowerCAmelCase_ : Tuple = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. """ '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Initialize our dataset. lowerCAmelCase_ : Union[str, Any] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , data_files=data_args.data_files , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # If we don't have a validation split, split off a percentage of train as validation. lowerCAmelCase_ : Optional[int] = None if '''validation''' in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , lowerCAmelCase_ ) and data_args.train_val_split > 0.0: lowerCAmelCase_ : List[str] = ds['''train'''].train_test_split(data_args.train_val_split ) lowerCAmelCase_ : Optional[int] = split['''train'''] lowerCAmelCase_ : List[str] = split['''test'''] # Load pretrained model and image processor # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowerCAmelCase_ : int = { '''cache_dir''': model_args.cache_dir, '''revision''': model_args.model_revision, '''use_auth_token''': True if model_args.use_auth_token else None, } if model_args.config_name: lowerCAmelCase_ : Optional[int] = ViTMAEConfig.from_pretrained(model_args.config_name , **lowerCAmelCase_ ) elif model_args.model_name_or_path: lowerCAmelCase_ : Union[str, Any] = ViTMAEConfig.from_pretrained(model_args.model_name_or_path , **lowerCAmelCase_ ) else: lowerCAmelCase_ : Any = ViTMAEConfig() logger.warning('''You are instantiating a new config instance from scratch.''' ) if model_args.config_overrides is not None: logger.info(f"""Overriding config: {model_args.config_overrides}""" ) config.update_from_string(model_args.config_overrides ) logger.info(f"""New config: {config}""" ) # adapt config config.update( { '''mask_ratio''': model_args.mask_ratio, '''norm_pix_loss''': model_args.norm_pix_loss, } ) # create image processor if model_args.image_processor_name: lowerCAmelCase_ : Union[str, Any] = ViTImageProcessor.from_pretrained(model_args.image_processor_name , **lowerCAmelCase_ ) elif model_args.model_name_or_path: lowerCAmelCase_ : int = ViTImageProcessor.from_pretrained(model_args.model_name_or_path , **lowerCAmelCase_ ) else: lowerCAmelCase_ : int = ViTImageProcessor() # create model if model_args.model_name_or_path: lowerCAmelCase_ : str = ViTMAEForPreTraining.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path ) , config=lowerCAmelCase_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info('''Training new model from scratch''' ) lowerCAmelCase_ : str = ViTMAEForPreTraining(lowerCAmelCase_ ) if training_args.do_train: lowerCAmelCase_ : Union[str, Any] = ds['''train'''].column_names else: lowerCAmelCase_ : int = ds['''validation'''].column_names if data_args.image_column_name is not None: lowerCAmelCase_ : Dict = data_args.image_column_name elif "image" in column_names: lowerCAmelCase_ : List[str] = '''image''' elif "img" in column_names: lowerCAmelCase_ : Union[str, Any] = '''img''' else: lowerCAmelCase_ : List[str] = column_names[0] # transformations as done in original MAE paper # source: https://github.com/facebookresearch/mae/blob/main/main_pretrain.py if "shortest_edge" in image_processor.size: lowerCAmelCase_ : Optional[int] = image_processor.size['''shortest_edge'''] else: lowerCAmelCase_ : str = (image_processor.size['''height'''], image_processor.size['''width''']) lowerCAmelCase_ : List[str] = Compose( [ Lambda(lambda lowerCAmelCase_ : img.convert('''RGB''' ) if img.mode != "RGB" else img ), RandomResizedCrop(lowerCAmelCase_ , scale=(0.2, 1.0) , interpolation=InterpolationMode.BICUBIC ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean , std=image_processor.image_std ), ] ) def preprocess_images(lowerCAmelCase_ ): lowerCAmelCase_ : Union[str, Any] = [transforms(lowerCAmelCase_ ) for image in examples[image_column_name]] return examples if training_args.do_train: if "train" not in ds: raise ValueError('''--do_train requires a train dataset''' ) if data_args.max_train_samples is not None: lowerCAmelCase_ : List[str] = ds['''train'''].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(lowerCAmelCase_ ) if training_args.do_eval: if "validation" not in ds: raise ValueError('''--do_eval requires a validation dataset''' ) if data_args.max_eval_samples is not None: lowerCAmelCase_ : Optional[Any] = ( ds['''validation'''].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(lowerCAmelCase_ ) # Compute absolute learning rate lowerCAmelCase_ : Tuple = ( training_args.train_batch_size * training_args.gradient_accumulation_steps * training_args.world_size ) if training_args.base_learning_rate is not None: lowerCAmelCase_ : Tuple = training_args.base_learning_rate * total_train_batch_size / 256 # Initialize our trainer lowerCAmelCase_ : Optional[int] = Trainer( model=lowerCAmelCase_ , args=lowerCAmelCase_ , train_dataset=ds['''train'''] if training_args.do_train else None , eval_dataset=ds['''validation'''] if training_args.do_eval else None , tokenizer=lowerCAmelCase_ , data_collator=lowerCAmelCase_ , ) # Training if training_args.do_train: lowerCAmelCase_ : str = None if training_args.resume_from_checkpoint is not None: lowerCAmelCase_ : Any = training_args.resume_from_checkpoint elif last_checkpoint is not None: lowerCAmelCase_ : Any = last_checkpoint lowerCAmelCase_ : List[Any] = trainer.train(resume_from_checkpoint=lowerCAmelCase_ ) trainer.save_model() trainer.log_metrics('''train''' , train_result.metrics ) trainer.save_metrics('''train''' , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: lowerCAmelCase_ : Tuple = trainer.evaluate() trainer.log_metrics('''eval''' , lowerCAmelCase_ ) trainer.save_metrics('''eval''' , lowerCAmelCase_ ) # Write model card and (optionally) push to hub lowerCAmelCase_ : Optional[int] = { '''tasks''': '''masked-auto-encoding''', '''dataset''': data_args.dataset_name, '''tags''': ['''masked-auto-encoding'''], } if training_args.push_to_hub: trainer.push_to_hub(**lowerCAmelCase_ ) else: trainer.create_model_card(**lowerCAmelCase_ ) def lowerCAmelCase ( lowerCAmelCase_ )-> str: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
262
def lowerCAmelCase ( lowerCAmelCase_ = 10**9 )-> int: lowerCAmelCase_ : List[Any] = 1 lowerCAmelCase_ : Optional[int] = 2 lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : str = 0 lowerCAmelCase_ : str = 0 while perimeter <= max_perimeter: perimeters_sum += perimeter prev_value += 2 * value value += prev_value lowerCAmelCase_ : Any = 2 * value + 2 if i % 2 == 0 else 2 * value - 2 i += 1 return perimeters_sum if __name__ == "__main__": print(f"""{solution() = }""")
262
1
'''simple docstring''' import itertools import os import random import tempfile import unittest import numpy as np from transformers import TvltFeatureExtractor, is_datasets_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch if is_datasets_available(): from datasets import load_dataset _SCREAMING_SNAKE_CASE : int = random.Random() def UpperCamelCase_( snake_case : Optional[Any] , snake_case : int=1.0 , snake_case : Optional[int]=None , snake_case : List[Any]=None ): '''simple docstring''' if rng is None: snake_case_ = global_rng snake_case_ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class _snake_case ( unittest.TestCase ): def __init__( self , a__ , a__=7 , a__=400 , a__=2_000 , a__=2_048 , a__=128 , a__=1 , a__=512 , a__=30 , a__=44_100 , ) -> str: '''simple docstring''' snake_case_ = parent snake_case_ = batch_size snake_case_ = min_seq_length snake_case_ = max_seq_length snake_case_ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) snake_case_ = spectrogram_length snake_case_ = feature_size snake_case_ = num_audio_channels snake_case_ = hop_length snake_case_ = chunk_length snake_case_ = sampling_rate def lowerCAmelCase__ ( self ) -> int: '''simple docstring''' return { "spectrogram_length": self.spectrogram_length, "feature_size": self.feature_size, "num_audio_channels": self.num_audio_channels, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "sampling_rate": self.sampling_rate, } def lowerCAmelCase__ ( self , a__=False , a__=False ) -> Optional[Any]: '''simple docstring''' def _flatten(a__ ): return list(itertools.chain(*A_ ) ) if equal_length: snake_case_ = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size snake_case_ = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: snake_case_ = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class _snake_case ( _SCREAMING_SNAKE_CASE , unittest.TestCase ): lowerCAmelCase_ : Any = TvltFeatureExtractor def lowerCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' snake_case_ = TvltFeatureExtractionTester(self ) def lowerCAmelCase__ ( self ) -> List[str]: '''simple docstring''' snake_case_ = self.feature_extraction_class(**self.feat_extract_dict ) self.assertTrue(hasattr(A_ , "spectrogram_length" ) ) self.assertTrue(hasattr(A_ , "feature_size" ) ) self.assertTrue(hasattr(A_ , "num_audio_channels" ) ) self.assertTrue(hasattr(A_ , "hop_length" ) ) self.assertTrue(hasattr(A_ , "chunk_length" ) ) self.assertTrue(hasattr(A_ , "sampling_rate" ) ) def lowerCAmelCase__ ( self ) -> Optional[int]: '''simple docstring''' snake_case_ = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: snake_case_ = feat_extract_first.save_pretrained(A_ )[0] check_json_file_has_correct_format(A_ ) snake_case_ = self.feature_extraction_class.from_pretrained(A_ ) snake_case_ = feat_extract_first.to_dict() snake_case_ = feat_extract_second.to_dict() snake_case_ = dict_first.pop("mel_filters" ) snake_case_ = dict_second.pop("mel_filters" ) self.assertTrue(np.allclose(A_ , A_ ) ) self.assertEqual(A_ , A_ ) def lowerCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' snake_case_ = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: snake_case_ = os.path.join(A_ , "feat_extract.json" ) feat_extract_first.to_json_file(A_ ) snake_case_ = self.feature_extraction_class.from_json_file(A_ ) snake_case_ = feat_extract_first.to_dict() snake_case_ = feat_extract_second.to_dict() snake_case_ = dict_first.pop("mel_filters" ) snake_case_ = dict_second.pop("mel_filters" ) self.assertTrue(np.allclose(A_ , A_ ) ) self.assertEqual(A_ , A_ ) def lowerCAmelCase__ ( self ) -> Union[str, Any]: '''simple docstring''' snake_case_ = self.feature_extraction_class(**self.feat_extract_dict ) # create three inputs of length 800, 1000, and 1200 snake_case_ = [floats_list((1, x) )[0] for x in range(800 , 1_400 , 200 )] snake_case_ = [np.asarray(A_ ) for speech_input in speech_inputs] # Test not batched input snake_case_ = feature_extractor(np_speech_inputs[0] , return_tensors="np" , sampling_rate=44_100 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test batched snake_case_ = feature_extractor(A_ , return_tensors="np" , sampling_rate=44_100 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test audio masking snake_case_ = feature_extractor( A_ , return_tensors="np" , sampling_rate=44_100 , mask_audio=A_ ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) # Test 2-D numpy arrays are batched. snake_case_ = [floats_list((1, x) )[0] for x in (800, 800, 800)] snake_case_ = np.asarray(A_ ) snake_case_ = feature_extractor(A_ , return_tensors="np" , sampling_rate=44_100 ).audio_values self.assertTrue(encoded_audios.ndim == 4 ) self.assertTrue(encoded_audios.shape[-1] == feature_extractor.feature_size ) self.assertTrue(encoded_audios.shape[-2] <= feature_extractor.spectrogram_length ) self.assertTrue(encoded_audios.shape[-3] == feature_extractor.num_channels ) def lowerCAmelCase__ ( self , a__ ) -> Optional[int]: '''simple docstring''' snake_case_ = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech snake_case_ = ds.sort("id" ).select(range(A_ ) )[:num_samples]["audio"] return [x["array"] for x in speech_samples] def lowerCAmelCase__ ( self ) -> Optional[Any]: '''simple docstring''' snake_case_ = self._load_datasamples(1 ) snake_case_ = TvltFeatureExtractor() snake_case_ = feature_extractor(A_ , return_tensors="pt" ).audio_values self.assertEquals(audio_values.shape , (1, 1, 192, 128) ) snake_case_ = torch.tensor([[-0.3_0_3_2, -0.2_7_0_8], [-0.4_4_3_4, -0.4_0_0_7]] ) self.assertTrue(torch.allclose(audio_values[0, 0, :2, :2] , A_ , atol=1e-4 ) )
371
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device _SCREAMING_SNAKE_CASE : Any = False class _snake_case ( unittest.TestCase ): pass @slow @require_torch_gpu class _snake_case ( unittest.TestCase ): def lowerCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' snake_case_ = VersatileDiffusionImageVariationPipeline.from_pretrained("shi-labs/versatile-diffusion" ) pipe.to(a__ ) pipe.set_progress_bar_config(disable=a__ ) snake_case_ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg" ) snake_case_ = torch.manual_seed(0 ) snake_case_ = pipe( image=a__ , generator=a__ , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" , ).images snake_case_ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) snake_case_ = np.array([0.0_4_4_1, 0.0_4_6_9, 0.0_5_0_7, 0.0_5_7_5, 0.0_6_3_2, 0.0_6_5_0, 0.0_8_6_5, 0.0_9_0_9, 0.0_9_4_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
92
0
"""simple docstring""" import gc import unittest from parameterized import parameterized from diffusers import FlaxUNetaDConditionModel from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp @slow @require_flax class _UpperCAmelCase ( unittest.TestCase): def _snake_case ( self : Any , lowercase_ : Optional[int] , lowercase_ : str ): return f"gaussian_noise_s={seed}_shape={'_'.join([str(lowercase_ ) for s in shape] )}.npy" def _snake_case ( self : Optional[Any] ): # clean up the VRAM after each test super().tearDown() gc.collect() def _snake_case ( self : Optional[int] , lowercase_ : Union[str, Any]=0 , lowercase_ : Optional[Any]=(4, 4, 64, 64) , lowercase_ : Any=False ): snake_case_ : List[str] = jnp.bfloataa if fpaa else jnp.floataa snake_case_ : List[Any] = jnp.array(load_hf_numpy(self.get_file_format(lowercase_ , lowercase_ ) ) , dtype=lowercase_ ) return image def _snake_case ( self : List[Any] , lowercase_ : Optional[Any]=False , lowercase_ : Optional[int]="CompVis/stable-diffusion-v1-4" ): snake_case_ : int = jnp.bfloataa if fpaa else jnp.floataa snake_case_ : Any = """bf16""" if fpaa else None snake_case_ : Union[str, Any] = FlaxUNetaDConditionModel.from_pretrained( lowercase_ , subfolder='''unet''' , dtype=lowercase_ , revision=lowercase_ ) return model, params def _snake_case ( self : int , lowercase_ : int=0 , lowercase_ : List[str]=(4, 77, 768) , lowercase_ : List[Any]=False ): snake_case_ : int = jnp.bfloataa if fpaa else jnp.floataa snake_case_ : Any = jnp.array(load_hf_numpy(self.get_file_format(lowercase_ , lowercase_ ) ) , dtype=lowercase_ ) return hidden_states @parameterized.expand( [ # fmt: off [83, 4, [-0.23_23, -0.13_04, 0.08_13, -0.30_93, -0.09_19, -0.15_71, -0.11_25, -0.58_06]], [17, 0.55, [-0.08_31, -0.24_43, 0.09_01, -0.09_19, 0.33_96, 0.01_03, -0.37_43, 0.07_01]], [8, 0.89, [-0.48_63, 0.08_59, 0.08_75, -0.16_58, 0.91_99, -0.01_14, 0.48_39, 0.46_39]], [3, 1000, [-0.56_49, 0.24_02, -0.55_18, 0.12_48, 1.13_28, -0.24_43, -0.03_25, -1.00_78]], # fmt: on ] ) def _snake_case ( self : List[str] , lowercase_ : Any , lowercase_ : Tuple , lowercase_ : Optional[Any] ): snake_case_ : int = self.get_unet_model(model_id='''CompVis/stable-diffusion-v1-4''' , fpaa=lowercase_ ) snake_case_ : str = self.get_latents(lowercase_ , fpaa=lowercase_ ) snake_case_ : Optional[Any] = self.get_encoder_hidden_states(lowercase_ , fpaa=lowercase_ ) snake_case_ : List[str] = model.apply( {'''params''': params} , lowercase_ , jnp.array(lowercase_ , dtype=jnp.intaa ) , encoder_hidden_states=lowercase_ , ).sample assert sample.shape == latents.shape snake_case_ : Optional[int] = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) snake_case_ : Union[str, Any] = jnp.array(lowercase_ , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, in the same hardware assert jnp.allclose(lowercase_ , lowercase_ , atol=1E-2 ) @parameterized.expand( [ # fmt: off [83, 4, [0.15_14, 0.08_07, 0.16_24, 0.10_16, -0.18_96, 0.02_63, 0.06_77, 0.23_10]], [17, 0.55, [0.11_64, -0.02_16, 0.01_70, 0.15_89, -0.31_20, 0.10_05, -0.05_81, -0.14_58]], [8, 0.89, [-0.17_58, -0.01_69, 0.10_04, -0.14_11, 0.13_12, 0.11_03, -0.19_96, 0.21_39]], [3, 1000, [0.12_14, 0.03_52, -0.07_31, -0.15_62, -0.09_94, -0.09_06, -0.23_40, -0.05_39]], # fmt: on ] ) def _snake_case ( self : Optional[int] , lowercase_ : int , lowercase_ : str , lowercase_ : Optional[int] ): snake_case_ : Dict = self.get_unet_model(model_id='''stabilityai/stable-diffusion-2''' , fpaa=lowercase_ ) snake_case_ : Union[str, Any] = self.get_latents(lowercase_ , shape=(4, 4, 96, 96) , fpaa=lowercase_ ) snake_case_ : List[Any] = self.get_encoder_hidden_states(lowercase_ , shape=(4, 77, 1024) , fpaa=lowercase_ ) snake_case_ : str = model.apply( {'''params''': params} , lowercase_ , jnp.array(lowercase_ , dtype=jnp.intaa ) , encoder_hidden_states=lowercase_ , ).sample assert sample.shape == latents.shape snake_case_ : List[str] = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten()) ) , dtype=jnp.floataa ) snake_case_ : Dict = jnp.array(lowercase_ , dtype=jnp.floataa ) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, on the same hardware assert jnp.allclose(lowercase_ , lowercase_ , atol=1E-2 )
264
import math def SCREAMING_SNAKE_CASE__ ( lowercase ) -> int: if not isinstance(lowercase ,lowercase ): snake_case : List[Any] = f"""Input value of [number={number}] must be an integer""" raise TypeError(lowercase ) if number < 1: snake_case : int = f"""Input value of [number={number}] must be > 0""" raise ValueError(lowercase ) elif number == 1: return 3 elif number == 2: return 5 else: snake_case : Any = int(math.log(number // 3 ,2 ) ) + 2 snake_case : List[Any] = [3, 5] snake_case : Optional[int] = 2 snake_case : Union[str, Any] = 3 for block in range(1 ,lowercase ): for _ in range(lowercase ): proth_list.append(2 ** (block + 1) + proth_list[proth_index - 1] ) proth_index += 1 increment *= 2 return proth_list[number - 1] if __name__ == "__main__": import doctest doctest.testmod() for number in range(1_1): lowerCamelCase : Optional[Any] = 0 try: lowerCamelCase : Tuple = proth(number) except ValueError: print(f"""ValueError: there is no {number}th Proth number""") continue print(f"""The {number}th Proth number: {value}""")
124
0
"""simple docstring""" import baseaa def lowerCamelCase_ (UpperCamelCase__ : str ): return baseaa.baaencode(string.encode('''utf-8''' ) ) def lowerCamelCase_ (UpperCamelCase__ : bytes ): return baseaa.baadecode(UpperCamelCase__ ).decode('''utf-8''' ) if __name__ == "__main__": _lowerCAmelCase :Any = 'Hello World!' _lowerCAmelCase :List[str] = baseaa_encode(test) print(encoded) _lowerCAmelCase :Any = baseaa_decode(encoded) print(decoded)
363
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax.numpy as jnp from jax import random from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .scheduling_utils_flax import FlaxSchedulerMixin @flax.struct.dataclass class _UpperCAmelCase : '''simple docstring''' a__ =None a__ =None a__ =None # sigma(t_i) @classmethod def __lowerCAmelCase ( cls ) -> Tuple: return cls() @dataclass class _UpperCAmelCase ( a ): '''simple docstring''' a__ =42 a__ =42 a__ =42 class _UpperCAmelCase ( a ,a ): '''simple docstring''' @property def __lowerCAmelCase ( self ) -> Optional[int]: return True @register_to_config def __init__( self , A = 0.02 , A = 1_0_0 , A = 1.007 , A = 8_0 , A = 0.05 , A = 5_0 , ) -> Any: pass def __lowerCAmelCase ( self ) -> int: return KarrasVeSchedulerState.create() def __lowerCAmelCase ( self , A , A , A = () ) -> KarrasVeSchedulerState: _UpperCAmelCase : Tuple = jnp.arange(0 , A )[::-1].copy() _UpperCAmelCase : Dict = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in timesteps ] return state.replace( num_inference_steps=A , schedule=jnp.array(A , dtype=jnp.floataa ) , timesteps=A , ) def __lowerCAmelCase ( self , A , A , A , A , ) -> Tuple[jnp.ndarray, float]: if self.config.s_min <= sigma <= self.config.s_max: _UpperCAmelCase : Optional[Any] = min(self.config.s_churn / state.num_inference_steps , 2**0.5 - 1 ) else: _UpperCAmelCase : Tuple = 0 # sample eps ~ N(0, S_noise^2 * I) _UpperCAmelCase : Any = random.split(A , num=1 ) _UpperCAmelCase : Optional[Any] = self.config.s_noise * random.normal(key=A , shape=sample.shape ) _UpperCAmelCase : str = sigma + gamma * sigma _UpperCAmelCase : str = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def __lowerCAmelCase ( self , A , A , A , A , A , A = True , ) -> Union[FlaxKarrasVeOutput, Tuple]: _UpperCAmelCase : Dict = sample_hat + sigma_hat * model_output _UpperCAmelCase : int = (sample_hat - pred_original_sample) / sigma_hat _UpperCAmelCase : Any = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=A , derivative=A , state=A ) def __lowerCAmelCase ( self , A , A , A , A , A , A , A , A = True , ) -> Union[FlaxKarrasVeOutput, Tuple]: _UpperCAmelCase : Union[str, Any] = sample_prev + sigma_prev * model_output _UpperCAmelCase : List[Any] = (sample_prev - pred_original_sample) / sigma_prev _UpperCAmelCase : Union[str, Any] = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative, state) return FlaxKarrasVeOutput(prev_sample=A , derivative=A , state=A ) def __lowerCAmelCase ( self , A , A , A , A ) -> Any: raise NotImplementedError()
68
0
from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL import torch from transformers import CLIPImageProcessor, CLIPVisionModel from ...models import PriorTransformer from ...pipelines import DiffusionPipeline from ...schedulers import HeunDiscreteScheduler from ...utils import ( BaseOutput, is_accelerate_available, logging, randn_tensor, replace_example_docstring, ) from .renderer import ShapERenderer UpperCAmelCase__ = logging.get_logger(__name__) # pylint: disable=invalid-name UpperCAmelCase__ = "\n Examples:\n ```py\n >>> from PIL import Image\n >>> import torch\n >>> from diffusers import DiffusionPipeline\n >>> from diffusers.utils import export_to_gif, load_image\n\n >>> device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n\n >>> repo = \"openai/shap-e-img2img\"\n >>> pipe = DiffusionPipeline.from_pretrained(repo, torch_dtype=torch.float16)\n >>> pipe = pipe.to(device)\n\n >>> guidance_scale = 3.0\n >>> image_url = \"https://hf.co/datasets/diffusers/docs-images/resolve/main/shap-e/corgi.png\"\n >>> image = load_image(image_url).convert(\"RGB\")\n\n >>> images = pipe(\n ... image,\n ... guidance_scale=guidance_scale,\n ... num_inference_steps=64,\n ... frame_size=256,\n ... ).images\n\n >>> gif_path = export_to_gif(images[0], \"corgi_3d.gif\")\n ```\n" @dataclass class lowercase_ ( lowercase ): '''simple docstring''' __snake_case = 42 class lowercase_ ( lowercase ): '''simple docstring''' def __init__( self : int , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : str , ) ->Optional[Any]: """simple docstring""" super().__init__() self.register_modules( prior=__UpperCAmelCase , image_encoder=__UpperCAmelCase , image_processor=__UpperCAmelCase , scheduler=__UpperCAmelCase , renderer=__UpperCAmelCase , ) def __lowerCAmelCase ( self : List[str] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : int , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Dict ) ->str: """simple docstring""" if latents is None: a = randn_tensor(__UpperCAmelCase , generator=__UpperCAmelCase , device=__UpperCAmelCase , dtype=__UpperCAmelCase ) else: if latents.shape != shape: raise ValueError(F"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) a = latents.to(__UpperCAmelCase ) a = latents * scheduler.init_noise_sigma return latents def __lowerCAmelCase ( self : List[Any] , __UpperCAmelCase : Any=0 ) ->str: """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) a = torch.device(F"""cuda:{gpu_id}""" ) a = [self.image_encoder, self.prior] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(__UpperCAmelCase , __UpperCAmelCase ) @property def __lowerCAmelCase ( self : Optional[int] ) ->int: """simple docstring""" if self.device != torch.device('''meta''' ) or not hasattr(self.image_encoder , '''_hf_hook''' ): return self.device for module in self.image_encoder.modules(): if ( hasattr(__UpperCAmelCase , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device def __lowerCAmelCase ( self : Optional[int] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Dict , __UpperCAmelCase : Dict , ) ->List[str]: """simple docstring""" if isinstance(__UpperCAmelCase , __UpperCAmelCase ) and isinstance(image[0] , torch.Tensor ): a = torch.cat(__UpperCAmelCase , axis=0 ) if image[0].ndim == 4 else torch.stack(__UpperCAmelCase , axis=0 ) if not isinstance(__UpperCAmelCase , torch.Tensor ): a = self.image_processor(__UpperCAmelCase , return_tensors='''pt''' ).pixel_values[0].unsqueeze(0 ) a = image.to(dtype=self.image_encoder.dtype , device=__UpperCAmelCase ) a = self.image_encoder(__UpperCAmelCase )['''last_hidden_state'''] a = image_embeds[:, 1:, :].contiguous() # batch_size, dim, 256 a = image_embeds.repeat_interleave(__UpperCAmelCase , dim=0 ) if do_classifier_free_guidance: a = torch.zeros_like(__UpperCAmelCase ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes a = torch.cat([negative_image_embeds, image_embeds] ) return image_embeds @torch.no_grad() @replace_example_docstring(__UpperCAmelCase ) def __call__( self : Optional[Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Any = 1 , __UpperCAmelCase : str = 25 , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : Union[str, Any] = 4.0 , __UpperCAmelCase : Tuple = 64 , __UpperCAmelCase : str = "pil" , __UpperCAmelCase : int = True , ) ->Dict: """simple docstring""" if isinstance(__UpperCAmelCase , PIL.Image.Image ): a = 1 elif isinstance(__UpperCAmelCase , torch.Tensor ): a = image.shape[0] elif isinstance(__UpperCAmelCase , __UpperCAmelCase ) and isinstance(image[0] , (torch.Tensor, PIL.Image.Image) ): a = len(__UpperCAmelCase ) else: raise ValueError( F"""`image` has to be of type `PIL.Image.Image`, `torch.Tensor`, `List[PIL.Image.Image]` or `List[torch.Tensor]` but is {type(__UpperCAmelCase )}""" ) a = self._execution_device a = batch_size * num_images_per_prompt a = guidance_scale > 1.0 a = self._encode_image(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) # prior self.scheduler.set_timesteps(__UpperCAmelCase , device=__UpperCAmelCase ) a = self.scheduler.timesteps a = self.prior.config.num_embeddings a = self.prior.config.embedding_dim a = self.prepare_latents( (batch_size, num_embeddings * embedding_dim) , image_embeds.dtype , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , self.scheduler , ) # YiYi notes: for testing only to match ldm, we can directly create a latents with desired shape: batch_size, num_embeddings, embedding_dim a = latents.reshape(latents.shape[0] , __UpperCAmelCase , __UpperCAmelCase ) for i, t in enumerate(self.progress_bar(__UpperCAmelCase ) ): # expand the latents if we are doing classifier free guidance a = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents a = self.scheduler.scale_model_input(__UpperCAmelCase , __UpperCAmelCase ) a = self.prior( __UpperCAmelCase , timestep=__UpperCAmelCase , proj_embedding=__UpperCAmelCase , ).predicted_image_embedding # remove the variance a , a = noise_pred.split( scaled_model_input.shape[2] , dim=2 ) # batch_size, num_embeddings, embedding_dim if do_classifier_free_guidance is not None: a , a = noise_pred.chunk(2 ) a = noise_pred_uncond + guidance_scale * (noise_pred - noise_pred_uncond) a = self.scheduler.step( __UpperCAmelCase , timestep=__UpperCAmelCase , sample=__UpperCAmelCase , ).prev_sample if output_type == "latent": return ShapEPipelineOutput(images=__UpperCAmelCase ) a = [] for i, latent in enumerate(__UpperCAmelCase ): print() a = self.renderer.decode( latent[None, :] , __UpperCAmelCase , size=__UpperCAmelCase , ray_batch_size=4_096 , n_coarse_samples=64 , n_fine_samples=128 , ) images.append(__UpperCAmelCase ) a = torch.stack(__UpperCAmelCase ) if output_type not in ["np", "pil"]: raise ValueError(F"""Only the output types `pil` and `np` are supported not output_type={output_type}""" ) a = images.cpu().numpy() if output_type == "pil": a = [self.numpy_to_pil(__UpperCAmelCase ) for image in images] # Offload last model to CPU if hasattr(self , '''final_offload_hook''' ) and self.final_offload_hook is not None: self.final_offload_hook.offload() if not return_dict: return (images,) return ShapEPipelineOutput(images=__UpperCAmelCase )
0
"""simple docstring""" import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ = { "vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_config_file": "tokenizer_config.json", } SCREAMING_SNAKE_CASE__ = { "vocab_file": {"facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json"}, "merges_file": {"facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt"}, "tokenizer_config_file": { "facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json" }, } SCREAMING_SNAKE_CASE__ = {"facebook/blenderbot-3B": 128} @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def lowerCAmelCase__ ( ) -> str: """simple docstring""" snake_case = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) snake_case = bs[:] snake_case = 0 for b in range(2**8 ): if b not in bs: bs.append(_UpperCamelCase ) cs.append(2**8 + n ) n += 1 snake_case = [chr(_UpperCamelCase ) for n in cs] return dict(zip(_UpperCamelCase , _UpperCamelCase ) ) def lowerCAmelCase__ ( _UpperCamelCase : List[str] ) -> Union[str, Any]: """simple docstring""" snake_case = set() snake_case = word[0] for char in word[1:]: pairs.add((prev_char, char) ) snake_case = char return pairs class lowerCAmelCase_ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase : str = VOCAB_FILES_NAMES _lowerCAmelCase : str = PRETRAINED_VOCAB_FILES_MAP _lowerCAmelCase : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCAmelCase : Optional[Any] = ["""input_ids""", """attention_mask"""] def __init__( self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase="replace" , lowerCAmelCase="<s>" , lowerCAmelCase="</s>" , lowerCAmelCase="</s>" , lowerCAmelCase="<s>" , lowerCAmelCase="<unk>" , lowerCAmelCase="<pad>" , lowerCAmelCase="<mask>" , lowerCAmelCase=False , **lowerCAmelCase , ): """simple docstring""" snake_case = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else bos_token snake_case = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else eos_token snake_case = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else sep_token snake_case = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else cls_token snake_case = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else unk_token snake_case = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else pad_token # Mask token behave like a normal word, i.e. include the space before it snake_case = AddedToken(lowerCAmelCase , lstrip=lowerCAmelCase , rstrip=lowerCAmelCase ) if isinstance(lowerCAmelCase , lowerCAmelCase ) else mask_token super().__init__( errors=lowerCAmelCase , bos_token=lowerCAmelCase , eos_token=lowerCAmelCase , unk_token=lowerCAmelCase , sep_token=lowerCAmelCase , cls_token=lowerCAmelCase , pad_token=lowerCAmelCase , mask_token=lowerCAmelCase , add_prefix_space=lowerCAmelCase , **lowerCAmelCase , ) with open(lowerCAmelCase , encoding='utf-8' ) as vocab_handle: snake_case = json.load(lowerCAmelCase ) snake_case = {v: k for k, v in self.encoder.items()} snake_case = errors # how to handle errors in decoding snake_case = bytes_to_unicode() snake_case = {v: k for k, v in self.byte_encoder.items()} with open(lowerCAmelCase , encoding='utf-8' ) as merges_handle: snake_case = merges_handle.read().split('\n' )[1:-1] snake_case = [tuple(merge.split() ) for merge in bpe_merges] snake_case = dict(zip(lowerCAmelCase , range(len(lowerCAmelCase ) ) ) ) snake_case = {} snake_case = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions snake_case = re.compile(R'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property # Copied from transformers.models.roberta.tokenization_roberta.RobertaTokenizer.vocab_size with Roberta->Blenderbot, RoBERTa->Blenderbot def snake_case ( self ): """simple docstring""" return len(self.encoder ) def snake_case ( self ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def snake_case ( self , lowerCAmelCase ): """simple docstring""" if token in self.cache: return self.cache[token] snake_case = tuple(lowerCAmelCase ) snake_case = get_pairs(lowerCAmelCase ) if not pairs: return token while True: snake_case = min(lowerCAmelCase , key=lambda lowerCAmelCase : self.bpe_ranks.get(lowerCAmelCase , float('inf' ) ) ) if bigram not in self.bpe_ranks: break snake_case ,snake_case = bigram snake_case = [] snake_case = 0 while i < len(lowerCAmelCase ): try: snake_case = word.index(lowerCAmelCase , lowerCAmelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) snake_case = j if word[i] == first and i < len(lowerCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 snake_case = tuple(lowerCAmelCase ) snake_case = new_word if len(lowerCAmelCase ) == 1: break else: snake_case = get_pairs(lowerCAmelCase ) snake_case = ' '.join(lowerCAmelCase ) snake_case = word return word def snake_case ( self , lowerCAmelCase ): """simple docstring""" snake_case = [] for token in re.findall(self.pat , lowerCAmelCase ): snake_case = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCAmelCase ).split(' ' ) ) return bpe_tokens def snake_case ( self , lowerCAmelCase ): """simple docstring""" return self.encoder.get(lowerCAmelCase , self.encoder.get(self.unk_token ) ) def snake_case ( self , lowerCAmelCase ): """simple docstring""" return self.decoder.get(lowerCAmelCase ) def snake_case ( self , lowerCAmelCase ): """simple docstring""" snake_case = ''.join(lowerCAmelCase ) snake_case = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' , errors=self.errors ) return text def snake_case ( self , lowerCAmelCase , lowerCAmelCase = None ): """simple docstring""" if not os.path.isdir(lowerCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return snake_case = os.path.join( lowerCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) snake_case = os.path.join( lowerCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCAmelCase , ensure_ascii=lowerCAmelCase ) + '\n' ) snake_case = 0 with open(lowerCAmelCase , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCAmelCase : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ' Please check that the tokenizer is not corrupted!' ) snake_case = token_index writer.write(' '.join(lowerCAmelCase ) + '\n' ) index += 1 return vocab_file, merge_file def snake_case ( self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase , token_ids_a=lowerCAmelCase , already_has_special_tokens=lowerCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase )) + [1] return [1] + ([0] * len(lowerCAmelCase )) + [1, 1] + ([0] * len(lowerCAmelCase )) + [1] def snake_case ( self , lowerCAmelCase , lowerCAmelCase = None ): """simple docstring""" snake_case = [self.sep_token_id] snake_case = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def snake_case ( self , lowerCAmelCase , lowerCAmelCase=False , **lowerCAmelCase ): """simple docstring""" snake_case = kwargs.pop('add_prefix_space' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCAmelCase ) > 0 and not text[0].isspace()): snake_case = ' ' + text return (text, kwargs) def snake_case ( self , lowerCAmelCase , lowerCAmelCase = None ): """simple docstring""" return token_ids_a + [self.eos_token_id] def snake_case ( self , lowerCAmelCase ): """simple docstring""" snake_case = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(' ' + text ) else: # Generated responses should contain them already. inputs.append(lowerCAmelCase ) snake_case = ' '.join(lowerCAmelCase ) snake_case = self.encode(lowerCAmelCase ) if len(lowerCAmelCase ) > self.model_max_length: snake_case = input_ids[-self.model_max_length :] logger.warning(F"""Trimmed input from conversation as it was longer than {self.model_max_length} tokens.""" ) return input_ids
150
0
"""simple docstring""" import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaProcessor, is_apex_available, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('1.6'): _snake_case = True from torch.cuda.amp import autocast _snake_case = logging.getLogger(__name__) def lowerCAmelCase__ ( UpperCamelCase__=None , UpperCamelCase__=None ): '''simple docstring''' return field(default_factory=lambda: default , metadata=UpperCamelCase__ ) @dataclass class UpperCamelCase : UpperCamelCase : str = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) UpperCamelCase : Optional[str] = field( default=snake_case_ , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) UpperCamelCase : Optional[bool] = field( default=snake_case_ , metadata={'''help''': '''Whether to freeze the feature extractor layers of the model.'''} ) UpperCamelCase : Optional[float] = field( default=0.1 , metadata={'''help''': '''The dropout ratio for the attention probabilities.'''} ) UpperCamelCase : Optional[float] = field( default=0.1 , metadata={'''help''': '''The dropout ratio for activations inside the fully connected layer.'''} ) UpperCamelCase : Optional[float] = field( default=0.1 , metadata={ '''help''': '''The dropout probabilitiy for all fully connected layers in the embeddings, encoder, and pooler.''' } , ) UpperCamelCase : Optional[float] = field( default=0.1 , metadata={'''help''': '''The dropout probabilitiy for all 1D convolutional layers in feature extractor.'''} , ) UpperCamelCase : Optional[float] = field( default=0.0_5 , metadata={ '''help''': ( '''Propability of each feature vector along the time axis to be chosen as the start of the vector''' '''span to be masked. Approximately ``mask_time_prob * sequence_length // mask_time_length`` feature''' '''vectors will be masked along the time axis. This is only relevant if ``apply_spec_augment is True``.''' ) } , ) UpperCamelCase : Optional[float] = field(default=0.0 , metadata={'''help''': '''The LayerDrop probability.'''} ) @dataclass class UpperCamelCase : UpperCamelCase : Optional[str] = field( default=snake_case_ , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) UpperCamelCase : Optional[str] = field( default='''train+validation''' , metadata={ '''help''': '''The name of the training data set split to use (via the datasets library). Defaults to \'train\'''' } , ) UpperCamelCase : bool = field( default=snake_case_ , metadata={'''help''': '''Overwrite the cached preprocessed datasets or not.'''} ) UpperCamelCase : Optional[int] = field( default=snake_case_ , metadata={'''help''': '''The number of processes to use for the preprocessing.'''} , ) UpperCamelCase : Optional[int] = field( default=snake_case_ , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of training examples to this ''' '''value if set.''' ) } , ) UpperCamelCase : Optional[int] = field( default=snake_case_ , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of validation examples to this ''' '''value if set.''' ) } , ) UpperCamelCase : List[str] = list_field( default=[''',''', '''?''', '''.''', '''!''', '''-''', ''';''', ''':''', '''""''', '''%''', '''\'''', '''"''', '''�'''] , metadata={'''help''': '''A list of characters to remove from the transcripts.'''} , ) @dataclass class UpperCamelCase : UpperCamelCase : WavaVecaProcessor UpperCamelCase : Union[bool, str] = True UpperCamelCase : Optional[int] = None UpperCamelCase : Optional[int] = None UpperCamelCase : Optional[int] = None UpperCamelCase : Optional[int] = None def __call__( self : List[Any] , UpperCAmelCase__ : List[Dict[str, Union[List[int], torch.Tensor]]] ) -> Dict[str, torch.Tensor]: # split inputs and labels since they have to be of different lenghts and need # different padding methods _a : Union[str, Any] = [{"""input_values""": feature["""input_values"""]} for feature in features] _a : str = [{"""input_ids""": feature["""labels"""]} for feature in features] _a : Optional[Any] = self.processor.pad( UpperCAmelCase__ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="""pt""" , ) _a : List[str] = self.processor.pad( labels=UpperCAmelCase__ , padding=self.padding , max_length=self.max_length_labels , pad_to_multiple_of=self.pad_to_multiple_of_labels , return_tensors="""pt""" , ) # replace padding with -100 to ignore loss correctly _a : Union[str, Any] = labels_batch["""input_ids"""].masked_fill(labels_batch.attention_mask.ne(1 ) , -100 ) _a : Any = labels return batch class UpperCamelCase ( snake_case_ ): def _lowercase ( self : Any , UpperCAmelCase__ : nn.Module , UpperCAmelCase__ : Dict[str, Union[torch.Tensor, Any]] ) -> torch.Tensor: model.train() _a : Optional[Any] = self._prepare_inputs(UpperCAmelCase__ ) if self.use_amp: with autocast(): _a : Optional[int] = self.compute_loss(UpperCAmelCase__ , UpperCAmelCase__ ) else: _a : Union[str, Any] = self.compute_loss(UpperCAmelCase__ , UpperCAmelCase__ ) if self.args.n_gpu > 1: if model.module.config.ctc_loss_reduction == "mean": _a : List[str] = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": _a : List[Any] = loss.sum() / (inputs["""labels"""] >= 0).sum() else: raise ValueError(f"""{model.config.ctc_loss_reduction} is not valid. Choose one of ['mean', 'sum']""" ) if self.args.gradient_accumulation_steps > 1: _a : Any = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(UpperCAmelCase__ ).backward() elif self.use_apex: with amp.scale_loss(UpperCAmelCase__ , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(UpperCAmelCase__ ) else: loss.backward() return loss.detach() def lowerCAmelCase__ ( ): '''simple docstring''' _a : Union[str, Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _a : List[Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _a : List[str] = parser.parse_args_into_dataclasses() # Detecting last checkpoint. _a : Union[str, Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _a : Union[str, Any] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"""Output directory ({training_args.output_dir}) already exists and is not empty. """ """Use --overwrite_output_dir to overcome.""" ) elif last_checkpoint is not None: logger.info( F"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ """the `--output_dir` or add `--overwrite_output_dir` to train from scratch.""" ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( F"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + F"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() logger.info("""Training/evaluation parameters %s""" , UpperCamelCase__ ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: _a : str = datasets.load_dataset( """common_voice""" , data_args.dataset_config_name , split=data_args.train_split_name ) _a : int = datasets.load_dataset("""common_voice""" , data_args.dataset_config_name , split="""test""" ) # Create and save tokenizer _a : str = F"""[{''.join(data_args.chars_to_ignore )}]""" def remove_special_characters(UpperCamelCase__ ): _a : int = re.sub(UpperCamelCase__ , """""" , batch["""sentence"""] ).lower() + """ """ return batch _a : List[Any] = train_dataset.map(UpperCamelCase__ , remove_columns=["""sentence"""] ) _a : Union[str, Any] = eval_dataset.map(UpperCamelCase__ , remove_columns=["""sentence"""] ) def extract_all_chars(UpperCamelCase__ ): _a : Any = """ """.join(batch["""text"""] ) _a : List[Any] = list(set(UpperCamelCase__ ) ) return {"vocab": [vocab], "all_text": [all_text]} _a : Optional[Any] = train_dataset.map( UpperCamelCase__ , batched=UpperCamelCase__ , batch_size=-1 , keep_in_memory=UpperCamelCase__ , remove_columns=train_dataset.column_names , ) _a : Tuple = train_dataset.map( UpperCamelCase__ , batched=UpperCamelCase__ , batch_size=-1 , keep_in_memory=UpperCamelCase__ , remove_columns=eval_dataset.column_names , ) _a : Tuple = list(set(vocab_train["""vocab"""][0] ) | set(vocab_test["""vocab"""][0] ) ) _a : str = {v: k for k, v in enumerate(UpperCamelCase__ )} _a : Dict = vocab_dict[""" """] del vocab_dict[" "] _a : Any = len(UpperCamelCase__ ) _a : Union[str, Any] = len(UpperCamelCase__ ) with open("""vocab.json""" , """w""" ) as vocab_file: json.dump(UpperCamelCase__ , UpperCamelCase__ ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _a : Union[str, Any] = WavaVecaCTCTokenizer( """vocab.json""" , unk_token="""[UNK]""" , pad_token="""[PAD]""" , word_delimiter_token="""|""" , ) _a : Optional[Any] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6_0_0_0 , padding_value=0.0 , do_normalize=UpperCamelCase__ , return_attention_mask=UpperCamelCase__ ) _a : Dict = WavaVecaProcessor(feature_extractor=UpperCamelCase__ , tokenizer=UpperCamelCase__ ) _a : Any = WavaVecaForCTC.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , activation_dropout=model_args.activation_dropout , attention_dropout=model_args.attention_dropout , hidden_dropout=model_args.hidden_dropout , feat_proj_dropout=model_args.feat_proj_dropout , mask_time_prob=model_args.mask_time_prob , gradient_checkpointing=training_args.gradient_checkpointing , layerdrop=model_args.layerdrop , ctc_loss_reduction="""mean""" , pad_token_id=processor.tokenizer.pad_token_id , vocab_size=len(processor.tokenizer ) , ) if data_args.max_train_samples is not None: _a : Any = min(len(UpperCamelCase__ ) , data_args.max_train_samples ) _a : Optional[int] = train_dataset.select(range(UpperCamelCase__ ) ) if data_args.max_val_samples is not None: _a : Dict = eval_dataset.select(range(data_args.max_val_samples ) ) _a : List[Any] = torchaudio.transforms.Resample(4_8_0_0_0 , 1_6_0_0_0 ) # Preprocessing the datasets. # We need to read the aduio files as arrays and tokenize the targets. def speech_file_to_array_fn(UpperCamelCase__ ): _a : str = torchaudio.load(batch["""path"""] ) _a : Tuple = resampler(UpperCamelCase__ ).squeeze().numpy() _a : List[Any] = 1_6_0_0_0 _a : Optional[int] = batch["""text"""] return batch _a : Tuple = train_dataset.map( UpperCamelCase__ , remove_columns=train_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) _a : int = eval_dataset.map( UpperCamelCase__ , remove_columns=eval_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) def prepare_dataset(UpperCamelCase__ ): # check that all files have the correct sampling rate assert ( len(set(batch["""sampling_rate"""] ) ) == 1 ), F"""Make sure all inputs have the same sampling rate of {processor.feature_extractor.sampling_rate}.""" _a : str = processor( audio=batch["""speech"""] , text=batch["""target_text"""] , sampling_rate=batch["""sampling_rate"""][0] ) batch.update(UpperCamelCase__ ) return batch _a : Optional[int] = train_dataset.map( UpperCamelCase__ , remove_columns=train_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=UpperCamelCase__ , num_proc=data_args.preprocessing_num_workers , ) _a : str = eval_dataset.map( UpperCamelCase__ , remove_columns=eval_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=UpperCamelCase__ , num_proc=data_args.preprocessing_num_workers , ) # Metric _a : str = datasets.load_metric("""wer""" ) def compute_metrics(UpperCamelCase__ ): _a : List[Any] = pred.predictions _a : int = np.argmax(UpperCamelCase__ , axis=-1 ) _a : Any = processor.tokenizer.pad_token_id _a : List[Any] = processor.batch_decode(UpperCamelCase__ ) # we do not want to group tokens when computing the metrics _a : int = processor.batch_decode(pred.label_ids , group_tokens=UpperCamelCase__ ) _a : Any = wer_metric.compute(predictions=UpperCamelCase__ , references=UpperCamelCase__ ) return {"wer": wer} if model_args.freeze_feature_extractor: model.freeze_feature_extractor() # Data collator _a : Optional[Any] = DataCollatorCTCWithPadding(processor=UpperCamelCase__ , padding=UpperCamelCase__ ) # Initialize our Trainer _a : Any = CTCTrainer( model=UpperCamelCase__ , data_collator=UpperCamelCase__ , args=UpperCamelCase__ , compute_metrics=UpperCamelCase__ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=processor.feature_extractor , ) # Training if training_args.do_train: if last_checkpoint is not None: _a : Any = last_checkpoint elif os.path.isdir(model_args.model_name_or_path ): _a : Tuple = model_args.model_name_or_path else: _a : List[str] = None # Save the feature_extractor and the tokenizer if is_main_process(training_args.local_rank ): processor.save_pretrained(training_args.output_dir ) _a : Optional[Any] = trainer.train(resume_from_checkpoint=UpperCamelCase__ ) trainer.save_model() _a : Optional[int] = train_result.metrics _a : List[str] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(UpperCamelCase__ ) ) _a : Optional[int] = min(UpperCamelCase__ , len(UpperCamelCase__ ) ) trainer.log_metrics("""train""" , UpperCamelCase__ ) trainer.save_metrics("""train""" , UpperCamelCase__ ) trainer.save_state() # Evaluation _a : List[Any] = {} if training_args.do_eval: logger.info("""*** Evaluate ***""" ) _a : str = trainer.evaluate() _a : Optional[int] = data_args.max_val_samples if data_args.max_val_samples is not None else len(UpperCamelCase__ ) _a : Dict = min(UpperCamelCase__ , len(UpperCamelCase__ ) ) trainer.log_metrics("""eval""" , UpperCamelCase__ ) trainer.save_metrics("""eval""" , UpperCamelCase__ ) return results if __name__ == "__main__": main()
351
"""simple docstring""" import argparse import dataclasses import json import logging import os import shutil from typing import List, Optional import datasets from accelerate import Accelerator from datasets import load_dataset from finetuning import finetune from tqdm.auto import tqdm import transformers from transformers import AutoConfig, set_seed from transformers.trainer_utils import IntervalStrategy _snake_case = logging.getLogger(__name__) _snake_case = 'pytorch_model.bin' @dataclasses.dataclass class UpperCamelCase : UpperCamelCase : str = dataclasses.field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models.'''} ) UpperCamelCase : Optional[str] = dataclasses.field( default=snake_case_ , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co.'''} , ) @dataclasses.dataclass class UpperCamelCase : UpperCamelCase : str = dataclasses.field(metadata={'''help''': '''A csv or a json file containing the training data.'''} ) UpperCamelCase : str = dataclasses.field(metadata={'''help''': '''A csv or a json file containing the data to predict on.'''} ) UpperCamelCase : Optional[str] = dataclasses.field( default=snake_case_ , metadata={'''help''': '''A csv or a json file containing the validation data.'''} ) UpperCamelCase : Optional[str] = dataclasses.field( default=snake_case_ , metadata={'''help''': '''The name of the task to train on.'''} , ) UpperCamelCase : Optional[List[str]] = dataclasses.field( default=snake_case_ , metadata={'''help''': '''The list of labels for the task.'''} ) @dataclasses.dataclass class UpperCamelCase : UpperCamelCase : str = dataclasses.field( metadata={'''help''': '''The output directory where the model predictions and checkpoints will be written.'''} ) UpperCamelCase : Optional[str] = dataclasses.field( default='''accuracy''' , metadata={'''help''': '''The evaluation metric used for the task.'''} ) UpperCamelCase : Optional[str] = dataclasses.field( default='''no''' , metadata={ '''help''': '''The evaluation strategy to adopt during training. Possible values are: ["no", "step", "epoch]''' } , ) UpperCamelCase : Optional[int] = dataclasses.field( default=10 , metadata={'''help''': '''Number of evaluation calls with no improvement after which training will be stopped.'''} , ) UpperCamelCase : Optional[float] = dataclasses.field( default=0.0 , metadata={ '''help''': '''How much the specified evaluation metric must improve to satisfy early stopping conditions.''' } , ) UpperCamelCase : Optional[bool] = dataclasses.field( default=snake_case_ , metadata={'''help''': '''Whether to filter the pseudo-labeled data based on the confidence score.'''} , ) UpperCamelCase : Optional[bool] = dataclasses.field( default=snake_case_ , metadata={'''help''': '''Whether to filter the pseudo-labeled data based on the validation performance.'''} , ) UpperCamelCase : Optional[bool] = dataclasses.field( default=snake_case_ , metadata={'''help''': '''Whether to fine-tune on labeled data after pseudo training.'''} , ) UpperCamelCase : Optional[float] = dataclasses.field( default=0.0 , metadata={'''help''': '''Confidence threshold for pseudo-labeled data filtering.'''} , ) UpperCamelCase : Optional[int] = dataclasses.field( default=100 , metadata={'''help''': '''Number of evaluation calls with no improvement after which training will be stopped.'''} , ) UpperCamelCase : Optional[int] = dataclasses.field( default=snake_case_ , metadata={'''help''': '''Random seed for initialization.'''} , ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' _a : Optional[int] = datasets.concatenate_datasets([infer_input, infer_output] , axis=1 ) if args.do_filter_by_confidence: _a : Union[str, Any] = dataset.filter(lambda UpperCamelCase__ : example["probability"] > args.confidence_threshold ) if args.do_filter_by_val_performance: assert eval_result >= 0.0 and eval_result <= 1.0 _a : Any = int(eval_result * len(UpperCamelCase__ ) ) print(UpperCamelCase__ ) _a : str = dataset.sort("""probability""" , reverse=UpperCamelCase__ ) _a : Any = dataset.select(range(UpperCamelCase__ ) ) _a : Tuple = dataset.remove_columns(["""label""", """probability"""] ) _a : Optional[Any] = dataset.rename_column("""prediction""" , """label""" ) _a : Dict = dataset.map(lambda UpperCamelCase__ : {"label": idalabel[example["label"]]} ) _a : Union[str, Any] = dataset.shuffle(seed=args.seed ) _a : Optional[int] = os.path.join(UpperCamelCase__ , F"""train_pseudo.{args.data_file_extension}""" ) if args.data_file_extension == "csv": dataset.to_csv(UpperCamelCase__ , index=UpperCamelCase__ ) else: dataset.to_json(UpperCamelCase__ ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , **UpperCamelCase__ ): '''simple docstring''' _a : Optional[int] = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO , ) logger.info(accelerator.state ) # Setup logging, we only want one process per machine to log things on the # screen. accelerator.is_local_main_process is only True for one process per # machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() _a : Dict = STModelArguments(model_name_or_path=UpperCamelCase__ ) _a : Union[str, Any] = STDataArguments(train_file=UpperCamelCase__ , infer_file=UpperCamelCase__ ) _a : Any = STTrainingArguments(output_dir=UpperCamelCase__ ) _a : Any = argparse.Namespace() for arg_class in (model_args, data_args, training_args): for key, value in vars(UpperCamelCase__ ).items(): setattr(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) for key, value in kwargs.items(): if hasattr(UpperCamelCase__ , UpperCamelCase__ ): setattr(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # Sanity checks _a : Union[str, Any] = {} _a : Tuple = None # You need to provide the training data and the data to predict on assert args.train_file is not None assert args.infer_file is not None _a : int = args.train_file _a : List[Any] = args.infer_file if args.evaluation_strategy != IntervalStrategy.NO.value: assert args.eval_file is not None _a : Union[str, Any] = args.eval_file for key in data_files: _a : Optional[Any] = data_files[key].split(""".""" )[-1] assert extension in ["csv", "json"], F"""`{key}_file` should be a csv or a json file.""" if args.data_file_extension is None: _a : str = extension else: assert extension == args.data_file_extension, F"""`{key}_file` should be a {args.data_file_extension} file`.""" assert ( args.eval_metric in datasets.list_metrics() ), F"""{args.eval_metric} not in the list of supported metrics {datasets.list_metrics()}.""" # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed ) logger.info("""Creating the initial data directory for self-training...""" ) _a : Tuple = F"""{args.output_dir}/self-train_iter-{{}}""".format _a : Dict = data_dir_format(0 ) if accelerator.is_main_process: if args.output_dir is not None: os.makedirs(args.output_dir , exist_ok=UpperCamelCase__ ) os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) accelerator.wait_for_everyone() _a : str = None _a : int = None _a : str = 0 _a : List[Any] = False # Show the progress bar _a : List[Any] = tqdm(range(args.max_selftrain_iterations ) , disable=not accelerator.is_local_main_process ) # Self-train for iteration in range(0 , int(args.max_selftrain_iterations ) ): _a : Union[str, Any] = data_dir_format(UpperCamelCase__ ) assert os.path.exists(UpperCamelCase__ ) # Stage 1: initial fine-tuning for iteration = 0 or pseudo-training for # iteration > 0 _a : str = os.path.join(UpperCamelCase__ , """stage-1""" ) _a : Tuple = { """accelerator""": accelerator, """model_name_or_path""": args.model_name_or_path, """cache_dir""": args.cache_dir, """do_train""": True, """train_file""": data_files["""train"""] if iteration == 0 else data_files["""train_pseudo"""], """do_eval""": True if args.eval_file is not None else False, """eval_file""": data_files["""eval"""], """do_predict""": True, """infer_file""": data_files["""infer"""], """task_name""": args.task_name, """label_list""": args.label_list, """output_dir""": current_output_dir, """eval_metric""": args.eval_metric, """evaluation_strategy""": args.evaluation_strategy, """early_stopping_patience""": args.early_stopping_patience, """early_stopping_threshold""": args.early_stopping_threshold, """seed""": args.seed, } # Add additional training arguments for key, value in kwargs.items(): if key not in arguments_dict and not hasattr(UpperCamelCase__ , UpperCamelCase__ ): arguments_dict.update({key: value} ) _a : int = os.path.join(UpperCamelCase__ , """best-checkpoint""" , UpperCamelCase__ ) if os.path.exists(UpperCamelCase__ ): logger.info( """Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 1.""" , UpperCamelCase__ , UpperCamelCase__ , ) else: logger.info("""***** Running self-training: iteration: %d, stage: 1 *****""" , UpperCamelCase__ ) finetune(**UpperCamelCase__ ) accelerator.wait_for_everyone() assert os.path.exists(UpperCamelCase__ ) logger.info("""Self-training job completed: iteration: %d, stage: 1.""" , UpperCamelCase__ ) if iteration > 0 and args.finetune_on_labeled_data: # Stage 2 (optional): fine-tuning on the original labeled data _a : Dict = os.path.join(UpperCamelCase__ , """best-checkpoint""" ) _a : List[str] = os.path.join(UpperCamelCase__ , """stage-2""" ) # Update arguments_dict _a : int = model_path _a : Dict = data_files["""train"""] _a : int = current_output_dir _a : Any = os.path.join(UpperCamelCase__ , """best-checkpoint""" , UpperCamelCase__ ) if os.path.exists(UpperCamelCase__ ): logger.info( """Found existing model checkpoint at %s. Skipping self-training: iteration: %d, stage: 2.""" , UpperCamelCase__ , UpperCamelCase__ , ) else: logger.info("""***** Running self-training: iteration: %d, stage: 2 *****""" , UpperCamelCase__ ) finetune(**UpperCamelCase__ ) accelerator.wait_for_everyone() assert os.path.exists(UpperCamelCase__ ) logger.info("""Self-training job completed: iteration: %d, stage: 2.""" , UpperCamelCase__ ) _a : List[Any] = iteration _a : int = data_dir_format(iteration + 1 ) _a : Dict = AutoConfig.from_pretrained(os.path.join(UpperCamelCase__ , """best-checkpoint""" ) ) _a : Union[str, Any] = config.idalabel _a : Any = os.path.join(UpperCamelCase__ , """eval_results_best-checkpoint.json""" ) _a : Any = os.path.join(UpperCamelCase__ , """test_results_best-checkpoint.json""" ) assert os.path.exists(UpperCamelCase__ ) with open(UpperCamelCase__ , """r""" ) as f: _a : Tuple = float(json.load(UpperCamelCase__ )[args.eval_metric] ) _a : Dict = os.path.join(UpperCamelCase__ , """infer_output_best-checkpoint.csv""" ) assert os.path.exists(UpperCamelCase__ ) # Loading the dataset from local csv or json files. _a : List[Any] = load_dataset(args.data_file_extension , data_files={"""data""": data_files["""infer"""]} )["""data"""] _a : Any = load_dataset("""csv""" , data_files={"""data""": infer_output_file} )["""data"""] if accelerator.is_main_process: os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) shutil.copy(UpperCamelCase__ , os.path.join(UpperCamelCase__ , F"""eval_results_iter-{iteration}.json""" ) ) if os.path.exists(UpperCamelCase__ ): shutil.copy(UpperCamelCase__ , os.path.join(UpperCamelCase__ , F"""test_results_iter-{iteration}.json""" ) ) create_pseudo_labeled_data(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) accelerator.wait_for_everyone() _a : List[str] = os.path.join(UpperCamelCase__ , F"""train_pseudo.{args.data_file_extension}""" ) if args.evaluation_strategy != IntervalStrategy.NO.value: _a : Any = eval_result if best_iteration is None: _a : Union[str, Any] = new_iteration _a : str = new_eval_result else: if new_eval_result - best_eval_result > args.early_stopping_threshold: _a : Union[str, Any] = new_iteration _a : List[str] = new_eval_result _a : Optional[Any] = 0 else: if new_eval_result == best_eval_result: _a : Tuple = new_iteration _a : List[Any] = new_eval_result early_stopping_patience_counter += 1 if early_stopping_patience_counter >= args.early_stopping_patience: _a : Union[str, Any] = True progress_bar.update(1 ) if should_training_stop: break if best_iteration is not None: # Save the best iteration logger.info("""Best iteration: %d""" , UpperCamelCase__ ) logger.info("""Best evaluation result: %s = %f""" , args.eval_metric , UpperCamelCase__ ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(UpperCamelCase__ , F"""eval_results_iter-{iteration}.json""" ) , os.path.join(UpperCamelCase__ , """eval_results_best-iteration.json""" ) , ) else: # Assume that the last iteration is the best logger.info("""Best iteration: %d""" , args.max_selftrain_iterations - 1 ) logger.info("""Best evaluation result: %s = %f""" , args.eval_metric , UpperCamelCase__ ) accelerator.wait_for_everyone() if accelerator.is_main_process: shutil.copy( os.path.join(UpperCamelCase__ , F"""eval_results_iter-{args.max_selftrain_iterations - 1}.json""" ) , os.path.join(UpperCamelCase__ , """eval_results_best-iteration.json""" ) , )
324
0
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTConfig, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __UpperCAmelCase = logging.get_logger(__name__) def __A ( lowerCamelCase_ ): """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = MobileViTConfig() # size of the architecture if "mobilevit_s" in mobilevit_name: SCREAMING_SNAKE_CASE : Dict = [1_44, 1_92, 2_40] SCREAMING_SNAKE_CASE : Optional[Any] = [16, 32, 64, 96, 1_28, 1_60, 6_40] elif "mobilevit_xs" in mobilevit_name: SCREAMING_SNAKE_CASE : List[Any] = [96, 1_20, 1_44] SCREAMING_SNAKE_CASE : Optional[Any] = [16, 32, 48, 64, 80, 96, 3_84] elif "mobilevit_xxs" in mobilevit_name: SCREAMING_SNAKE_CASE : Tuple = [64, 80, 96] SCREAMING_SNAKE_CASE : List[str] = [16, 16, 24, 48, 64, 80, 3_20] SCREAMING_SNAKE_CASE : str = 0.05 SCREAMING_SNAKE_CASE : List[Any] = 2.0 if mobilevit_name.startswith("""deeplabv3_""" ): SCREAMING_SNAKE_CASE : Any = 5_12 SCREAMING_SNAKE_CASE : Dict = 16 SCREAMING_SNAKE_CASE : Optional[int] = 21 SCREAMING_SNAKE_CASE : Optional[int] = """pascal-voc-id2label.json""" else: SCREAMING_SNAKE_CASE : Dict = 10_00 SCREAMING_SNAKE_CASE : str = """imagenet-1k-id2label.json""" SCREAMING_SNAKE_CASE : Tuple = """huggingface/label-files""" SCREAMING_SNAKE_CASE : Optional[Any] = json.load(open(hf_hub_download(__lowercase , __lowercase , repo_type="""dataset""" ) , """r""" ) ) SCREAMING_SNAKE_CASE : List[str] = {int(__lowercase ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE : List[Any] = idalabel SCREAMING_SNAKE_CASE : Any = {v: k for k, v in idalabel.items()} return config def __A ( lowerCamelCase_ , lowerCamelCase_=False ): """simple docstring""" for i in range(1 , 6 ): if f'''layer_{i}.''' in name: SCREAMING_SNAKE_CASE : Tuple = name.replace(f'''layer_{i}.''' , f'''encoder.layer.{i - 1}.''' ) if "conv_1." in name: SCREAMING_SNAKE_CASE : str = name.replace("""conv_1.""" , """conv_stem.""" ) if ".block." in name: SCREAMING_SNAKE_CASE : List[str] = name.replace(""".block.""" , """.""" ) if "exp_1x1" in name: SCREAMING_SNAKE_CASE : str = name.replace("""exp_1x1""" , """expand_1x1""" ) if "red_1x1" in name: SCREAMING_SNAKE_CASE : Optional[Any] = name.replace("""red_1x1""" , """reduce_1x1""" ) if ".local_rep.conv_3x3." in name: SCREAMING_SNAKE_CASE : Optional[int] = name.replace(""".local_rep.conv_3x3.""" , """.conv_kxk.""" ) if ".local_rep.conv_1x1." in name: SCREAMING_SNAKE_CASE : Union[str, Any] = name.replace(""".local_rep.conv_1x1.""" , """.conv_1x1.""" ) if ".norm." in name: SCREAMING_SNAKE_CASE : List[str] = name.replace(""".norm.""" , """.normalization.""" ) if ".conv." in name: SCREAMING_SNAKE_CASE : Dict = name.replace(""".conv.""" , """.convolution.""" ) if ".conv_proj." in name: SCREAMING_SNAKE_CASE : Union[str, Any] = name.replace(""".conv_proj.""" , """.conv_projection.""" ) for i in range(0 , 2 ): for j in range(0 , 4 ): if f'''.{i}.{j}.''' in name: SCREAMING_SNAKE_CASE : Any = name.replace(f'''.{i}.{j}.''' , f'''.{i}.layer.{j}.''' ) for i in range(2 , 6 ): for j in range(0 , 4 ): if f'''.{i}.{j}.''' in name: SCREAMING_SNAKE_CASE : Any = name.replace(f'''.{i}.{j}.''' , f'''.{i}.''' ) if "expand_1x1" in name: SCREAMING_SNAKE_CASE : List[str] = name.replace("""expand_1x1""" , """downsampling_layer.expand_1x1""" ) if "conv_3x3" in name: SCREAMING_SNAKE_CASE : Tuple = name.replace("""conv_3x3""" , """downsampling_layer.conv_3x3""" ) if "reduce_1x1" in name: SCREAMING_SNAKE_CASE : Any = name.replace("""reduce_1x1""" , """downsampling_layer.reduce_1x1""" ) for i in range(2 , 5 ): if f'''.global_rep.{i}.weight''' in name: SCREAMING_SNAKE_CASE : Optional[Any] = name.replace(f'''.global_rep.{i}.weight''' , """.layernorm.weight""" ) if f'''.global_rep.{i}.bias''' in name: SCREAMING_SNAKE_CASE : List[Any] = name.replace(f'''.global_rep.{i}.bias''' , """.layernorm.bias""" ) if ".global_rep." in name: SCREAMING_SNAKE_CASE : Tuple = name.replace(""".global_rep.""" , """.transformer.""" ) if ".pre_norm_mha.0." in name: SCREAMING_SNAKE_CASE : List[str] = name.replace(""".pre_norm_mha.0.""" , """.layernorm_before.""" ) if ".pre_norm_mha.1.out_proj." in name: SCREAMING_SNAKE_CASE : Any = name.replace(""".pre_norm_mha.1.out_proj.""" , """.attention.output.dense.""" ) if ".pre_norm_ffn.0." in name: SCREAMING_SNAKE_CASE : Dict = name.replace(""".pre_norm_ffn.0.""" , """.layernorm_after.""" ) if ".pre_norm_ffn.1." in name: SCREAMING_SNAKE_CASE : Union[str, Any] = name.replace(""".pre_norm_ffn.1.""" , """.intermediate.dense.""" ) if ".pre_norm_ffn.4." in name: SCREAMING_SNAKE_CASE : Dict = name.replace(""".pre_norm_ffn.4.""" , """.output.dense.""" ) if ".transformer." in name: SCREAMING_SNAKE_CASE : Dict = name.replace(""".transformer.""" , """.transformer.layer.""" ) if ".aspp_layer." in name: SCREAMING_SNAKE_CASE : Optional[Any] = name.replace(""".aspp_layer.""" , """.""" ) if ".aspp_pool." in name: SCREAMING_SNAKE_CASE : Optional[int] = name.replace(""".aspp_pool.""" , """.""" ) if "seg_head." in name: SCREAMING_SNAKE_CASE : Optional[int] = name.replace("""seg_head.""" , """segmentation_head.""" ) if "segmentation_head.classifier.classifier." in name: SCREAMING_SNAKE_CASE : Dict = name.replace("""segmentation_head.classifier.classifier.""" , """segmentation_head.classifier.""" ) if "classifier.fc." in name: SCREAMING_SNAKE_CASE : Any = name.replace("""classifier.fc.""" , """classifier.""" ) elif (not base_model) and ("segmentation_head." not in name): SCREAMING_SNAKE_CASE : Optional[int] = """mobilevit.""" + name return name def __A ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=False ): """simple docstring""" if base_model: SCREAMING_SNAKE_CASE : int = """""" else: SCREAMING_SNAKE_CASE : Tuple = """mobilevit.""" for key in orig_state_dict.copy().keys(): SCREAMING_SNAKE_CASE : int = orig_state_dict.pop(__lowercase ) if key[:8] == "encoder.": SCREAMING_SNAKE_CASE : Tuple = key[8:] if "qkv" in key: SCREAMING_SNAKE_CASE : List[Any] = key.split(""".""" ) SCREAMING_SNAKE_CASE : Optional[int] = int(key_split[0][6:] ) - 1 SCREAMING_SNAKE_CASE : List[Any] = int(key_split[3] ) SCREAMING_SNAKE_CASE : List[str] = model.get_submodule(f'''{model_prefix}encoder.layer.{layer_num}''' ) SCREAMING_SNAKE_CASE : Tuple = layer.transformer.layer[transformer_num].attention.attention.all_head_size SCREAMING_SNAKE_CASE : str = ( f'''{model_prefix}encoder.layer.{layer_num}.transformer.layer.{transformer_num}.attention.attention.''' ) if "weight" in key: SCREAMING_SNAKE_CASE : Dict = val[:dim, :] SCREAMING_SNAKE_CASE : Optional[int] = val[dim : dim * 2, :] SCREAMING_SNAKE_CASE : Tuple = val[-dim:, :] else: SCREAMING_SNAKE_CASE : List[str] = val[:dim] SCREAMING_SNAKE_CASE : Optional[int] = val[dim : dim * 2] SCREAMING_SNAKE_CASE : List[Any] = val[-dim:] else: SCREAMING_SNAKE_CASE : int = val return orig_state_dict def __A ( ): """simple docstring""" SCREAMING_SNAKE_CASE : str = """http://images.cocodataset.org/val2017/000000039769.jpg""" SCREAMING_SNAKE_CASE : List[Any] = Image.open(requests.get(__lowercase , stream=__lowercase ).raw ) return im @torch.no_grad() def __A ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=False ): """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = get_mobilevit_config(__lowercase ) # load original state_dict SCREAMING_SNAKE_CASE : List[str] = torch.load(__lowercase , map_location="""cpu""" ) # load 🤗 model if mobilevit_name.startswith("""deeplabv3_""" ): SCREAMING_SNAKE_CASE : str = MobileViTForSemanticSegmentation(__lowercase ).eval() else: SCREAMING_SNAKE_CASE : int = MobileViTForImageClassification(__lowercase ).eval() SCREAMING_SNAKE_CASE : Optional[int] = convert_state_dict(__lowercase , __lowercase ) model.load_state_dict(__lowercase ) # Check outputs on an image, prepared by MobileViTImageProcessor SCREAMING_SNAKE_CASE : Dict = MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 32 ) SCREAMING_SNAKE_CASE : Union[str, Any] = image_processor(images=prepare_img() , return_tensors="""pt""" ) SCREAMING_SNAKE_CASE : List[str] = model(**__lowercase ) SCREAMING_SNAKE_CASE : int = outputs.logits if mobilevit_name.startswith("""deeplabv3_""" ): assert logits.shape == (1, 21, 32, 32) if mobilevit_name == "deeplabv3_mobilevit_s": SCREAMING_SNAKE_CASE : Any = torch.tensor( [ [[6.2_065, 6.1_292, 6.2_070], [6.1_079, 6.1_254, 6.1_747], [6.0_042, 6.1_071, 6.1_034]], [[-6.9_253, -6.8_653, -7.0_398], [-7.3_218, -7.3_983, -7.3_670], [-7.1_961, -7.2_482, -7.1_569]], [[-4.4_723, -4.4_348, -4.3_769], [-5.3_629, -5.4_632, -5.4_598], [-5.1_587, -5.3_402, -5.5_059]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xs": SCREAMING_SNAKE_CASE : Union[str, Any] = torch.tensor( [ [[5.4_449, 5.5_733, 5.6_314], [5.1_815, 5.3_930, 5.5_963], [5.1_656, 5.4_333, 5.4_853]], [[-9.4_423, -9.7_766, -9.6_714], [-9.1_581, -9.5_720, -9.5_519], [-9.1_006, -9.6_458, -9.5_703]], [[-7.7_721, -7.3_716, -7.1_583], [-8.4_599, -8.0_624, -7.7_944], [-8.4_172, -7.8_366, -7.5_025]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xxs": SCREAMING_SNAKE_CASE : Any = torch.tensor( [ [[6.9_811, 6.9_743, 7.3_123], [7.1_777, 7.1_931, 7.3_938], [7.5_633, 7.8_050, 7.8_901]], [[-10.5_536, -10.2_332, -10.2_924], [-10.2_336, -9.8_624, -9.5_964], [-10.8_840, -10.8_158, -10.6_659]], [[-3.4_938, -3.0_631, -2.8_620], [-3.4_205, -2.8_135, -2.6_875], [-3.4_179, -2.7_945, -2.8_750]], ] ) else: raise ValueError(f'''Unknown mobilevit_name: {mobilevit_name}''' ) assert torch.allclose(logits[0, :3, :3, :3] , __lowercase , atol=1E-4 ) else: assert logits.shape == (1, 10_00) if mobilevit_name == "mobilevit_s": SCREAMING_SNAKE_CASE : List[Any] = torch.tensor([-0.9_866, 0.2_392, -1.1_241] ) elif mobilevit_name == "mobilevit_xs": SCREAMING_SNAKE_CASE : List[str] = torch.tensor([-2.4_761, -0.9_399, -1.9_587] ) elif mobilevit_name == "mobilevit_xxs": SCREAMING_SNAKE_CASE : int = torch.tensor([-1.9_364, -1.2_327, -0.4_653] ) else: raise ValueError(f'''Unknown mobilevit_name: {mobilevit_name}''' ) assert torch.allclose(logits[0, :3] , __lowercase , atol=1E-4 ) Path(__lowercase ).mkdir(exist_ok=__lowercase ) print(f'''Saving model {mobilevit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(__lowercase ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__lowercase ) if push_to_hub: SCREAMING_SNAKE_CASE : Any = { """mobilevit_s""": """mobilevit-small""", """mobilevit_xs""": """mobilevit-x-small""", """mobilevit_xxs""": """mobilevit-xx-small""", """deeplabv3_mobilevit_s""": """deeplabv3-mobilevit-small""", """deeplabv3_mobilevit_xs""": """deeplabv3-mobilevit-x-small""", """deeplabv3_mobilevit_xxs""": """deeplabv3-mobilevit-xx-small""", } print("""Pushing to the hub...""" ) SCREAMING_SNAKE_CASE : Optional[Any] = model_mapping[mobilevit_name] image_processor.push_to_hub(__lowercase , organization="""apple""" ) model.push_to_hub(__lowercase , organization="""apple""" ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--mobilevit_name""", default="""mobilevit_s""", type=str, help=( """Name of the MobileViT model you\'d like to convert. Should be one of \'mobilevit_s\', \'mobilevit_xs\',""" """ \'mobilevit_xxs\', \'deeplabv3_mobilevit_s\', \'deeplabv3_mobilevit_xs\', \'deeplabv3_mobilevit_xxs\'.""" ), ) parser.add_argument( """--checkpoint_path""", required=True, type=str, help="""Path to the original state dict (.pt file).""" ) parser.add_argument( """--pytorch_dump_folder_path""", required=True, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) __UpperCAmelCase = parser.parse_args() convert_movilevit_checkpoint( args.mobilevit_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
323
'''simple docstring''' from __future__ import annotations import unittest from transformers import RoFormerConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerModel, ) from transformers.models.roformer.modeling_tf_roformer import ( TFRoFormerSelfAttention, TFRoFormerSinusoidalPositionalEmbedding, ) class _UpperCAmelCase : """simple docstring""" def __init__( self : Tuple , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Tuple=13 , __UpperCAmelCase : Optional[int]=7 , __UpperCAmelCase : int=True , __UpperCAmelCase : str=True , __UpperCAmelCase : Union[str, Any]=True , __UpperCAmelCase : str=True , __UpperCAmelCase : List[str]=99 , __UpperCAmelCase : List[str]=32 , __UpperCAmelCase : Union[str, Any]=2 , __UpperCAmelCase : List[str]=4 , __UpperCAmelCase : Optional[Any]=37 , __UpperCAmelCase : Any="gelu" , __UpperCAmelCase : Optional[Any]=0.1 , __UpperCAmelCase : int=0.1 , __UpperCAmelCase : Dict=512 , __UpperCAmelCase : List[Any]=16 , __UpperCAmelCase : List[str]=2 , __UpperCAmelCase : Optional[Any]=0.02 , __UpperCAmelCase : int=3 , __UpperCAmelCase : Dict=4 , __UpperCAmelCase : str=None , ): '''simple docstring''' _A = parent _A = 13 _A = 7 _A = True _A = True _A = True _A = True _A = 99 _A = 32 _A = 2 _A = 4 _A = 37 _A = "gelu" _A = 0.1 _A = 0.1 _A = 512 _A = 16 _A = 2 _A = 0.02 _A = 3 _A = 4 _A = None def lowerCAmelCase ( self : Dict ): '''simple docstring''' _A = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _A = None if self.use_input_mask: _A = random_attention_mask([self.batch_size, self.seq_length] ) _A = None if self.use_token_type_ids: _A = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _A = None _A = None _A = None if self.use_labels: _A = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _A = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _A = ids_tensor([self.batch_size] , self.num_choices ) _A = RoFormerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=__UpperCAmelCase , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCAmelCase ( self : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : Dict , __UpperCAmelCase : Any , __UpperCAmelCase : str , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Union[str, Any] ): '''simple docstring''' _A = TFRoFormerModel(config=__UpperCAmelCase ) _A = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _A = [input_ids, input_mask] _A = model(__UpperCAmelCase ) _A = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase ( self : int , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : Tuple , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Any , __UpperCAmelCase : List[Any] ): '''simple docstring''' _A = True _A = TFRoFormerForCausalLM(config=__UpperCAmelCase ) _A = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _A = model(__UpperCAmelCase )["logits"] self.parent.assertListEqual( list(prediction_scores.numpy().shape ) , [self.batch_size, self.seq_length, self.vocab_size] ) def lowerCAmelCase ( self : int , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Tuple , __UpperCAmelCase : str , __UpperCAmelCase : Tuple , __UpperCAmelCase : Dict , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : str ): '''simple docstring''' _A = TFRoFormerForMaskedLM(config=__UpperCAmelCase ) _A = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _A = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase ( self : List[Any] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Union[str, Any] ): '''simple docstring''' _A = self.num_labels _A = TFRoFormerForSequenceClassification(config=__UpperCAmelCase ) _A = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _A = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase ( self : List[Any] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : List[Any] ): '''simple docstring''' _A = self.num_choices _A = TFRoFormerForMultipleChoice(config=__UpperCAmelCase ) _A = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) _A = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) _A = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) _A = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } _A = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCAmelCase ( self : Dict , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : List[str] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Optional[int] ): '''simple docstring''' _A = self.num_labels _A = TFRoFormerForTokenClassification(config=__UpperCAmelCase ) _A = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _A = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCAmelCase ( self : List[Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Dict , __UpperCAmelCase : int , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Any , __UpperCAmelCase : int , __UpperCAmelCase : int ): '''simple docstring''' _A = TFRoFormerForQuestionAnswering(config=__UpperCAmelCase ) _A = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _A = model(__UpperCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCAmelCase ( self : Union[str, Any] ): '''simple docstring''' _A = self.prepare_config_and_inputs() ( ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ) = config_and_inputs _A = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class _UpperCAmelCase ( snake_case_ , snake_case_ , unittest.TestCase ): """simple docstring""" snake_case = ( ( TFRoFormerModel, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerForMultipleChoice, ) if is_tf_available() else () ) snake_case = ( { '''feature-extraction''': TFRoFormerModel, '''fill-mask''': TFRoFormerForMaskedLM, '''question-answering''': TFRoFormerForQuestionAnswering, '''text-classification''': TFRoFormerForSequenceClassification, '''text-generation''': TFRoFormerForCausalLM, '''token-classification''': TFRoFormerForTokenClassification, '''zero-shot''': TFRoFormerForSequenceClassification, } if is_tf_available() else {} ) snake_case = False snake_case = False def lowerCAmelCase ( self : int , __UpperCAmelCase : Dict , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : Optional[int] ): '''simple docstring''' if pipeline_test_casse_name == "TextGenerationPipelineTests": return True return False def lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' _A = TFRoFormerModelTester(self ) _A = ConfigTester(self , config_class=__UpperCAmelCase , hidden_size=37 ) def lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def lowerCAmelCase ( self : Any ): '''simple docstring''' _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCAmelCase ) def lowerCAmelCase ( self : Union[str, Any] ): '''simple docstring''' _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head(*__UpperCAmelCase ) def lowerCAmelCase ( self : Optional[Any] ): '''simple docstring''' _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCAmelCase ) def lowerCAmelCase ( self : List[Any] ): '''simple docstring''' _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCAmelCase ) def lowerCAmelCase ( self : str ): '''simple docstring''' _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCAmelCase ) def lowerCAmelCase ( self : Any ): '''simple docstring''' _A = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCAmelCase ) @slow def lowerCAmelCase ( self : Dict ): '''simple docstring''' _A = TFRoFormerModel.from_pretrained("junnyu/roformer_chinese_base" ) self.assertIsNotNone(__UpperCAmelCase ) @require_tf class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def lowerCAmelCase ( self : Optional[int] ): '''simple docstring''' _A = TFRoFormerForMaskedLM.from_pretrained("junnyu/roformer_chinese_base" ) _A = tf.constant([[0, 1, 2, 3, 4, 5]] ) _A = model(__UpperCAmelCase )[0] # TODO Replace vocab size _A = 50000 _A = [1, 6, vocab_size] self.assertEqual(output.shape , __UpperCAmelCase ) print(output[:, :3, :3] ) # TODO Replace values below with what was printed above. _A = tf.constant( [ [ [-0.12053341, -1.0264901, 0.29221946], [-1.5133783, 0.197433, 0.15190607], [-5.0135403, -3.900256, -0.84038764], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , __UpperCAmelCase , atol=1E-4 ) @require_tf class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" snake_case = 1E-4 def lowerCAmelCase ( self : List[str] ): '''simple docstring''' _A = tf.constant([[4, 10]] ) _A = TFRoFormerSinusoidalPositionalEmbedding(num_positions=6 , embedding_dim=6 ) _A = emba(input_ids.shape ) _A = tf.constant( [[0.0000, 0.0000, 0.0000, 1.0000, 1.0000, 1.0000], [0.8415, 0.0464, 0.0022, 0.5403, 0.9989, 1.0000]] ) tf.debugging.assert_near(__UpperCAmelCase , __UpperCAmelCase , atol=self.tolerance ) def lowerCAmelCase ( self : Tuple ): '''simple docstring''' _A = tf.constant( [ [0.0000, 0.0000, 0.0000, 0.0000, 0.0000], [0.8415, 0.8219, 0.8020, 0.7819, 0.7617], [0.9093, 0.9364, 0.9581, 0.9749, 0.9870], ] ) _A = TFRoFormerSinusoidalPositionalEmbedding(num_positions=512 , embedding_dim=512 ) emba([2, 16, 512] ) _A = emba.weight[:3, :5] tf.debugging.assert_near(__UpperCAmelCase , __UpperCAmelCase , atol=self.tolerance ) @require_tf class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" snake_case = 1E-4 def lowerCAmelCase ( self : str ): '''simple docstring''' _A = tf.reshape(tf.range(2 * 12 * 16 * 64 , dtype=tf.floataa ) , shape=(2, 12, 16, 64) ) / 100 _A = -tf.reshape(tf.range(2 * 12 * 16 * 64 , dtype=tf.floataa ) , shape=(2, 12, 16, 64) ) / 100 _A = TFRoFormerSinusoidalPositionalEmbedding(num_positions=32 , embedding_dim=64 ) _A = embed_positions([2, 16, 768] )[None, None, :, :] _A , _A = TFRoFormerSelfAttention.apply_rotary_position_embeddings( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) _A = tf.constant( [ [0.0000, 0.0100, 0.0200, 0.0300, 0.0400, 0.0500, 0.0600, 0.0700], [-0.2012, 0.8897, 0.0263, 0.9401, 0.2074, 0.9463, 0.3481, 0.9343], [-1.7057, 0.6271, -1.2145, 1.3897, -0.6303, 1.7647, -0.1173, 1.8985], [-2.1731, -1.6397, -2.7358, 0.2854, -2.1840, 1.7183, -1.3018, 2.4871], [0.2717, -3.6173, -2.9206, -2.1988, -3.6638, 0.3858, -2.9155, 2.2980], [3.9859, -2.1580, -0.7984, -4.4904, -4.1181, -2.0252, -4.4782, 1.1253], ] ) _A = tf.constant( [ [0.0000, -0.0100, -0.0200, -0.0300, -0.0400, -0.0500, -0.0600, -0.0700], [0.2012, -0.8897, -0.0263, -0.9401, -0.2074, -0.9463, -0.3481, -0.9343], [1.7057, -0.6271, 1.2145, -1.3897, 0.6303, -1.7647, 0.1173, -1.8985], [2.1731, 1.6397, 2.7358, -0.2854, 2.1840, -1.7183, 1.3018, -2.4871], [-0.2717, 3.6173, 2.9206, 2.1988, 3.6638, -0.3858, 2.9155, -2.2980], [-3.9859, 2.1580, 0.7984, 4.4904, 4.1181, 2.0252, 4.4782, -1.1253], ] ) tf.debugging.assert_near(query_layer[0, 0, :6, :8] , __UpperCAmelCase , atol=self.tolerance ) tf.debugging.assert_near(key_layer[0, 0, :6, :8] , __UpperCAmelCase , atol=self.tolerance )
79
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available A : List[Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : int = ["MLukeTokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mluke import MLukeTokenizer else: import sys A : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
357
"""simple docstring""" import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() A : Tuple = logging.get_logger(__name__) A : Union[str, Any] = "https://openaipublic.azureedge.net/jukebox/models/" A : Tuple = { "jukebox-1b-lyrics": [ "5b/vqvae.pth.tar", "5b/prior_level_0.pth.tar", "5b/prior_level_1.pth.tar", "1b_lyrics/prior_level_2.pth.tar", ], "jukebox-5b-lyrics": [ "5b/vqvae.pth.tar", "5b/prior_level_0.pth.tar", "5b/prior_level_1.pth.tar", "5b_lyrics/prior_level_2.pth.tar", ], } def _lowerCamelCase ( _UpperCamelCase ): '''simple docstring''' if key.endswith(".model.1.bias" ) and len(key.split("." ) ) > 10: __lowerCAmelCase = key.replace(".model.1.bias" , ".conv1d_1.bias" ) elif key.endswith(".model.1.weight" ) and len(key.split("." ) ) > 10: __lowerCAmelCase = key.replace(".model.1.weight" , ".conv1d_1.weight" ) elif key.endswith(".model.3.bias" ) and len(key.split("." ) ) > 10: __lowerCAmelCase = key.replace(".model.3.bias" , ".conv1d_2.bias" ) elif key.endswith(".model.3.weight" ) and len(key.split("." ) ) > 10: __lowerCAmelCase = key.replace(".model.3.weight" , ".conv1d_2.weight" ) if "conditioner_blocks.0." in key: __lowerCAmelCase = key.replace("conditioner_blocks.0" , "conditioner_blocks" ) if "prime_prior" in key: __lowerCAmelCase = key.replace("prime_prior" , "encoder" ) if ".emb." in key and "total" not in key and "absolute" not in key and "relative" not in key: __lowerCAmelCase = key.replace(".emb." , "." ) if key.endswith("k" ): # replace vqvae.X.k with vqvae.X.codebook return key.replace(".k" , ".codebook" ) if "y_emb." in key: return key.replace("y_emb." , "metadata_embedding." ) if "x_emb.emb." in key: __lowerCAmelCase = key.replace("0.x_emb.emb" , "embed_tokens" ) if "prime_state_ln" in key: return key.replace("prime_state_ln" , "encoder.final_layer_norm" ) if ".ln" in key: return key.replace(".ln" , ".layer_norm" ) if "_ln" in key: return key.replace("_ln" , "_layer_norm" ) if "prime_state_proj" in key: return key.replace("prime_state_proj" , "encoder.proj_in" ) if "prime_x_out" in key: return key.replace("prime_x_out" , "encoder.lm_head" ) if "prior.x_out" in key: return key.replace("x_out" , "fc_proj_out" ) if "x_emb" in key: return key.replace("x_emb" , "embed_tokens" ) return key def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = {} import re __lowerCAmelCase = re.compile(R"encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)" ) __lowerCAmelCase = re.compile( R"encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) __lowerCAmelCase = re.compile(R"encoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)" ) __lowerCAmelCase = re.compile(R"decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)" ) __lowerCAmelCase = re.compile( R"decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) __lowerCAmelCase = re.compile(R"decoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)" ) __lowerCAmelCase = re.compile(R"conditioner_blocks.(\d*).cond.model.(\d*).(\d).(bias|weight)" ) __lowerCAmelCase = re.compile( R"conditioner_blocks.(\d*).cond.model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) __lowerCAmelCase = re.compile(R"conditioner_blocks.(\d*).cond.model.(\d*).(bias|weight)" ) for original_key, value in state_dict.items(): # rename vqvae.encoder keys if re_encoder_block_conv_in.fullmatch(_UpperCamelCase ): __lowerCAmelCase = re_encoder_block_conv_in.match(_UpperCamelCase ) __lowerCAmelCase = regex_match.groups() __lowerCAmelCase = int(groups[2] ) * 2 + int(groups[3] ) __lowerCAmelCase = f"encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.{groups[-1]}" __lowerCAmelCase = re_encoder_block_conv_in.sub(_UpperCamelCase , _UpperCamelCase ) elif re_encoder_block_resnet.fullmatch(_UpperCamelCase ): __lowerCAmelCase = re_encoder_block_resnet.match(_UpperCamelCase ) __lowerCAmelCase = regex_match.groups() __lowerCAmelCase = int(groups[2] ) * 2 + int(groups[3] ) __lowerCAmelCase = {"1": 1, "3": 2}[groups[-2]] __lowerCAmelCase = f"encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}." __lowerCAmelCase = f"resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}" __lowerCAmelCase = prefix + resnet_block __lowerCAmelCase = re_encoder_block_resnet.sub(_UpperCamelCase , _UpperCamelCase ) elif re_encoder_block_proj_out.fullmatch(_UpperCamelCase ): __lowerCAmelCase = re_encoder_block_proj_out.match(_UpperCamelCase ) __lowerCAmelCase = regex_match.groups() __lowerCAmelCase = f"encoders.{groups[0]}.level_blocks.{groups[1]}.proj_out.{groups[-1]}" __lowerCAmelCase = re_encoder_block_proj_out.sub(_UpperCamelCase , _UpperCamelCase ) # rename vqvae.decoder keys elif re_decoder_block_conv_out.fullmatch(_UpperCamelCase ): __lowerCAmelCase = re_decoder_block_conv_out.match(_UpperCamelCase ) __lowerCAmelCase = regex_match.groups() __lowerCAmelCase = int(groups[2] ) * 2 + int(groups[3] ) - 2 __lowerCAmelCase = f"decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.{groups[-1]}" __lowerCAmelCase = re_decoder_block_conv_out.sub(_UpperCamelCase , _UpperCamelCase ) elif re_decoder_block_resnet.fullmatch(_UpperCamelCase ): __lowerCAmelCase = re_decoder_block_resnet.match(_UpperCamelCase ) __lowerCAmelCase = regex_match.groups() __lowerCAmelCase = int(groups[2] ) * 2 + int(groups[3] ) - 2 __lowerCAmelCase = {"1": 1, "3": 2}[groups[-2]] __lowerCAmelCase = f"decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}." __lowerCAmelCase = f"resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}" __lowerCAmelCase = prefix + resnet_block __lowerCAmelCase = re_decoder_block_resnet.sub(_UpperCamelCase , _UpperCamelCase ) elif re_decoder_block_proj_in.fullmatch(_UpperCamelCase ): __lowerCAmelCase = re_decoder_block_proj_in.match(_UpperCamelCase ) __lowerCAmelCase = regex_match.groups() __lowerCAmelCase = f"decoders.{groups[0]}.level_blocks.{groups[1]}.proj_in.{groups[-1]}" __lowerCAmelCase = re_decoder_block_proj_in.sub(_UpperCamelCase , _UpperCamelCase ) # rename prior cond.model to upsampler.upsample_block and resnet elif re_prior_cond_conv_out.fullmatch(_UpperCamelCase ): __lowerCAmelCase = re_prior_cond_conv_out.match(_UpperCamelCase ) __lowerCAmelCase = regex_match.groups() __lowerCAmelCase = int(groups[1] ) * 2 + int(groups[2] ) - 2 __lowerCAmelCase = f"conditioner_blocks.upsampler.upsample_block.{block_index}.{groups[-1]}" __lowerCAmelCase = re_prior_cond_conv_out.sub(_UpperCamelCase , _UpperCamelCase ) elif re_prior_cond_resnet.fullmatch(_UpperCamelCase ): __lowerCAmelCase = re_prior_cond_resnet.match(_UpperCamelCase ) __lowerCAmelCase = regex_match.groups() __lowerCAmelCase = int(groups[1] ) * 2 + int(groups[2] ) - 2 __lowerCAmelCase = {"1": 1, "3": 2}[groups[-2]] __lowerCAmelCase = f"conditioner_blocks.upsampler.upsample_block.{block_index}." __lowerCAmelCase = f"resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}" __lowerCAmelCase = prefix + resnet_block __lowerCAmelCase = re_prior_cond_resnet.sub(_UpperCamelCase , _UpperCamelCase ) elif re_prior_cond_proj_in.fullmatch(_UpperCamelCase ): __lowerCAmelCase = re_prior_cond_proj_in.match(_UpperCamelCase ) __lowerCAmelCase = regex_match.groups() __lowerCAmelCase = f"conditioner_blocks.upsampler.proj_in.{groups[-1]}" __lowerCAmelCase = re_prior_cond_proj_in.sub(_UpperCamelCase , _UpperCamelCase ) # keep original key else: __lowerCAmelCase = original_key __lowerCAmelCase = replace_key(_UpperCamelCase ) if f"{key_prefix}.{key}" not in model_state_dict or key is None: print(f"failed converting {original_key} to {key}, does not match" ) # handle missmatched shape elif value.shape != model_state_dict[f"{key_prefix}.{key}"].shape: __lowerCAmelCase = model_state_dict[f"{key_prefix}.{key}"] print(f"{original_key}-> {key} : \nshape {val.shape} and { value.shape}, do not match" ) __lowerCAmelCase = original_key __lowerCAmelCase = original_key __lowerCAmelCase = value return new_dict @torch.no_grad() def _lowerCamelCase ( _UpperCamelCase=None , _UpperCamelCase=None ): '''simple docstring''' for file in MODEL_MAPPING[model_name]: if not os.path.isfile(f"{pytorch_dump_folder_path}/{file.split('/' )[-1]}" ): __lowerCAmelCase = requests.get(f"{PREFIX}{file}" , allow_redirects=_UpperCamelCase ) os.makedirs(f"{pytorch_dump_folder_path}/" , exist_ok=_UpperCamelCase ) open(f"{pytorch_dump_folder_path}/{file.split('/' )[-1]}" , "wb" ).write(r.content ) __lowerCAmelCase = MODEL_MAPPING[model_name.split("/" )[-1]] __lowerCAmelCase = JukeboxConfig.from_pretrained(_UpperCamelCase ) __lowerCAmelCase = JukeboxModel(_UpperCamelCase ) __lowerCAmelCase = [] __lowerCAmelCase = {} for i, dict_name in enumerate(_UpperCamelCase ): __lowerCAmelCase = torch.load(f"{pytorch_dump_folder_path}/{dict_name.split('/' )[-1]}" )["model"] __lowerCAmelCase = {} for k in old_dic.keys(): if k.endswith(".b" ): __lowerCAmelCase = old_dic[k] elif k.endswith(".w" ): __lowerCAmelCase = old_dic[k] elif "level_2" not in dict_name and "cond.model." in k: __lowerCAmelCase = old_dic[k] else: __lowerCAmelCase = old_dic[k] __lowerCAmelCase = "vqvae" if i == 0 else f"priors.{3 - i}" __lowerCAmelCase = fix_jukebox_keys(_UpperCamelCase , model.state_dict() , _UpperCamelCase , _UpperCamelCase ) weight_dict.append(_UpperCamelCase ) __lowerCAmelCase = weight_dict.pop(0 ) model.vqvae.load_state_dict(_UpperCamelCase ) for i in range(len(_UpperCamelCase ) ): model.priors[i].load_state_dict(weight_dict[2 - i] ) Path(_UpperCamelCase ).mkdir(exist_ok=_UpperCamelCase ) with open(f"{pytorch_dump_folder_path}/mapping.json" , "w" ) as txtfile: json.dump(_UpperCamelCase , _UpperCamelCase ) print(f"Saving model {model_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(_UpperCamelCase ) return weight_dict if __name__ == "__main__": A : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="jukebox-5b-lyrics", type=str, help="Name of the model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default="jukebox-5b-lyrics-converted", type=str, help="Path to the output PyTorch model directory.", ) A : Union[str, Any] = parser.parse_args() convert_openai_checkpoint(args.model_name, args.pytorch_dump_folder_path)
259
0
'''simple docstring''' import os import sys import warnings from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen from ..table import array_cast from ..utils.file_utils import is_local_path from ..utils.py_utils import first_non_null_value, no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: import PIL.Image from .features import FeatureType a__ : Optional[List[str]] = None a__ : Dict = '<' if sys.byteorder == 'little' else '>' # Origin: https://github.com/python-pillow/Pillow/blob/698951e19e19972aeed56df686868f1329981c12/src/PIL/Image.py#L3126 minus "|i1" which values are not preserved correctly when saving and loading an image a__ : Any = [ np.dtype('|b1'), np.dtype('|u1'), np.dtype('<u2'), np.dtype('>u2'), np.dtype('<i2'), np.dtype('>i2'), np.dtype('<u4'), np.dtype('>u4'), np.dtype('<i4'), np.dtype('>i4'), np.dtype('<f4'), np.dtype('>f4'), np.dtype('<f8'), np.dtype('>f8'), ] @dataclass class lowercase_ : __UpperCAmelCase = True __UpperCAmelCase = None # Automatically constructed __UpperCAmelCase = "PIL.Image.Image" __UpperCAmelCase = pa.struct({'bytes': pa.binary(), 'path': pa.string()} ) __UpperCAmelCase = field(default='Image' , init=a__ , repr=a__ ) def __call__( self ): return self.pa_type def __a ( self , a ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) if isinstance(a , a ): UpperCamelCase__ = np.array(a ) if isinstance(a , a ): return {"path": value, "bytes": None} elif isinstance(a , a ): return {"path": None, "bytes": value} elif isinstance(a , np.ndarray ): # convert the image array to PNG/TIFF bytes return encode_np_array(a ) elif isinstance(a , PIL.Image.Image ): # convert the PIL image to bytes (default format is PNG/TIFF) return encode_pil_image(a ) elif value.get("path" ) is not None and os.path.isfile(value["path"] ): # we set "bytes": None to not duplicate the data if they're already available locally return {"bytes": None, "path": value.get("path" )} elif value.get("bytes" ) is not None or value.get("path" ) is not None: # store the image bytes, and path is used to infer the image format using the file extension return {"bytes": value.get("bytes" ), "path": value.get("path" )} else: raise ValueError( f'''An image sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.''' ) def __a ( self , a , a=None ): if not self.decode: raise RuntimeError("Decoding is disabled for this feature. Please use Image(decode=True) instead." ) if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support decoding images, please install 'Pillow'." ) if token_per_repo_id is None: UpperCamelCase__ = {} UpperCamelCase__ , UpperCamelCase__ = value["path"], value["bytes"] if bytes_ is None: if path is None: raise ValueError(f'''An image should have one of \'path\' or \'bytes\' but both are None in {value}.''' ) else: if is_local_path(a ): UpperCamelCase__ = PIL.Image.open(a ) else: UpperCamelCase__ = path.split("::" )[-1] try: UpperCamelCase__ = string_to_dict(a , config.HUB_DATASETS_URL )["repo_id"] UpperCamelCase__ = token_per_repo_id.get(a ) except ValueError: UpperCamelCase__ = None with xopen(a , "rb" , use_auth_token=a ) as f: UpperCamelCase__ = BytesIO(f.read() ) UpperCamelCase__ = PIL.Image.open(bytes_ ) else: UpperCamelCase__ = PIL.Image.open(BytesIO(bytes_ ) ) image.load() # to avoid "Too many open files" errors return image def __a ( self ): from .features import Value return ( self if self.decode else { "bytes": Value("binary" ), "path": Value("string" ), } ) def __a ( self , a ): if pa.types.is_string(storage.type ): UpperCamelCase__ = pa.array([None] * len(a ) , type=pa.binary() ) UpperCamelCase__ = pa.StructArray.from_arrays([bytes_array, storage] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): UpperCamelCase__ = pa.array([None] * len(a ) , type=pa.string() ) UpperCamelCase__ = pa.StructArray.from_arrays([storage, path_array] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("bytes" ) >= 0: UpperCamelCase__ = storage.field("bytes" ) else: UpperCamelCase__ = pa.array([None] * len(a ) , type=pa.binary() ) if storage.type.get_field_index("path" ) >= 0: UpperCamelCase__ = storage.field("path" ) else: UpperCamelCase__ = pa.array([None] * len(a ) , type=pa.string() ) UpperCamelCase__ = pa.StructArray.from_arrays([bytes_array, path_array] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_list(storage.type ): UpperCamelCase__ = pa.array( [encode_np_array(np.array(a ) )["bytes"] if arr is not None else None for arr in storage.to_pylist()] , type=pa.binary() , ) UpperCamelCase__ = pa.array([None] * len(a ) , type=pa.string() ) UpperCamelCase__ = pa.StructArray.from_arrays( [bytes_array, path_array] , ["bytes", "path"] , mask=bytes_array.is_null() ) return array_cast(a , self.pa_type ) def __a ( self , a ): @no_op_if_value_is_null def path_to_bytes(a ): with xopen(a , "rb" ) as f: UpperCamelCase__ = f.read() return bytes_ UpperCamelCase__ = pa.array( [ (path_to_bytes(x["path"] ) if x["bytes"] is None else x["bytes"]) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) UpperCamelCase__ = pa.array( [os.path.basename(a ) if path is not None else None for path in storage.field("path" ).to_pylist()] , type=pa.string() , ) UpperCamelCase__ = pa.StructArray.from_arrays([bytes_array, path_array] , ["bytes", "path"] , mask=bytes_array.is_null() ) return array_cast(a , self.pa_type ) def _UpperCamelCase ( ) -> List[str]: '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) global _IMAGE_COMPRESSION_FORMATS if _IMAGE_COMPRESSION_FORMATS is None: PIL.Image.init() UpperCamelCase__ = list(set(PIL.Image.OPEN.keys() ) & set(PIL.Image.SAVE.keys() ) ) return _IMAGE_COMPRESSION_FORMATS def _UpperCamelCase ( __A ) -> bytes: '''simple docstring''' UpperCamelCase__ = BytesIO() if image.format in list_image_compression_formats(): UpperCamelCase__ = image.format else: UpperCamelCase__ = "PNG" if image.mode in ["1", "L", "LA", "RGB", "RGBA"] else "TIFF" image.save(__A , format=__A ) return buffer.getvalue() def _UpperCamelCase ( __A ) -> dict: '''simple docstring''' if hasattr(__A , "filename" ) and image.filename != "": return {"path": image.filename, "bytes": None} else: return {"path": None, "bytes": image_to_bytes(__A )} def _UpperCamelCase ( __A ) -> dict: '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) UpperCamelCase__ = array.dtype UpperCamelCase__ = dtype.byteorder if dtype.byteorder != "=" else _NATIVE_BYTEORDER UpperCamelCase__ = dtype.kind UpperCamelCase__ = dtype.itemsize UpperCamelCase__ = None # Multi-channel array case (only np.dtype("|u1") is allowed) if array.shape[2:]: UpperCamelCase__ = np.dtype("|u1" ) if dtype_kind not in ["u", "i"]: raise TypeError( F'''Unsupported array dtype {dtype} for image encoding. Only {dest_dtype} is supported for multi-channel arrays.''' ) if dtype is not dest_dtype: warnings.warn(F'''Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'''' ) # Exact match elif dtype in _VALID_IMAGE_ARRAY_DTPYES: UpperCamelCase__ = dtype else: # Downcast the type within the kind (np.can_cast(from_type, to_type, casting="same_kind") doesn't behave as expected, so do it manually) while dtype_itemsize >= 1: UpperCamelCase__ = dtype_byteorder + dtype_kind + str(__A ) UpperCamelCase__ = np.dtype(__A ) if dest_dtype in _VALID_IMAGE_ARRAY_DTPYES: warnings.warn(F'''Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'''' ) break else: dtype_itemsize //= 2 if dest_dtype is None: raise TypeError( F'''Cannot convert dtype {dtype} to a valid image dtype. Valid image dtypes: {_VALID_IMAGE_ARRAY_DTPYES}''' ) UpperCamelCase__ = PIL.Image.fromarray(array.astype(__A ) ) return {"path": None, "bytes": image_to_bytes(__A )} def _UpperCamelCase ( __A ) -> List[dict]: '''simple docstring''' if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) if objs: UpperCamelCase__ , UpperCamelCase__ = first_non_null_value(__A ) if isinstance(__A , __A ): return [{"path": obj, "bytes": None} if obj is not None else None for obj in objs] if isinstance(__A , np.ndarray ): UpperCamelCase__ = no_op_if_value_is_null(__A ) return [obj_to_image_dict_func(__A ) for obj in objs] elif isinstance(__A , PIL.Image.Image ): UpperCamelCase__ = no_op_if_value_is_null(__A ) return [obj_to_image_dict_func(__A ) for obj in objs] else: return objs else: return objs
80
"""simple docstring""" import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class A_ : """simple docstring""" def UpperCAmelCase__ ( self :Any ) -> List[str]: torch.manual_seed(0 ) UpperCAmelCase = TaEncoderModel.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) UpperCAmelCase = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) UpperCAmelCase = UNetaDConditionModel( sample_size=32 , layers_per_block=1 , block_out_channels=[32, 64] , down_block_types=[ 'ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D', ] , mid_block_type='UNetMidBlock2DSimpleCrossAttn' , up_block_types=['SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'] , in_channels=3 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type='text' , addition_embed_type_num_heads=2 , cross_attention_norm='group_norm' , resnet_time_scale_shift='scale_shift' , act_fn='gelu' , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) UpperCAmelCase = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule='squaredcos_cap_v2' , beta_start=0.0001 , beta_end=0.02 , thresholding=lowercase_ , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type='epsilon' , variance_type='learned_range' , ) torch.manual_seed(0 ) UpperCAmelCase = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def UpperCAmelCase__ ( self :List[Any] ) -> Any: torch.manual_seed(0 ) UpperCAmelCase = TaEncoderModel.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) UpperCAmelCase = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) UpperCAmelCase = UNetaDConditionModel( sample_size=32 , layers_per_block=[1, 2] , block_out_channels=[32, 64] , down_block_types=[ 'ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D', ] , mid_block_type='UNetMidBlock2DSimpleCrossAttn' , up_block_types=['SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'] , in_channels=6 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type='text' , addition_embed_type_num_heads=2 , cross_attention_norm='group_norm' , resnet_time_scale_shift='scale_shift' , act_fn='gelu' , class_embed_type='timestep' , mid_block_scale_factor=1.414 , time_embedding_act_fn='gelu' , time_embedding_dim=32 , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) UpperCAmelCase = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule='squaredcos_cap_v2' , beta_start=0.0001 , beta_end=0.02 , thresholding=lowercase_ , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type='epsilon' , variance_type='learned_range' , ) torch.manual_seed(0 ) UpperCAmelCase = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule='squaredcos_cap_v2' , beta_start=0.0001 , beta_end=0.02 , ) torch.manual_seed(0 ) UpperCAmelCase = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def UpperCAmelCase__ ( self :List[str] ) -> str: UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = self.pipeline_class(**lowercase_ ) pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) UpperCAmelCase = self.get_dummy_inputs(lowercase_ ) UpperCAmelCase = inputs['prompt'] UpperCAmelCase = inputs['generator'] UpperCAmelCase = inputs['num_inference_steps'] UpperCAmelCase = inputs['output_type'] if "image" in inputs: UpperCAmelCase = inputs['image'] else: UpperCAmelCase = None if "mask_image" in inputs: UpperCAmelCase = inputs['mask_image'] else: UpperCAmelCase = None if "original_image" in inputs: UpperCAmelCase = inputs['original_image'] else: UpperCAmelCase = None UpperCAmelCase , UpperCAmelCase = pipe.encode_prompt(lowercase_ ) # inputs with prompt converted to embeddings UpperCAmelCase = { 'prompt_embeds': prompt_embeds, 'negative_prompt_embeds': negative_prompt_embeds, 'generator': generator, 'num_inference_steps': num_inference_steps, 'output_type': output_type, } if image is not None: UpperCAmelCase = image if mask_image is not None: UpperCAmelCase = mask_image if original_image is not None: UpperCAmelCase = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(lowercase_ , lowercase_ , lowercase_ ) UpperCAmelCase = pipe(**lowercase_ )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(lowercase_ ) UpperCAmelCase = self.pipeline_class.from_pretrained(lowercase_ ) pipe_loaded.to(lowercase_ ) pipe_loaded.set_progress_bar_config(disable=lowercase_ ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(lowercase_ , lowercase_ ) is None , f"""`{optional_component}` did not stay set to None after loading.""" , ) UpperCAmelCase = self.get_dummy_inputs(lowercase_ ) UpperCAmelCase = inputs['generator'] UpperCAmelCase = inputs['num_inference_steps'] UpperCAmelCase = inputs['output_type'] # inputs with prompt converted to embeddings UpperCAmelCase = { 'prompt_embeds': prompt_embeds, 'negative_prompt_embeds': negative_prompt_embeds, 'generator': generator, 'num_inference_steps': num_inference_steps, 'output_type': output_type, } if image is not None: UpperCAmelCase = image if mask_image is not None: UpperCAmelCase = mask_image if original_image is not None: UpperCAmelCase = original_image UpperCAmelCase = pipe_loaded(**lowercase_ )[0] UpperCAmelCase = np.abs(to_np(lowercase_ ) - to_np(lowercase_ ) ).max() self.assertLess(lowercase_ , 1E-4 ) def UpperCAmelCase__ ( self :List[Any] ) -> str: UpperCAmelCase = self.get_dummy_components() UpperCAmelCase = self.pipeline_class(**lowercase_ ) pipe.to(lowercase_ ) pipe.set_progress_bar_config(disable=lowercase_ ) UpperCAmelCase = self.get_dummy_inputs(lowercase_ ) UpperCAmelCase = pipe(**lowercase_ )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(lowercase_ ) UpperCAmelCase = self.pipeline_class.from_pretrained(lowercase_ ) pipe_loaded.to(lowercase_ ) pipe_loaded.set_progress_bar_config(disable=lowercase_ ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests UpperCAmelCase = self.get_dummy_inputs(lowercase_ ) UpperCAmelCase = pipe_loaded(**lowercase_ )[0] UpperCAmelCase = np.abs(to_np(lowercase_ ) - to_np(lowercase_ ) ).max() self.assertLess(lowercase_ , 1E-4 )
78
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) snake_case_ : List[str] = {'configuration_reformer': ['REFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ReformerConfig']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Optional[Any] = ['ReformerTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Dict = ['ReformerTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Union[str, Any] = [ 'REFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'ReformerAttention', 'ReformerForMaskedLM', 'ReformerForQuestionAnswering', 'ReformerForSequenceClassification', 'ReformerLayer', 'ReformerModel', 'ReformerModelWithLMHead', 'ReformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_reformer import REFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ReformerConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_reformer import ReformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_reformer_fast import ReformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_reformer import ( REFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ReformerAttention, ReformerForMaskedLM, ReformerForQuestionAnswering, ReformerForSequenceClassification, ReformerLayer, ReformerModel, ReformerModelWithLMHead, ReformerPreTrainedModel, ) else: import sys snake_case_ : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
236
'''simple docstring''' import tempfile import unittest from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from transformers.testing_utils import ( is_torch_available, require_optimum, require_torch, slow, ) if is_torch_available(): import torch @require_torch @require_optimum @slow class lowercase__ ( unittest.TestCase ): def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Any = 'hf-internal-testing/tiny-random-t5' _UpperCamelCase : str = AutoTokenizer.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : int = AutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : List[str] = tokenizer('This is me' ,return_tensors='pt' ) _UpperCamelCase : str = model.to_bettertransformer() self.assertTrue(any('BetterTransformer' in mod.__class__.__name__ for _, mod in model.named_modules() ) ) _UpperCamelCase : Optional[Any] = model.generate(**lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = model.reverse_bettertransformer() self.assertFalse(any('BetterTransformer' in mod.__class__.__name__ for _, mod in model.named_modules() ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase__ ) _UpperCamelCase : List[Any] = AutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ ) self.assertFalse( any('BetterTransformer' in mod.__class__.__name__ for _, mod in model_reloaded.named_modules() ) ) _UpperCamelCase : Optional[Any] = model_reloaded.generate(**lowerCamelCase__ ) self.assertTrue(torch.allclose(lowerCamelCase__ ,lowerCamelCase__ ) ) def UpperCamelCase_ ( self : Any ): '''simple docstring''' _UpperCamelCase : List[Any] = 'hf-internal-testing/tiny-random-t5' _UpperCamelCase : List[Any] = AutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : List[str] = model.to_bettertransformer() with tempfile.TemporaryDirectory() as tmpdirname: with self.assertRaises(lowerCamelCase__ ): model.save_pretrained(lowerCamelCase__ ) _UpperCamelCase : str = model.reverse_bettertransformer() model.save_pretrained(lowerCamelCase__ )
236
1
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class lowercase_ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): A__ : Tuple = TextToVideoSDPipeline A__ : Any = TEXT_TO_IMAGE_PARAMS A__ : List[str] = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. A__ : Union[str, Any] = frozenset( [ """num_inference_steps""", """generator""", """latents""", """return_dict""", """callback""", """callback_steps""", ] ) def lowerCamelCase_ ( self ): """simple docstring""" torch.manual_seed(0 ) UpperCamelCase_ = UNetaDConditionModel( block_out_channels=(3_2, 6_4, 6_4, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=3_2 , attention_head_dim=4 , ) UpperCamelCase_ = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=__UpperCamelCase , set_alpha_to_one=__UpperCamelCase , ) torch.manual_seed(0 ) UpperCamelCase_ = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) UpperCamelCase_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act="""gelu""" , projection_dim=5_1_2 , ) UpperCamelCase_ = CLIPTextModel(__UpperCamelCase ) UpperCamelCase_ = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) UpperCamelCase_ = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase=0 ): """simple docstring""" if str(__UpperCamelCase ).startswith("""mps""" ): UpperCamelCase_ = torch.manual_seed(__UpperCamelCase ) else: UpperCamelCase_ = torch.Generator(device=__UpperCamelCase ).manual_seed(__UpperCamelCase ) UpperCamelCase_ = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCamelCase_ = self.get_dummy_components() UpperCamelCase_ = TextToVideoSDPipeline(**__UpperCamelCase ) UpperCamelCase_ = sd_pipe.to(__UpperCamelCase ) sd_pipe.set_progress_bar_config(disable=__UpperCamelCase ) UpperCamelCase_ = self.get_dummy_inputs(__UpperCamelCase ) UpperCamelCase_ = """np""" UpperCamelCase_ = sd_pipe(**__UpperCamelCase ).frames UpperCamelCase_ = frames[0][-3:, -3:, -1] assert frames[0].shape == (6_4, 6_4, 3) UpperCamelCase_ = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def lowerCamelCase_ ( self ): """simple docstring""" self._test_attention_slicing_forward_pass(test_mean_pixel_difference=__UpperCamelCase , expected_max_diff=3e-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def lowerCamelCase_ ( self ): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=__UpperCamelCase , expected_max_diff=1e-2 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def lowerCamelCase_ ( self ): """simple docstring""" pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def lowerCamelCase_ ( self ): """simple docstring""" pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def lowerCamelCase_ ( self ): """simple docstring""" pass def lowerCamelCase_ ( self ): """simple docstring""" return super().test_progress_bar() @slow @skip_mps class lowercase_ ( unittest.TestCase ): def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy""" ) UpperCamelCase_ = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) UpperCamelCase_ = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) UpperCamelCase_ = pipe.to("""cuda""" ) UpperCamelCase_ = """Spiderman is surfing""" UpperCamelCase_ = torch.Generator(device="""cpu""" ).manual_seed(0 ) UpperCamelCase_ = pipe(__UpperCamelCase , generator=__UpperCamelCase , num_inference_steps=2_5 , output_type="""pt""" ).frames UpperCamelCase_ = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2 def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy""" ) UpperCamelCase_ = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) UpperCamelCase_ = pipe.to("""cuda""" ) UpperCamelCase_ = """Spiderman is surfing""" UpperCamelCase_ = torch.Generator(device="""cpu""" ).manual_seed(0 ) UpperCamelCase_ = pipe(__UpperCamelCase , generator=__UpperCamelCase , num_inference_steps=2 , output_type="""pt""" ).frames UpperCamelCase_ = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2
122
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, Encoder, VectorQuantizer @dataclass class lowercase_ ( __SCREAMING_SNAKE_CASE ): A__ : torch.FloatTensor class lowercase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): @register_to_config def __init__( self , __UpperCamelCase = 3 , __UpperCamelCase = 3 , __UpperCamelCase = ("DownEncoderBlock2D",) , __UpperCamelCase = ("UpDecoderBlock2D",) , __UpperCamelCase = (6_4,) , __UpperCamelCase = 1 , __UpperCamelCase = "silu" , __UpperCamelCase = 3 , __UpperCamelCase = 3_2 , __UpperCamelCase = 2_5_6 , __UpperCamelCase = 3_2 , __UpperCamelCase = None , __UpperCamelCase = 0.18_215 , __UpperCamelCase = "group" , ): """simple docstring""" super().__init__() # pass init params to Encoder UpperCamelCase_ = Encoder( in_channels=__UpperCamelCase , out_channels=__UpperCamelCase , down_block_types=__UpperCamelCase , block_out_channels=__UpperCamelCase , layers_per_block=__UpperCamelCase , act_fn=__UpperCamelCase , norm_num_groups=__UpperCamelCase , double_z=__UpperCamelCase , ) UpperCamelCase_ = vq_embed_dim if vq_embed_dim is not None else latent_channels UpperCamelCase_ = nn.Convad(__UpperCamelCase , __UpperCamelCase , 1 ) UpperCamelCase_ = VectorQuantizer(__UpperCamelCase , __UpperCamelCase , beta=0.25 , remap=__UpperCamelCase , sane_index_shape=__UpperCamelCase ) UpperCamelCase_ = nn.Convad(__UpperCamelCase , __UpperCamelCase , 1 ) # pass init params to Decoder UpperCamelCase_ = Decoder( in_channels=__UpperCamelCase , out_channels=__UpperCamelCase , up_block_types=__UpperCamelCase , block_out_channels=__UpperCamelCase , layers_per_block=__UpperCamelCase , act_fn=__UpperCamelCase , norm_num_groups=__UpperCamelCase , norm_type=__UpperCamelCase , ) @apply_forward_hook def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase = True ): """simple docstring""" UpperCamelCase_ = self.encoder(__UpperCamelCase ) UpperCamelCase_ = self.quant_conv(__UpperCamelCase ) if not return_dict: return (h,) return VQEncoderOutput(latents=__UpperCamelCase ) @apply_forward_hook def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase = False , __UpperCamelCase = True ): """simple docstring""" if not force_not_quantize: UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = self.quantize(__UpperCamelCase ) else: UpperCamelCase_ = h UpperCamelCase_ = self.post_quant_conv(__UpperCamelCase ) UpperCamelCase_ = self.decoder(__UpperCamelCase , quant if self.config.norm_type == """spatial""" else None ) if not return_dict: return (dec,) return DecoderOutput(sample=__UpperCamelCase ) def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase = True ): """simple docstring""" UpperCamelCase_ = sample UpperCamelCase_ = self.encode(__UpperCamelCase ).latents UpperCamelCase_ = self.decode(__UpperCamelCase ).sample if not return_dict: return (dec,) return DecoderOutput(sample=__UpperCamelCase )
122
1
from ..utils import DummyObject, requires_backends class a ( metaclass=__snake_case ): snake_case_ = ["transformers", "torch", "note_seq"] def __init__( self : Dict , *lowercase_ : Optional[int] , **lowercase_ : Union[str, Any] ): requires_backends(self , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def A_ ( cls : Any , *lowercase_ : Union[str, Any] , **lowercase_ : List[str] ): requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def A_ ( cls : Any , *lowercase_ : List[str] , **lowercase_ : Optional[Any] ): requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] )
352
'''simple docstring''' def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ) -> int: '''simple docstring''' if exponent == 1: return base if exponent % 2 == 0: snake_case_ = _modexpt(__UpperCAmelCase, exponent // 2, __UpperCAmelCase ) % modulo_value return (x * x) % modulo_value else: return (base * _modexpt(__UpperCAmelCase, exponent - 1, __UpperCAmelCase )) % modulo_value def __magic_name__ ( __UpperCAmelCase = 1777, __UpperCAmelCase = 1855, __UpperCAmelCase = 8 ) -> int: '''simple docstring''' snake_case_ = base for _ in range(1, __UpperCAmelCase ): snake_case_ = _modexpt(__UpperCAmelCase, __UpperCAmelCase, 10**digits ) return result if __name__ == "__main__": print(f'''{solution() = }''')
72
0
"""simple docstring""" import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModel, PreTrainedModel from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) def UpperCAmelCase ( UpperCAmelCase , UpperCAmelCase ) -> int: snake_case_ = nn.functional.normalize(UpperCAmelCase ) snake_case_ = nn.functional.normalize(UpperCAmelCase ) return torch.mm(UpperCAmelCase , normalized_text_embeds.t() ) class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = CLIPConfig SCREAMING_SNAKE_CASE_ = ["CLIPEncoderLayer"] def __init__( self, lowerCAmelCase__) -> Optional[int]: super().__init__(lowerCAmelCase__) snake_case_ = CLIPVisionModel(config.vision_config) snake_case_ = nn.Linear(config.vision_config.hidden_size, config.projection_dim, bias=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(17, config.projection_dim), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(3, config.projection_dim), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(17), requires_grad=lowerCAmelCase__) snake_case_ = nn.Parameter(torch.ones(3), requires_grad=lowerCAmelCase__) @torch.no_grad() def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Tuple: snake_case_ = self.vision_model(lowerCAmelCase__)[1] # pooled_output snake_case_ = self.visual_projection(lowerCAmelCase__) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 snake_case_ = cosine_distance(lowerCAmelCase__, self.special_care_embeds).cpu().float().numpy() snake_case_ = cosine_distance(lowerCAmelCase__, self.concept_embeds).cpu().float().numpy() snake_case_ = [] snake_case_ = image_embeds.shape[0] for i in range(lowerCAmelCase__): snake_case_ = {'special_scores': {}, 'special_care': [], 'concept_scores': {}, 'bad_concepts': []} # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign images snake_case_ = 0.0 for concept_idx in range(len(special_cos_dist[0])): snake_case_ = special_cos_dist[i][concept_idx] snake_case_ = self.special_care_embeds_weights[concept_idx].item() snake_case_ = round(concept_cos - concept_threshold + adjustment, 3) if result_img["special_scores"][concept_idx] > 0: result_img["special_care"].append({concept_idx, result_img['special_scores'][concept_idx]}) snake_case_ = 0.01 for concept_idx in range(len(cos_dist[0])): snake_case_ = cos_dist[i][concept_idx] snake_case_ = self.concept_embeds_weights[concept_idx].item() snake_case_ = round(concept_cos - concept_threshold + adjustment, 3) if result_img["concept_scores"][concept_idx] > 0: result_img["bad_concepts"].append(lowerCAmelCase__) result.append(lowerCAmelCase__) snake_case_ = [len(res['bad_concepts']) > 0 for res in result] return images, has_nsfw_concepts @torch.no_grad() def a_ ( self, lowerCAmelCase__, lowerCAmelCase__) -> Optional[int]: snake_case_ = self.vision_model(lowerCAmelCase__)[1] # pooled_output snake_case_ = self.visual_projection(lowerCAmelCase__) snake_case_ = cosine_distance(lowerCAmelCase__, self.special_care_embeds) snake_case_ = cosine_distance(lowerCAmelCase__, self.concept_embeds) # increase this value to create a stronger `nsfw` filter # at the cost of increasing the possibility of filtering benign images snake_case_ = 0.0 snake_case_ = special_cos_dist - self.special_care_embeds_weights + adjustment # special_scores = special_scores.round(decimals=3) snake_case_ = torch.any(special_scores > 0, dim=1) snake_case_ = special_care * 0.01 snake_case_ = special_adjustment.unsqueeze(1).expand(-1, cos_dist.shape[1]) snake_case_ = (cos_dist - self.concept_embeds_weights) + special_adjustment # concept_scores = concept_scores.round(decimals=3) snake_case_ = torch.any(concept_scores > 0, dim=1) return images, has_nsfw_concepts
69
"""simple docstring""" import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = (DPMSolverSinglestepScheduler,) SCREAMING_SNAKE_CASE_ = (("num_inference_steps", 2_5),) def a_ ( self, **lowerCAmelCase__) -> int: snake_case_ = { 'num_train_timesteps': 1000, 'beta_start': 0.0001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'solver_order': 2, 'prediction_type': 'epsilon', 'thresholding': False, 'sample_max_value': 1.0, 'algorithm_type': 'dpmsolver++', 'solver_type': 'midpoint', 'lambda_min_clipped': -float('inf'), 'variance_type': None, } config.update(**lowerCAmelCase__) return config def a_ ( self, lowerCAmelCase__=0, **lowerCAmelCase__) -> List[Any]: snake_case_ = dict(self.forward_default_kwargs) snake_case_ = kwargs.pop('num_inference_steps', lowerCAmelCase__) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals snake_case_ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCAmelCase__) snake_case_ = scheduler_class.from_pretrained(lowerCAmelCase__) new_scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals snake_case_ = dummy_past_residuals[: new_scheduler.config.solver_order] snake_case_ , snake_case_ = sample, sample for t in range(lowerCAmelCase__, time_step + scheduler.config.solver_order + 1): snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample snake_case_ = new_scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1e-5, "Scheduler outputs are not identical" def a_ ( self) -> Union[str, Any]: pass def a_ ( self, lowerCAmelCase__=0, **lowerCAmelCase__) -> int: snake_case_ = dict(self.forward_default_kwargs) snake_case_ = kwargs.pop('num_inference_steps', lowerCAmelCase__) snake_case_ = self.dummy_sample snake_case_ = 0.1 * sample snake_case_ = [residual + 0.2, residual + 0.15, residual + 0.10] for scheduler_class in self.scheduler_classes: snake_case_ = self.get_scheduler_config() snake_case_ = scheduler_class(**lowerCAmelCase__) scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residuals (must be after setting timesteps) snake_case_ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowerCAmelCase__) snake_case_ = scheduler_class.from_pretrained(lowerCAmelCase__) # copy over dummy past residuals new_scheduler.set_timesteps(lowerCAmelCase__) # copy over dummy past residual (must be after setting timesteps) snake_case_ = dummy_past_residuals[: new_scheduler.config.solver_order] snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample snake_case_ = new_scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, **lowerCAmelCase__).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1e-5, "Scheduler outputs are not identical" def a_ ( self, lowerCAmelCase__=None, **lowerCAmelCase__) -> Union[str, Any]: if scheduler is None: snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(**lowerCAmelCase__) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = 10 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter scheduler.set_timesteps(lowerCAmelCase__) for i, t in enumerate(scheduler.timesteps): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample return sample def a_ ( self) -> List[Any]: snake_case_ = DPMSolverSinglestepScheduler(**self.get_scheduler_config()) snake_case_ = 50 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter scheduler.set_timesteps(lowerCAmelCase__) # make sure that the first t is uneven for i, t in enumerate(scheduler.timesteps[3:]): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2574) < 1e-3 def a_ ( self) -> Dict: for timesteps in [25, 50, 100, 999, 1000]: self.check_over_configs(num_train_timesteps=lowerCAmelCase__) def a_ ( self) -> Optional[Any]: # make sure that iterating over schedulers with same config names gives same results # for defaults snake_case_ = DPMSolverSinglestepScheduler(**self.get_scheduler_config()) snake_case_ = self.full_loop(scheduler=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 snake_case_ = DEISMultistepScheduler.from_config(scheduler.config) snake_case_ = DPMSolverMultistepScheduler.from_config(scheduler.config) snake_case_ = UniPCMultistepScheduler.from_config(scheduler.config) snake_case_ = DPMSolverSinglestepScheduler.from_config(scheduler.config) snake_case_ = self.full_loop(scheduler=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 def a_ ( self) -> str: self.check_over_configs(thresholding=lowerCAmelCase__) for order in [1, 2, 3]: for solver_type in ["midpoint", "heun"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=lowerCAmelCase__, prediction_type=lowerCAmelCase__, sample_max_value=lowerCAmelCase__, algorithm_type='dpmsolver++', solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, ) def a_ ( self) -> Tuple: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCAmelCase__) def a_ ( self) -> Optional[int]: for algorithm_type in ["dpmsolver", "dpmsolver++"]: for solver_type in ["midpoint", "heun"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, prediction_type=lowerCAmelCase__, algorithm_type=lowerCAmelCase__, ) snake_case_ = self.full_loop( solver_order=lowerCAmelCase__, solver_type=lowerCAmelCase__, prediction_type=lowerCAmelCase__, algorithm_type=lowerCAmelCase__, ) assert not torch.isnan(lowerCAmelCase__).any(), "Samples have nan numbers" def a_ ( self) -> Optional[Any]: self.check_over_configs(lower_order_final=lowerCAmelCase__) self.check_over_configs(lower_order_final=lowerCAmelCase__) def a_ ( self) -> Any: self.check_over_configs(lambda_min_clipped=-float('inf')) self.check_over_configs(lambda_min_clipped=-5.1) def a_ ( self) -> Any: self.check_over_configs(variance_type=lowerCAmelCase__) self.check_over_configs(variance_type='learned_range') def a_ ( self) -> List[Any]: for num_inference_steps in [1, 2, 3, 5, 10, 50, 100, 999, 1000]: self.check_over_forward(num_inference_steps=lowerCAmelCase__, time_step=0) def a_ ( self) -> int: snake_case_ = self.full_loop() snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2791) < 1e-3 def a_ ( self) -> Dict: snake_case_ = self.full_loop(use_karras_sigmas=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.2248) < 1e-3 def a_ ( self) -> Union[str, Any]: snake_case_ = self.full_loop(prediction_type='v_prediction') snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.1453) < 1e-3 def a_ ( self) -> Optional[Any]: snake_case_ = self.full_loop(prediction_type='v_prediction', use_karras_sigmas=lowerCAmelCase__) snake_case_ = torch.mean(torch.abs(lowerCAmelCase__)) assert abs(result_mean.item() - 0.0649) < 1e-3 def a_ ( self) -> Optional[int]: snake_case_ = self.scheduler_classes[0] snake_case_ = self.get_scheduler_config(thresholding=lowerCAmelCase__, dynamic_thresholding_ratio=0) snake_case_ = scheduler_class(**lowerCAmelCase__) snake_case_ = 10 snake_case_ = self.dummy_model() snake_case_ = self.dummy_sample_deter.half() scheduler.set_timesteps(lowerCAmelCase__) for i, t in enumerate(scheduler.timesteps): snake_case_ = model(lowerCAmelCase__, lowerCAmelCase__) snake_case_ = scheduler.step(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__).prev_sample assert sample.dtype == torch.floataa
69
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _UpperCAmelCase : int = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = { """junnyu/roformer_chinese_small""": """https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json""", """junnyu/roformer_chinese_base""": """https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json""", """junnyu/roformer_chinese_char_small""": ( """https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json""" ), """junnyu/roformer_chinese_char_base""": ( """https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json""" ), """junnyu/roformer_small_discriminator""": ( """https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json""" ), """junnyu/roformer_small_generator""": ( """https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json""" ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class lowercase ( lowercase_ ): __SCREAMING_SNAKE_CASE : List[Any] = '''roformer''' def __init__( self , snake_case=5_0000 , snake_case=None , snake_case=768 , snake_case=12 , snake_case=12 , snake_case=3072 , snake_case="gelu" , snake_case=0.1 , snake_case=0.1 , snake_case=1536 , snake_case=2 , snake_case=0.02 , snake_case=1e-1_2 , snake_case=0 , snake_case=False , snake_case=True , **snake_case , ): super().__init__(pad_token_id=snake_case , **snake_case ) snake_case_ = vocab_size snake_case_ = hidden_size if embedding_size is None else embedding_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = hidden_act snake_case_ = intermediate_size snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = initializer_range snake_case_ = layer_norm_eps snake_case_ = rotary_value snake_case_ = use_cache class lowercase ( lowercase_ ): @property def a ( self ): if self.task == "multiple-choice": snake_case_ = {0: 'batch', 1: 'choice', 2: 'sequence'} else: snake_case_ = {0: 'batch', 1: 'sequence'} snake_case_ = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('token_type_ids', dynamic_axis), ] )
200
def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' return int((input_a, input_a).count(0 ) != 0 ) def __lowerCamelCase ( ): '''simple docstring''' assert nand_gate(0 , 0 ) == 1 assert nand_gate(0 , 1 ) == 1 assert nand_gate(1 , 0 ) == 1 assert nand_gate(1 , 1 ) == 0 if __name__ == "__main__": print(nand_gate(0, 0)) print(nand_gate(0, 1)) print(nand_gate(1, 0)) print(nand_gate(1, 1))
200
1
"""simple docstring""" import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets __lowerCAmelCase : Optional[Any] ="""\ @inproceedings{kakwani2020indicnlpsuite, title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}}, author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar}, year={2020}, booktitle={Findings of EMNLP}, } """ __lowerCAmelCase : List[Any] ="""\ IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te. """ __lowerCAmelCase : List[Any] =""" Compute IndicGLUE evaluation metric associated to each IndicGLUE dataset. Args: predictions: list of predictions to score (as int64), except for 'cvit-mkb-clsr' where each prediction is a vector (of float32). references: list of ground truth labels corresponding to the predictions (as int64), except for 'cvit-mkb-clsr' where each reference is a vector (of float32). Returns: depending on the IndicGLUE subset, one or several of: \"accuracy\": Accuracy \"f1\": F1 score \"precision\": Precision@10 Examples: >>> indic_glue_metric = datasets.load_metric('indic_glue', 'wnli') # 'wnli' or any of [\"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\"] >>> references = [0, 1] >>> predictions = [0, 1] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> indic_glue_metric = datasets.load_metric('indic_glue', 'wiki-ner') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> indic_glue_metric = datasets.load_metric('indic_glue', 'cvit-mkb-clsr') >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]] >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'precision@10': 1.0} """ def UpperCAmelCase__ ( lowerCAmelCase__ :Optional[Any] , lowerCAmelCase__ :Any ) -> Optional[Any]: '''simple docstring''' return float((preds == labels).mean() ) def UpperCAmelCase__ ( lowerCAmelCase__ :List[str] , lowerCAmelCase__ :Tuple ) -> str: '''simple docstring''' lowercase = simple_accuracy(lowerCAmelCase__ , lowerCAmelCase__ ) lowercase = float(fa_score(y_true=lowerCAmelCase__ , y_pred=lowerCAmelCase__ ) ) return { "accuracy": acc, "f1": fa, } def UpperCAmelCase__ ( lowerCAmelCase__ :List[Any] , lowerCAmelCase__ :Optional[int] ) -> Optional[int]: '''simple docstring''' lowercase = np.array(lowerCAmelCase__ ) lowercase = np.array(lowerCAmelCase__ ) lowercase = en_sentvecs.shape[0] # mean centering lowercase = en_sentvecs - np.mean(lowerCAmelCase__ , axis=0 ) lowercase = in_sentvecs - np.mean(lowerCAmelCase__ , axis=0 ) lowercase = cdist(lowerCAmelCase__ , lowerCAmelCase__ , """cosine""" ) lowercase = np.array(range(lowerCAmelCase__ ) ) lowercase = sim.argsort(axis=1 )[:, :1_0] lowercase = np.any(preds == actual[:, None] , axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _A ( datasets.Metric ): def A__ ( self ): """simple docstring""" if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( """You should supply a configuration name selected in """ """[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", """ """\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", """ """\"wiki-ner\"]""" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int64""" ) if self.config_name != """cvit-mkb-clsr""" else datasets.Sequence(datasets.Value("""float32""" ) ), """references""": datasets.Value("""int64""" ) if self.config_name != """cvit-mkb-clsr""" else datasets.Sequence(datasets.Value("""float32""" ) ), } ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" if self.config_name != """cvit-mkb-clsr""" else None , ) def A__ ( self , __lowerCAmelCase , __lowerCAmelCase ): """simple docstring""" if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(__lowerCAmelCase , __lowerCAmelCase )} elif self.config_name in ["wiki-ner"]: return acc_and_fa(__lowerCAmelCase , __lowerCAmelCase ) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(__lowerCAmelCase , __lowerCAmelCase )} else: raise KeyError( """You should supply a configuration name selected in """ """[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", """ """\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", """ """\"wiki-ner\"]""" )
197
"""simple docstring""" from scipy.stats import pearsonr import datasets __lowerCAmelCase : List[Any] =""" Pearson correlation coefficient and p-value for testing non-correlation. The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. """ __lowerCAmelCase : Optional[int] =""" Args: predictions (`list` of `int`): Predicted class labels, as returned by a model. references (`list` of `int`): Ground truth labels. return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`. Returns: pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation. p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities. Examples: Example 1-A simple example using only predictions and references. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5]) >>> print(round(results['pearsonr'], 2)) -0.74 Example 2-The same as Example 1, but that also returns the `p-value`. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True) >>> print(sorted(list(results.keys()))) ['p-value', 'pearsonr'] >>> print(round(results['pearsonr'], 2)) -0.74 >>> print(round(results['p-value'], 2)) 0.15 """ __lowerCAmelCase : str =""" @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, Ilhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Antonio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _A ( datasets.Metric ): def A__ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) , reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html"""] , ) def A__ ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False ): """simple docstring""" if return_pvalue: lowercase = pearsonr(__lowerCAmelCase , __lowerCAmelCase ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(__lowerCAmelCase , __lowerCAmelCase )[0] )}
197
1
'''simple docstring''' def __snake_case( _lowerCAmelCase ) -> Optional[int]: snake_case__ : Any = min(_lowerCAmelCase ) # min() finds the minimum value snake_case__ : str = max(_lowerCAmelCase ) # max() finds the maximum value snake_case__ : Tuple = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size snake_case__ : Optional[int] = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. snake_case__ : Optional[Any] = 0 for count in range(_lowerCAmelCase ): while holes[count] > 0: holes[count] -= 1 snake_case__ : Any = count + min_val i += 1 def __snake_case( ) -> List[Any]: snake_case__ : str = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(_lowerCAmelCase ) print("""Sorted order is:""" , """ """.join(_lowerCAmelCase ) ) if __name__ == "__main__": main()
368
'''simple docstring''' import tempfile import unittest from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from transformers.testing_utils import ( is_torch_available, require_optimum, require_torch, slow, ) if is_torch_available(): import torch @require_torch @require_optimum @slow class UpperCAmelCase_ ( unittest.TestCase ): """simple docstring""" def lowerCamelCase ( self : int ): snake_case__ : List[str] = """hf-internal-testing/tiny-random-t5""" snake_case__ : Any = AutoTokenizer.from_pretrained(snake_case_ ) snake_case__ : Optional[Any] = AutoModelForSeqaSeqLM.from_pretrained(snake_case_ ) snake_case__ : Union[str, Any] = tokenizer("""This is me""" , return_tensors="""pt""" ) snake_case__ : str = model.to_bettertransformer() self.assertTrue(any("""BetterTransformer""" in mod.__class__.__name__ for _, mod in model.named_modules() ) ) snake_case__ : Optional[int] = model.generate(**snake_case_ ) snake_case__ : Any = model.reverse_bettertransformer() self.assertFalse(any("""BetterTransformer""" in mod.__class__.__name__ for _, mod in model.named_modules() ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(snake_case_ ) snake_case__ : int = AutoModelForSeqaSeqLM.from_pretrained(snake_case_ ) self.assertFalse( any("""BetterTransformer""" in mod.__class__.__name__ for _, mod in model_reloaded.named_modules() ) ) snake_case__ : Optional[Any] = model_reloaded.generate(**snake_case_ ) self.assertTrue(torch.allclose(snake_case_ , snake_case_ ) ) def lowerCamelCase ( self : List[Any] ): snake_case__ : Optional[Any] = """hf-internal-testing/tiny-random-t5""" snake_case__ : List[str] = AutoModelForSeqaSeqLM.from_pretrained(snake_case_ ) snake_case__ : int = model.to_bettertransformer() with tempfile.TemporaryDirectory() as tmpdirname: with self.assertRaises(snake_case_ ): model.save_pretrained(snake_case_ ) snake_case__ : int = model.reverse_bettertransformer() model.save_pretrained(snake_case_ )
43
0
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(_UpperCAmelCase ) , 'Tatoeba directory does not exist.' ) class A ( unittest.TestCase ): """simple docstring""" @cached_property def snake_case__ ( self : str )-> Union[str, Any]: '''simple docstring''' A__ = tempfile.mkdtemp() return TatoebaConverter(save_dir=lowercase_ ) @slow def snake_case__ ( self : List[str] )-> Optional[int]: '''simple docstring''' self.resolver.convert_models(['heb-eng'] ) @slow def snake_case__ ( self : int )-> Optional[int]: '''simple docstring''' A__ , A__ = self.resolver.write_model_card('opus-mt-he-en',dry_run=lowercase_ ) assert mmeta["long_pair"] == "heb-eng"
7
from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { "t5-small": "https://huggingface.co/t5-small/resolve/main/config.json", "t5-base": "https://huggingface.co/t5-base/resolve/main/config.json", "t5-large": "https://huggingface.co/t5-large/resolve/main/config.json", "t5-3b": "https://huggingface.co/t5-3b/resolve/main/config.json", "t5-11b": "https://huggingface.co/t5-11b/resolve/main/config.json", } class A ( _UpperCAmelCase ): """simple docstring""" lowerCamelCase = 't5' lowerCamelCase = ['past_key_values'] lowerCamelCase = {'hidden_size': 'd_model', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers'} def __init__( self : Union[str, Any],lowercase_ : int=3_2_1_2_8,lowercase_ : int=5_1_2,lowercase_ : List[str]=6_4,lowercase_ : Tuple=2_0_4_8,lowercase_ : Any=6,lowercase_ : List[str]=None,lowercase_ : Union[str, Any]=8,lowercase_ : int=3_2,lowercase_ : Dict=1_2_8,lowercase_ : Optional[int]=0.1,lowercase_ : List[str]=1E-6,lowercase_ : Tuple=1.0,lowercase_ : Any="relu",lowercase_ : Union[str, Any]=True,lowercase_ : Optional[Any]=True,lowercase_ : int=0,lowercase_ : str=1,**lowercase_ : str,)-> Any: '''simple docstring''' A__ = vocab_size A__ = d_model A__ = d_kv A__ = d_ff A__ = num_layers A__ = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry A__ = num_heads A__ = relative_attention_num_buckets A__ = relative_attention_max_distance A__ = dropout_rate A__ = layer_norm_epsilon A__ = initializer_factor A__ = feed_forward_proj A__ = use_cache A__ = self.feed_forward_proj.split('-' ) A__ = act_info[-1] A__ = act_info[0] == 'gated' if len(lowercase_ ) > 1 and act_info[0] != "gated" or len(lowercase_ ) > 2: raise ValueError( F'`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.' 'Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ' '\'gated-gelu\' or \'relu\'' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": A__ = 'gelu_new' super().__init__( pad_token_id=lowercase_,eos_token_id=lowercase_,is_encoder_decoder=lowercase_,**lowercase_,) class A ( _UpperCAmelCase ): """simple docstring""" @property def snake_case__ ( self : Tuple )-> Mapping[str, Mapping[int, str]]: '''simple docstring''' A__ = { 'input_ids': {0: 'batch', 1: 'encoder_sequence'}, 'attention_mask': {0: 'batch', 1: 'encoder_sequence'}, } if self.use_past: A__ = 'past_encoder_sequence + sequence' A__ = {0: 'batch'} A__ = {0: 'batch', 1: 'past_decoder_sequence + sequence'} else: A__ = {0: 'batch', 1: 'decoder_sequence'} A__ = {0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(lowercase_,direction='inputs' ) return common_inputs @property def snake_case__ ( self : Any )-> int: '''simple docstring''' return 1_3
7
1
'''simple docstring''' import qiskit def _lowerCAmelCase ( __snake_case : int = 2 ) -> qiskit.result.counts.Counts: __A : Optional[Any] = qubits # Using Aer's simulator __A : int = qiskit.Aer.get_backend('aer_simulator' ) # Creating a Quantum Circuit acting on the q register __A : Tuple = qiskit.QuantumCircuit(__snake_case , __snake_case ) # Adding a H gate on qubit 0 (now q0 in superposition) circuit.h(0 ) for i in range(1 , __snake_case ): # Adding CX (CNOT) gate circuit.cx(i - 1 , __snake_case ) # Mapping the quantum measurement to the classical bits circuit.measure(list(range(__snake_case ) ) , list(range(__snake_case ) ) ) # Now measuring any one qubit would affect other qubits to collapse # their super position and have same state as the measured one. # Executing the circuit on the simulator __A : Optional[int] = qiskit.execute(__snake_case , __snake_case , shots=10_00 ) return job.result().get_counts(__snake_case ) if __name__ == "__main__": print(f"""Total count for various states are: {quantum_entanglement(3)}""")
190
'''simple docstring''' import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel lowercase__ : Optional[int] = HfApi() lowercase__ : Dict = {} # fmt: off lowercase__ : List[str] = torch.tensor([ -0.7515, -1.6883, 0.2420, 0.0300, 0.6347, 1.3433, -1.1743, -3.7467, 1.2342, -2.2485, 0.4636, 0.8076, -0.7991, 0.3969, 0.8498, 0.9189, -1.8887, -3.3522, 0.7639, 0.2040, 0.6271, -2.7148, -1.6316, 3.0839, 0.3186, 0.2721, -0.9759, -1.2461, 2.6257, 1.3557 ]) lowercase__ : Tuple = torch.tensor([ -2.3639, -2.5344, 0.0054, -0.6674, 1.5990, 1.0158, 0.3124, -2.1436, 1.8795, -2.5429, -0.1566, -0.3973, 1.2490, 2.6447, 1.2283, -0.5208, -2.8154, -3.5119, 2.3838, 1.2033, 1.7201, -2.1256, -1.4576, 2.7948, 2.4204, -0.9752, -1.2546, 0.8027, 3.2758, 3.1365 ]) lowercase__ : Optional[Any] = torch.tensor([ -0.6531, -0.6891, -0.3172, -0.5375, -0.9140, -0.5367, -0.1175, -0.7869, -0.3808, -0.4513, -0.2098, -0.0083, 0.3183, 0.5140, 0.2247, -0.1304, -0.1302, -0.2802, -0.2084, -0.2025, -0.4967, -0.4873, -0.0861, 0.6925, 0.0250, 0.1290, -0.1543, 0.6316, 1.0460, 1.4943 ]) lowercase__ : List[Any] = torch.tensor([ 0.0911, 0.1107, 0.0182, 0.0435, -0.0805, -0.0608, 0.0381, 0.2172, -0.0280, 0.1327, -0.0299, -0.0255, -0.0050, -0.1170, -0.1046, 0.0309, 0.1367, 0.1728, -0.0533, -0.0748, -0.0534, 0.1624, 0.0384, -0.1805, -0.0707, 0.0642, 0.0220, -0.0134, -0.1333, -0.1505 ]) lowercase__ : Dict = torch.tensor([ 0.1321, 0.1337, 0.0440, 0.0622, -0.0591, -0.0370, 0.0503, 0.2133, -0.0177, 0.1415, -0.0116, -0.0112, 0.0044, -0.0980, -0.0789, 0.0395, 0.1502, 0.1785, -0.0488, -0.0514, -0.0404, 0.1539, 0.0454, -0.1559, -0.0665, 0.0659, 0.0383, -0.0005, -0.1266, -0.1386 ]) lowercase__ : Optional[int] = torch.tensor([ 0.1154, 0.1218, 0.0307, 0.0526, -0.0711, -0.0541, 0.0366, 0.2078, -0.0267, 0.1317, -0.0226, -0.0193, -0.0014, -0.1055, -0.0902, 0.0330, 0.1391, 0.1709, -0.0562, -0.0693, -0.0560, 0.1482, 0.0381, -0.1683, -0.0681, 0.0661, 0.0331, -0.0046, -0.1268, -0.1431 ]) lowercase__ : List[Any] = torch.tensor([ 0.1192, 0.1240, 0.0414, 0.0606, -0.0557, -0.0412, 0.0430, 0.2042, -0.0200, 0.1385, -0.0115, -0.0132, 0.0017, -0.0965, -0.0802, 0.0398, 0.1433, 0.1747, -0.0458, -0.0533, -0.0407, 0.1545, 0.0419, -0.1574, -0.0645, 0.0626, 0.0341, -0.0010, -0.1199, -0.1390 ]) lowercase__ : List[str] = torch.tensor([ 0.1075, 0.1074, 0.0205, 0.0431, -0.0774, -0.0607, 0.0298, 0.2042, -0.0320, 0.1267, -0.0281, -0.0250, -0.0064, -0.1091, -0.0946, 0.0290, 0.1328, 0.1650, -0.0580, -0.0738, -0.0586, 0.1440, 0.0337, -0.1746, -0.0712, 0.0605, 0.0250, -0.0099, -0.1316, -0.1473 ]) lowercase__ : Dict = torch.tensor([ -1.4572, -2.0481, -0.0414, -0.6005, 1.4136, 0.5848, 0.4028, -2.7330, 1.2212, -2.1228, 0.2155, 0.4039, 0.7662, 2.0535, 0.7477, -0.3243, -2.1758, -2.7648, 1.6947, 0.7026, 1.2338, -1.6078, -0.8682, 2.2810, 1.8574, -0.5718, -0.5586, -0.0186, 2.3415, 2.1251]) lowercase__ : Optional[int] = torch.tensor([ -1.3690, -1.9720, -0.4090, -0.6966, 1.4660, 0.9938, -0.1385, -2.7324, 0.7736, -1.8917, 0.2923, 0.4293, 0.1693, 1.4112, 1.1887, -0.3181, -2.2160, -2.6381, 1.3170, 0.8163, 0.9240, -1.6544, -0.6099, 2.5259, 1.6430, -0.9090, -0.9392, -0.0126, 2.4268, 2.3266 ]) lowercase__ : List[str] = torch.tensor([ -1.3525, -1.9628, -0.3956, -0.6860, 1.4664, 1.0014, -0.1259, -2.7212, 0.7772, -1.8811, 0.2996, 0.4388, 0.1704, 1.4029, 1.1701, -0.3027, -2.2053, -2.6287, 1.3350, 0.8131, 0.9274, -1.6292, -0.6098, 2.5131, 1.6505, -0.8958, -0.9298, -0.0151, 2.4257, 2.3355 ]) lowercase__ : Optional[int] = torch.tensor([ -2.0585, -2.7897, -0.2850, -0.8940, 1.9052, 0.5702, 0.6345, -3.8959, 1.5932, -3.2319, 0.1974, 0.0287, 1.7566, 2.6543, 0.8387, -0.5351, -3.2736, -4.3375, 2.9029, 1.6390, 1.4640, -2.1701, -1.9013, 2.9341, 3.4981, -0.6255, -1.1644, -0.1591, 3.7097, 3.2066 ]) lowercase__ : int = torch.tensor([ -2.3139, -2.5594, -0.0197, -0.6785, 1.7001, 1.1606, 0.3075, -2.1740, 1.8071, -2.5630, -0.0926, -0.3811, 1.2116, 2.6246, 1.2731, -0.5398, -2.8153, -3.6140, 2.3893, 1.3262, 1.6258, -2.1856, -1.3267, 2.8395, 2.3779, -1.0623, -1.2468, 0.8959, 3.3367, 3.2243 ]) lowercase__ : int = torch.tensor([ -2.0628, -2.7667, -0.2089, -0.8263, 2.0539, 0.5992, 0.6495, -3.8336, 1.6025, -3.2817, 0.1721, -0.0633, 1.7516, 2.7039, 0.8100, -0.5908, -3.2113, -4.4343, 2.9257, 1.3632, 1.5562, -2.1489, -1.9894, 3.0560, 3.3396, -0.7328, -1.0417, 0.0383, 3.7093, 3.2343 ]) lowercase__ : List[Any] = torch.tensor([ -1.4574, -2.0569, -0.0473, -0.6117, 1.4018, 0.5769, 0.4129, -2.7344, 1.2241, -2.1397, 0.2000, 0.3937, 0.7616, 2.0453, 0.7324, -0.3391, -2.1746, -2.7744, 1.6963, 0.6921, 1.2187, -1.6172, -0.8877, 2.2439, 1.8471, -0.5839, -0.5605, -0.0464, 2.3250, 2.1219 ]) # fmt: on lowercase__ : str = api.list_models(filter='''diffusers''') for mod in models: if "google" in mod.author or mod.modelId == "CompVis/ldm-celebahq-256": lowercase__ : int = '''/home/patrick/google_checkpoints/''' + mod.modelId.split('''/''')[-1] print(f"""Started running {mod.modelId}!!!""") if mod.modelId.startswith('''CompVis'''): lowercase__ : Optional[Any] = UNetaDModel.from_pretrained(local_checkpoint, subfolder='''unet''') else: lowercase__ : Tuple = UNetaDModel.from_pretrained(local_checkpoint) torch.manual_seed(0) random.seed(0) lowercase__ : List[str] = torch.randn(1, model.config.in_channels, model.config.sample_size, model.config.sample_size) lowercase__ : int = torch.tensor([10] * noise.shape[0]) with torch.no_grad(): lowercase__ : Tuple = model(noise, time_step).sample assert torch.allclose( logits[0, 0, 0, :30], results['''_'''.join('''_'''.join(mod.modelId.split('''/''')).split('''-'''))], atol=1e-3 ) print(f"""{mod.modelId} has passed successfully!!!""")
190
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A_ :Any = {'''configuration_xglm''': ['''XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XGLMConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :Any = ['''XGLMTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :Optional[int] = ['''XGLMTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :Optional[Any] = [ '''XGLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XGLMForCausalLM''', '''XGLMModel''', '''XGLMPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :Tuple = [ '''FlaxXGLMForCausalLM''', '''FlaxXGLMModel''', '''FlaxXGLMPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :Tuple = [ '''TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXGLMForCausalLM''', '''TFXGLMModel''', '''TFXGLMPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys A_ :Dict = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
71
import random def A ( a_ ,a_ ,a_ = False ) -> dict: __UpperCamelCase : dict ={i: [] for i in range(a_ )} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(a_ ) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(a_ ): for j in range(i + 1 ,a_ ): if random.random() < probability: graph[i].append(a_ ) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(a_ ) return graph def A ( a_ ) -> dict: return { i: [j for j in range(a_ ) if i != j] for i in range(a_ ) } if __name__ == "__main__": import doctest doctest.testmod()
71
1
from abc import ABC, abstractmethod from typing import List, Optional class snake_case__ (A__ ): """simple docstring""" def __init__( self ) -> int: """simple docstring""" self.test() def SCREAMING_SNAKE_CASE__( self ) -> Optional[Any]: """simple docstring""" a__ : Optional[Any] = 0 a__ : List[str] = False while not completed: if counter == 1: self.reset() a__ : Dict = self.advance() if not self.does_advance(__lowercase ): raise Exception( """Custom Constraint is not defined correctly. self.does_advance(self.advance()) must be true.""" ) a__ , a__ , a__ : str = self.update(__lowercase ) counter += 1 if counter > 1_0_0_0_0: raise Exception("""update() does not fulfill the constraint.""" ) if self.remaining() != 0: raise Exception("""Custom Constraint is not defined correctly.""" ) @abstractmethod def SCREAMING_SNAKE_CASE__( self ) -> List[str]: """simple docstring""" raise NotImplementedError( F'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) @abstractmethod def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> Tuple: """simple docstring""" raise NotImplementedError( F'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) @abstractmethod def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> int: """simple docstring""" raise NotImplementedError( F'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) @abstractmethod def SCREAMING_SNAKE_CASE__( self ) -> Optional[Any]: """simple docstring""" raise NotImplementedError( F'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) @abstractmethod def SCREAMING_SNAKE_CASE__( self ) -> str: """simple docstring""" raise NotImplementedError( F'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) @abstractmethod def SCREAMING_SNAKE_CASE__( self , __lowercase=False ) -> Dict: """simple docstring""" raise NotImplementedError( F'''{self.__class__} is an abstract class. Only classes inheriting this class can be called.''' ) class snake_case__ (A__ ): """simple docstring""" def __init__( self , __lowercase ) -> Union[str, Any]: """simple docstring""" super(__lowercase , self ).__init__() if not isinstance(__lowercase , __lowercase ) or len(__lowercase ) == 0: raise ValueError(F'''`token_ids` has to be a non-empty list, but is {token_ids}.''' ) if any((not isinstance(__lowercase , __lowercase ) or token_id < 0) for token_id in token_ids ): raise ValueError(F'''Each list in `token_ids` has to be a list of positive integers, but is {token_ids}.''' ) a__ : List[Any] = token_ids a__ : List[Any] = len(self.token_ids ) a__ : str = -1 # the index of the currently fulfilled step a__ : List[str] = False def SCREAMING_SNAKE_CASE__( self ) -> List[Any]: """simple docstring""" if self.completed: return None return self.token_ids[self.fulfilled_idx + 1] def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> Tuple: """simple docstring""" if not isinstance(__lowercase , __lowercase ): raise ValueError(F'''`token_id` has to be an `int`, but is {token_id} of type {type(__lowercase )}''' ) if self.completed: return False return token_id == self.token_ids[self.fulfilled_idx + 1] def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> str: """simple docstring""" if not isinstance(__lowercase , __lowercase ): raise ValueError(F'''`token_id` has to be an `int`, but is {token_id} of type {type(__lowercase )}''' ) a__ : Optional[int] = False a__ : str = False a__ : str = False if self.does_advance(__lowercase ): self.fulfilled_idx += 1 a__ : Tuple = True if self.fulfilled_idx == (self.seqlen - 1): a__ : Dict = True a__ : Optional[Any] = completed else: # failed to make progress. a__ : List[str] = True self.reset() return stepped, completed, reset def SCREAMING_SNAKE_CASE__( self ) -> Tuple: """simple docstring""" a__ : List[str] = False a__ : int = 0 def SCREAMING_SNAKE_CASE__( self ) -> Tuple: """simple docstring""" return self.seqlen - (self.fulfilled_idx + 1) def SCREAMING_SNAKE_CASE__( self , __lowercase=False ) -> Tuple: """simple docstring""" a__ : Tuple = PhrasalConstraint(self.token_ids ) if stateful: a__ : Optional[Any] = self.seqlen a__ : int = self.fulfilled_idx a__ : Any = self.completed return new_constraint class snake_case__ : """simple docstring""" def __init__( self , __lowercase , __lowercase=True ) -> Any: """simple docstring""" a__ : List[Any] = max([len(__lowercase ) for one in nested_token_ids] ) a__ : Any = {} for token_ids in nested_token_ids: a__ : int = root for tidx, token_id in enumerate(__lowercase ): if token_id not in level: a__ : List[Any] = {} a__ : Union[str, Any] = level[token_id] if no_subsets and self.has_subsets(__lowercase , __lowercase ): raise ValueError( """Each list in `nested_token_ids` can't be a complete subset of another list, but is""" F''' {nested_token_ids}.''' ) a__ : Any = root def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> Optional[int]: """simple docstring""" a__ : Any = self.trie for current_token in current_seq: a__ : Optional[Any] = start[current_token] a__ : Union[str, Any] = list(start.keys() ) return next_tokens def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> Tuple: """simple docstring""" a__ : int = self.next_tokens(__lowercase ) return len(__lowercase ) == 0 def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> Union[str, Any]: """simple docstring""" a__ : Optional[int] = list(root.values() ) if len(__lowercase ) == 0: return 1 else: return sum([self.count_leaves(__lowercase ) for nn in next_nodes] ) def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase ) -> str: """simple docstring""" a__ : str = self.count_leaves(__lowercase ) return len(__lowercase ) != leaf_count class snake_case__ (A__ ): """simple docstring""" def __init__( self , __lowercase ) -> Any: """simple docstring""" super(__lowercase , self ).__init__() if not isinstance(__lowercase , __lowercase ) or len(__lowercase ) == 0: raise ValueError(F'''`nested_token_ids` has to be a non-empty list, but is {nested_token_ids}.''' ) if any(not isinstance(__lowercase , __lowercase ) for token_ids in nested_token_ids ): raise ValueError(F'''`nested_token_ids` has to be a list of lists, but is {nested_token_ids}.''' ) if any( any((not isinstance(__lowercase , __lowercase ) or token_id < 0) for token_id in token_ids ) for token_ids in nested_token_ids ): raise ValueError( F'''Each list in `nested_token_ids` has to be a list of positive integers, but is {nested_token_ids}.''' ) a__ : Tuple = DisjunctiveTrie(__lowercase ) a__ : Any = nested_token_ids a__ : List[Any] = self.trie.max_height a__ : str = [] a__ : Optional[int] = False def SCREAMING_SNAKE_CASE__( self ) -> Optional[int]: """simple docstring""" a__ : Union[str, Any] = self.trie.next_tokens(self.current_seq ) if len(__lowercase ) == 0: return None else: return token_list def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> int: """simple docstring""" if not isinstance(__lowercase , __lowercase ): raise ValueError(F'''`token_id` is supposed to be type `int`, but is {token_id} of type {type(__lowercase )}''' ) a__ : Any = self.trie.next_tokens(self.current_seq ) return token_id in next_tokens def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> Dict: """simple docstring""" if not isinstance(__lowercase , __lowercase ): raise ValueError(F'''`token_id` is supposed to be type `int`, but is {token_id} of type {type(__lowercase )}''' ) a__ : Any = False a__ : List[Any] = False a__ : List[Any] = False if self.does_advance(__lowercase ): self.current_seq.append(__lowercase ) a__ : Tuple = True else: a__ : Union[str, Any] = True self.reset() a__ : Dict = self.trie.reached_leaf(self.current_seq ) a__ : Union[str, Any] = completed return stepped, completed, reset def SCREAMING_SNAKE_CASE__( self ) -> Any: """simple docstring""" a__ : Tuple = False a__ : Optional[int] = [] def SCREAMING_SNAKE_CASE__( self ) -> Union[str, Any]: """simple docstring""" if self.completed: # since this can be completed without reaching max height return 0 else: return self.seqlen - len(self.current_seq ) def SCREAMING_SNAKE_CASE__( self , __lowercase=False ) -> Optional[int]: """simple docstring""" a__ : str = DisjunctiveConstraint(self.token_ids ) if stateful: a__ : List[Any] = self.seqlen a__ : Optional[Any] = self.current_seq a__ : Optional[Any] = self.completed return new_constraint class snake_case__ : """simple docstring""" def __init__( self , __lowercase ) -> Optional[Any]: """simple docstring""" a__ : List[str] = constraints # max # of steps required to fulfill a given constraint a__ : str = max([c.seqlen for c in constraints] ) a__ : int = len(__lowercase ) a__ : int = False self.init_state() def SCREAMING_SNAKE_CASE__( self ) -> int: """simple docstring""" a__ : List[str] = [] a__ : List[Any] = None a__ : Optional[int] = [constraint.copy(stateful=__lowercase ) for constraint in self.constraints] def SCREAMING_SNAKE_CASE__( self ) -> str: """simple docstring""" a__ : Optional[int] = 0 if self.inprogress_constraint: # extra points for having a constraint mid-fulfilled add += self.max_seqlen - self.inprogress_constraint.remaining() return (len(self.complete_constraints ) * self.max_seqlen) + add def SCREAMING_SNAKE_CASE__( self ) -> Optional[Any]: """simple docstring""" a__ : Any = [] if self.inprogress_constraint is None: for constraint in self.pending_constraints: # "pending" == "unfulfilled yet" a__ : List[Any] = constraint.advance() if isinstance(__lowercase , __lowercase ): token_list.append(__lowercase ) elif isinstance(__lowercase , __lowercase ): token_list.extend(__lowercase ) else: a__ : List[str] = self.inprogress_constraint.advance() if isinstance(__lowercase , __lowercase ): token_list.append(__lowercase ) elif isinstance(__lowercase , __lowercase ): token_list.extend(__lowercase ) if len(__lowercase ) == 0: return None else: return token_list def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> Union[str, Any]: """simple docstring""" self.init_state() if token_ids is not None: for token in token_ids: # completes or steps **one** constraint a__ , a__ : Any = self.add(__lowercase ) # the entire list of constraints are fulfilled if self.completed: break def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> Any: """simple docstring""" if not isinstance(__lowercase , __lowercase ): raise ValueError(F'''`token_id` should be an `int`, but is `{token_id}`.''' ) a__ , a__ : Optional[Any] = False, False if self.completed: a__ : List[str] = True a__ : Tuple = False return complete, stepped if self.inprogress_constraint is not None: # In the middle of fulfilling a constraint. If the `token_id` *does* makes an incremental progress to current # job, simply update the state a__ , a__ , a__ : int = self.inprogress_constraint.update(__lowercase ) if reset: # 1. If the next token breaks the progress, then we must restart. # e.g. constraint = "I love pies" and sequence so far is "I love" but `token_id` == "books". # But that doesn't mean we self.init_state(), since we only reset the state for this particular # constraint, not the full list of constraints. self.pending_constraints.append(self.inprogress_constraint.copy(stateful=__lowercase ) ) a__ : List[Any] = None if complete: # 2. If the next token completes the constraint, move it to completed list, set # inprogress to None. If there are no pending constraints either, then this full list of constraints # is complete. self.complete_constraints.append(self.inprogress_constraint ) a__ : int = None if len(self.pending_constraints ) == 0: # we're done! a__ : Optional[Any] = True else: # Not in the middle of fulfilling a constraint. So does this `token_id` helps us step towards any of our list # of constraints? for cidx, pending_constraint in enumerate(self.pending_constraints ): if pending_constraint.does_advance(__lowercase ): a__ , a__ , a__ : Optional[Any] = pending_constraint.update(__lowercase ) if not stepped: raise Exception( """`constraint.update(token_id)` is not yielding incremental progress, """ """even though `constraint.does_advance(token_id)` is true.""" ) if complete: self.complete_constraints.append(__lowercase ) a__ : List[Any] = None if not complete and stepped: a__ : Optional[Any] = pending_constraint if complete or stepped: # If we made any progress at all, then it's at least not a "pending constraint". a__ : Optional[int] = ( self.pending_constraints[:cidx] + self.pending_constraints[cidx + 1 :] ) if len(self.pending_constraints ) == 0 and self.inprogress_constraint is None: # If there's no longer any pending after this and no inprogress either, then we must be # complete. a__ : Optional[int] = True break # prevent accidentally stepping through multiple constraints with just one token. return complete, stepped def SCREAMING_SNAKE_CASE__( self , __lowercase=True ) -> Optional[Any]: """simple docstring""" a__ : int = ConstraintListState(self.constraints ) # we actually never though self.constraints objects # throughout this process. So it's at initialization state. if stateful: a__ : str = [ constraint.copy(stateful=__lowercase ) for constraint in self.complete_constraints ] if self.inprogress_constraint is not None: a__ : List[str] = self.inprogress_constraint.copy(stateful=__lowercase ) a__ : Optional[int] = [constraint.copy() for constraint in self.pending_constraints] return new_state
266
from __future__ import annotations import math def lowerCAmelCase_ ( _lowercase : float , _lowercase : int) -> float: """simple docstring""" a__ : Union[str, Any] = u for i in range(1 , _lowercase): a__ : Optional[int] = temp * (u - i) return temp def lowerCAmelCase_ ( ) -> None: """simple docstring""" a__ : Tuple = int(input("""enter the numbers of values: """)) a__ : list[list[float]] = [] for _ in range(_lowercase): y.append([]) for i in range(_lowercase): for j in range(_lowercase): y[i].append(_lowercase) a__ : Optional[Any] = 0 print("""enter the values of parameters in a list: """) a__ : List[Any] = list(map(_lowercase , input().split())) print("""enter the values of corresponding parameters: """) for i in range(_lowercase): a__ : Optional[Any] = float(input()) a__ : Tuple = int(input("""enter the value to interpolate: """)) a__ : int = (value - x[0]) / (x[1] - x[0]) # for calculating forward difference table for i in range(1 , _lowercase): for j in range(n - i): a__ : int = y[j + 1][i - 1] - y[j][i - 1] a__ : Optional[int] = y[0][0] for i in range(1 , _lowercase): summ += (ucal(_lowercase , _lowercase) * y[0][i]) / math.factorial(_lowercase) print(F'''the value at {value} is {summ}''') if __name__ == "__main__": main()
266
1
import numpy class UpperCAmelCase__ : """simple docstring""" def __init__( self : Any , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : numpy.ndarray ) -> Any: SCREAMING_SNAKE_CASE__ = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. SCREAMING_SNAKE_CASE__ = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. SCREAMING_SNAKE_CASE__ = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. SCREAMING_SNAKE_CASE__ = numpy.random.rand(3 , 1 ) # Real output values provided. SCREAMING_SNAKE_CASE__ = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. SCREAMING_SNAKE_CASE__ = numpy.zeros(output_array.shape ) def lowercase_ ( self : Optional[Any] ) -> Tuple: SCREAMING_SNAKE_CASE__ = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. SCREAMING_SNAKE_CASE__ = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. SCREAMING_SNAKE_CASE__ = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def lowercase_ ( self : Dict ) -> Any: SCREAMING_SNAKE_CASE__ = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) SCREAMING_SNAKE_CASE__ = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) SCREAMING_SNAKE_CASE__ = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def lowercase_ ( self : Any , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : int , __lowerCamelCase : bool ) -> Optional[int]: for iteration in range(1 , iterations + 1 ): SCREAMING_SNAKE_CASE__ = self.feedforward() self.back_propagation() if give_loss: SCREAMING_SNAKE_CASE__ = numpy.mean(numpy.square(output - self.feedforward() ) ) print(f'''Iteration {iteration} Loss: {loss}''' ) def lowercase_ ( self : Tuple , __lowerCamelCase : numpy.ndarray ) -> List[str]: SCREAMING_SNAKE_CASE__ = input_arr SCREAMING_SNAKE_CASE__ = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) SCREAMING_SNAKE_CASE__ = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) SCREAMING_SNAKE_CASE__ = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def UpperCAmelCase_ ( _A ): '''simple docstring''' return 1 / (1 + numpy.exp(-value )) def UpperCAmelCase_ ( _A ): '''simple docstring''' return (value) * (1 - (value)) def UpperCAmelCase_ ( ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. SCREAMING_SNAKE_CASE__ = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. SCREAMING_SNAKE_CASE__ = TwoHiddenLayerNeuralNetwork( input_array=_A , output_array=_A ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=_A , iterations=10 , give_loss=_A ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
314
'''simple docstring''' import math class __SCREAMING_SNAKE_CASE : """simple docstring""" def UpperCamelCase__ ( self : List[str] , __a : list[list[float]] , __a : list[int] ): _a = 0.0 _a = 0.0 for i in range(len(__a ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def UpperCamelCase__ ( self : List[Any] , __a : list[list[int | float]] , __a : list[int] , __a : int , __a : float ): for i in range(len(__a ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def _lowerCamelCase ( ) -> None: # Training Examples ( m, n ) _a = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) _a = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training _a = SelfOrganizingMap() _a = 3 _a = 0.5 for _ in range(lowercase ): for j in range(len(lowercase ) ): # training sample _a = training_samples[j] # Compute the winning vector _a = self_organizing_map.get_winner(lowercase , lowercase ) # Update the winning vector _a = self_organizing_map.update(lowercase , lowercase , lowercase , lowercase ) # classify test sample _a = [0, 0, 0, 1] _a = self_organizing_map.get_winner(lowercase , lowercase ) # results print(F'Clusters that the test sample belongs to : {winner}' ) print(F'Weights that have been trained : {weights}' ) # running the main() function if __name__ == "__main__": main()
63
0
'''simple docstring''' from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean lowercase : Union[str, Any] = 0 lowercase : str = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] lowercase : List[Any] = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right lowercase : List[str] = tuple[int, int] class A : def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ) -> None: """simple docstring""" A : Union[str, Any] = pos_x A : List[str] = pos_y A : int = (pos_y, pos_x) A : Optional[Any] = goal_x A : Any = goal_y A : int = g_cost A : List[str] = parent A : List[Any] = self.calculate_heuristic() A : Any = self.g_cost + self.h_cost def __lowerCAmelCase ( self ) -> float: """simple docstring""" A : Optional[Any] = self.pos_x - self.goal_x A : Dict = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(SCREAMING_SNAKE_CASE ) + abs(SCREAMING_SNAKE_CASE ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self , SCREAMING_SNAKE_CASE ) -> bool: """simple docstring""" return self.f_cost < other.f_cost class A : def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> List[str]: """simple docstring""" A : List[Any] = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , SCREAMING_SNAKE_CASE ) A : Optional[int] = Node(goal[1] , goal[0] , goal[1] , goal[0] , 99999 , SCREAMING_SNAKE_CASE ) A : Tuple = [self.start] A : list[Node] = [] A : Union[str, Any] = False def __lowerCAmelCase ( self ) -> list[TPosition]: """simple docstring""" while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() A : List[str] = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(SCREAMING_SNAKE_CASE ) self.closed_nodes.append(SCREAMING_SNAKE_CASE ) A : Optional[int] = self.get_successors(SCREAMING_SNAKE_CASE ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(SCREAMING_SNAKE_CASE ) else: # retrieve the best current path A : Union[str, Any] = self.open_nodes.pop(self.open_nodes.index(SCREAMING_SNAKE_CASE ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(SCREAMING_SNAKE_CASE ) else: self.open_nodes.append(SCREAMING_SNAKE_CASE ) return [self.start.pos] def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE ) -> list[Node]: """simple docstring""" A : str = [] for action in delta: A : Any = parent.pos_x + action[1] A : str = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(SCREAMING_SNAKE_CASE ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , SCREAMING_SNAKE_CASE , ) ) return successors def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE ) -> list[TPosition]: """simple docstring""" A : str = node A : str = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) A : Dict = current_node.parent path.reverse() return path class A : def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> None: """simple docstring""" A : int = AStar(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) A : Dict = AStar(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) A : List[Any] = False def __lowerCAmelCase ( self ) -> list[TPosition]: """simple docstring""" while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() A : Dict = self.fwd_astar.open_nodes.pop(0 ) A : Optional[Any] = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) self.fwd_astar.closed_nodes.append(SCREAMING_SNAKE_CASE ) self.bwd_astar.closed_nodes.append(SCREAMING_SNAKE_CASE ) A : Optional[Any] = current_bwd_node A : int = current_fwd_node A : Optional[int] = { self.fwd_astar: self.fwd_astar.get_successors(SCREAMING_SNAKE_CASE ), self.bwd_astar: self.bwd_astar.get_successors(SCREAMING_SNAKE_CASE ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(SCREAMING_SNAKE_CASE ) else: # retrieve the best current path A : str = astar.open_nodes.pop( astar.open_nodes.index(SCREAMING_SNAKE_CASE ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(SCREAMING_SNAKE_CASE ) else: astar.open_nodes.append(SCREAMING_SNAKE_CASE ) return [self.fwd_astar.start.pos] def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> list[TPosition]: """simple docstring""" A : Tuple = self.fwd_astar.retrace_path(SCREAMING_SNAKE_CASE ) A : int = self.bwd_astar.retrace_path(SCREAMING_SNAKE_CASE ) bwd_path.pop() bwd_path.reverse() A : Union[str, Any] = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] lowercase : Union[str, Any] = (0, 0) lowercase : int = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) lowercase : str = time.time() lowercase : int = AStar(init, goal) lowercase : int = a_star.search() lowercase : str = time.time() - start_time print(f'''AStar execution time = {end_time:f} seconds''') lowercase : List[Any] = time.time() lowercase : str = BidirectionalAStar(init, goal) lowercase : Optional[int] = time.time() - bd_start_time print(f'''BidirectionalAStar execution time = {bd_end_time:f} seconds''')
370
'''simple docstring''' import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class A : def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=sys.maxsize ) -> Union[str, Any]: """simple docstring""" A : Tuple = '''bilinear''' A : Optional[int] = max_size A : Dict = short_edge_length def __call__( self , SCREAMING_SNAKE_CASE ) -> Tuple: """simple docstring""" A : Tuple = [] for img in imgs: A, A : str = img.shape[:2] # later: provide list and randomly choose index for resize A : Union[str, Any] = np.random.randint(self.short_edge_length[0] , self.short_edge_length[1] + 1 ) if size == 0: return img A : int = size * 1.0 / min(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if h < w: A, A : Tuple = size, scale * w else: A, A : str = scale * h, size if max(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) > self.max_size: A : List[str] = self.max_size * 1.0 / max(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) A : Tuple = newh * scale A : int = neww * scale A : List[str] = int(neww + 0.5 ) A : int = int(newh + 0.5 ) if img.dtype == np.uinta: A : Dict = Image.fromarray(SCREAMING_SNAKE_CASE ) A : Optional[Any] = pil_image.resize((neww, newh) , PILImageResampling.BILINEAR ) A : str = np.asarray(SCREAMING_SNAKE_CASE ) else: A : Dict = img.permute(2 , 0 , 1 ).unsqueeze(0 ) # 3, 0, 1) # hw(c) -> nchw A : List[Any] = nn.functional.interpolate( SCREAMING_SNAKE_CASE , (newh, neww) , mode=self.interp_method , align_corners=SCREAMING_SNAKE_CASE ).squeeze(0 ) img_augs.append(SCREAMING_SNAKE_CASE ) return img_augs class A : def __init__( self , SCREAMING_SNAKE_CASE ) -> Dict: """simple docstring""" A : Any = ResizeShortestEdge([cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST] , cfg.INPUT.MAX_SIZE_TEST ) A : str = cfg.INPUT.FORMAT A : int = cfg.SIZE_DIVISIBILITY A : Optional[int] = cfg.PAD_VALUE A : Dict = cfg.INPUT.MAX_SIZE_TEST A : Optional[Any] = cfg.MODEL.DEVICE A : Dict = torch.tensor(cfg.MODEL.PIXEL_STD ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) A : Tuple = torch.tensor(cfg.MODEL.PIXEL_MEAN ).to(self.device ).view(len(cfg.MODEL.PIXEL_STD ) , 1 , 1 ) A : str = lambda SCREAMING_SNAKE_CASE : (x - self.pixel_mean) / self.pixel_std def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE ) -> Optional[Any]: """simple docstring""" A : Union[str, Any] = tuple(max(SCREAMING_SNAKE_CASE ) for s in zip(*[img.shape for img in images] ) ) A : List[str] = [im.shape[-2:] for im in images] A : Optional[Any] = [ nn.functional.pad( SCREAMING_SNAKE_CASE , [0, max_size[-1] - size[1], 0, max_size[-2] - size[0]] , value=self.pad_value , ) for size, im in zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ] return torch.stack(SCREAMING_SNAKE_CASE ), torch.tensor(SCREAMING_SNAKE_CASE ) def __call__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=False ) -> Union[str, Any]: """simple docstring""" with torch.no_grad(): if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): A : str = [images] if single_image: assert len(SCREAMING_SNAKE_CASE ) == 1 for i in range(len(SCREAMING_SNAKE_CASE ) ): if isinstance(images[i] , torch.Tensor ): images.insert(SCREAMING_SNAKE_CASE , images.pop(SCREAMING_SNAKE_CASE ).to(self.device ).float() ) elif not isinstance(images[i] , torch.Tensor ): images.insert( SCREAMING_SNAKE_CASE , torch.as_tensor(img_tensorize(images.pop(SCREAMING_SNAKE_CASE ) , input_format=self.input_format ) ) .to(self.device ) .float() , ) # resize smallest edge A : Tuple = torch.tensor([im.shape[:2] for im in images] ) A : Dict = self.aug(SCREAMING_SNAKE_CASE ) # transpose images and convert to torch tensors # images = [torch.as_tensor(i.astype("float32")).permute(2, 0, 1).to(self.device) for i in images] # now normalize before pad to avoid useless arithmetic A : Tuple = [self.normalizer(SCREAMING_SNAKE_CASE ) for x in images] # now pad them to do the following operations A, A : Optional[int] = self.pad(SCREAMING_SNAKE_CASE ) # Normalize if self.size_divisibility > 0: raise NotImplementedError() # pad A : Tuple = torch.true_divide(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if single_image: return images[0], sizes[0], scales_yx[0] else: return images, sizes, scales_yx def lowerCAmelCase_ ( snake_case__ , snake_case__ ): '''simple docstring''' boxes[:, 0::2] *= scale_yx[:, 1] boxes[:, 1::2] *= scale_yx[:, 0] return boxes def lowerCAmelCase_ ( snake_case__ , snake_case__ ): '''simple docstring''' assert torch.isfinite(snake_case__ ).all(), "Box tensor contains infinite or NaN!" A, A : str = box_size tensor[:, 0].clamp_(min=0 , max=snake_case__ ) tensor[:, 1].clamp_(min=0 , max=snake_case__ ) tensor[:, 2].clamp_(min=0 , max=snake_case__ ) tensor[:, 3].clamp_(min=0 , max=snake_case__ )
311
0
'''simple docstring''' import math import qiskit def lowerCamelCase (_SCREAMING_SNAKE_CASE : int = 1 , _SCREAMING_SNAKE_CASE : int = 1 , _SCREAMING_SNAKE_CASE : int = 1 ): if ( isinstance(__lowercase , __lowercase ) or isinstance(__lowercase , __lowercase ) or isinstance(__lowercase , __lowercase ) ): raise TypeError('inputs must be integers.' ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError('inputs must be positive.' ) if ( (math.floor(__lowercase ) != input_a) or (math.floor(__lowercase ) != input_a) or (math.floor(__lowercase ) != carry_in) ): raise ValueError('inputs must be exact integers.' ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError('inputs must be less or equal to 2.' ) # build registers __a : Tuple = qiskit.QuantumRegister(4 , 'qr' ) __a : Optional[int] = qiskit.ClassicalRegister(2 , 'cr' ) # list the entries __a : Tuple = [input_a, input_a, carry_in] __a : Any = qiskit.QuantumCircuit(__lowercase , __lowercase ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(__lowercase ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(__lowercase ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(__lowercase ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , __lowercase ) # measure the last two qbits __a : int = qiskit.Aer.get_backend('aer_simulator' ) __a : Union[str, Any] = qiskit.execute(__lowercase , __lowercase , shots=1_000 ) return job.result().get_counts(__lowercase ) if __name__ == "__main__": print(f'''Total sum count for state is: {quantum_full_adder(1, 1, 1)}''')
27
import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() _UpperCAmelCase = 2 class UpperCAmelCase : '''simple docstring''' def __init__( self , *, # begin keyword-only arguments lowercase="<s>" , lowercase="<pad>" , lowercase="</s>" , lowercase="<unk>" , lowercase=None , ): """simple docstring""" A_ , A_ , A_ , A_ : Tuple = bos, unk, pad, eos A_ : Optional[Any] = [] A_ : Dict = [] A_ : List[Any] = {} A_ : int = self.add_symbol(lowercase ) A_ : Union[str, Any] = self.add_symbol(lowercase ) A_ : Union[str, Any] = self.add_symbol(lowercase ) A_ : Any = self.add_symbol(lowercase ) if extra_special_symbols: for s in extra_special_symbols: self.add_symbol(lowercase ) A_ : Tuple = len(self.symbols ) def __eq__( self , lowercase ): """simple docstring""" return self.indices == other.indices def __getitem__( self , lowercase ): """simple docstring""" if idx < len(self.symbols ): return self.symbols[idx] return self.unk_word def __len__( self ): """simple docstring""" return len(self.symbols ) def __contains__( self , lowercase ): """simple docstring""" return sym in self.indices @classmethod def lowerCAmelCase_ ( cls , lowercase ): """simple docstring""" A_ : int = cls() d.add_from_file(lowercase ) return d def lowerCAmelCase_ ( self , lowercase , lowercase=1 , lowercase=False ): """simple docstring""" if word in self.indices and not overwrite: A_ : List[Any] = self.indices[word] A_ : List[str] = self.count[idx] + n return idx else: A_ : int = len(self.symbols ) A_ : Optional[Any] = idx self.symbols.append(lowercase ) self.count.append(lowercase ) return idx def lowerCAmelCase_ ( self , lowercase ): """simple docstring""" return 0 def lowerCAmelCase_ ( self , lowercase ): """simple docstring""" if isinstance(lowercase , lowercase ): try: with open(lowercase , 'r' , encoding='utf-8' ) as fd: self.add_from_file(lowercase ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception('Incorrect encoding detected in {}, please rebuild the dataset'.format(lowercase ) ) return A_ : Any = f.readlines() A_ : List[Any] = self._load_meta(lowercase ) for line in lines[indices_start_line:]: try: A_ , A_ : int = line.rstrip().rsplit(' ' , 1 ) if field == "#fairseq:overwrite": A_ : Optional[int] = True A_ , A_ : str = line.rsplit(' ' , 1 ) else: A_ : Optional[int] = False A_ : Optional[int] = int(lowercase ) A_ : Tuple = line if word in self and not overwrite: raise RuntimeError( 'Duplicate word found when loading Dictionary: \'{}\'. ' 'Duplicate words can overwrite earlier ones by adding the ' '#fairseq:overwrite flag at the end of the corresponding row ' 'in the dictionary file. If using the Camembert model, please ' 'download an updated copy of the model file.'.format(lowercase ) ) self.add_symbol(lowercase , n=lowercase , overwrite=lowercase ) except ValueError: raise ValueError('Incorrect dictionary format, expected \'<token> <cnt> [flags]\'' ) def UpperCamelCase ( __lowercase : Any ): '''simple docstring''' A_ : Optional[Any] = dict((re.sub(r'@@$' ,'' ,__lowercase ), v) if k.endswith('@@' ) else (re.sub(r'$' ,'</w>' ,__lowercase ), v) for k, v in d.items() ) A_ : Optional[Any] = '<s> <pad> </s> <unk>'.split() # restore the special tokens for k in keep_keys: del da[f'''{k}</w>'''] A_ : Union[str, Any] = d[k] # restore return da def UpperCamelCase ( __lowercase : Any ,__lowercase : str ): '''simple docstring''' if not os.path.exists(__lowercase ): raise ValueError(f'''path {biogpt_checkpoint_path} does not exist!''' ) os.makedirs(__lowercase ,exist_ok=__lowercase ) print(f'''Writing results to {pytorch_dump_folder_path}''' ) # handle various types of models A_ : Optional[Any] = os.path.join(__lowercase ,'checkpoint.pt' ) if not os.path.isfile(__lowercase ): raise ValueError(f'''path to the file {checkpoint_file} does not exist!''' ) A_ : Any = torch.load(__lowercase ,map_location='cpu' ) A_ : str = chkpt['cfg']['model'] # dicts A_ : Any = os.path.join(__lowercase ,'dict.txt' ) if not os.path.isfile(__lowercase ): raise ValueError(f'''path to the file {dict_file} does not exist!''' ) A_ : Optional[int] = Dictionary.load(__lowercase ) A_ : Union[str, Any] = rewrite_dict_keys(src_dict.indices ) A_ : List[Any] = len(__lowercase ) A_ : Tuple = os.path.join(__lowercase ,VOCAB_FILES_NAMES['vocab_file'] ) print(f'''Generating {src_vocab_file} of {src_vocab_size} records''' ) with open(__lowercase ,'w' ,encoding='utf-8' ) as f: f.write(json.dumps(__lowercase ,ensure_ascii=__lowercase ,indent=__lowercase ) ) # merges_file (bpecodes) A_ : List[Any] = os.path.join(__lowercase ,'bpecodes' ) if not os.path.isfile(__lowercase ): raise ValueError(f'''path to the file {bpecodes_file} does not exist!''' ) A_ : Optional[Any] = os.path.join(__lowercase ,VOCAB_FILES_NAMES['merges_file'] ) shutil.copyfile(__lowercase ,__lowercase ) # model config A_ : Dict = os.path.join(__lowercase ,'config.json' ) A_ : List[Any] = { 'activation_dropout': args['activation_dropout'], 'architectures': ['BioGptForCausalLM'], 'attention_probs_dropout_prob': args['attention_dropout'], 'bos_token_id': 0, 'eos_token_id': 2, 'hidden_act': args['activation_fn'], 'hidden_dropout_prob': args['dropout'], 'hidden_size': args['decoder_embed_dim'], 'initializer_range': 0.02, 'intermediate_size': args['decoder_ffn_embed_dim'], 'layer_norm_eps': 1e-1_2, 'layerdrop': args['decoder_layerdrop'], 'max_position_embeddings': args['max_target_positions'], 'model_type': 'biogpt', 'num_attention_heads': args['decoder_attention_heads'], 'num_hidden_layers': args['decoder_layers'], 'pad_token_id': 1, 'scale_embedding': not args['no_scale_embedding'], 'tie_word_embeddings': args['share_decoder_input_output_embed'], 'vocab_size': src_vocab_size, } # good hparam defaults to start with print(f'''Generating {biogpt_model_config_file}''' ) with open(__lowercase ,'w' ,encoding='utf-8' ) as f: f.write(json.dumps(__lowercase ,ensure_ascii=__lowercase ,indent=__lowercase ) ) # tokenizer config A_ : List[Any] = os.path.join(__lowercase ,__lowercase ) A_ : Dict = { 'bos_token': '<s>', 'eos_token': '</s>', 'model_max_length': 10_24, 'pad_token': '<pad>', 'special_tokens_map_file': None, 'tokenizer_class': 'BioGptTokenizer', 'unk_token': '<unk>', } print(f'''Generating {biogpt_tokenizer_config_file}''' ) with open(__lowercase ,'w' ,encoding='utf-8' ) as f: f.write(json.dumps(__lowercase ,ensure_ascii=__lowercase ,indent=__lowercase ) ) # model A_ : Any = chkpt['model'] # remove unneeded keys A_ : List[Any] = [ 'decoder.version', ] for k in ignore_keys: model_state_dict.pop(__lowercase ,__lowercase ) A_ : int = list(model_state_dict.keys() ) for layer_name in layer_names: if layer_name.endswith('output_projection.weight' ): A_ : Union[str, Any] = model_state_dict.pop(__lowercase ) else: A_ : str = model_state_dict.pop(__lowercase ) A_ : Optional[int] = BioGptConfig.from_pretrained(__lowercase ) A_ : List[Any] = BioGptForCausalLM(__lowercase ) # check that it loads ok model_new.load_state_dict(__lowercase ) # save A_ : List[str] = os.path.join(__lowercase ,__lowercase ) print(f'''Generating {pytorch_weights_dump_path}''' ) torch.save(__lowercase ,__lowercase ) print('Conversion is done!' ) if __name__ == "__main__": _UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--biogpt_checkpoint_path""", default=None, type=str, required=True, help=( """Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,""" """ bpecodes, etc.""" ), ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) _UpperCAmelCase = parser.parse_args() convert_biogpt_checkpoint_to_pytorch(args.biogpt_checkpoint_path, args.pytorch_dump_folder_path)
140
0
"""simple docstring""" class __snake_case : """simple docstring""" def __init__( self , __lowerCamelCase , __lowerCamelCase ): '''simple docstring''' __A : str = name __A : Optional[int] = val def __str__( self ): '''simple docstring''' return F"""{self.__class__.__name__}({self.name}, {self.val})""" def __lt__( self , __lowerCamelCase ): '''simple docstring''' return self.val < other.val class __snake_case : """simple docstring""" def __init__( self , __lowerCamelCase ): '''simple docstring''' __A : Union[str, Any] = {} __A : List[str] = {} __A : Dict = self.build_heap(__lowerCamelCase ) def __getitem__( self , __lowerCamelCase ): '''simple docstring''' return self.get_value(__lowerCamelCase ) def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' return (idx - 1) // 2 def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' return idx * 2 + 1 def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' return idx * 2 + 2 def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' return self.heap_dict[key] def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' __A : int = len(__lowerCamelCase ) - 1 __A : List[str] = self.get_parent_idx(__lowerCamelCase ) for idx, i in enumerate(__lowerCamelCase ): __A : List[Any] = idx __A : str = i.val for i in range(__lowerCamelCase , -1 , -1 ): self.sift_down(__lowerCamelCase , __lowerCamelCase ) return array def UpperCamelCase__( self , __lowerCamelCase , __lowerCamelCase ): '''simple docstring''' while True: __A : Dict = self.get_left_child_idx(__lowerCamelCase ) # noqa: E741 __A : List[str] = self.get_right_child_idx(__lowerCamelCase ) __A : Any = idx if l < len(__lowerCamelCase ) and array[l] < array[idx]: __A : str = l if r < len(__lowerCamelCase ) and array[r] < array[smallest]: __A : Tuple = r if smallest != idx: __A , __A : Dict = array[smallest], array[idx] ( ( __A ) , ( __A ) , ) : List[Any] = ( self.idx_of_element[array[smallest]], self.idx_of_element[array[idx]], ) __A : Dict = smallest else: break def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' __A : Dict = self.get_parent_idx(__lowerCamelCase ) while p >= 0 and self.heap[p] > self.heap[idx]: __A , __A : List[str] = self.heap[idx], self.heap[p] __A , __A : Tuple = ( self.idx_of_element[self.heap[idx]], self.idx_of_element[self.heap[p]], ) __A : Tuple = p __A : Tuple = self.get_parent_idx(__lowerCamelCase ) def UpperCamelCase__( self ): '''simple docstring''' return self.heap[0] def UpperCamelCase__( self ): '''simple docstring''' __A , __A : Optional[int] = self.heap[-1], self.heap[0] __A , __A : List[Any] = ( self.idx_of_element[self.heap[-1]], self.idx_of_element[self.heap[0]], ) __A : Any = self.heap.pop() del self.idx_of_element[x] self.sift_down(0 , self.heap ) return x def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' self.heap.append(__lowerCamelCase ) __A : Optional[Any] = len(self.heap ) - 1 __A : str = node.val self.sift_up(len(self.heap ) - 1 ) def UpperCamelCase__( self ): '''simple docstring''' return len(self.heap ) == 0 def UpperCamelCase__( self , __lowerCamelCase , __lowerCamelCase ): '''simple docstring''' assert ( self.heap[self.idx_of_element[node]].val > new_value ), "newValue must be less that current value" __A : int = new_value __A : Optional[int] = new_value self.sift_up(self.idx_of_element[node] ) a_ = Node("""R""", -1) a_ = Node("""B""", 6) a_ = Node("""A""", 3) a_ = Node("""X""", 1) a_ = Node("""E""", 4) # Use one of these two ways to generate Min-Heap # Generating Min-Heap from array a_ = MinHeap([r, b, a, x, e]) # Generating Min-Heap by Insert method # myMinHeap.insert(a) # myMinHeap.insert(b) # myMinHeap.insert(x) # myMinHeap.insert(r) # myMinHeap.insert(e) # Before print("""Min Heap - before decrease key""") for i in my_min_heap.heap: print(i) print("""Min Heap - After decrease key of node [B -> -17]""") my_min_heap.decrease_key(b, -17) # After for i in my_min_heap.heap: print(i) if __name__ == "__main__": import doctest doctest.testmod()
356
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available a_ = { """configuration_tapas""": ["""TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TapasConfig"""], """tokenization_tapas""": ["""TapasTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ """TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST""", """TapasForMaskedLM""", """TapasForQuestionAnswering""", """TapasForSequenceClassification""", """TapasModel""", """TapasPreTrainedModel""", """load_tf_weights_in_tapas""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ """TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFTapasForMaskedLM""", """TFTapasForQuestionAnswering""", """TFTapasForSequenceClassification""", """TFTapasModel""", """TFTapasPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_tapas import TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP, TapasConfig from .tokenization_tapas import TapasTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tapas import ( TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasPreTrainedModel, load_tf_weights_in_tapas, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_tapas import ( TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TFTapasForMaskedLM, TFTapasForQuestionAnswering, TFTapasForSequenceClassification, TFTapasModel, TFTapasPreTrainedModel, ) else: import sys a_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
291
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase_ : List[Any] = {"configuration_ibert": ["IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "IBertConfig", "IBertOnnxConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : str = [ "IBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "IBertForMaskedLM", "IBertForMultipleChoice", "IBertForQuestionAnswering", "IBertForSequenceClassification", "IBertForTokenClassification", "IBertModel", "IBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys UpperCAmelCase_ : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
91
'''simple docstring''' import os _SCREAMING_SNAKE_CASE : int = {"I": 1, "V": 5, "X": 10, "L": 50, "C": 100, "D": 500, "M": 1000} def UpperCamelCase_( snake_case : str ): '''simple docstring''' snake_case_ = 0 snake_case_ = 0 while index < len(snake_case ) - 1: snake_case_ = SYMBOLS[numerals[index]] snake_case_ = SYMBOLS[numerals[index + 1]] if current_value < next_value: total_value -= current_value else: total_value += current_value index += 1 total_value += SYMBOLS[numerals[index]] return total_value def UpperCamelCase_( snake_case : int ): '''simple docstring''' snake_case_ = "" snake_case_ = num // 1_0_0_0 numerals += m_count * "M" num %= 1_0_0_0 snake_case_ = num // 1_0_0 if c_count == 9: numerals += "CM" c_count -= 9 elif c_count == 4: numerals += "CD" c_count -= 4 if c_count >= 5: numerals += "D" c_count -= 5 numerals += c_count * "C" num %= 1_0_0 snake_case_ = num // 1_0 if x_count == 9: numerals += "XC" x_count -= 9 elif x_count == 4: numerals += "XL" x_count -= 4 if x_count >= 5: numerals += "L" x_count -= 5 numerals += x_count * "X" num %= 1_0 if num == 9: numerals += "IX" num -= 9 elif num == 4: numerals += "IV" num -= 4 if num >= 5: numerals += "V" num -= 5 numerals += num * "I" return numerals def UpperCamelCase_( snake_case : str = "/p089_roman.txt" ): '''simple docstring''' snake_case_ = 0 with open(os.path.dirname(snake_case ) + roman_numerals_filename ) as filea: snake_case_ = filea.readlines() for line in lines: snake_case_ = line.strip() snake_case_ = parse_roman_numerals(snake_case ) snake_case_ = generate_roman_numerals(snake_case ) savings += len(snake_case ) - len(snake_case ) return savings if __name__ == "__main__": print(F"{solution() = }")
85
0
a_ : Dict = '\n# Transformers 설치 방법\n! pip install transformers datasets\n# 마지막 릴리스 대신 소스에서 설치하려면, 위 명령을 주석으로 바꾸고 아래 명령을 해제하세요.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' a_ : Optional[Any] = [{'type': 'code', 'content': INSTALL_CONTENT}] a_ : Tuple = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
327
import baseaa def lowerCamelCase__ (_UpperCAmelCase): return baseaa.aaaencode(string.encode('utf-8')) def lowerCamelCase__ (_UpperCAmelCase): return baseaa.aaadecode(_UpperCAmelCase).decode('utf-8') if __name__ == "__main__": import doctest doctest.testmod()
327
1
def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : int ): if not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): raise ValueError('multiplicative_persistence() only accepts integral values' ) if num < 0: raise ValueError('multiplicative_persistence() does not accept negative values' ) __UpperCamelCase =0 __UpperCamelCase =str(SCREAMING_SNAKE_CASE__ ) while len(SCREAMING_SNAKE_CASE__ ) != 1: __UpperCamelCase =[int(SCREAMING_SNAKE_CASE__ ) for i in num_string] __UpperCamelCase =1 for i in range(0 , len(SCREAMING_SNAKE_CASE__ ) ): total *= numbers[i] __UpperCamelCase =str(SCREAMING_SNAKE_CASE__ ) steps += 1 return steps def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : int ): if not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): raise ValueError('additive_persistence() only accepts integral values' ) if num < 0: raise ValueError('additive_persistence() does not accept negative values' ) __UpperCamelCase =0 __UpperCamelCase =str(SCREAMING_SNAKE_CASE__ ) while len(SCREAMING_SNAKE_CASE__ ) != 1: __UpperCamelCase =[int(SCREAMING_SNAKE_CASE__ ) for i in num_string] __UpperCamelCase =0 for i in range(0 , len(SCREAMING_SNAKE_CASE__ ) ): total += numbers[i] __UpperCamelCase =str(SCREAMING_SNAKE_CASE__ ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
62
import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) _A = logging.getLogger(__name__) def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : List[Any] ): __UpperCamelCase =np.argmax(SCREAMING_SNAKE_CASE__ , axis=1 ) return np.sum(outputs == labels ) def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : List[Any] ): with open(SCREAMING_SNAKE_CASE__ , encoding='utf_8' ) as f: __UpperCamelCase =csv.reader(SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =[] next(SCREAMING_SNAKE_CASE__ ) # skip the first line for line in tqdm(SCREAMING_SNAKE_CASE__ ): output.append((' '.join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Dict ): __UpperCamelCase =[] for dataset in encoded_datasets: __UpperCamelCase =len(SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =np.zeros((n_batch, 2, input_len) , dtype=np.intaa ) __UpperCamelCase =np.zeros((n_batch, 2) , dtype=np.intaa ) __UpperCamelCase =np.full((n_batch, 2, input_len) , fill_value=-1_00 , dtype=np.intaa ) __UpperCamelCase =np.zeros((n_batch,) , dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(SCREAMING_SNAKE_CASE__ ): __UpperCamelCase =[start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __UpperCamelCase =[start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __UpperCamelCase =with_conta __UpperCamelCase =with_conta __UpperCamelCase =len(SCREAMING_SNAKE_CASE__ ) - 1 __UpperCamelCase =len(SCREAMING_SNAKE_CASE__ ) - 1 __UpperCamelCase =with_conta __UpperCamelCase =with_conta __UpperCamelCase =mc_label __UpperCamelCase =(input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(SCREAMING_SNAKE_CASE__ ) for t in all_inputs ) ) return tensor_datasets def _UpperCAmelCase ( ): __UpperCamelCase =argparse.ArgumentParser() parser.add_argument('--model_name' , type=SCREAMING_SNAKE_CASE__ , default='openai-gpt' , help='pretrained model name' ) parser.add_argument('--do_train' , action='store_true' , help='Whether to run training.' ) parser.add_argument('--do_eval' , action='store_true' , help='Whether to run eval on the dev set.' ) parser.add_argument( '--output_dir' , default=SCREAMING_SNAKE_CASE__ , type=SCREAMING_SNAKE_CASE__ , required=SCREAMING_SNAKE_CASE__ , help='The output directory where the model predictions and checkpoints will be written.' , ) parser.add_argument('--train_dataset' , type=SCREAMING_SNAKE_CASE__ , default='' ) parser.add_argument('--eval_dataset' , type=SCREAMING_SNAKE_CASE__ , default='' ) parser.add_argument('--seed' , type=SCREAMING_SNAKE_CASE__ , default=42 ) parser.add_argument('--num_train_epochs' , type=SCREAMING_SNAKE_CASE__ , default=3 ) parser.add_argument('--train_batch_size' , type=SCREAMING_SNAKE_CASE__ , default=8 ) parser.add_argument('--eval_batch_size' , type=SCREAMING_SNAKE_CASE__ , default=16 ) parser.add_argument('--adam_epsilon' , default=1E-8 , type=SCREAMING_SNAKE_CASE__ , help='Epsilon for Adam optimizer.' ) parser.add_argument('--max_grad_norm' , type=SCREAMING_SNAKE_CASE__ , default=1 ) parser.add_argument( '--max_steps' , default=-1 , type=SCREAMING_SNAKE_CASE__ , help=( 'If > 0: set total number of training steps to perform. Override num_train_epochs.' ) , ) parser.add_argument( '--gradient_accumulation_steps' , type=SCREAMING_SNAKE_CASE__ , default=1 , help='Number of updates steps to accumulate before performing a backward/update pass.' , ) parser.add_argument('--learning_rate' , type=SCREAMING_SNAKE_CASE__ , default=6.25E-5 ) parser.add_argument('--warmup_steps' , default=0 , type=SCREAMING_SNAKE_CASE__ , help='Linear warmup over warmup_steps.' ) parser.add_argument('--lr_schedule' , type=SCREAMING_SNAKE_CASE__ , default='warmup_linear' ) parser.add_argument('--weight_decay' , type=SCREAMING_SNAKE_CASE__ , default=0.01 ) parser.add_argument('--lm_coef' , type=SCREAMING_SNAKE_CASE__ , default=0.9 ) parser.add_argument('--n_valid' , type=SCREAMING_SNAKE_CASE__ , default=3_74 ) parser.add_argument('--server_ip' , type=SCREAMING_SNAKE_CASE__ , default='' , help='Can be used for distant debugging.' ) parser.add_argument('--server_port' , type=SCREAMING_SNAKE_CASE__ , default='' , help='Can be used for distant debugging.' ) __UpperCamelCase =parser.parse_args() print(SCREAMING_SNAKE_CASE__ ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print('Waiting for debugger attach' ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=SCREAMING_SNAKE_CASE__ ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) __UpperCamelCase =torch.device('cuda' if torch.cuda.is_available() else 'cpu' ) __UpperCamelCase =torch.cuda.device_count() logger.info('device: {}, n_gpu {}'.format(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) if not args.do_train and not args.do_eval: raise ValueError('At least one of `do_train` or `do_eval` must be True.' ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset __UpperCamelCase =['_start_', '_delimiter_', '_classify_'] __UpperCamelCase =OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(SCREAMING_SNAKE_CASE__ ) ) model.to(SCREAMING_SNAKE_CASE__ ) # Load and encode the datasets def tokenize_and_encode(SCREAMING_SNAKE_CASE__ : str ): if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(SCREAMING_SNAKE_CASE__ ) ) elif isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): return obj return [tokenize_and_encode(SCREAMING_SNAKE_CASE__ ) for o in obj] logger.info('Encoding dataset...' ) __UpperCamelCase =load_rocstories_dataset(args.train_dataset ) __UpperCamelCase =load_rocstories_dataset(args.eval_dataset ) __UpperCamelCase =(train_dataset, eval_dataset) __UpperCamelCase =tokenize_and_encode(SCREAMING_SNAKE_CASE__ ) # Compute the max input length for the Transformer __UpperCamelCase =model.config.n_positions // 2 - 2 __UpperCamelCase =max( len(story[:max_length] ) + max(len(conta[:max_length] ) , len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) __UpperCamelCase =min(SCREAMING_SNAKE_CASE__ , model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders __UpperCamelCase =pre_process_datasets(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , *SCREAMING_SNAKE_CASE__ ) __UpperCamelCase , __UpperCamelCase =tensor_datasets[0], tensor_datasets[1] __UpperCamelCase =TensorDataset(*SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =RandomSampler(SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =DataLoader(SCREAMING_SNAKE_CASE__ , sampler=SCREAMING_SNAKE_CASE__ , batch_size=args.train_batch_size ) __UpperCamelCase =TensorDataset(*SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =SequentialSampler(SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =DataLoader(SCREAMING_SNAKE_CASE__ , sampler=SCREAMING_SNAKE_CASE__ , batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: __UpperCamelCase =args.max_steps __UpperCamelCase =args.max_steps // (len(SCREAMING_SNAKE_CASE__ ) // args.gradient_accumulation_steps) + 1 else: __UpperCamelCase =len(SCREAMING_SNAKE_CASE__ ) // args.gradient_accumulation_steps * args.num_train_epochs __UpperCamelCase =list(model.named_parameters() ) __UpperCamelCase =['bias', 'LayerNorm.bias', 'LayerNorm.weight'] __UpperCamelCase =[ { 'params': [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], 'weight_decay': args.weight_decay, }, {'params': [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], 'weight_decay': 0.0}, ] __UpperCamelCase =AdamW(SCREAMING_SNAKE_CASE__ , lr=args.learning_rate , eps=args.adam_epsilon ) __UpperCamelCase =get_linear_schedule_with_warmup( SCREAMING_SNAKE_CASE__ , num_warmup_steps=args.warmup_steps , num_training_steps=SCREAMING_SNAKE_CASE__ ) if args.do_train: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase =0, 0, None model.train() for _ in trange(int(args.num_train_epochs ) , desc='Epoch' ): __UpperCamelCase =0 __UpperCamelCase =0 __UpperCamelCase =tqdm(SCREAMING_SNAKE_CASE__ , desc='Training' ) for step, batch in enumerate(SCREAMING_SNAKE_CASE__ ): __UpperCamelCase =tuple(t.to(SCREAMING_SNAKE_CASE__ ) for t in batch ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase =batch __UpperCamelCase =model(SCREAMING_SNAKE_CASE__ , mc_token_ids=SCREAMING_SNAKE_CASE__ , lm_labels=SCREAMING_SNAKE_CASE__ , mc_labels=SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() __UpperCamelCase =( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 __UpperCamelCase ='Training loss: {:.2e} lr: {:.2e}'.format(SCREAMING_SNAKE_CASE__ , scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer __UpperCamelCase =model.module if hasattr(SCREAMING_SNAKE_CASE__ , 'module' ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` __UpperCamelCase =os.path.join(args.output_dir , SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =os.path.join(args.output_dir , SCREAMING_SNAKE_CASE__ ) torch.save(model_to_save.state_dict() , SCREAMING_SNAKE_CASE__ ) model_to_save.config.to_json_file(SCREAMING_SNAKE_CASE__ ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned __UpperCamelCase =OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) __UpperCamelCase =OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(SCREAMING_SNAKE_CASE__ ) if args.do_eval: model.eval() __UpperCamelCase , __UpperCamelCase =0, 0 __UpperCamelCase , __UpperCamelCase =0, 0 for batch in tqdm(SCREAMING_SNAKE_CASE__ , desc='Evaluating' ): __UpperCamelCase =tuple(t.to(SCREAMING_SNAKE_CASE__ ) for t in batch ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase =batch with torch.no_grad(): __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase =model( SCREAMING_SNAKE_CASE__ , mc_token_ids=SCREAMING_SNAKE_CASE__ , lm_labels=SCREAMING_SNAKE_CASE__ , mc_labels=SCREAMING_SNAKE_CASE__ ) __UpperCamelCase =mc_logits.detach().cpu().numpy() __UpperCamelCase =mc_labels.to('cpu' ).numpy() __UpperCamelCase =accuracy(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 __UpperCamelCase =eval_loss / nb_eval_steps __UpperCamelCase =eval_accuracy / nb_eval_examples __UpperCamelCase =tr_loss / nb_tr_steps if args.do_train else None __UpperCamelCase ={'eval_loss': eval_loss, 'eval_accuracy': eval_accuracy, 'train_loss': train_loss} __UpperCamelCase =os.path.join(args.output_dir , 'eval_results.txt' ) with open(SCREAMING_SNAKE_CASE__ , 'w' ) as writer: logger.info('***** Eval results *****' ) for key in sorted(result.keys() ): logger.info(' %s = %s' , SCREAMING_SNAKE_CASE__ , str(result[key] ) ) writer.write('%s = %s\n' % (key, str(result[key] )) ) if __name__ == "__main__": main()
62
1
'''simple docstring''' import numpy as np def _A ( A__ , A__ , A__ = 1e-12 , A__ = 100 , ): """simple docstring""" assert np.shape(A__ )[0] == np.shape(A__ )[1] # Ensure proper dimensionality. assert np.shape(A__ )[0] == np.shape(A__ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(A__ ) == np.iscomplexobj(A__ ) __lowercase = np.iscomplexobj(A__ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(A__ , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. __lowercase = False __lowercase = 0 __lowercase = 0 __lowercase = 1e12 while not convergence: # Multiple matrix by the vector. __lowercase = np.dot(A__ , A__ ) # Normalize the resulting output vector. __lowercase = w / np.linalg.norm(A__ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) __lowercase = vector.conj().T if is_complex else vector.T __lowercase = np.dot(A__ , np.dot(A__ , A__ ) ) # Check convergence. __lowercase = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: __lowercase = True __lowercase = lambda_ if is_complex: __lowercase = np.real(lambda_ ) return lambda_, vector def _A ( ): """simple docstring""" __lowercase = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) __lowercase = np.array([41, 4, 20] ) __lowercase = real_input_matrix.astype(np.complexaaa ) __lowercase = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T __lowercase = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": __lowercase = real_input_matrix __lowercase = real_vector elif problem_type == "complex": __lowercase = complex_input_matrix __lowercase = complex_vector # Our implementation. __lowercase , __lowercase = power_iteration(A__ , A__ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). __lowercase , __lowercase = np.linalg.eigh(A__ ) # Last eigenvalue is the maximum one. __lowercase = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. __lowercase = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(A__ ) - np.abs(A__ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
52
'''simple docstring''' lowerCAmelCase__ = { "joule": 1.0, "kilojoule": 1000, "megajoule": 100_0000, "gigajoule": 10_0000_0000, "wattsecond": 1.0, "watthour": 3600, "kilowatthour": 360_0000, "newtonmeter": 1.0, "calorie_nutr": 4186.8, "kilocalorie_nutr": 418_6800.00, "electronvolt": 1.602_176_634e-19, "britishthermalunit_it": 1055.0_5585, "footpound": 1.355_818, } def _A ( A__ , A__ , A__ ): """simple docstring""" if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: __lowercase = ( F"Incorrect 'from_type' or 'to_type' value: {from_type!r}, {to_type!r}\n" F"Valid values are: {', '.join(A__ )}" ) raise ValueError(A__ ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
52
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available A_ : int = { "configuration_xlm": ["XLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLMConfig", "XLMOnnxConfig"], "tokenization_xlm": ["XLMTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : List[str] = [ "XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XLMForMultipleChoice", "XLMForQuestionAnswering", "XLMForQuestionAnsweringSimple", "XLMForSequenceClassification", "XLMForTokenClassification", "XLMModel", "XLMPreTrainedModel", "XLMWithLMHeadModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Optional[int] = [ "TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLMForMultipleChoice", "TFXLMForQuestionAnsweringSimple", "TFXLMForSequenceClassification", "TFXLMForTokenClassification", "TFXLMMainLayer", "TFXLMModel", "TFXLMPreTrainedModel", "TFXLMWithLMHeadModel", ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys A_ : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
165
"""simple docstring""" import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel A_ : Dict = HfApi() A_ : List[str] = {} # fmt: off A_ : Dict = torch.tensor([ -0.75_15, -1.68_83, 0.24_20, 0.03_00, 0.63_47, 1.34_33, -1.17_43, -3.74_67, 1.23_42, -2.24_85, 0.46_36, 0.80_76, -0.79_91, 0.39_69, 0.84_98, 0.91_89, -1.88_87, -3.35_22, 0.76_39, 0.20_40, 0.62_71, -2.71_48, -1.63_16, 3.08_39, 0.31_86, 0.27_21, -0.97_59, -1.24_61, 2.62_57, 1.35_57 ]) A_ : List[Any] = torch.tensor([ -2.36_39, -2.53_44, 0.00_54, -0.66_74, 1.59_90, 1.01_58, 0.31_24, -2.14_36, 1.87_95, -2.54_29, -0.15_66, -0.39_73, 1.24_90, 2.64_47, 1.22_83, -0.52_08, -2.81_54, -3.51_19, 2.38_38, 1.20_33, 1.72_01, -2.12_56, -1.45_76, 2.79_48, 2.42_04, -0.97_52, -1.25_46, 0.80_27, 3.27_58, 3.13_65 ]) A_ : str = torch.tensor([ -0.65_31, -0.68_91, -0.31_72, -0.53_75, -0.91_40, -0.53_67, -0.11_75, -0.78_69, -0.38_08, -0.45_13, -0.20_98, -0.00_83, 0.31_83, 0.51_40, 0.22_47, -0.13_04, -0.13_02, -0.28_02, -0.20_84, -0.20_25, -0.49_67, -0.48_73, -0.08_61, 0.69_25, 0.02_50, 0.12_90, -0.15_43, 0.63_16, 1.04_60, 1.49_43 ]) A_ : List[Any] = torch.tensor([ 0.09_11, 0.11_07, 0.01_82, 0.04_35, -0.08_05, -0.06_08, 0.03_81, 0.21_72, -0.02_80, 0.13_27, -0.02_99, -0.02_55, -0.00_50, -0.11_70, -0.10_46, 0.03_09, 0.13_67, 0.17_28, -0.05_33, -0.07_48, -0.05_34, 0.16_24, 0.03_84, -0.18_05, -0.07_07, 0.06_42, 0.02_20, -0.01_34, -0.13_33, -0.15_05 ]) A_ : Tuple = torch.tensor([ 0.13_21, 0.13_37, 0.04_40, 0.06_22, -0.05_91, -0.03_70, 0.05_03, 0.21_33, -0.01_77, 0.14_15, -0.01_16, -0.01_12, 0.00_44, -0.09_80, -0.07_89, 0.03_95, 0.15_02, 0.17_85, -0.04_88, -0.05_14, -0.04_04, 0.15_39, 0.04_54, -0.15_59, -0.06_65, 0.06_59, 0.03_83, -0.00_05, -0.12_66, -0.13_86 ]) A_ : List[str] = torch.tensor([ 0.11_54, 0.12_18, 0.03_07, 0.05_26, -0.07_11, -0.05_41, 0.03_66, 0.20_78, -0.02_67, 0.13_17, -0.02_26, -0.01_93, -0.00_14, -0.10_55, -0.09_02, 0.03_30, 0.13_91, 0.17_09, -0.05_62, -0.06_93, -0.05_60, 0.14_82, 0.03_81, -0.16_83, -0.06_81, 0.06_61, 0.03_31, -0.00_46, -0.12_68, -0.14_31 ]) A_ : List[Any] = torch.tensor([ 0.11_92, 0.12_40, 0.04_14, 0.06_06, -0.05_57, -0.04_12, 0.04_30, 0.20_42, -0.02_00, 0.13_85, -0.01_15, -0.01_32, 0.00_17, -0.09_65, -0.08_02, 0.03_98, 0.14_33, 0.17_47, -0.04_58, -0.05_33, -0.04_07, 0.15_45, 0.04_19, -0.15_74, -0.06_45, 0.06_26, 0.03_41, -0.00_10, -0.11_99, -0.13_90 ]) A_ : Dict = torch.tensor([ 0.10_75, 0.10_74, 0.02_05, 0.04_31, -0.07_74, -0.06_07, 0.02_98, 0.20_42, -0.03_20, 0.12_67, -0.02_81, -0.02_50, -0.00_64, -0.10_91, -0.09_46, 0.02_90, 0.13_28, 0.16_50, -0.05_80, -0.07_38, -0.05_86, 0.14_40, 0.03_37, -0.17_46, -0.07_12, 0.06_05, 0.02_50, -0.00_99, -0.13_16, -0.14_73 ]) A_ : Tuple = torch.tensor([ -1.45_72, -2.04_81, -0.04_14, -0.60_05, 1.41_36, 0.58_48, 0.40_28, -2.73_30, 1.22_12, -2.12_28, 0.21_55, 0.40_39, 0.76_62, 2.05_35, 0.74_77, -0.32_43, -2.17_58, -2.76_48, 1.69_47, 0.70_26, 1.23_38, -1.60_78, -0.86_82, 2.28_10, 1.85_74, -0.57_18, -0.55_86, -0.01_86, 2.34_15, 2.12_51]) A_ : str = torch.tensor([ -1.36_90, -1.97_20, -0.40_90, -0.69_66, 1.46_60, 0.99_38, -0.13_85, -2.73_24, 0.77_36, -1.89_17, 0.29_23, 0.42_93, 0.16_93, 1.41_12, 1.18_87, -0.31_81, -2.21_60, -2.63_81, 1.31_70, 0.81_63, 0.92_40, -1.65_44, -0.60_99, 2.52_59, 1.64_30, -0.90_90, -0.93_92, -0.01_26, 2.42_68, 2.32_66 ]) A_ : str = torch.tensor([ -1.35_25, -1.96_28, -0.39_56, -0.68_60, 1.46_64, 1.00_14, -0.12_59, -2.72_12, 0.77_72, -1.88_11, 0.29_96, 0.43_88, 0.17_04, 1.40_29, 1.17_01, -0.30_27, -2.20_53, -2.62_87, 1.33_50, 0.81_31, 0.92_74, -1.62_92, -0.60_98, 2.51_31, 1.65_05, -0.89_58, -0.92_98, -0.01_51, 2.42_57, 2.33_55 ]) A_ : int = torch.tensor([ -2.05_85, -2.78_97, -0.28_50, -0.89_40, 1.90_52, 0.57_02, 0.63_45, -3.89_59, 1.59_32, -3.23_19, 0.19_74, 0.02_87, 1.75_66, 2.65_43, 0.83_87, -0.53_51, -3.27_36, -4.33_75, 2.90_29, 1.63_90, 1.46_40, -2.17_01, -1.90_13, 2.93_41, 3.49_81, -0.62_55, -1.16_44, -0.15_91, 3.70_97, 3.20_66 ]) A_ : int = torch.tensor([ -2.31_39, -2.55_94, -0.01_97, -0.67_85, 1.70_01, 1.16_06, 0.30_75, -2.17_40, 1.80_71, -2.56_30, -0.09_26, -0.38_11, 1.21_16, 2.62_46, 1.27_31, -0.53_98, -2.81_53, -3.61_40, 2.38_93, 1.32_62, 1.62_58, -2.18_56, -1.32_67, 2.83_95, 2.37_79, -1.06_23, -1.24_68, 0.89_59, 3.33_67, 3.22_43 ]) A_ : str = torch.tensor([ -2.06_28, -2.76_67, -0.20_89, -0.82_63, 2.05_39, 0.59_92, 0.64_95, -3.83_36, 1.60_25, -3.28_17, 0.17_21, -0.06_33, 1.75_16, 2.70_39, 0.81_00, -0.59_08, -3.21_13, -4.43_43, 2.92_57, 1.36_32, 1.55_62, -2.14_89, -1.98_94, 3.05_60, 3.33_96, -0.73_28, -1.04_17, 0.03_83, 3.70_93, 3.23_43 ]) A_ : Optional[int] = torch.tensor([ -1.45_74, -2.05_69, -0.04_73, -0.61_17, 1.40_18, 0.57_69, 0.41_29, -2.73_44, 1.22_41, -2.13_97, 0.20_00, 0.39_37, 0.76_16, 2.04_53, 0.73_24, -0.33_91, -2.17_46, -2.77_44, 1.69_63, 0.69_21, 1.21_87, -1.61_72, -0.88_77, 2.24_39, 1.84_71, -0.58_39, -0.56_05, -0.04_64, 2.32_50, 2.12_19 ]) # fmt: on A_ : List[str] = api.list_models(filter="diffusers") for mod in models: if "google" in mod.author or mod.modelId == "CompVis/ldm-celebahq-256": A_ : Dict = "/home/patrick/google_checkpoints/" + mod.modelId.split("/")[-1] print(F'Started running {mod.modelId}!!!') if mod.modelId.startswith("CompVis"): A_ : int = UNetaDModel.from_pretrained(local_checkpoint, subfolder="unet") else: A_ : Optional[int] = UNetaDModel.from_pretrained(local_checkpoint) torch.manual_seed(0) random.seed(0) A_ : Any = torch.randn(1, model.config.in_channels, model.config.sample_size, model.config.sample_size) A_ : int = torch.tensor([10] * noise.shape[0]) with torch.no_grad(): A_ : Optional[int] = model(noise, time_step).sample assert torch.allclose( logits[0, 0, 0, :30], results["_".join("_".join(mod.modelId.split("/")).split("-"))], atol=1E-3 ) print(F'{mod.modelId} has passed successfully!!!')
165
1
"""simple docstring""" import argparse import torch from huggingface_hub import hf_hub_download from transformers import AutoTokenizer, RobertaPreLayerNormConfig, RobertaPreLayerNormForMaskedLM from transformers.utils import logging logging.set_verbosity_info() __UpperCamelCase : List[Any] = logging.get_logger(__name__) def __SCREAMING_SNAKE_CASE ( A_ , A_ ): lowerCAmelCase__ : str = RobertaPreLayerNormConfig.from_pretrained( A_ , architectures=['''RobertaPreLayerNormForMaskedLM'''] ) # convert state_dict lowerCAmelCase__ : Dict = torch.load(hf_hub_download(repo_id=A_ , filename='''pytorch_model.bin''' ) ) lowerCAmelCase__ : Union[str, Any] = {} for tensor_key, tensor_value in original_state_dict.items(): # The transformer implementation gives the model a unique name, rather than overwiriting 'roberta' if tensor_key.startswith('''roberta.''' ): lowerCAmelCase__ : int = '''roberta_prelayernorm.''' + tensor_key[len('''roberta.''' ) :] # The original implementation contains weights which are not used, remove them from the state_dict if tensor_key.endswith('''.self.LayerNorm.weight''' ) or tensor_key.endswith('''.self.LayerNorm.bias''' ): continue lowerCAmelCase__ : Any = tensor_value lowerCAmelCase__ : List[Any] = RobertaPreLayerNormForMaskedLM.from_pretrained( pretrained_model_name_or_path=A_ , config=A_ , state_dict=A_ ) model.save_pretrained(A_ ) # convert tokenizer lowerCAmelCase__ : Optional[Any] = AutoTokenizer.from_pretrained(A_ ) tokenizer.save_pretrained(A_ ) if __name__ == "__main__": __UpperCamelCase : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint-repo''', default=None, type=str, required=True, help='''Path the official PyTorch dump, e.g. \'andreasmadsen/efficient_mlm_m0.40\'.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __UpperCamelCase : Dict = parser.parse_args() convert_roberta_prelayernorm_checkpoint_to_pytorch(args.checkpoint_repo, args.pytorch_dump_folder_path)
74
"""simple docstring""" from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer __UpperCamelCase : Union[str, Any] = logging.get_logger(__name__) __UpperCamelCase : List[str] = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } __UpperCamelCase : Tuple = { '''vocab_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json''' }, '''merges_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt''' }, '''tokenizer_config_file''': { '''facebook/blenderbot_small-90M''': ( '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json''' ) }, } __UpperCamelCase : List[Any] = { '''facebook/blenderbot_small-90M''': 5_1_2, } class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = BlenderbotSmallTokenizer def __init__( self : Dict ,lowercase_ : Dict=None ,lowercase_ : Union[str, Any]=None ,lowercase_ : Any="<|endoftext|>" ,lowercase_ : Optional[Any]="<|endoftext|>" ,lowercase_ : Dict="<|endoftext|>" ,lowercase_ : Optional[int]=False ,lowercase_ : Union[str, Any]=True ,**lowercase_ : Union[str, Any] ,): super().__init__( ByteLevelBPETokenizer( vocab=lowercase_ ,merges=lowercase_ ,add_prefix_space=lowercase_ ,trim_offsets=lowercase_ ,) ,bos_token=lowercase_ ,eos_token=lowercase_ ,unk_token=lowercase_ ,**lowercase_ ,) lowerCAmelCase__ : Tuple = add_prefix_space def __lowerCAmelCase ( self : Optional[Any] ,lowercase_ : Optional[int] ,lowercase_ : int=None ): lowerCAmelCase__ : Dict = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def __lowerCAmelCase ( self : List[Any] ,lowercase_ : List[int] ,lowercase_ : Optional[List[int]] = None ): lowerCAmelCase__ : Union[str, Any] = [self.sep_token_id] lowerCAmelCase__ : Dict = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
74
1
'''simple docstring''' import numpy as np import torch from torch.utils.data import Dataset from utils import logger class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" def __init__( self : Tuple , lowerCAmelCase__ : int , lowerCAmelCase__ : str ) -> str: '''simple docstring''' _UpperCamelCase = params _UpperCamelCase = np.array(lowerCAmelCase__ ) _UpperCamelCase = np.array([len(lowerCAmelCase__ ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : Optional[Any] , lowerCAmelCase__ : Dict ) -> Optional[Any]: '''simple docstring''' return (self.token_ids[index], self.lengths[index]) def __len__( self : Union[str, Any] ) -> Tuple: '''simple docstring''' return len(self.lengths ) def snake_case__ ( self : Dict ) -> Dict: '''simple docstring''' assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def snake_case__ ( self : Tuple ) -> str: '''simple docstring''' _UpperCamelCase = self.params.max_model_input_size _UpperCamelCase = self.lengths > max_len logger.info(f"""Splitting {sum(lowerCAmelCase__ )} too long sequences.""" ) def divide_chunks(lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Optional[Any] ): return [l[i : i + n] for i in range(0 , len(lowerCAmelCase__ ) , lowerCAmelCase__ )] _UpperCamelCase = [] _UpperCamelCase = [] if self.params.mlm: _UpperCamelCase , _UpperCamelCase = self.params.special_tok_ids['''cls_token'''], self.params.special_tok_ids['''sep_token'''] else: _UpperCamelCase , _UpperCamelCase = self.params.special_tok_ids['''bos_token'''], self.params.special_tok_ids['''eos_token'''] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: _UpperCamelCase = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: _UpperCamelCase = np.insert(lowerCAmelCase__ , 0 , lowerCAmelCase__ ) if sub_s[-1] != sep_id: _UpperCamelCase = np.insert(lowerCAmelCase__ , len(lowerCAmelCase__ ) , lowerCAmelCase__ ) assert len(lowerCAmelCase__ ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(lowerCAmelCase__ ) new_tok_ids.extend(lowerCAmelCase__ ) new_lengths.extend([len(lowerCAmelCase__ ) for l in sub_seqs] ) _UpperCamelCase = np.array(lowerCAmelCase__ ) _UpperCamelCase = np.array(lowerCAmelCase__ ) def snake_case__ ( self : List[Any] ) -> Dict: '''simple docstring''' _UpperCamelCase = len(self ) _UpperCamelCase = self.lengths > 11 _UpperCamelCase = self.token_ids[indices] _UpperCamelCase = self.lengths[indices] _UpperCamelCase = len(self ) logger.info(f"""Remove {init_size - new_size} too short (<=11 tokens) sequences.""" ) def snake_case__ ( self : List[str] ) -> List[Any]: '''simple docstring''' if "unk_token" not in self.params.special_tok_ids: return else: _UpperCamelCase = self.params.special_tok_ids['''unk_token'''] _UpperCamelCase = len(self ) _UpperCamelCase = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) _UpperCamelCase = (unk_occs / self.lengths) < 0.5 _UpperCamelCase = self.token_ids[indices] _UpperCamelCase = self.lengths[indices] _UpperCamelCase = len(self ) logger.info(f"""Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).""" ) def snake_case__ ( self : Any ) -> int: '''simple docstring''' if not self.params.is_master: return logger.info(f"""{len(self )} sequences""" ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def snake_case__ ( self : List[Any] , lowerCAmelCase__ : Dict ) -> List[Any]: '''simple docstring''' _UpperCamelCase = [t[0] for t in batch] _UpperCamelCase = [t[1] for t in batch] assert len(lowerCAmelCase__ ) == len(lowerCAmelCase__ ) # Max for paddings _UpperCamelCase = max(lowerCAmelCase__ ) # Pad token ids if self.params.mlm: _UpperCamelCase = self.params.special_tok_ids['''pad_token'''] else: _UpperCamelCase = self.params.special_tok_ids['''unk_token'''] _UpperCamelCase = [list(t.astype(lowerCAmelCase__ ) ) + [pad_idx] * (max_seq_len_ - len(lowerCAmelCase__ )) for t in token_ids] assert len(tk_ ) == len(lowerCAmelCase__ ) assert all(len(lowerCAmelCase__ ) == max_seq_len_ for t in tk_ ) _UpperCamelCase = torch.tensor(tk_ ) # (bs, max_seq_len_) _UpperCamelCase = torch.tensor(lowerCAmelCase__ ) # (bs) return tk_t, lg_t
324
'''simple docstring''' import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home lowercase__ : Union[str, Any] = HUGGINGFACE_HUB_CACHE lowercase__ : int = 'config.json' lowercase__ : Optional[int] = 'diffusion_pytorch_model.bin' lowercase__ : List[str] = 'diffusion_flax_model.msgpack' lowercase__ : str = 'model.onnx' lowercase__ : Optional[int] = 'diffusion_pytorch_model.safetensors' lowercase__ : List[str] = 'weights.pb' lowercase__ : str = 'https://huggingface.co' lowercase__ : str = default_cache_path lowercase__ : Optional[int] = 'diffusers_modules' lowercase__ : Optional[int] = os.getenv('HF_MODULES_CACHE', os.path.join(hf_cache_home, 'modules')) lowercase__ : Tuple = ['fp16', 'non-ema'] lowercase__ : int = '.self_attn'
324
1
"""simple docstring""" import re def _lowercase ( __lowerCAmelCase ) -> bool: SCREAMING_SNAKE_CASE__ : Tuple = re.compile( r"""^(?:0|94|\+94|0{2}94)""" r"""7(0|1|2|4|5|6|7|8)""" r"""(-| |)""" r"""\d{7}$""" ) return bool(re.search(__lowerCAmelCase , __lowerCAmelCase ) ) if __name__ == "__main__": a :Optional[int] = "0094702343221" print(is_sri_lankan_phone_number(phone))
56
"""simple docstring""" def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> int: return number | (1 << position) def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> int: return number & ~(1 << position) def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> int: return number ^ (1 << position) def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> bool: return ((number >> position) & 1) == 1 def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> int: return int((number & (1 << position)) != 0 ) if __name__ == "__main__": import doctest doctest.testmod()
56
1
import os import sys import unittest SCREAMING_SNAKE_CASE_:Tuple = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, create_dummy_object, find_backend, read_init # noqa: E402 # Align TRANSFORMERS_PATH in check_dummies with the current path SCREAMING_SNAKE_CASE_:List[Any] = os.path.join(git_repo_path, """src""", """diffusers""") class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' def _lowerCAmelCase ( self ): A : int = find_backend(""" if not is_torch_available():""" ) self.assertEqual(lowerCamelCase__, """torch""" ) # backend_with_underscore = find_backend(" if not is_tensorflow_text_available():") # self.assertEqual(backend_with_underscore, "tensorflow_text") A : Optional[int] = find_backend(""" if not (is_torch_available() and is_transformers_available()):""" ) self.assertEqual(lowerCamelCase__, """torch_and_transformers""" ) # double_backend_with_underscore = find_backend( # " if not (is_sentencepiece_available() and is_tensorflow_text_available()):" # ) # self.assertEqual(double_backend_with_underscore, "sentencepiece_and_tensorflow_text") A : Tuple = find_backend( """ if not (is_torch_available() and is_transformers_available() and is_onnx_available()):""" ) self.assertEqual(lowerCamelCase__, """torch_and_transformers_and_onnx""" ) def _lowerCAmelCase ( self ): A : List[Any] = read_init() # We don't assert on the exact list of keys to allow for smooth grow of backend-specific objects self.assertIn("""torch""", lowerCamelCase__ ) self.assertIn("""torch_and_transformers""", lowerCamelCase__ ) self.assertIn("""flax_and_transformers""", lowerCamelCase__ ) self.assertIn("""torch_and_transformers_and_onnx""", lowerCamelCase__ ) # Likewise, we can't assert on the exact content of a key self.assertIn("""UNet2DModel""", objects["""torch"""] ) self.assertIn("""FlaxUNet2DConditionModel""", objects["""flax"""] ) self.assertIn("""StableDiffusionPipeline""", objects["""torch_and_transformers"""] ) self.assertIn("""FlaxStableDiffusionPipeline""", objects["""flax_and_transformers"""] ) self.assertIn("""LMSDiscreteScheduler""", objects["""torch_and_scipy"""] ) self.assertIn("""OnnxStableDiffusionPipeline""", objects["""torch_and_transformers_and_onnx"""] ) def _lowerCAmelCase ( self ): A : Optional[Any] = create_dummy_object("""CONSTANT""", """'torch'""" ) self.assertEqual(lowerCamelCase__, """\nCONSTANT = None\n""" ) A : List[str] = create_dummy_object("""function""", """'torch'""" ) self.assertEqual( lowerCamelCase__, """\ndef function(*args, **kwargs):\n requires_backends(function, 'torch')\n""" ) A : Dict = """ class FakeClass(metaclass=DummyObject): _backends = 'torch' def __init__(self, *args, **kwargs): requires_backends(self, 'torch') @classmethod def from_config(cls, *args, **kwargs): requires_backends(cls, 'torch') @classmethod def from_pretrained(cls, *args, **kwargs): requires_backends(cls, 'torch') """ A : List[str] = create_dummy_object("""FakeClass""", """'torch'""" ) self.assertEqual(lowerCamelCase__, lowerCamelCase__ ) def _lowerCAmelCase ( self ): A : Optional[int] = """# This file is autogenerated by the command `make fix-copies`, do not edit. from ..utils import DummyObject, requires_backends CONSTANT = None def function(*args, **kwargs): requires_backends(function, [\"torch\"]) class FakeClass(metaclass=DummyObject): _backends = [\"torch\"] def __init__(self, *args, **kwargs): requires_backends(self, [\"torch\"]) @classmethod def from_config(cls, *args, **kwargs): requires_backends(cls, [\"torch\"]) @classmethod def from_pretrained(cls, *args, **kwargs): requires_backends(cls, [\"torch\"]) """ A : Any = create_dummy_files({"""torch""": ["""CONSTANT""", """function""", """FakeClass"""]} ) self.assertEqual(dummy_files["""torch"""], lowerCamelCase__ )
116
from __future__ import annotations from math import ceil, floor, sqrt def __UpperCamelCase ( _lowerCAmelCase = 200_0000 ) -> int: """simple docstring""" A : list[int] = [0] A : int for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target A : int = 0 # the area corresponding to the grid that gives the product closest to target A : int = 0 # an estimate of b, using the quadratic formula A : float # the largest integer less than b_estimate A : int # the largest integer less than b_estimate A : int # the triangle number corresponding to b_floor A : int # the triangle number corresponding to b_ceil A : int for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): A : Union[str, Any] = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 A : List[Any] = floor(_lowerCAmelCase ) A : Tuple = ceil(_lowerCAmelCase ) A : int = triangle_numbers[b_floor] A : Dict = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): A : Optional[int] = triangle_b_first_guess * triangle_a A : Optional[int] = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): A : Tuple = triangle_b_second_guess * triangle_a A : Tuple = idx_a * b_ceil return area if __name__ == "__main__": print(F"""{solution() = }""")
116
1
import argparse import os import re import packaging.version lowerCAmelCase_ = "examples/" lowerCAmelCase_ = { "examples": (re.compile(R'^check_min_version\(\"[^\"]+\"\)\s*$', re.MULTILINE), "check_min_version(\"VERSION\")\n"), "init": (re.compile(R'^__version__\s+=\s+\"([^\"]+)\"\s*$', re.MULTILINE), "__version__ = \"VERSION\"\n"), "setup": (re.compile(R'^(\s*)version\s*=\s*\"[^\"]+\",', re.MULTILINE), R"\1version=\"VERSION\","), "doc": (re.compile(R'^(\s*)release\s*=\s*\"[^\"]+\"$', re.MULTILINE), "release = \"VERSION\"\n"), } lowerCAmelCase_ = { "init": "src/transformers/__init__.py", "setup": "setup.py", } lowerCAmelCase_ = "README.md" def snake_case( __magic_name__ , __magic_name__ , __magic_name__ ) -> Optional[Any]: '''simple docstring''' with open(lowerCAmelCase__ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowercase : List[Any] = f.read() lowercase , lowercase : List[Any] = REPLACE_PATTERNS[pattern] lowercase : str = replace.replace('''VERSION''' , lowerCAmelCase__ ) lowercase : List[str] = re_pattern.sub(lowerCAmelCase__ , lowerCAmelCase__ ) with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(lowerCAmelCase__ ) def snake_case( __magic_name__ ) -> Optional[Any]: '''simple docstring''' for folder, directories, fnames in os.walk(lowerCAmelCase__ ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(lowerCAmelCase__ , lowerCAmelCase__ ) , lowerCAmelCase__ , pattern='''examples''' ) def snake_case( __magic_name__ , __magic_name__=False ) -> List[Any]: '''simple docstring''' for pattern, fname in REPLACE_FILES.items(): update_version_in_file(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) if not patch: update_version_in_examples(lowerCAmelCase__ ) def snake_case( ) -> str: '''simple docstring''' lowercase : Any = '''🤗 Transformers currently provides the following architectures''' lowercase : Tuple = '''1. Want to contribute a new model?''' with open(lowerCAmelCase__ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowercase : Optional[int] = f.readlines() # Find the start of the list. lowercase : Tuple = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 lowercase : List[Any] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): lowercase : str = lines[index].replace( '''https://huggingface.co/docs/transformers/main/model_doc''' , '''https://huggingface.co/docs/transformers/model_doc''' , ) index += 1 with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(lowerCAmelCase__ ) def snake_case( ) -> Optional[int]: '''simple docstring''' with open(REPLACE_FILES['''init'''] , '''r''' ) as f: lowercase : str = f.read() lowercase : Union[str, Any] = REPLACE_PATTERNS['''init'''][0].search(lowerCAmelCase__ ).groups()[0] return packaging.version.parse(lowerCAmelCase__ ) def snake_case( __magic_name__=False ) -> Any: '''simple docstring''' lowercase : int = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: lowercase : List[Any] = default_version.base_version elif patch: lowercase : Tuple = F"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: lowercase : Tuple = F"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. lowercase : Union[str, Any] = input(F"""Which version are you releasing? [{default_version}]""" ) if len(lowerCAmelCase__ ) == 0: lowercase : Tuple = default_version print(F"""Updating version to {version}.""" ) global_version_update(lowerCAmelCase__ , patch=lowerCAmelCase__ ) if not patch: print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() def snake_case( ) -> Optional[Any]: '''simple docstring''' lowercase : Optional[int] = get_version() lowercase : Dict = F"""{current_version.major}.{current_version.minor + 1}.0.dev0""" lowercase : Optional[Any] = current_version.base_version # Check with the user we got that right. lowercase : int = input(F"""Which version are we developing now? [{dev_version}]""" ) if len(lowerCAmelCase__ ) == 0: lowercase : Any = dev_version print(F"""Updating version to {version}.""" ) global_version_update(lowerCAmelCase__ ) print('''Cleaning main README, don\'t forget to run `make fix-copies`.''' ) clean_main_ref_in_model_list() if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() parser.add_argument('--post_release', action='store_true', help='Whether this is pre or post release.') parser.add_argument('--patch', action='store_true', help='Whether or not this is a patch release.') lowerCAmelCase_ = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('Nothing to do after a patch :-)') else: post_release_work()
366
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) lowerCAmelCase_ = '\\n Text data.\n Second line of data.' lowerCAmelCase_ = 'file' @pytest.fixture(scope='''session''' ) def snake_case( __magic_name__ ) -> List[str]: '''simple docstring''' lowercase : Any = tmp_path_factory.mktemp('''data''' ) / (FILE_PATH + '''.zstd''') lowercase : List[Any] = bytes(__magic_name__ , '''utf-8''' ) with zstd.open(__magic_name__ , '''wb''' ) as f: f.write(__magic_name__ ) return path @pytest.fixture def snake_case( __magic_name__ ) -> List[Any]: '''simple docstring''' with open(os.path.join(tmpfs.local_root_dir , __magic_name__ ) , '''w''' ) as f: f.write(__magic_name__ ) return FILE_PATH @pytest.mark.parametrize('''compression_format''' , ['''gzip''', '''xz''', '''zstd'''] ) def snake_case( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> List[str]: '''simple docstring''' lowercase : Optional[int] = {'''gzip''': gz_file, '''xz''': xz_file, '''zstd''': zstd_path} lowercase : int = input_paths[compression_format] lowercase : Any = tmp_path / '''cache''' lowercase : int = DownloadConfig(cache_dir=__magic_name__ , extract_compressed_file=__magic_name__ ) lowercase : Optional[Any] = cached_path(__magic_name__ , download_config=__magic_name__ ) with open(__magic_name__ ) as f: lowercase : Optional[int] = f.read() with open(__magic_name__ ) as f: lowercase : str = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize('''default_extracted''' , [True, False] ) @pytest.mark.parametrize('''default_cache_dir''' , [True, False] ) def snake_case( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> List[str]: '''simple docstring''' lowercase : Dict = '''custom_cache''' lowercase : Union[str, Any] = '''custom_extracted_dir''' lowercase : str = tmp_path / '''custom_extracted_path''' if default_extracted: lowercase : Tuple = ('''downloads''' if default_cache_dir else custom_cache_dir, '''extracted''') else: monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_DIR''' , __magic_name__ ) monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_PATH''' , str(__magic_name__ ) ) lowercase : int = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) lowercase : List[str] = xz_file lowercase : Any = ( DownloadConfig(extract_compressed_file=__magic_name__ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=__magic_name__ ) ) lowercase : Optional[int] = cached_path(__magic_name__ , download_config=__magic_name__ ) assert Path(__magic_name__ ).parent.parts[-2:] == expected def snake_case( __magic_name__ ) -> List[Any]: '''simple docstring''' lowercase : Any = str(Path(__magic_name__ ).resolve() ) assert cached_path(__magic_name__ ) == text_file # relative path lowercase : Union[str, Any] = str(Path(__magic_name__ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(__magic_name__ ) == text_file def snake_case( __magic_name__ ) -> Union[str, Any]: '''simple docstring''' lowercase : List[Any] = str(tmp_path.resolve() / '''__missing_file__.txt''' ) with pytest.raises(__magic_name__ ): cached_path(__magic_name__ ) # relative path lowercase : Optional[int] = '''./__missing_file__.txt''' with pytest.raises(__magic_name__ ): cached_path(__magic_name__ ) def snake_case( __magic_name__ ) -> List[str]: '''simple docstring''' lowercase : List[str] = get_from_cache(F"""tmp://{tmpfs_file}""" ) with open(__magic_name__ ) as f: lowercase : List[Any] = f.read() assert output_file_content == FILE_CONTENT @patch('''datasets.config.HF_DATASETS_OFFLINE''' , __magic_name__ ) def snake_case( ) -> List[Any]: '''simple docstring''' with pytest.raises(__magic_name__ ): cached_path('''https://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , __magic_name__ ) def snake_case( __magic_name__ ) -> Tuple: '''simple docstring''' lowercase : Any = tmp_path_factory.mktemp('''data''' ) / '''file.html''' with pytest.raises(__magic_name__ ): http_get('''https://huggingface.co''' , temp_file=__magic_name__ ) with pytest.raises(__magic_name__ ): http_head('''https://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , __magic_name__ ) def snake_case( __magic_name__ ) -> Optional[Any]: '''simple docstring''' lowercase : Optional[Any] = tmp_path_factory.mktemp('''data''' ) / '''file.html''' with pytest.raises(__magic_name__ ): ftp_get('''ftp://huggingface.co''' , temp_file=__magic_name__ ) with pytest.raises(__magic_name__ ): ftp_head('''ftp://huggingface.co''' ) @patch('''datasets.config.HF_DATASETS_OFFLINE''' , __magic_name__ ) def snake_case( __magic_name__ ) -> List[str]: '''simple docstring''' lowercase : str = tmp_path_factory.mktemp('''data''' ) / '''file.html''' with pytest.raises(__magic_name__ ): fsspec_get('''s3://huggingface.co''' , temp_file=__magic_name__ ) with pytest.raises(__magic_name__ ): fsspec_head('''s3://huggingface.co''' )
116
0
def lowerCamelCase__ ( _a , _a , _a): return round(float(moles / volume) * nfactor) def lowerCamelCase__ ( _a , _a , _a): return round(float((moles * 0.0821 * temperature) / (volume))) def lowerCamelCase__ ( _a , _a , _a): return round(float((moles * 0.0821 * temperature) / (pressure))) def lowerCamelCase__ ( _a , _a , _a): return round(float((pressure * volume) / (0.0821 * moles))) if __name__ == "__main__": import doctest doctest.testmod()
76
def __UpperCamelCase ( _A = 1000000 ): lowerCAmelCase_ = 1 lowerCAmelCase_ = 1 lowerCAmelCase_ = {1: 1} for inputa in range(2 , _A ): lowerCAmelCase_ = 0 lowerCAmelCase_ = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: lowerCAmelCase_ = (3 * number) + 1 counter += 1 if inputa not in counters: lowerCAmelCase_ = counter if counter > pre_counter: lowerCAmelCase_ = inputa lowerCAmelCase_ = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
278
0
'''simple docstring''' import inspect import math import tempfile import unittest import numpy as np from transformers import ViTMAEConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMAEForPreTraining, ViTMAEModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class a : def __init__( self : Optional[Any] , lowercase_ : Tuple , lowercase_ : Tuple=13 , lowercase_ : Dict=30 , lowercase_ : Dict=2 , lowercase_ : Optional[int]=3 , lowercase_ : List[Any]=True , lowercase_ : Any=True , lowercase_ : str=32 , lowercase_ : Any=5 , lowercase_ : int=4 , lowercase_ : List[str]=37 , lowercase_ : Any="gelu" , lowercase_ : Optional[Any]=0.1 , lowercase_ : List[str]=0.1 , lowercase_ : str=10 , lowercase_ : Optional[Any]=0.02 , lowercase_ : List[str]=3 , lowercase_ : Union[str, Any]=0.6 , lowercase_ : List[Any]=None , ): snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = is_training snake_case_ = use_labels snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = mask_ratio snake_case_ = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) snake_case_ = (image_size // patch_size) ** 2 snake_case_ = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def A_ ( self : Optional[int] ): snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = self.get_config() return config, pixel_values, labels def A_ ( self : Optional[int] ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowercase_ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def A_ ( self : Any , lowercase_ : Any , lowercase_ : Union[str, Any] , lowercase_ : Dict ): snake_case_ = ViTMAEModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() snake_case_ = model(lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A_ ( self : List[str] , lowercase_ : Union[str, Any] , lowercase_ : List[str] , lowercase_ : Union[str, Any] ): snake_case_ = ViTMAEForPreTraining(lowercase_ ) model.to(lowercase_ ) model.eval() snake_case_ = model(lowercase_ ) snake_case_ = (self.image_size // self.patch_size) ** 2 snake_case_ = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images snake_case_ = 1 snake_case_ = ViTMAEForPreTraining(lowercase_ ) model.to(lowercase_ ) model.eval() snake_case_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case_ = model(lowercase_ ) snake_case_ = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def A_ ( self : Optional[int] ): snake_case_ = self.prepare_config_and_inputs() snake_case_ = config_and_inputs snake_case_ = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , unittest.TestCase ): snake_case_ = (ViTMAEModel, ViTMAEForPreTraining) if is_torch_available() else () snake_case_ = {"feature-extraction": ViTMAEModel} if is_torch_available() else {} snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def A_ ( self : Optional[Any] ): snake_case_ = ViTMAEModelTester(self ) snake_case_ = ConfigTester(self , config_class=lowercase_ , has_text_modality=lowercase_ , hidden_size=37 ) def A_ ( self : Dict ): self.config_tester.run_common_tests() @unittest.skip(reason='''ViTMAE does not use inputs_embeds''' ) def A_ ( self : Optional[int] ): pass def A_ ( self : List[str] ): snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(lowercase_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase_ , nn.Linear ) ) def A_ ( self : Optional[Any] ): snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(lowercase_ ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowercase_ ) def A_ ( self : Tuple ): snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) def A_ ( self : Optional[int] ): snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase_ ) def A_ ( self : Optional[int] , lowercase_ : Dict , lowercase_ : List[Any] , lowercase_ : Optional[Any] ): # make masks reproducible np.random.seed(2 ) snake_case_ = int((pt_model.config.image_size // pt_model.config.patch_size) ** 2 ) snake_case_ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) snake_case_ = torch.from_numpy(lowercase_ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument snake_case_ = pt_noise super().check_pt_tf_models(lowercase_ , lowercase_ , lowercase_ ) def A_ ( self : int ): snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(lowercase_ ) model.to(lowercase_ ) model.eval() # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(lowercase_ , lowercase_ ) ) snake_case_ = outputs[0].cpu().numpy() snake_case_ = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowercase_ ) snake_case_ = model_class.from_pretrained(lowercase_ ) model.to(lowercase_ ) # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(lowercase_ , lowercase_ ) ) # Make sure we don't have nans snake_case_ = after_outputs[0].cpu().numpy() snake_case_ = 0 snake_case_ = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowercase_ , 1e-5 ) @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.''' ) def A_ ( self : Optional[int] ): pass @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.''' ) def A_ ( self : Union[str, Any] ): pass @unittest.skip( reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.''' ) def A_ ( self : Optional[Any] ): pass @unittest.skip(reason='''ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load''' ) def A_ ( self : Tuple ): pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def A_ ( self : List[Any] ): pass @slow def A_ ( self : List[str] ): for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = ViTMAEModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) def __magic_name__ ( ) -> Tuple: '''simple docstring''' snake_case_ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class a ( unittest.TestCase ): @cached_property def A_ ( self : Union[str, Any] ): return ViTImageProcessor.from_pretrained('''facebook/vit-mae-base''' ) if is_vision_available() else None @slow def A_ ( self : int ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) snake_case_ = ViTMAEForPreTraining.from_pretrained('''facebook/vit-mae-base''' ).to(lowercase_ ) snake_case_ = self.default_image_processor snake_case_ = prepare_img() snake_case_ = image_processor(images=lowercase_ , return_tensors='''pt''' ).to(lowercase_ ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) snake_case_ = ViTMAEConfig() snake_case_ = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) snake_case_ = np.random.uniform(size=(1, num_patches) ) # forward pass with torch.no_grad(): snake_case_ = model(**lowercase_ , noise=torch.from_numpy(lowercase_ ).to(device=lowercase_ ) ) # verify the logits snake_case_ = torch.Size((1, 196, 768) ) self.assertEqual(outputs.logits.shape , lowercase_ ) snake_case_ = torch.tensor( [[-0.0548, -1.7023, -0.9325], [0.3721, -0.5670, -0.2233], [0.8235, -1.3878, -0.3524]] ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , expected_slice.to(lowercase_ ) , atol=1e-4 ) )
350
'''simple docstring''' import pytest import datasets # Import fixture modules as plugins a : int = ['tests.fixtures.files', 'tests.fixtures.hub', 'tests.fixtures.fsspec'] def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> Tuple: '''simple docstring''' for item in items: if any(marker in item.keywords for marker in ['''integration''', '''unit'''] ): continue item.add_marker(pytest.mark.unit ) def __magic_name__ ( __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' config.addinivalue_line('''markers''', '''torchaudio_latest: mark test to run with torchaudio>=0.12''' ) @pytest.fixture(autouse=__UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> List[Any]: '''simple docstring''' snake_case_ = tmp_path_factory.getbasetemp() / '''cache''' snake_case_ = test_hf_cache_home / '''datasets''' snake_case_ = test_hf_cache_home / '''metrics''' snake_case_ = test_hf_cache_home / '''modules''' monkeypatch.setattr('''datasets.config.HF_DATASETS_CACHE''', str(__UpperCAmelCase ) ) monkeypatch.setattr('''datasets.config.HF_METRICS_CACHE''', str(__UpperCAmelCase ) ) monkeypatch.setattr('''datasets.config.HF_MODULES_CACHE''', str(__UpperCAmelCase ) ) snake_case_ = test_hf_datasets_cache / '''downloads''' monkeypatch.setattr('''datasets.config.DOWNLOADED_DATASETS_PATH''', str(__UpperCAmelCase ) ) snake_case_ = test_hf_datasets_cache / '''downloads''' / '''extracted''' monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_PATH''', str(__UpperCAmelCase ) ) @pytest.fixture(autouse=__UpperCAmelCase, scope='''session''' ) def __magic_name__ ( ) -> List[Any]: '''simple docstring''' datasets.disable_progress_bar() @pytest.fixture(autouse=__UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' monkeypatch.setattr('''datasets.config.HF_UPDATE_DOWNLOAD_COUNTS''', __UpperCAmelCase ) @pytest.fixture def __magic_name__ ( __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' monkeypatch.setattr('''sqlalchemy.util.deprecations.SILENCE_UBER_WARNING''', __UpperCAmelCase )
72
0
'''simple docstring''' from __future__ import annotations import pandas as pd def _A ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =[0] * no_of_processes __lowercase =[0] * no_of_processes # Copy the burst time into remaining_time[] for i in range(_lowerCAmelCase ): __lowercase =burst_time[i] __lowercase =0 __lowercase =0 __lowercase =999_999_999 __lowercase =0 __lowercase =False # Process until all processes are completed while complete != no_of_processes: for j in range(_lowerCAmelCase ): if arrival_time[j] <= increment_time and remaining_time[j] > 0: if remaining_time[j] < minm: __lowercase =remaining_time[j] __lowercase =j __lowercase =True if not check: increment_time += 1 continue remaining_time[short] -= 1 __lowercase =remaining_time[short] if minm == 0: __lowercase =999_999_999 if remaining_time[short] == 0: complete += 1 __lowercase =False # Find finish time of current process __lowercase =increment_time + 1 # Calculate waiting time __lowercase =finish_time - arrival_time[short] __lowercase =finar - burst_time[short] if waiting_time[short] < 0: __lowercase =0 # Increment time increment_time += 1 return waiting_time def _A ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =[0] * no_of_processes for i in range(_lowerCAmelCase ): __lowercase =burst_time[i] + waiting_time[i] return turn_around_time def _A ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =0 __lowercase =0 for i in range(_lowerCAmelCase ): __lowercase =total_waiting_time + waiting_time[i] __lowercase =total_turn_around_time + turn_around_time[i] print(f"""Average waiting time = {total_waiting_time / no_of_processes:.5f}""" ) print('Average turn around time =' , total_turn_around_time / no_of_processes ) if __name__ == "__main__": print("""Enter how many process you want to analyze""") lowerCamelCase = int(input()) lowerCamelCase = [0] * no_of_processes lowerCamelCase = [0] * no_of_processes lowerCamelCase = list(range(1, no_of_processes + 1)) for i in range(no_of_processes): print("""Enter the arrival time and burst time for process:--""" + str(i + 1)) lowerCamelCase , lowerCamelCase = map(int, input().split()) lowerCamelCase = calculate_waitingtime(arrival_time, burst_time, no_of_processes) lowerCamelCase = burst_time lowerCamelCase = no_of_processes lowerCamelCase = waiting_time lowerCamelCase = calculate_turnaroundtime(bt, n, wt) calculate_average_times(waiting_time, turn_around_time, no_of_processes) lowerCamelCase = pd.DataFrame( list(zip(processes, burst_time, arrival_time, waiting_time, turn_around_time)), columns=[ """Process""", """BurstTime""", """ArrivalTime""", """WaitingTime""", """TurnAroundTime""", ], ) # Printing the dataFrame pd.set_option("""display.max_rows""", fcfs.shape[0] + 1) print(fcfs)
166
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_field, _ask_options, _convert_compute_environment # noqa: F401 from .sagemaker import get_sagemaker_input lowerCamelCase = """Launches a series of prompts to create and save a `default_config.yaml` configuration file for your training system. Should always be ran first on your machine""" def _A ( ): """simple docstring""" __lowercase =_ask_options( 'In which compute environment are you running?' , ['This machine', 'AWS (Amazon SageMaker)'] , _convert_compute_environment , ) if compute_environment == ComputeEnvironment.AMAZON_SAGEMAKER: __lowercase =get_sagemaker_input() else: __lowercase =get_cluster_input() return config def _A ( _lowerCAmelCase=None ): """simple docstring""" if subparsers is not None: __lowercase =subparsers.add_parser('config' , description=_lowerCAmelCase ) else: __lowercase =argparse.ArgumentParser('Accelerate config command' , description=_lowerCAmelCase ) parser.add_argument( '--config_file' , default=_lowerCAmelCase , help=( 'The path to use to store the config file. Will default to a file named default_config.yaml in the cache ' 'location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ' 'such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ' 'with \'huggingface\'.' ) , ) if subparsers is not None: parser.set_defaults(func=_lowerCAmelCase ) return parser def _A ( _lowerCAmelCase ): """simple docstring""" __lowercase =get_user_input() if args.config_file is not None: __lowercase =args.config_file else: if not os.path.isdir(_lowerCAmelCase ): os.makedirs(_lowerCAmelCase ) __lowercase =default_yaml_config_file if config_file.endswith('.json' ): config.to_json_file(_lowerCAmelCase ) else: config.to_yaml_file(_lowerCAmelCase ) print(f"""accelerate configuration saved at {config_file}""" ) def _A ( ): """simple docstring""" __lowercase =config_command_parser() __lowercase =parser.parse_args() config_command(_lowerCAmelCase ) if __name__ == "__main__": main()
166
1
'''simple docstring''' import os from typing import Optional import fsspec from fsspec.archive import AbstractArchiveFileSystem from fsspec.utils import DEFAULT_BLOCK_SIZE class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" _SCREAMING_SNAKE_CASE = """""" _SCREAMING_SNAKE_CASE = ( None # protocol passed in prefix to the url. ex: "gzip", for gzip://file.txt::http://foo.bar/file.txt.gz ) _SCREAMING_SNAKE_CASE = None # compression type in fsspec. ex: "gzip" _SCREAMING_SNAKE_CASE = None # extension of the filename to strip. ex: "".gz" to get file.txt from file.txt.gz def __init__( self : int , UpperCamelCase__ : str = "" , UpperCamelCase__ : Optional[str] = None , UpperCamelCase__ : Optional[dict] = None , **UpperCamelCase__ : List[str] ): """simple docstring""" super().__init__(self , **UpperCamelCase__ ) # always open as "rb" since fsspec can then use the TextIOWrapper to make it work for "r" mode UpperCamelCase = fsspec.open( UpperCamelCase__ , mode='rb' , protocol=UpperCamelCase__ , compression=self.compression , client_kwargs={ 'requote_redirect_url': False, # see https://github.com/huggingface/datasets/pull/5459 'trust_env': True, # Enable reading proxy env variables. **(target_options or {}).pop('client_kwargs' , {} ), # To avoid issues if it was already passed. } , **(target_options or {}) , ) UpperCamelCase = os.path.basename(self.file.path.split('::' )[0] ) UpperCamelCase = ( self.compressed_name[: self.compressed_name.rindex('.' )] if '.' in self.compressed_name else self.compressed_name ) UpperCamelCase = None @classmethod def A ( cls : int , UpperCamelCase__ : List[str] ): """simple docstring""" return super()._strip_protocol(UpperCamelCase__ ).lstrip('/' ) def A ( self : Optional[int] ): """simple docstring""" if self.dir_cache is None: UpperCamelCase = {**self.file.fs.info(self.file.path ), 'name': self.uncompressed_name} UpperCamelCase = {f['name']: f} def A ( self : Any , UpperCamelCase__ : str ): """simple docstring""" return self.file.open().read() def A ( self : List[str] , UpperCamelCase__ : str , UpperCamelCase__ : str = "rb" , UpperCamelCase__ : Union[str, Any]=None , UpperCamelCase__ : Any=True , UpperCamelCase__ : List[Any]=None , **UpperCamelCase__ : Dict , ): """simple docstring""" UpperCamelCase = self._strip_protocol(UpperCamelCase__ ) if mode != "rb": raise ValueError(f"""Tried to read with mode {mode} on file {self.file.path} opened with mode 'rb'""" ) return self.file.open() class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" _SCREAMING_SNAKE_CASE = """bz2""" _SCREAMING_SNAKE_CASE = """bz2""" _SCREAMING_SNAKE_CASE = """.bz2""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" _SCREAMING_SNAKE_CASE = """gzip""" _SCREAMING_SNAKE_CASE = """gzip""" _SCREAMING_SNAKE_CASE = """.gz""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" _SCREAMING_SNAKE_CASE = """lz4""" _SCREAMING_SNAKE_CASE = """lz4""" _SCREAMING_SNAKE_CASE = """.lz4""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" _SCREAMING_SNAKE_CASE = """xz""" _SCREAMING_SNAKE_CASE = """xz""" _SCREAMING_SNAKE_CASE = """.xz""" class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" _SCREAMING_SNAKE_CASE = """zstd""" _SCREAMING_SNAKE_CASE = """zstd""" _SCREAMING_SNAKE_CASE = """.zst""" def __init__( self : List[str] , UpperCamelCase__ : str , UpperCamelCase__ : str = "rb" , UpperCamelCase__ : Optional[str] = None , UpperCamelCase__ : Optional[dict] = None , UpperCamelCase__ : int = DEFAULT_BLOCK_SIZE , **UpperCamelCase__ : Dict , ): """simple docstring""" super().__init__( fo=UpperCamelCase__ , mode=UpperCamelCase__ , target_protocol=UpperCamelCase__ , target_options=UpperCamelCase__ , block_size=UpperCamelCase__ , **UpperCamelCase__ , ) # We need to wrap the zstd decompressor to avoid this error in fsspec==2021.7.0 and zstandard==0.15.2: # # File "/Users/user/.virtualenvs/hf-datasets/lib/python3.7/site-packages/fsspec/core.py", line 145, in open # out.close = close # AttributeError: 'zstd.ZstdDecompressionReader' object attribute 'close' is read-only # # see https://github.com/intake/filesystem_spec/issues/725 UpperCamelCase = self.file.__enter__ class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self : Optional[int] , UpperCamelCase__ : Optional[int] ): """simple docstring""" UpperCamelCase = file_ def __enter__( self : Tuple ): """simple docstring""" self._file.__enter__() return self def __exit__( self : int , *UpperCamelCase__ : Tuple , **UpperCamelCase__ : Tuple ): """simple docstring""" self._file.__exit__(*UpperCamelCase__ , **UpperCamelCase__ ) def __iter__( self : Dict ): """simple docstring""" return iter(self._file ) def A ( self : str ): """simple docstring""" return next(self._file ) def __getattr__( self : Optional[Any] , UpperCamelCase__ : Union[str, Any] ): """simple docstring""" return getattr(self._file , UpperCamelCase__ ) def fixed_enter(*UpperCamelCase__ : List[str] , **UpperCamelCase__ : Any ): return WrappedFile(_enter(*UpperCamelCase__ , **UpperCamelCase__ ) ) UpperCamelCase = fixed_enter
353
'''simple docstring''' import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def __lowerCamelCase ( A__ , A__ , A__ ) -> Optional[int]: """simple docstring""" UpperCamelCase = ('dense.weight', 'attention.self.query', 'attention.self.key', 'attention.self.value') UpperCamelCase = ( ('layer.', 'layer_'), ('word_embeddings.weight', 'word_embeddings'), ('position_embeddings.weight', 'position_embeddings'), ('token_type_embeddings.weight', 'token_type_embeddings'), ('.', '/'), ('LayerNorm/weight', 'LayerNorm/gamma'), ('LayerNorm/bias', 'LayerNorm/beta'), ('weight', 'kernel'), ) if not os.path.isdir(A__ ): os.makedirs(A__ ) UpperCamelCase = model.state_dict() def to_tf_var_name(A__ ): for patt, repl in iter(A__ ): UpperCamelCase = name.replace(A__ , A__ ) return F"""bert/{name}""" def create_tf_var(A__ , A__ , A__ ): UpperCamelCase = tf.dtypes.as_dtype(tensor.dtype ) UpperCamelCase = tf.get_variable(dtype=A__ , shape=tensor.shape , name=A__ , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(A__ ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: UpperCamelCase = to_tf_var_name(A__ ) UpperCamelCase = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): UpperCamelCase = torch_tensor.T UpperCamelCase = create_tf_var(tensor=A__ , name=A__ , session=A__ ) tf.keras.backend.set_value(A__ , A__ ) UpperCamelCase = session.run(A__ ) print(F"""Successfully created {tf_name}: {np.allclose(A__ , A__ )}""" ) UpperCamelCase = tf.train.Saver(tf.trainable_variables() ) saver.save(A__ , os.path.join(A__ , model_name.replace('-' , '_' ) + '.ckpt' ) ) def __lowerCamelCase ( A__=None ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = argparse.ArgumentParser() parser.add_argument('--model_name' , type=A__ , required=A__ , help='model name e.g. bert-base-uncased' ) parser.add_argument( '--cache_dir' , type=A__ , default=A__ , required=A__ , help='Directory containing pytorch model' ) parser.add_argument('--pytorch_model_path' , type=A__ , required=A__ , help='/path/to/<pytorch-model-name>.bin' ) parser.add_argument('--tf_cache_dir' , type=A__ , required=A__ , help='Directory in which to save tensorflow model' ) UpperCamelCase = parser.parse_args(A__ ) UpperCamelCase = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=A__ , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
249
0
import unittest from datasets import load_dataset from transformers import BloomTokenizerFast from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A_ ( _a , unittest.TestCase ): '''simple docstring''' a__ = None a__ = BloomTokenizerFast a__ = BloomTokenizerFast a__ = True a__ = False a__ = "tokenizer_file" a__ = {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"} def lowerCAmelCase_ (self ) -> List[str]: super().setUp() __UpperCAmelCase = BloomTokenizerFast.from_pretrained('''bigscience/tokenizer''' ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase_ (self , **lowercase__ ) -> str: kwargs.update(self.special_tokens_map ) return BloomTokenizerFast.from_pretrained(self.tmpdirname , **lowercase__ ) def lowerCAmelCase_ (self ) -> Optional[Any]: __UpperCAmelCase = self.get_rust_tokenizer() __UpperCAmelCase = ['''The quick brown fox</s>''', '''jumps over the lazy dog</s>'''] __UpperCAmelCase = [[2_175, 23_714, 73_173, 144_252, 2], [77, 132_619, 3_478, 368, 109_586, 35_433, 2]] __UpperCAmelCase = tokenizer.batch_encode_plus(lowercase__ )['''input_ids'''] self.assertListEqual(lowercase__ , lowercase__ ) __UpperCAmelCase = tokenizer.batch_decode(lowercase__ ) self.assertListEqual(lowercase__ , lowercase__ ) def lowerCAmelCase_ (self , lowercase__=6 ) -> Dict: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(lowercase__ , **lowercase__ ) # tokenizer_r.pad_token = None # Hotfixing padding = None # Simple input __UpperCAmelCase = '''This is a simple input''' __UpperCAmelCase = ['''This is a simple input 1''', '''This is a simple input 2'''] __UpperCAmelCase = ('''This is a simple input''', '''This is a pair''') __UpperCAmelCase = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests try: tokenizer_r.encode(lowercase__ , max_length=lowercase__ ) tokenizer_r.encode_plus(lowercase__ , max_length=lowercase__ ) tokenizer_r.batch_encode_plus(lowercase__ , max_length=lowercase__ ) tokenizer_r.encode(lowercase__ , max_length=lowercase__ ) tokenizer_r.batch_encode_plus(lowercase__ , max_length=lowercase__ ) except ValueError: self.fail('''Bloom Tokenizer should be able to deal with padding''' ) __UpperCAmelCase = None # Hotfixing padding = None self.assertRaises(lowercase__ , tokenizer_r.encode , lowercase__ , max_length=lowercase__ , padding='''max_length''' ) # Simple input self.assertRaises(lowercase__ , tokenizer_r.encode_plus , lowercase__ , max_length=lowercase__ , padding='''max_length''' ) # Simple input self.assertRaises( lowercase__ , tokenizer_r.batch_encode_plus , lowercase__ , max_length=lowercase__ , padding='''max_length''' , ) # Pair input self.assertRaises(lowercase__ , tokenizer_r.encode , lowercase__ , max_length=lowercase__ , padding='''max_length''' ) # Pair input self.assertRaises(lowercase__ , tokenizer_r.encode_plus , lowercase__ , max_length=lowercase__ , padding='''max_length''' ) # Pair input self.assertRaises( lowercase__ , tokenizer_r.batch_encode_plus , lowercase__ , max_length=lowercase__ , padding='''max_length''' , ) def lowerCAmelCase_ (self ) -> List[Any]: __UpperCAmelCase = self.get_rust_tokenizer() __UpperCAmelCase = load_dataset('''xnli''' , '''all_languages''' , split='''test''' , streaming=lowercase__ ) __UpperCAmelCase = next(iter(lowercase__ ) )['''premise'''] # pick up one data __UpperCAmelCase = list(sample_data.values() ) __UpperCAmelCase = list(map(tokenizer.encode , lowercase__ ) ) __UpperCAmelCase = [tokenizer.decode(lowercase__ , clean_up_tokenization_spaces=lowercase__ ) for x in output_tokens] self.assertListEqual(lowercase__ , lowercase__ ) def lowerCAmelCase_ (self ) -> str: # The test has to be overriden because BLOOM uses ALiBi positional embeddings that does not have # any sequence length constraints. This test of the parent class will fail since it relies on the # maximum sequence length of the positoonal embeddings. self.assertGreaterEqual(len(self.tokenizer_class.pretrained_vocab_files_map ) , 1 ) self.assertGreaterEqual(len(list(self.tokenizer_class.pretrained_vocab_files_map.values() )[0] ) , 1 )
333
def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: '''simple docstring''' __UpperCAmelCase = [[] for _ in range(SCREAMING_SNAKE_CASE )] __UpperCAmelCase = key - 1 if key <= 0: raise ValueError('''Height of grid can\'t be 0 or negative''' ) if key == 1 or len(SCREAMING_SNAKE_CASE ) <= key: return input_string for position, character in enumerate(SCREAMING_SNAKE_CASE ): __UpperCAmelCase = position % (lowest * 2) # puts it in bounds __UpperCAmelCase = min(SCREAMING_SNAKE_CASE , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append(SCREAMING_SNAKE_CASE ) __UpperCAmelCase = [''''''.join(SCREAMING_SNAKE_CASE ) for row in temp_grid] __UpperCAmelCase = ''''''.join(SCREAMING_SNAKE_CASE ) return output_string def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> str: '''simple docstring''' __UpperCAmelCase = [] __UpperCAmelCase = key - 1 if key <= 0: raise ValueError('''Height of grid can\'t be 0 or negative''' ) if key == 1: return input_string __UpperCAmelCase = [[] for _ in range(SCREAMING_SNAKE_CASE )] # generates template for position in range(len(SCREAMING_SNAKE_CASE ) ): __UpperCAmelCase = position % (lowest * 2) # puts it in bounds __UpperCAmelCase = min(SCREAMING_SNAKE_CASE , lowest * 2 - num ) # creates zigzag pattern temp_grid[num].append('''*''' ) __UpperCAmelCase = 0 for row in temp_grid: # fills in the characters __UpperCAmelCase = input_string[counter : counter + len(SCREAMING_SNAKE_CASE )] grid.append(list(SCREAMING_SNAKE_CASE ) ) counter += len(SCREAMING_SNAKE_CASE ) __UpperCAmelCase = '''''' # reads as zigzag for position in range(len(SCREAMING_SNAKE_CASE ) ): __UpperCAmelCase = position % (lowest * 2) # puts it in bounds __UpperCAmelCase = min(SCREAMING_SNAKE_CASE , lowest * 2 - num ) # creates zigzag pattern output_string += grid[num][0] grid[num].pop(0 ) return output_string def __a ( SCREAMING_SNAKE_CASE ) -> dict[int, str]: '''simple docstring''' __UpperCAmelCase = {} for key_guess in range(1 , len(SCREAMING_SNAKE_CASE ) ): # tries every key __UpperCAmelCase = decrypt(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return results if __name__ == "__main__": import doctest doctest.testmod()
333
1
import itertools import random import unittest import numpy as np from transformers import is_speech_available from transformers.testing_utils import require_torch, require_torchaudio from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import SpeechaTextFeatureExtractor a_ = random.Random() def __lowercase ( snake_case_ : Dict ,snake_case_ : List[Any]=1.0 ,snake_case_ : Tuple=None ,snake_case_ : List[Any]=None ) ->Union[str, Any]: '''simple docstring''' if rng is None: __A : Tuple = global_rng __A : Tuple = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class __snake_case ( unittest.TestCase ): """simple docstring""" def __init__( self , __lowerCamelCase , __lowerCamelCase=7 , __lowerCamelCase=400 , __lowerCamelCase=2000 , __lowerCamelCase=24 , __lowerCamelCase=24 , __lowerCamelCase=0.0 , __lowerCamelCase=1_6000 , __lowerCamelCase=True , __lowerCamelCase=True , ): '''simple docstring''' __A : Tuple = parent __A : Union[str, Any] = batch_size __A : Optional[Any] = min_seq_length __A : Any = max_seq_length __A : List[str] = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __A : Optional[int] = feature_size __A : str = num_mel_bins __A : str = padding_value __A : Any = sampling_rate __A : int = return_attention_mask __A : Optional[int] = do_normalize def UpperCamelCase__( self ): '''simple docstring''' return { "feature_size": self.feature_size, "num_mel_bins": self.num_mel_bins, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def UpperCamelCase__( self , __lowerCamelCase=False , __lowerCamelCase=False ): '''simple docstring''' def _flatten(__lowerCamelCase ): return list(itertools.chain(*__lowerCamelCase ) ) if equal_length: __A : Optional[Any] = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __A : str = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __A : Optional[Any] = [np.asarray(__lowerCamelCase ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __snake_case ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): """simple docstring""" _lowerCamelCase = SpeechaTextFeatureExtractor if is_speech_available() else None def UpperCamelCase__( self ): '''simple docstring''' __A : List[Any] = SpeechaTextFeatureExtractionTester(self ) def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' self.assertTrue(np.all(np.mean(__lowerCamelCase , axis=0 ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(__lowerCamelCase , axis=0 ) - 1 ) < 1e-3 ) ) def UpperCamelCase__( self ): '''simple docstring''' __A : str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __A : Tuple = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __A : int = [np.asarray(__lowerCamelCase ) for speech_input in speech_inputs] # Test feature size __A : str = feature_extractor(__lowerCamelCase , padding=__lowerCamelCase , return_tensors='''np''' ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.feature_size ) # Test not batched input __A : str = feature_extractor(speech_inputs[0] , return_tensors='''np''' ).input_features __A : List[Any] = feature_extractor(np_speech_inputs[0] , return_tensors='''np''' ).input_features self.assertTrue(np.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-3 ) ) # Test batched __A : Optional[Any] = feature_extractor(__lowerCamelCase , return_tensors='''np''' ).input_features __A : Tuple = feature_extractor(__lowerCamelCase , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(__lowerCamelCase , __lowerCamelCase ): self.assertTrue(np.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. __A : Union[str, Any] = [floats_list((1, x) )[0] for x in (800, 800, 800)] __A : List[Any] = np.asarray(__lowerCamelCase ) __A : Union[str, Any] = feature_extractor(__lowerCamelCase , return_tensors='''np''' ).input_features __A : Union[str, Any] = feature_extractor(__lowerCamelCase , return_tensors='''np''' ).input_features for enc_seq_a, enc_seq_a in zip(__lowerCamelCase , __lowerCamelCase ): self.assertTrue(np.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-3 ) ) def UpperCamelCase__( self ): '''simple docstring''' __A : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __A : Optional[int] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __A : Any = ['''longest''', '''max_length''', '''do_not_pad'''] __A : Optional[Any] = [None, 16, None] for max_length, padding in zip(__lowerCamelCase , __lowerCamelCase ): __A : Dict = feature_extractor( __lowerCamelCase , padding=__lowerCamelCase , max_length=__lowerCamelCase , return_attention_mask=__lowerCamelCase ) __A : Dict = inputs.input_features __A : List[str] = inputs.attention_mask __A : Any = [np.sum(__lowerCamelCase ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def UpperCamelCase__( self ): '''simple docstring''' __A : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __A : Dict = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __A : Optional[Any] = ['''longest''', '''max_length''', '''do_not_pad'''] __A : str = [None, 16, None] for max_length, padding in zip(__lowerCamelCase , __lowerCamelCase ): __A : Union[str, Any] = feature_extractor( __lowerCamelCase , max_length=__lowerCamelCase , padding=__lowerCamelCase , return_tensors='''np''' , return_attention_mask=__lowerCamelCase ) __A : str = inputs.input_features __A : List[Any] = inputs.attention_mask __A : Dict = [np.sum(__lowerCamelCase ) for x in attention_mask] self._check_zero_mean_unit_variance(input_features[0][: fbank_feat_lengths[0]] ) self.assertTrue(input_features[0][fbank_feat_lengths[0] :].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_features[1][: fbank_feat_lengths[1]] ) self.assertTrue(input_features[0][fbank_feat_lengths[1] :].sum() < 1e-6 ) self._check_zero_mean_unit_variance(input_features[2][: fbank_feat_lengths[2]] ) def UpperCamelCase__( self ): '''simple docstring''' __A : List[str] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __A : Optional[int] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __A : Optional[int] = feature_extractor( __lowerCamelCase , padding='''max_length''' , max_length=4 , truncation=__lowerCamelCase , return_tensors='''np''' , return_attention_mask=__lowerCamelCase , ) __A : List[Any] = inputs.input_features __A : Optional[int] = inputs.attention_mask __A : List[str] = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1] ) self._check_zero_mean_unit_variance(input_features[2] ) def UpperCamelCase__( self ): '''simple docstring''' __A : int = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __A : Tuple = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __A : Any = feature_extractor( __lowerCamelCase , padding='''longest''' , max_length=4 , truncation=__lowerCamelCase , return_tensors='''np''' , return_attention_mask=__lowerCamelCase , ) __A : Optional[int] = inputs.input_features __A : List[Any] = inputs.attention_mask __A : Optional[Any] = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 4, 24) ) __A : Optional[Any] = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] __A : Tuple = feature_extractor( __lowerCamelCase , padding='''longest''' , max_length=16 , truncation=__lowerCamelCase , return_tensors='''np''' , return_attention_mask=__lowerCamelCase , ) __A : Union[str, Any] = inputs.input_features __A : List[str] = inputs.attention_mask __A : Optional[Any] = np.sum(attention_mask == 1 , axis=1 ) self._check_zero_mean_unit_variance(input_features[0, : fbank_feat_lengths[0]] ) self._check_zero_mean_unit_variance(input_features[1, : fbank_feat_lengths[1]] ) self._check_zero_mean_unit_variance(input_features[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertEqual(input_features.shape , (3, 6, 24) ) def UpperCamelCase__( self ): '''simple docstring''' import torch __A : Optional[int] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __A : Optional[int] = np.random.rand(100 , 32 ).astype(np.floataa ) __A : Union[str, Any] = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __A : Tuple = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''np''' ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __A : Optional[Any] = feature_extractor.pad([{'''input_features''': inputs}] , return_tensors='''pt''' ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' from datasets import load_dataset __A : Optional[int] = load_dataset('''hf-internal-testing/librispeech_asr_dummy''' , '''clean''' , split='''validation''' ) # automatic decoding with librispeech __A : Optional[int] = ds.sort('''id''' ).select(range(__lowerCamelCase ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def UpperCamelCase__( self ): '''simple docstring''' __A : Optional[Any] = np.array([ -1.5_7_4_5, -1.7_7_1_3, -1.7_0_2_0, -1.6_0_6_9, -1.2_2_5_0, -1.1_1_0_5, -0.9_0_7_2, -0.8_2_4_1, -1.2_3_1_0, -0.8_0_9_8, -0.3_3_2_0, -0.4_1_0_1, -0.7_9_8_5, -0.4_9_9_6, -0.8_2_1_3, -0.9_1_2_8, -1.0_4_2_0, -1.1_2_8_6, -1.0_4_4_0, -0.7_9_9_9, -0.8_4_0_5, -1.2_2_7_5, -1.5_4_4_3, -1.4_6_2_5, ] ) # fmt: on __A : Dict = self._load_datasamples(1 ) __A : str = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __A : Dict = feature_extractor(__lowerCamelCase , return_tensors='''pt''' ).input_features self.assertEquals(input_features.shape , (1, 584, 24) ) self.assertTrue(np.allclose(input_features[0, 0, :30] , __lowerCamelCase , atol=1e-4 ) )
352
"""simple docstring""" import doctest import glob import importlib import inspect import os import re from contextlib import contextmanager from functools import wraps from unittest.mock import patch import numpy as np import pytest from absl.testing import parameterized import datasets from datasets import load_metric from .utils import for_all_test_methods, local, slow # mark all tests as integration a_ = pytest.mark.integration a_ = {"""comet"""} a_ = importlib.util.find_spec("""fairseq""") is not None a_ = {"""code_eval"""} a_ = os.name == """nt""" a_ = {"""bertscore""", """frugalscore""", """perplexity"""} a_ = importlib.util.find_spec("""transformers""") is not None def __lowercase ( snake_case_ : str ) ->Any: '''simple docstring''' @wraps(snake_case_ ) def wrapper(self : List[Any] ,snake_case_ : int ): if not _has_fairseq and metric_name in REQUIRE_FAIRSEQ: self.skipTest('''"test requires Fairseq"''' ) else: test_case(self ,snake_case_ ) return wrapper def __lowercase ( snake_case_ : int ) ->str: '''simple docstring''' @wraps(snake_case_ ) def wrapper(self : List[Any] ,snake_case_ : List[str] ): if not _has_transformers and metric_name in REQUIRE_TRANSFORMERS: self.skipTest('''"test requires transformers"''' ) else: test_case(self ,snake_case_ ) return wrapper def __lowercase ( snake_case_ : Union[str, Any] ) ->Tuple: '''simple docstring''' @wraps(snake_case_ ) def wrapper(self : int ,snake_case_ : Dict ): if _on_windows and metric_name in UNSUPPORTED_ON_WINDOWS: self.skipTest('''"test not supported on Windows"''' ) else: test_case(self ,snake_case_ ) return wrapper def __lowercase ( ) ->Tuple: '''simple docstring''' __A : int = [metric_dir.split(os.sep )[-2] for metric_dir in glob.glob('''./metrics/*/''' )] return [{"testcase_name": x, "metric_name": x} for x in metrics if x != "gleu"] # gleu is unfinished @parameterized.named_parameters(get_local_metric_names() ) @for_all_test_methods( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) @local class __snake_case ( parameterized.TestCase ): """simple docstring""" _lowerCamelCase = {} _lowerCamelCase = None @pytest.mark.filterwarnings('''ignore:metric_module_factory is deprecated:FutureWarning''' ) @pytest.mark.filterwarnings('''ignore:load_metric is deprecated:FutureWarning''' ) def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' __A : int = '''[...]''' __A : Any = importlib.import_module( datasets.load.metric_module_factory(os.path.join('''metrics''' , __lowerCamelCase ) ).module_path ) __A : str = datasets.load.import_main_class(metric_module.__name__ , dataset=__lowerCamelCase ) # check parameters __A : Optional[int] = inspect.signature(metric._compute ).parameters self.assertTrue(all(p.kind != p.VAR_KEYWORD for p in parameters.values() ) ) # no **kwargs # run doctest with self.patch_intensive_calls(__lowerCamelCase , metric_module.__name__ ): with self.use_local_metrics(): try: __A : Tuple = doctest.testmod(__lowerCamelCase , verbose=__lowerCamelCase , raise_on_error=__lowerCamelCase ) except doctest.UnexpectedException as e: raise e.exc_info[1] # raise the exception that doctest caught self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @slow def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' __A : Any = '''[...]''' __A : Union[str, Any] = importlib.import_module( datasets.load.metric_module_factory(os.path.join('''metrics''' , __lowerCamelCase ) ).module_path ) # run doctest with self.use_local_metrics(): __A : Union[str, Any] = doctest.testmod(__lowerCamelCase , verbose=__lowerCamelCase , raise_on_error=__lowerCamelCase ) self.assertEqual(results.failed , 0 ) self.assertGreater(results.attempted , 1 ) @contextmanager def UpperCamelCase__( self , __lowerCamelCase , __lowerCamelCase ): '''simple docstring''' if metric_name in self.INTENSIVE_CALLS_PATCHER: with self.INTENSIVE_CALLS_PATCHER[metric_name](__lowerCamelCase ): yield else: yield @contextmanager def UpperCamelCase__( self ): '''simple docstring''' def load_local_metric(__lowerCamelCase , *__lowerCamelCase , **__lowerCamelCase ): return load_metric(os.path.join('''metrics''' , __lowerCamelCase ) , *__lowerCamelCase , **__lowerCamelCase ) with patch('''datasets.load_metric''' ) as mock_load_metric: __A : List[Any] = load_local_metric yield @classmethod def UpperCamelCase__( cls , __lowerCamelCase ): '''simple docstring''' def wrapper(__lowerCamelCase ): __A : Any = contextmanager(__lowerCamelCase ) __A : Optional[Any] = patcher return patcher return wrapper @LocalMetricTest.register_intensive_calls_patcher('''bleurt''' ) def __lowercase ( snake_case_ : Tuple ) ->int: '''simple docstring''' import tensorflow.compat.va as tf from bleurt.score import Predictor tf.flags.DEFINE_string('''sv''' ,'''''' ,'''''' ) # handle pytest cli flags class __snake_case ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def UpperCamelCase__( self , __lowerCamelCase ): '''simple docstring''' assert len(input_dict['''input_ids'''] ) == 2 return np.array([1.0_3, 1.0_4] ) # mock predict_fn which is supposed to do a forward pass with a bleurt model with patch('''bleurt.score._create_predictor''' ) as mock_create_predictor: __A : List[str] = MockedPredictor() yield @LocalMetricTest.register_intensive_calls_patcher('''bertscore''' ) def __lowercase ( snake_case_ : List[str] ) ->Dict: '''simple docstring''' import torch def bert_cos_score_idf(snake_case_ : Union[str, Any] ,snake_case_ : List[str] ,*snake_case_ : List[str] ,**snake_case_ : Dict ): return torch.tensor([[1.0, 1.0, 1.0]] * len(snake_case_ ) ) # mock get_model which is supposed to do download a bert model # mock bert_cos_score_idf which is supposed to do a forward pass with a bert model with patch('''bert_score.scorer.get_model''' ), patch( '''bert_score.scorer.bert_cos_score_idf''' ) as mock_bert_cos_score_idf: __A : str = bert_cos_score_idf yield @LocalMetricTest.register_intensive_calls_patcher('''comet''' ) def __lowercase ( snake_case_ : Optional[int] ) ->List[Any]: '''simple docstring''' def load_from_checkpoint(snake_case_ : str ): class __snake_case : """simple docstring""" def UpperCamelCase__( self , __lowerCamelCase , *__lowerCamelCase , **__lowerCamelCase ): '''simple docstring''' assert len(__lowerCamelCase ) == 2 __A : str = [0.1_9, 0.9_2] return scores, sum(__lowerCamelCase ) / len(__lowerCamelCase ) return Model() # mock load_from_checkpoint which is supposed to do download a bert model # mock load_from_checkpoint which is supposed to do download a bert model with patch('''comet.download_model''' ) as mock_download_model: __A : int = None with patch('''comet.load_from_checkpoint''' ) as mock_load_from_checkpoint: __A : Dict = load_from_checkpoint yield def __lowercase ( ) ->str: '''simple docstring''' __A : Optional[Any] = load_metric(os.path.join('''metrics''' ,'''seqeval''' ) ) __A : Optional[int] = '''ERROR''' __A : str = F"""Scheme should be one of [IOB1, IOB2, IOE1, IOE2, IOBES, BILOU], got {wrong_scheme}""" with pytest.raises(snake_case_ ,match=re.escape(snake_case_ ) ): metric.compute(predictions=[] ,references=[] ,scheme=snake_case_ )
291
0
"""simple docstring""" import unittest from transformers import DebertaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, ) from transformers.models.deberta.modeling_deberta import DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST class snake_case ( __lowerCAmelCase ): def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=7 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=99 , __UpperCAmelCase=32 , __UpperCAmelCase=5 , __UpperCAmelCase=4 , __UpperCAmelCase=37 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=5_12 , __UpperCAmelCase=16 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=False , __UpperCAmelCase=True , __UpperCAmelCase="None" , __UpperCAmelCase=3 , __UpperCAmelCase=4 , __UpperCAmelCase=None , ) ->List[str]: a_ = parent a_ = batch_size a_ = seq_length a_ = is_training a_ = use_input_mask a_ = use_token_type_ids a_ = use_labels a_ = vocab_size a_ = hidden_size a_ = num_hidden_layers a_ = num_attention_heads a_ = intermediate_size a_ = hidden_act a_ = hidden_dropout_prob a_ = attention_probs_dropout_prob a_ = max_position_embeddings a_ = type_vocab_size a_ = type_sequence_label_size a_ = initializer_range a_ = num_labels a_ = num_choices a_ = relative_attention a_ = position_biased_input a_ = pos_att_type a_ = scope def UpperCAmelCase__ ( self) ->Optional[int]: a_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) a_ = None if self.use_input_mask: a_ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2) a_ = None if self.use_token_type_ids: a_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) a_ = None a_ = None a_ = None if self.use_labels: a_ = ids_tensor([self.batch_size] , self.type_sequence_label_size) a_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) a_ = ids_tensor([self.batch_size] , self.num_choices) a_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase__ ( self) ->Optional[int]: return DebertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def UpperCAmelCase__ ( self) ->Optional[int]: a_ = self.get_config() a_ = 3_00 return config def UpperCAmelCase__ ( self , __UpperCAmelCase) ->List[str]: self.parent.assertListEqual(list(result.loss.size()) , []) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) ->Optional[Any]: a_ = DebertaModel(config=_UpperCAmelCase) model.to(_UpperCAmelCase) model.eval() a_ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase)[0] a_ = model(_UpperCAmelCase , token_type_ids=_UpperCAmelCase)[0] a_ = model(_UpperCAmelCase)[0] self.parent.assertListEqual(list(sequence_output.size()) , [self.batch_size, self.seq_length, self.hidden_size]) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) ->Any: a_ = DebertaForMaskedLM(config=_UpperCAmelCase) model.to(_UpperCAmelCase) model.eval() a_ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , labels=_UpperCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) ->int: a_ = self.num_labels a_ = DebertaForSequenceClassification(_UpperCAmelCase) model.to(_UpperCAmelCase) model.eval() a_ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , labels=_UpperCAmelCase) self.parent.assertListEqual(list(result.logits.size()) , [self.batch_size, self.num_labels]) self.check_loss_output(_UpperCAmelCase) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) ->Union[str, Any]: a_ = self.num_labels a_ = DebertaForTokenClassification(config=_UpperCAmelCase) model.to(_UpperCAmelCase) model.eval() a_ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , labels=_UpperCAmelCase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase) ->Dict: a_ = DebertaForQuestionAnswering(config=_UpperCAmelCase) model.to(_UpperCAmelCase) model.eval() a_ = model( _UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , start_positions=_UpperCAmelCase , end_positions=_UpperCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length)) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length)) def UpperCAmelCase__ ( self) ->Union[str, Any]: a_ = self.prepare_config_and_inputs() ( ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ( a_ ) , ) = config_and_inputs a_ = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class snake_case ( __lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): a_ : Any = ( ( DebertaModel, DebertaForMaskedLM, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaForQuestionAnswering, ) if is_torch_available() else () ) a_ : Union[str, Any] = ( { '''feature-extraction''': DebertaModel, '''fill-mask''': DebertaForMaskedLM, '''question-answering''': DebertaForQuestionAnswering, '''text-classification''': DebertaForSequenceClassification, '''token-classification''': DebertaForTokenClassification, '''zero-shot''': DebertaForSequenceClassification, } if is_torch_available() else {} ) a_ : Optional[int] = True a_ : Union[str, Any] = False a_ : Union[str, Any] = False a_ : str = False a_ : Optional[Any] = False def UpperCAmelCase__ ( self) ->List[str]: a_ = DebertaModelTester(self) a_ = ConfigTester(self , config_class=_UpperCAmelCase , hidden_size=37) def UpperCAmelCase__ ( self) ->List[Any]: self.config_tester.run_common_tests() def UpperCAmelCase__ ( self) ->int: a_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*_UpperCAmelCase) def UpperCAmelCase__ ( self) ->List[str]: a_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*_UpperCAmelCase) def UpperCAmelCase__ ( self) ->Dict: a_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*_UpperCAmelCase) def UpperCAmelCase__ ( self) ->Optional[int]: a_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*_UpperCAmelCase) def UpperCAmelCase__ ( self) ->Optional[Any]: a_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*_UpperCAmelCase) @slow def UpperCAmelCase__ ( self) ->Any: for model_name in DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a_ = DebertaModel.from_pretrained(_UpperCAmelCase) self.assertIsNotNone(_UpperCAmelCase) @require_torch @require_sentencepiece @require_tokenizers class snake_case ( unittest.TestCase ): @unittest.skip(reason="Model not available yet") def UpperCAmelCase__ ( self) ->List[Any]: pass @slow def UpperCAmelCase__ ( self) ->List[Any]: a_ = DebertaModel.from_pretrained("microsoft/deberta-base") a_ = torch.tensor([[0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2]]) a_ = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]) with torch.no_grad(): a_ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase)[0] # compare the actual values for a slice. a_ = torch.tensor( [[[-0.5_986, -0.8_055, -0.8_462], [1.4_484, -0.9_348, -0.8_059], [0.3_123, 0.0_032, -1.4_131]]]) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _UpperCAmelCase , atol=1E-4) , F'''{output[:, 1:4, 1:4]}''')
243
"""simple docstring""" from manim import * class lowercase_ ( __lowerCAmelCase ): '''simple docstring''' def lowerCAmelCase_ ( self : Dict ): _A = Rectangle(height=0.5 , width=0.5 ) _A = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) _A = Rectangle(height=0.25 , width=0.25 ) _A = [mem.copy() for i in range(6 )] _A = [mem.copy() for i in range(6 )] _A = VGroup(*_UpperCAmelCase ).arrange(_UpperCAmelCase , buff=0 ) _A = VGroup(*_UpperCAmelCase ).arrange(_UpperCAmelCase , buff=0 ) _A = VGroup(_UpperCAmelCase , _UpperCAmelCase ).arrange(_UpperCAmelCase , buff=0 ) _A = Text('CPU' , font_size=24 ) _A = Group(_UpperCAmelCase , _UpperCAmelCase ).arrange(_UpperCAmelCase , buff=0.5 , aligned_edge=_UpperCAmelCase ) cpu.move_to([-2.5, -0.5, 0] ) self.add(_UpperCAmelCase ) _A = [mem.copy() for i in range(4 )] _A = VGroup(*_UpperCAmelCase ).arrange(_UpperCAmelCase , buff=0 ) _A = Text('GPU' , font_size=24 ) _A = Group(_UpperCAmelCase , _UpperCAmelCase ).arrange(_UpperCAmelCase , buff=0.5 , aligned_edge=_UpperCAmelCase ) gpu.move_to([-1, -1, 0] ) self.add(_UpperCAmelCase ) _A = [mem.copy() for i in range(6 )] _A = VGroup(*_UpperCAmelCase ).arrange(_UpperCAmelCase , buff=0 ) _A = Text('Model' , font_size=24 ) _A = Group(_UpperCAmelCase , _UpperCAmelCase ).arrange(_UpperCAmelCase , buff=0.5 , aligned_edge=_UpperCAmelCase ) model.move_to([3, -1.0, 0] ) self.add(_UpperCAmelCase ) _A = [] _A = [] for i, rect in enumerate(_UpperCAmelCase ): _A = fill.copy().set_fill(_UpperCAmelCase , opacity=0.8 ) target.move_to(_UpperCAmelCase ) model_arr.append(_UpperCAmelCase ) _A = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0.0 ).set_fill(_UpperCAmelCase , opacity=0.8 ) cpu_target.move_to(cpu_left_col_base[i] ) model_cpu_arr.append(_UpperCAmelCase ) self.add(*_UpperCAmelCase , *_UpperCAmelCase ) _A = [meta_mem.copy() for i in range(6 )] _A = [meta_mem.copy() for i in range(6 )] _A = VGroup(*_UpperCAmelCase ).arrange(_UpperCAmelCase , buff=0 ) _A = VGroup(*_UpperCAmelCase ).arrange(_UpperCAmelCase , buff=0 ) _A = VGroup(_UpperCAmelCase , _UpperCAmelCase ).arrange(_UpperCAmelCase , buff=0 ) _A = Text('Disk' , font_size=24 ) _A = Group(_UpperCAmelCase , _UpperCAmelCase ).arrange(_UpperCAmelCase , buff=0.5 , aligned_edge=_UpperCAmelCase ) disk.move_to([-4, -1.25, 0] ) self.add(_UpperCAmelCase , _UpperCAmelCase ) _A = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) _A = MarkupText( F'''<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model''' , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(_UpperCAmelCase , _UpperCAmelCase ) _A = MarkupText( F'''<span fgcolor=\'{BLUE}\'>●</span> Checkpoint''' , font_size=18 , ) blue_text.next_to(_UpperCAmelCase , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(_UpperCAmelCase ) _A = MarkupText( F'''Now watch as an input is passed through the model\nand how the memory is utilized and handled.''' , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(_UpperCAmelCase ) ) _A = Square(0.3 ) input.set_fill(_UpperCAmelCase , opacity=1.0 ) input.set_stroke(width=0.0 ) input.next_to(model_base[0] , _UpperCAmelCase , buff=0.5 ) self.play(Write(_UpperCAmelCase ) ) input.generate_target() input.target.next_to(model_arr[0] , direction=_UpperCAmelCase , buff=0.02 ) self.play(MoveToTarget(_UpperCAmelCase ) ) self.play(FadeOut(_UpperCAmelCase ) ) _A = Arrow(start=_UpperCAmelCase , end=_UpperCAmelCase , color=_UpperCAmelCase , buff=0.5 ) a.next_to(model_arr[0].get_left() , _UpperCAmelCase , buff=0.2 ) model_cpu_arr[0].generate_target() model_cpu_arr[0].target.move_to(gpu_rect[0] ) _A = MarkupText( F'''As the input reaches a layer, the hook triggers\nand weights are moved from the CPU\nto the GPU and back.''' , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(_UpperCAmelCase , run_time=3 ) ) _A = {'run_time': 1, 'fade_in': True, 'fade_out': True, 'buff': 0.02} self.play( Write(_UpperCAmelCase ) , Circumscribe(model_arr[0] , color=_UpperCAmelCase , **_UpperCAmelCase ) , Circumscribe(model_cpu_arr[0] , color=_UpperCAmelCase , **_UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=_UpperCAmelCase , **_UpperCAmelCase ) , ) self.play(MoveToTarget(model_cpu_arr[0] ) ) _A = a.copy() for i in range(6 ): a_c.next_to(model_arr[i].get_right() + 0.02 , _UpperCAmelCase , buff=0.2 ) input.generate_target() input.target.move_to(model_arr[i].get_right() + 0.02 ) _A = AnimationGroup( FadeOut(_UpperCAmelCase , run_time=0.5 ) , MoveToTarget(_UpperCAmelCase , run_time=0.5 ) , FadeIn(_UpperCAmelCase , run_time=0.5 ) , lag_ratio=0.2 ) self.play(_UpperCAmelCase ) model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[i] ) if i < 5: model_cpu_arr[i + 1].generate_target() model_cpu_arr[i + 1].target.move_to(gpu_rect[0] ) if i >= 1: _A = 0.7 self.play( Circumscribe(model_arr[i] , **_UpperCAmelCase ) , Circumscribe(cpu_left_col_base[i] , **_UpperCAmelCase ) , Circumscribe(cpu_left_col_base[i + 1] , color=_UpperCAmelCase , **_UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=_UpperCAmelCase , **_UpperCAmelCase ) , Circumscribe(model_arr[i + 1] , color=_UpperCAmelCase , **_UpperCAmelCase ) , ) if i < 1: self.play( MoveToTarget(model_cpu_arr[i] ) , MoveToTarget(model_cpu_arr[i + 1] ) , ) else: self.play( MoveToTarget(model_cpu_arr[i] , run_time=0.7 ) , MoveToTarget(model_cpu_arr[i + 1] , run_time=0.7 ) , ) else: model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[-1] ) input.generate_target() input.target.next_to(model_arr[-1].get_right() , RIGHT + 0.02 , buff=0.2 ) self.play( Circumscribe(model_arr[-1] , color=_UpperCAmelCase , **_UpperCAmelCase ) , Circumscribe(cpu_left_col_base[-1] , color=_UpperCAmelCase , **_UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=_UpperCAmelCase , **_UpperCAmelCase ) , ) self.play(MoveToTarget(model_cpu_arr[i] ) ) _A = a_c _A = a_c.copy() input.generate_target() input.target.next_to(model_base[-1] , RIGHT + 0.02 , buff=0.5 ) self.play( FadeOut(_UpperCAmelCase ) , FadeOut(_UpperCAmelCase , run_time=0.5 ) , ) _A = MarkupText(F'''Inference on a model too large for GPU memory\nis successfully completed.''' , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(_UpperCAmelCase , run_time=3 ) , MoveToTarget(_UpperCAmelCase ) ) self.wait()
315
0
import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation import ( MaxLengthCriteria, MaxNewTokensCriteria, MaxTimeCriteria, StoppingCriteriaList, validate_stopping_criteria, ) @require_torch class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self , __lowercase) -> List[str]: __UpperCamelCase :List[Any] = 3 __UpperCamelCase :Optional[Any] = 250 __UpperCamelCase :Union[str, Any] = ids_tensor((batch_size, length) , lowercase_) __UpperCamelCase :Dict = torch.ones((batch_size, length) , device=lowercase_ , dtype=torch.float) / length return input_ids, scores def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase , __UpperCamelCase :Dict = self._get_tensors(5) __UpperCamelCase :Dict = StoppingCriteriaList( [ MaxLengthCriteria(max_length=10), MaxTimeCriteria(max_time=0.1), ]) self.assertFalse(criteria(lowercase_ , lowercase_)) __UpperCamelCase , __UpperCamelCase :str = self._get_tensors(9) self.assertFalse(criteria(lowercase_ , lowercase_)) __UpperCamelCase , __UpperCamelCase :Any = self._get_tensors(10) self.assertTrue(criteria(lowercase_ , lowercase_)) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :str = MaxLengthCriteria(max_length=10) __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self._get_tensors(5) self.assertFalse(criteria(lowercase_ , lowercase_)) __UpperCamelCase , __UpperCamelCase :List[str] = self._get_tensors(9) self.assertFalse(criteria(lowercase_ , lowercase_)) __UpperCamelCase , __UpperCamelCase :Tuple = self._get_tensors(10) self.assertTrue(criteria(lowercase_ , lowercase_)) def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Optional[Any] = MaxNewTokensCriteria(start_length=5 , max_new_tokens=5) __UpperCamelCase , __UpperCamelCase :Optional[int] = self._get_tensors(5) self.assertFalse(criteria(lowercase_ , lowercase_)) __UpperCamelCase , __UpperCamelCase :int = self._get_tensors(9) self.assertFalse(criteria(lowercase_ , lowercase_)) __UpperCamelCase , __UpperCamelCase :List[Any] = self._get_tensors(10) self.assertTrue(criteria(lowercase_ , lowercase_)) __UpperCamelCase :Tuple = StoppingCriteriaList([criteria]) self.assertEqual(criteria_list.max_length , 10) def UpperCamelCase__ ( self) -> Any: __UpperCamelCase , __UpperCamelCase :int = self._get_tensors(5) __UpperCamelCase :Any = MaxTimeCriteria(max_time=0.1) self.assertFalse(criteria(lowercase_ , lowercase_)) __UpperCamelCase :Any = MaxTimeCriteria(max_time=0.1 , initial_timestamp=time.time() - 0.2) self.assertTrue(criteria(lowercase_ , lowercase_)) def UpperCamelCase__ ( self) -> Tuple: validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10)]) , 10) with self.assertWarns(lowercase_): validate_stopping_criteria(StoppingCriteriaList([MaxLengthCriteria(10)]) , 11) __UpperCamelCase :Union[str, Any] = validate_stopping_criteria(StoppingCriteriaList() , 11) self.assertEqual(len(lowercase_) , 1)
364
import argparse import hashlib import os import urllib import warnings import torch from torch import nn from tqdm import tqdm from transformers import WhisperConfig, WhisperForConditionalGeneration __lowercase = { '''tiny.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d3dd57d32accea0b295c96e26691aa14d8822fac7d9d27d5dc00b4ca2826dd03/tiny.en.pt''', '''tiny''': '''https://openaipublic.azureedge.net/main/whisper/models/65147644a518d12f04e32d6f3b26facc3f8dd46e5390956a9424a650c0ce22b9/tiny.pt''', '''base.en''': '''https://openaipublic.azureedge.net/main/whisper/models/25a8566e1d0c1e2231d1c762132cd20e0f96a85d16145c3a00adf5d1ac670ead/base.en.pt''', '''base''': '''https://openaipublic.azureedge.net/main/whisper/models/ed3a0b6b1c0edf879ad9b11b1af5a0e6ab5db9205f891f668f8b0e6c6326e34e/base.pt''', '''small.en''': '''https://openaipublic.azureedge.net/main/whisper/models/f953ad0fd29cacd07d5a9eda5624af0f6bcf2258be67c92b79389873d91e0872/small.en.pt''', '''small''': '''https://openaipublic.azureedge.net/main/whisper/models/9ecf779972d90ba49c06d968637d720dd632c55bbf19d441fb42bf17a411e794/small.pt''', '''medium.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d7440d1dc186f76616474e0ff0b3b6b879abc9d1a4926b7adfa41db2d497ab4f/medium.en.pt''', '''medium''': '''https://openaipublic.azureedge.net/main/whisper/models/345ae4da62f9b3d59415adc60127b97c714f32e89e936602e85993674d08dcb1/medium.pt''', '''large''': '''https://openaipublic.azureedge.net/main/whisper/models/e4b87e7e0bf463eb8e6956e646f1e277e901512310def2c24bf0e11bd3c28e9a/large.pt''', '''large-v2''': '''https://openaipublic.azureedge.net/main/whisper/models/81f7c96c852ee8fc832187b0132e569d6c3065a3252ed18e56effd0b6a73e524/large-v2.pt''', } def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :int = ['''layers''', '''blocks'''] for k in ignore_keys: state_dict.pop(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) __lowercase = { '''blocks''': '''layers''', '''mlp.0''': '''fc1''', '''mlp.2''': '''fc2''', '''mlp_ln''': '''final_layer_norm''', '''.attn.query''': '''.self_attn.q_proj''', '''.attn.key''': '''.self_attn.k_proj''', '''.attn.value''': '''.self_attn.v_proj''', '''.attn_ln''': '''.self_attn_layer_norm''', '''.attn.out''': '''.self_attn.out_proj''', '''.cross_attn.query''': '''.encoder_attn.q_proj''', '''.cross_attn.key''': '''.encoder_attn.k_proj''', '''.cross_attn.value''': '''.encoder_attn.v_proj''', '''.cross_attn_ln''': '''.encoder_attn_layer_norm''', '''.cross_attn.out''': '''.encoder_attn.out_proj''', '''decoder.ln.''': '''decoder.layer_norm.''', '''encoder.ln.''': '''encoder.layer_norm.''', '''token_embedding''': '''embed_tokens''', '''encoder.positional_embedding''': '''encoder.embed_positions.weight''', '''decoder.positional_embedding''': '''decoder.embed_positions.weight''', '''ln_post''': '''layer_norm''', } def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[str] = list(s_dict.keys() ) for key in keys: __UpperCamelCase :Dict = key for k, v in WHISPER_MAPPING.items(): if k in key: __UpperCamelCase :str = new_key.replace(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) print(f"""{key} -> {new_key}""" ) __UpperCamelCase :Any = s_dict.pop(SCREAMING_SNAKE_CASE ) return s_dict def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase , __UpperCamelCase :List[Any] = emb.weight.shape __UpperCamelCase :Any = nn.Linear(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , bias=SCREAMING_SNAKE_CASE ) __UpperCamelCase :str = emb.weight.data return lin_layer def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' os.makedirs(SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) __UpperCamelCase :int = os.path.basename(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[Any] = url.split('''/''' )[-2] __UpperCamelCase :Tuple = os.path.join(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if os.path.exists(SCREAMING_SNAKE_CASE ) and not os.path.isfile(SCREAMING_SNAKE_CASE ): raise RuntimeError(f"""{download_target} exists and is not a regular file""" ) if os.path.isfile(SCREAMING_SNAKE_CASE ): __UpperCamelCase :List[str] = open(SCREAMING_SNAKE_CASE , '''rb''' ).read() if hashlib.shaaaa(SCREAMING_SNAKE_CASE ).hexdigest() == expected_shaaaa: return model_bytes else: warnings.warn(f"""{download_target} exists, but the SHA256 checksum does not match; re-downloading the file""" ) with urllib.request.urlopen(SCREAMING_SNAKE_CASE ) as source, open(SCREAMING_SNAKE_CASE , '''wb''' ) as output: with tqdm( total=int(source.info().get('''Content-Length''' ) ) , ncols=80 , unit='''iB''' , unit_scale=SCREAMING_SNAKE_CASE , unit_divisor=1_024 ) as loop: while True: __UpperCamelCase :Optional[Any] = source.read(8_192 ) if not buffer: break output.write(SCREAMING_SNAKE_CASE ) loop.update(len(SCREAMING_SNAKE_CASE ) ) __UpperCamelCase :str = open(SCREAMING_SNAKE_CASE , '''rb''' ).read() if hashlib.shaaaa(SCREAMING_SNAKE_CASE ).hexdigest() != expected_shaaaa: raise RuntimeError( '''Model has been downloaded but the SHA256 checksum does not not match. Please retry loading the model.''' ) return model_bytes def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' if ".pt" not in checkpoint_path: __UpperCamelCase :Tuple = _download(_MODELS[checkpoint_path] ) else: __UpperCamelCase :Optional[int] = torch.load(SCREAMING_SNAKE_CASE , map_location='''cpu''' ) __UpperCamelCase :Union[str, Any] = original_checkpoint['''dims'''] __UpperCamelCase :List[Any] = original_checkpoint['''model_state_dict'''] __UpperCamelCase :Optional[Any] = state_dict['''decoder.token_embedding.weight'''] remove_ignore_keys_(SCREAMING_SNAKE_CASE ) rename_keys(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Dict = True __UpperCamelCase :Tuple = state_dict['''decoder.layers.0.fc1.weight'''].shape[0] __UpperCamelCase :Dict = WhisperConfig( vocab_size=dimensions['''n_vocab'''] , encoder_ffn_dim=SCREAMING_SNAKE_CASE , decoder_ffn_dim=SCREAMING_SNAKE_CASE , num_mel_bins=dimensions['''n_mels'''] , d_model=dimensions['''n_audio_state'''] , max_target_positions=dimensions['''n_text_ctx'''] , encoder_layers=dimensions['''n_audio_layer'''] , encoder_attention_heads=dimensions['''n_audio_head'''] , decoder_layers=dimensions['''n_text_layer'''] , decoder_attention_heads=dimensions['''n_text_state'''] , max_source_positions=dimensions['''n_audio_ctx'''] , ) __UpperCamelCase :str = WhisperForConditionalGeneration(SCREAMING_SNAKE_CASE ) __UpperCamelCase , __UpperCamelCase :Any = model.model.load_state_dict(SCREAMING_SNAKE_CASE , strict=SCREAMING_SNAKE_CASE ) if len(SCREAMING_SNAKE_CASE ) > 0 and not set(SCREAMING_SNAKE_CASE ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( '''Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,''' f""" but all the following weights are missing {missing}""" ) if tie_embeds: __UpperCamelCase :Optional[Any] = make_linear_from_emb(model.model.decoder.embed_tokens ) else: __UpperCamelCase :Union[str, Any] = proj_out_weights model.save_pretrained(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() # # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Patht to the downloaded checkpoints''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') __lowercase = parser.parse_args() convert_openai_whisper_to_tfms(args.checkpoint_path, args.pytorch_dump_folder_path)
105
0
from math import log from scipy.constants import Boltzmann, physical_constants lowercase : List[Any] = 300 # TEMPERATURE (unit = K) def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ) -> float: if donor_conc <= 0: raise ValueError("""Donor concentration should be positive""" ) elif acceptor_conc <= 0: raise ValueError("""Acceptor concentration should be positive""" ) elif intrinsic_conc <= 0: raise ValueError("""Intrinsic concentration should be positive""" ) elif donor_conc <= intrinsic_conc: raise ValueError( """Donor concentration should be greater than intrinsic concentration""" ) elif acceptor_conc <= intrinsic_conc: raise ValueError( """Acceptor concentration should be greater than intrinsic concentration""" ) else: return ( Boltzmann * T * log((donor_conc * acceptor_conc) / intrinsic_conc**2 ) / physical_constants["electron volt"][0] ) if __name__ == "__main__": import doctest doctest.testmod()
20
import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class lowerCAmelCase_ ( unittest.TestCase ): def __init__( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_=7, SCREAMING_SNAKE_CASE_=3, SCREAMING_SNAKE_CASE_=18, SCREAMING_SNAKE_CASE_=30, SCREAMING_SNAKE_CASE_=400, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=None, SCREAMING_SNAKE_CASE_=True, ) -> Union[str, Any]: UpperCamelCase : str = size if size is not None else {'height': 18, 'width': 18} UpperCamelCase : int = parent UpperCamelCase : List[Any] = batch_size UpperCamelCase : int = num_channels UpperCamelCase : Any = image_size UpperCamelCase : Optional[int] = min_resolution UpperCamelCase : Optional[Any] = max_resolution UpperCamelCase : Union[str, Any] = do_resize UpperCamelCase : List[Any] = size UpperCamelCase : int = do_normalize def snake_case_ ( self ) -> Tuple: return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.88_66_44_36_34_03_32_03, 0.66_18_82_93_69_54_49_83, 0.38_91_74_64_01_78_68_04], [-0.60_42_55_91_46_88_11_04, -0.0_22_95_00_88_60_52_84_69, 0.54_23_79_73_69_00_32_96], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class lowerCAmelCase_ ( a__ , unittest.TestCase ): UpperCAmelCase__ : Any = ImageGPTImageProcessor if is_vision_available() else None def snake_case_ ( self ) -> int: UpperCamelCase : str = ImageGPTImageProcessingTester(self ) @property def snake_case_ ( self ) -> Optional[Any]: return self.image_processor_tester.prepare_image_processor_dict() def snake_case_ ( self ) -> str: UpperCamelCase : int = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_, 'clusters' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_, 'do_resize' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_, 'size' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_, 'do_normalize' ) ) def snake_case_ ( self ) -> str: UpperCamelCase : Tuple = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size, {'height': 18, 'width': 18} ) UpperCamelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict, size=42 ) self.assertEqual(image_processor.size, {'height': 42, 'width': 42} ) def snake_case_ ( self ) -> List[str]: UpperCamelCase : str = self.image_processing_class(**self.image_processor_dict ) UpperCamelCase : int = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(SCREAMING_SNAKE_CASE_, obj[key] ) ) else: self.assertEqual(obj[key], SCREAMING_SNAKE_CASE_ ) def snake_case_ ( self ) -> Dict: UpperCamelCase : Tuple = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: UpperCamelCase : List[str] = os.path.join(SCREAMING_SNAKE_CASE_, 'image_processor.json' ) image_processor_first.to_json_file(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Tuple = self.image_processing_class.from_json_file(SCREAMING_SNAKE_CASE_ ).to_dict() UpperCamelCase : List[Any] = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(SCREAMING_SNAKE_CASE_, image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key], SCREAMING_SNAKE_CASE_ ) def snake_case_ ( self ) -> Optional[int]: UpperCamelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : str = self.image_processing_class.from_pretrained(SCREAMING_SNAKE_CASE_ ).to_dict() UpperCamelCase : Union[str, Any] = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(SCREAMING_SNAKE_CASE_, image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key], SCREAMING_SNAKE_CASE_ ) @unittest.skip('ImageGPT requires clusters at initialization' ) def snake_case_ ( self ) -> str: pass def UpperCamelCase ( ) -> int: UpperCamelCase : Optional[int] = load_dataset('hf-internal-testing/fixtures_image_utils' , split='test' ) UpperCamelCase : int = Image.open(dataset[4]['file'] ) UpperCamelCase : Optional[Any] = Image.open(dataset[5]['file'] ) UpperCamelCase : str = [imagea, imagea] return images @require_vision @require_torch class lowerCAmelCase_ ( unittest.TestCase ): @slow def snake_case_ ( self ) -> str: UpperCamelCase : List[str] = ImageGPTImageProcessor.from_pretrained('openai/imagegpt-small' ) UpperCamelCase : List[str] = prepare_images() # test non-batched UpperCamelCase : int = image_processing(images[0], return_tensors='pt' ) self.assertIsInstance(encoding.input_ids, torch.LongTensor ) self.assertEqual(encoding.input_ids.shape, (1, 1024) ) UpperCamelCase : Union[str, Any] = [306, 191, 191] self.assertEqual(encoding.input_ids[0, :3].tolist(), SCREAMING_SNAKE_CASE_ ) # test batched UpperCamelCase : Tuple = image_processing(SCREAMING_SNAKE_CASE_, return_tensors='pt' ) self.assertIsInstance(encoding.input_ids, torch.LongTensor ) self.assertEqual(encoding.input_ids.shape, (2, 1024) ) UpperCamelCase : Optional[Any] = [303, 13, 13] self.assertEqual(encoding.input_ids[1, -3:].tolist(), SCREAMING_SNAKE_CASE_ )
119
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a : Union[str, Any] = logging.get_logger(__name__) a : Any = { 'sail/poolformer_s12': 'https://huggingface.co/sail/poolformer_s12/resolve/main/config.json', # See all PoolFormer models at https://huggingface.co/models?filter=poolformer } class _a ( _lowerCAmelCase ): A = '''poolformer''' def __init__(self, SCREAMING_SNAKE_CASE_=3, SCREAMING_SNAKE_CASE_=16, SCREAMING_SNAKE_CASE_=16, SCREAMING_SNAKE_CASE_=3, SCREAMING_SNAKE_CASE_=4.0, SCREAMING_SNAKE_CASE_=[2, 2, 6, 2], SCREAMING_SNAKE_CASE_=[64, 128, 320, 512], SCREAMING_SNAKE_CASE_=[7, 3, 3, 3], SCREAMING_SNAKE_CASE_=[4, 2, 2, 2], SCREAMING_SNAKE_CASE_=[2, 1, 1, 1], SCREAMING_SNAKE_CASE_=4, SCREAMING_SNAKE_CASE_=0.0, SCREAMING_SNAKE_CASE_="gelu", SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=1E-5, SCREAMING_SNAKE_CASE_=0.0_2, **SCREAMING_SNAKE_CASE_, ) -> Any: UpperCAmelCase_: Any = num_channels UpperCAmelCase_: Optional[int] = patch_size UpperCAmelCase_: Any = stride UpperCAmelCase_: Any = padding UpperCAmelCase_: List[str] = pool_size UpperCAmelCase_: Optional[Any] = hidden_sizes UpperCAmelCase_: Tuple = mlp_ratio UpperCAmelCase_: Any = depths UpperCAmelCase_: Tuple = patch_sizes UpperCAmelCase_: Dict = strides UpperCAmelCase_: List[Any] = num_encoder_blocks UpperCAmelCase_: str = drop_path_rate UpperCAmelCase_: Optional[int] = hidden_act UpperCAmelCase_: Optional[Any] = use_layer_scale UpperCAmelCase_: Tuple = layer_scale_init_value UpperCAmelCase_: Optional[int] = initializer_range super().__init__(**SCREAMING_SNAKE_CASE_ ) class _a ( _lowerCAmelCase ): A = version.parse('''1.11''' ) @property def __snake_case (self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def __snake_case (self ) -> float: return 2E-3
82
a : Tuple = 'Tobias Carryer' from time import time class _a : def __init__(self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_=int(time() ) ) -> List[Any]: # noqa: B008 UpperCAmelCase_: List[str] = multiplier UpperCAmelCase_: Tuple = increment UpperCAmelCase_: Tuple = modulo UpperCAmelCase_: List[str] = seed def __snake_case (self ) -> Any: UpperCAmelCase_: List[str] = (self.multiplier * self.seed + self.increment) % self.modulo return self.seed if __name__ == "__main__": # Show the LCG in action. a : Optional[int] = LinearCongruentialGenerator(1_664_525, 1_013_904_223, 2 << 31) while True: print(lcg.next_number())
82
1
import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCAmelCase = 'src/transformers' # This is to make sure the transformers module imported is the one in the repo. lowerCAmelCase = direct_transformers_import(PATH_TO_TRANSFORMERS) lowerCAmelCase = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowerCAmelCase = re.compile(R'\[(.+?)\]\((https://huggingface\.co/.+?)\)') lowerCAmelCase = { 'DecisionTransformerConfig', 'EncoderDecoderConfig', 'MusicgenConfig', 'RagConfig', 'SpeechEncoderDecoderConfig', 'TimmBackboneConfig', 'VisionEncoderDecoderConfig', 'VisionTextDualEncoderConfig', 'LlamaConfig', } def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = None # source code of `config_class` lowercase__ = inspect.getsource(SCREAMING_SNAKE_CASE ) lowercase__ = _re_checkpoint.findall(SCREAMING_SNAKE_CASE ) # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` for ckpt_name, ckpt_link in checkpoints: # allow the link to end with `/` if ckpt_link.endswith('''/''' ): lowercase__ = ckpt_link[:-1] # verify the checkpoint name corresponds to the checkpoint link lowercase__ = f'https://huggingface.co/{ckpt_name}' if ckpt_link == ckpt_link_from_name: lowercase__ = ckpt_name break return checkpoint def _a ( ): """simple docstring""" lowercase__ = [] for config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in config_class.__module__: continue lowercase__ = get_checkpoint_from_config_class(SCREAMING_SNAKE_CASE ) lowercase__ = config_class.__name__ if checkpoint is None and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(SCREAMING_SNAKE_CASE ) if len(SCREAMING_SNAKE_CASE ) > 0: lowercase__ = '''\n'''.join(sorted(SCREAMING_SNAKE_CASE ) ) raise ValueError(f'The following configurations don\'t contain any valid checkpoint:\n{message}' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
110
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCAmelCase = { 'configuration_tapas': ['TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TapasConfig'], 'tokenization_tapas': ['TapasTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase = [ 'TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TapasForMaskedLM', 'TapasForQuestionAnswering', 'TapasForSequenceClassification', 'TapasModel', 'TapasPreTrainedModel', 'load_tf_weights_in_tapas', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase = [ 'TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFTapasForMaskedLM', 'TFTapasForQuestionAnswering', 'TFTapasForSequenceClassification', 'TFTapasModel', 'TFTapasPreTrainedModel', ] if TYPE_CHECKING: from .configuration_tapas import TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP, TapasConfig from .tokenization_tapas import TapasTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tapas import ( TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasPreTrainedModel, load_tf_weights_in_tapas, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_tapas import ( TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TFTapasForMaskedLM, TFTapasForQuestionAnswering, TFTapasForSequenceClassification, TFTapasModel, TFTapasPreTrainedModel, ) else: import sys lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
110
1
"""simple docstring""" from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=_lowercase) class UpperCAmelCase_ ( _lowercase): snake_case__ = field(default='''language-modeling''' , metadata={'''include_in_asdict_even_if_is_default''': True}) snake_case__ = Features({'''text''': Value('''string''')}) snake_case__ = Features({}) snake_case__ = '''text''' @property def _UpperCamelCase ( self : Any ) -> Dict[str, str]: return {self.text_column: "text"}
357
"""simple docstring""" from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput UpperCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name class UpperCAmelCase_ ( _lowercase , _lowercase): @register_to_config def __init__( self : Tuple , __UpperCamelCase : bool , __UpperCamelCase : Optional[int] = None , __UpperCamelCase : Optional[int] = None ) -> int: super().__init__() _UpperCamelCase = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" _UpperCamelCase = torch.zeros(__UpperCamelCase , __UpperCamelCase ) else: _UpperCamelCase = None _UpperCamelCase = torch.nn.Parameter(__UpperCamelCase ) class UpperCAmelCase_ ( _lowercase): snake_case__ = 42 snake_case__ = 42 snake_case__ = 42 snake_case__ = 42 snake_case__ = 42 snake_case__ = 42 def __init__( self : List[str] , __UpperCamelCase : VQModel , __UpperCamelCase : CLIPTextModel , __UpperCamelCase : CLIPTokenizer , __UpperCamelCase : TransformeraDModel , __UpperCamelCase : VQDiffusionScheduler , __UpperCamelCase : LearnedClassifierFreeSamplingEmbeddings , ) -> Optional[int]: super().__init__() self.register_modules( vqvae=__UpperCamelCase , transformer=__UpperCamelCase , text_encoder=__UpperCamelCase , tokenizer=__UpperCamelCase , scheduler=__UpperCamelCase , learned_classifier_free_sampling_embeddings=__UpperCamelCase , ) def _UpperCamelCase ( self : Optional[Any] , __UpperCamelCase : Dict , __UpperCamelCase : Dict , __UpperCamelCase : List[str] ) -> str: _UpperCamelCase = len(__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else 1 # get prompt text embeddings _UpperCamelCase = self.tokenizer( __UpperCamelCase , padding='''max_length''' , max_length=self.tokenizer.model_max_length , return_tensors='''pt''' , ) _UpperCamelCase = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: _UpperCamelCase = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( '''The following part of your input was truncated because CLIP can only handle sequences up to''' F''' {self.tokenizer.model_max_length} tokens: {removed_text}''' ) _UpperCamelCase = text_input_ids[:, : self.tokenizer.model_max_length] _UpperCamelCase = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 _UpperCamelCase = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=__UpperCamelCase ) # duplicate text embeddings for each generation per prompt _UpperCamelCase = prompt_embeds.repeat_interleave(__UpperCamelCase , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: _UpperCamelCase = self.learned_classifier_free_sampling_embeddings.embeddings _UpperCamelCase = negative_prompt_embeds.unsqueeze(0 ).repeat(__UpperCamelCase , 1 , 1 ) else: _UpperCamelCase = [''''''] * batch_size _UpperCamelCase = text_input_ids.shape[-1] _UpperCamelCase = self.tokenizer( __UpperCamelCase , padding='''max_length''' , max_length=__UpperCamelCase , truncation=__UpperCamelCase , return_tensors='''pt''' , ) _UpperCamelCase = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings _UpperCamelCase = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=__UpperCamelCase ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method _UpperCamelCase = negative_prompt_embeds.shape[1] _UpperCamelCase = negative_prompt_embeds.repeat(1 , __UpperCamelCase , 1 ) _UpperCamelCase = negative_prompt_embeds.view(batch_size * num_images_per_prompt , __UpperCamelCase , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes _UpperCamelCase = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self : List[str] , __UpperCamelCase : Union[str, List[str]] , __UpperCamelCase : int = 100 , __UpperCamelCase : float = 5.0 , __UpperCamelCase : float = 1.0 , __UpperCamelCase : int = 1 , __UpperCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __UpperCamelCase : Optional[torch.FloatTensor] = None , __UpperCamelCase : Optional[str] = "pil" , __UpperCamelCase : bool = True , __UpperCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , __UpperCamelCase : int = 1 , ) -> Union[ImagePipelineOutput, Tuple]: if isinstance(__UpperCamelCase , __UpperCamelCase ): _UpperCamelCase = 1 elif isinstance(__UpperCamelCase , __UpperCamelCase ): _UpperCamelCase = len(__UpperCamelCase ) else: raise ValueError(F'''`prompt` has to be of type `str` or `list` but is {type(__UpperCamelCase )}''' ) _UpperCamelCase = batch_size * num_images_per_prompt _UpperCamelCase = guidance_scale > 1.0 _UpperCamelCase = self._encode_prompt(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__UpperCamelCase , __UpperCamelCase ) or callback_steps <= 0) ): raise ValueError( F'''`callback_steps` has to be a positive integer but is {callback_steps} of type''' F''' {type(__UpperCamelCase )}.''' ) # get the initial completely masked latents unless the user supplied it _UpperCamelCase = (batch_size, self.transformer.num_latent_pixels) if latents is None: _UpperCamelCase = self.transformer.num_vector_embeds - 1 _UpperCamelCase = torch.full(__UpperCamelCase , __UpperCamelCase ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {latents_shape}''' ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( '''Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0,''' F''' {self.transformer.num_vector_embeds - 1} (inclusive).''' ) _UpperCamelCase = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(__UpperCamelCase , device=self.device ) _UpperCamelCase = self.scheduler.timesteps.to(self.device ) _UpperCamelCase = latents for i, t in enumerate(self.progress_bar(__UpperCamelCase ) ): # expand the sample if we are doing classifier free guidance _UpperCamelCase = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` _UpperCamelCase = self.transformer(__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , timestep=__UpperCamelCase ).sample if do_classifier_free_guidance: _UpperCamelCase , _UpperCamelCase = model_output.chunk(2 ) _UpperCamelCase = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(__UpperCamelCase , dim=1 , keepdim=__UpperCamelCase ) _UpperCamelCase = self.truncate(__UpperCamelCase , __UpperCamelCase ) # remove `log(0)`'s (`-inf`s) _UpperCamelCase = model_output.clamp(-70 ) # compute the previous noisy sample x_t -> x_t-1 _UpperCamelCase = self.scheduler.step(__UpperCamelCase , timestep=__UpperCamelCase , sample=__UpperCamelCase , generator=__UpperCamelCase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) _UpperCamelCase = self.vqvae.config.vq_embed_dim _UpperCamelCase = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) _UpperCamelCase = self.vqvae.quantize.get_codebook_entry(__UpperCamelCase , shape=__UpperCamelCase ) _UpperCamelCase = self.vqvae.decode(__UpperCamelCase , force_not_quantize=__UpperCamelCase ).sample _UpperCamelCase = (image / 2 + 0.5).clamp(0 , 1 ) _UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": _UpperCamelCase = self.numpy_to_pil(__UpperCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__UpperCamelCase ) def _UpperCamelCase ( self : Optional[Any] , __UpperCamelCase : torch.FloatTensor , __UpperCamelCase : float ) -> torch.FloatTensor: _UpperCamelCase , _UpperCamelCase = torch.sort(__UpperCamelCase , 1 , descending=__UpperCamelCase ) _UpperCamelCase = torch.exp(__UpperCamelCase ) _UpperCamelCase = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out _UpperCamelCase = torch.full_like(keep_mask[:, 0:1, :] , __UpperCamelCase ) _UpperCamelCase = torch.cat((all_true, keep_mask) , dim=1 ) _UpperCamelCase = keep_mask[:, :-1, :] _UpperCamelCase = keep_mask.gather(1 , indices.argsort(1 ) ) _UpperCamelCase = log_p_x_0.clone() _UpperCamelCase = -torch.inf # -inf = log(0) return rv
54
0
"""simple docstring""" import argparse import torch from transformers import GPTaConfig, GPTaModel, load_tf_weights_in_gpta from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowercase ( _snake_case : int , _snake_case : Optional[int] , _snake_case : List[Any] ) ->List[str]: """simple docstring""" if gpta_config_file == "": __snake_case : Dict = GPTaConfig() else: __snake_case : Optional[int] = GPTaConfig.from_json_file(_snake_case ) __snake_case : Optional[int] = GPTaModel(_snake_case ) # Load weights from numpy load_tf_weights_in_gpta(_snake_case , _snake_case , _snake_case ) # Save pytorch-model __snake_case : Optional[int] = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME __snake_case : str = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(f"""Save PyTorch model to {pytorch_weights_dump_path}""" ) torch.save(model.state_dict() , _snake_case ) print(f"""Save configuration file to {pytorch_config_dump_path}""" ) with open(_snake_case , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": SCREAMING_SNAKE_CASE : str = argparse.ArgumentParser() # Required parameters parser.add_argument( """--gpt2_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--gpt2_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained OpenAI model. \n""" """This specifies the model architecture.""" ), ) SCREAMING_SNAKE_CASE : int = parser.parse_args() convert_gpta_checkpoint_to_pytorch(args.gpta_checkpoint_path, args.gpta_config_file, args.pytorch_dump_folder_path)
102
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roberta import RobertaTokenizer __A = logging.get_logger(__name__) __A = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} __A = { "vocab_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/vocab.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/vocab.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/vocab.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/vocab.json", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/vocab.json", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/vocab.json" ), }, "merges_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/merges.txt", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/merges.txt", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/merges.txt", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/merges.txt", "roberta-base-openai-detector": "https://huggingface.co/roberta-base-openai-detector/resolve/main/merges.txt", "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/merges.txt" ), }, "tokenizer_file": { "roberta-base": "https://huggingface.co/roberta-base/resolve/main/tokenizer.json", "roberta-large": "https://huggingface.co/roberta-large/resolve/main/tokenizer.json", "roberta-large-mnli": "https://huggingface.co/roberta-large-mnli/resolve/main/tokenizer.json", "distilroberta-base": "https://huggingface.co/distilroberta-base/resolve/main/tokenizer.json", "roberta-base-openai-detector": ( "https://huggingface.co/roberta-base-openai-detector/resolve/main/tokenizer.json" ), "roberta-large-openai-detector": ( "https://huggingface.co/roberta-large-openai-detector/resolve/main/tokenizer.json" ), }, } __A = { "roberta-base": 512, "roberta-large": 512, "roberta-large-mnli": 512, "distilroberta-base": 512, "roberta-base-openai-detector": 512, "roberta-large-openai-detector": 512, } class snake_case ( __snake_case ): SCREAMING_SNAKE_CASE_ : str = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ : Optional[int] = ["""input_ids""", """attention_mask"""] SCREAMING_SNAKE_CASE_ : Any = RobertaTokenizer def __init__( self : Optional[int] , UpperCamelCase__ : int=None , UpperCamelCase__ : Union[str, Any]=None , UpperCamelCase__ : Optional[Any]=None , UpperCamelCase__ : int="replace" , UpperCamelCase__ : Union[str, Any]="<s>" , UpperCamelCase__ : List[Any]="</s>" , UpperCamelCase__ : Any="</s>" , UpperCamelCase__ : Union[str, Any]="<s>" , UpperCamelCase__ : str="<unk>" , UpperCamelCase__ : Optional[int]="<pad>" , UpperCamelCase__ : int="<mask>" , UpperCamelCase__ : Optional[int]=False , UpperCamelCase__ : Optional[Any]=True , **UpperCamelCase__ : Tuple , )-> Optional[int]: '''simple docstring''' super().__init__( UpperCamelCase__ , UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , errors=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ , **UpperCamelCase__ , ) __lowerCAmelCase: Tuple = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__()) if pre_tok_state.get("add_prefix_space" , UpperCamelCase__) != add_prefix_space: __lowerCAmelCase: str = getattr(UpperCamelCase__ , pre_tok_state.pop("type")) __lowerCAmelCase: Optional[int] = add_prefix_space __lowerCAmelCase: Dict = pre_tok_class(**UpperCamelCase__) __lowerCAmelCase: Any = add_prefix_space __lowerCAmelCase: int = "post_processor" __lowerCAmelCase: Optional[Any] = getattr(self.backend_tokenizer , UpperCamelCase__ , UpperCamelCase__) if tokenizer_component_instance: __lowerCAmelCase: Dict = json.loads(tokenizer_component_instance.__getstate__()) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: __lowerCAmelCase: List[Any] = tuple(state["sep"]) if "cls" in state: __lowerCAmelCase: str = tuple(state["cls"]) __lowerCAmelCase: str = False if state.get("add_prefix_space" , UpperCamelCase__) != add_prefix_space: __lowerCAmelCase: Optional[Any] = add_prefix_space __lowerCAmelCase: List[str] = True if state.get("trim_offsets" , UpperCamelCase__) != trim_offsets: __lowerCAmelCase: Any = trim_offsets __lowerCAmelCase: List[str] = True if changes_to_apply: __lowerCAmelCase: str = getattr(UpperCamelCase__ , state.pop("type")) __lowerCAmelCase: List[str] = component_class(**UpperCamelCase__) setattr(self.backend_tokenizer , UpperCamelCase__ , UpperCamelCase__) @property def lowercase_ ( self : List[str])-> str: '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet.") return None return str(self._mask_token) @mask_token.setter def lowercase_ ( self : Tuple , UpperCamelCase__ : Optional[int])-> Optional[Any]: '''simple docstring''' __lowerCAmelCase: List[Any] = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__) if isinstance(UpperCamelCase__ , UpperCamelCase__) else value __lowerCAmelCase: int = value def lowercase_ ( self : Any , *UpperCamelCase__ : str , **UpperCamelCase__ : List[Any])-> BatchEncoding: '''simple docstring''' __lowerCAmelCase: List[Any] = kwargs.get("is_split_into_words" , UpperCamelCase__) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*UpperCamelCase__ , **UpperCamelCase__) def lowercase_ ( self : int , *UpperCamelCase__ : Dict , **UpperCamelCase__ : List[str])-> BatchEncoding: '''simple docstring''' __lowerCAmelCase: Optional[Any] = kwargs.get("is_split_into_words" , UpperCamelCase__) assert self.add_prefix_space or not is_split_into_words, ( f"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs." ) return super()._encode_plus(*UpperCamelCase__ , **UpperCamelCase__) def lowercase_ ( self : Any , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None)-> Tuple[str]: '''simple docstring''' __lowerCAmelCase: str = self._tokenizer.model.save(UpperCamelCase__ , name=UpperCamelCase__) return tuple(UpperCamelCase__) def lowercase_ ( self : Dict , UpperCamelCase__ : Any , UpperCamelCase__ : Union[str, Any]=None)-> Optional[Any]: '''simple docstring''' __lowerCAmelCase: str = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def lowercase_ ( self : int , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None)-> List[int]: '''simple docstring''' __lowerCAmelCase: Optional[Any] = [self.sep_token_id] __lowerCAmelCase: str = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0]
217
0
'''simple docstring''' _SCREAMING_SNAKE_CASE = "Alexander Joslin" import operator as op from .stack import Stack def __lowerCamelCase ( __lowerCAmelCase : str ) -> int: snake_case = {"""*""": op.mul, """/""": op.truediv, """+""": op.add, """-""": op.sub} snake_case = Stack() snake_case = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(__lowerCAmelCase ) ) elif i in operators: # RULE 2 operator_stack.push(__lowerCAmelCase ) elif i == ")": # RULE 4 snake_case = operator_stack.peek() operator_stack.pop() snake_case = operand_stack.peek() operand_stack.pop() snake_case = operand_stack.peek() operand_stack.pop() snake_case = operators[opr](__lowerCAmelCase , __lowerCAmelCase ) operand_stack.push(__lowerCAmelCase ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": _SCREAMING_SNAKE_CASE = "(5 + ((4 * 2) * (2 + 3)))" # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
3
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : Any , __snake_case : Optional[Any] , __snake_case : List[Any]=7 , __snake_case : Optional[Any]=3 , __snake_case : str=18 , __snake_case : Union[str, Any]=30 , __snake_case : Union[str, Any]=4_00 , __snake_case : Optional[int]=True , __snake_case : Any=None , __snake_case : List[str]=True , )-> Optional[Any]: snake_case = size if size is not None else {"""height""": 18, """width""": 18} snake_case = parent snake_case = batch_size snake_case = num_channels snake_case = image_size snake_case = min_resolution snake_case = max_resolution snake_case = do_resize snake_case = size snake_case = apply_ocr def lowerCAmelCase ( self : List[Any] )-> List[str]: return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class _lowerCAmelCase ( A__ , unittest.TestCase ): """simple docstring""" snake_case_ = LayoutLMvaImageProcessor if is_pytesseract_available() else None def lowerCAmelCase ( self : int )-> Tuple: snake_case = LayoutLMvaImageProcessingTester(self ) @property def lowerCAmelCase ( self : Tuple )-> Tuple: return self.image_processor_tester.prepare_image_processor_dict() def lowerCAmelCase ( self : Union[str, Any] )-> Any: snake_case = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__snake_case , """do_resize""" ) ) self.assertTrue(hasattr(__snake_case , """size""" ) ) self.assertTrue(hasattr(__snake_case , """apply_ocr""" ) ) def lowerCAmelCase ( self : List[str] )-> List[Any]: snake_case = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 18} ) snake_case = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) def lowerCAmelCase ( self : Dict )-> Union[str, Any]: pass def lowerCAmelCase ( self : Tuple )-> Dict: # Initialize image_processing snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=__snake_case ) for image in image_inputs: self.assertIsInstance(__snake_case , Image.Image ) # Test not batched input snake_case = image_processing(image_inputs[0] , return_tensors="""pt""" ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) self.assertIsInstance(encoding.words , __snake_case ) self.assertIsInstance(encoding.boxes , __snake_case ) # Test batched snake_case = image_processing(__snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) def lowerCAmelCase ( self : int )-> str: # Initialize image_processing snake_case = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=__snake_case , numpify=__snake_case ) for image in image_inputs: self.assertIsInstance(__snake_case , np.ndarray ) # Test not batched input snake_case = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched snake_case = image_processing(__snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) def lowerCAmelCase ( self : List[Any] )-> Optional[Any]: # Initialize image_processing snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=__snake_case , torchify=__snake_case ) for image in image_inputs: self.assertIsInstance(__snake_case , torch.Tensor ) # Test not batched input snake_case = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) # Test batched snake_case = image_processing(__snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) , ) def lowerCAmelCase ( self : int )-> List[Any]: # with apply_OCR = True snake_case = LayoutLMvaImageProcessor() from datasets import load_dataset snake_case = load_dataset("""hf-internal-testing/fixtures_docvqa""" , split="""test""" ) snake_case = Image.open(ds[0]["""file"""] ).convert("""RGB""" ) snake_case = image_processing(__snake_case , return_tensors="""pt""" ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 2_24, 2_24) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 snake_case = [["""11:14""", """to""", """11:39""", """a.m""", """11:39""", """to""", """11:44""", """a.m.""", """11:44""", """a.m.""", """to""", """12:25""", """p.m.""", """12:25""", """to""", """12:58""", """p.m.""", """12:58""", """to""", """4:00""", """p.m.""", """2:00""", """to""", """5:00""", """p.m.""", """Coffee""", """Break""", """Coffee""", """will""", """be""", """served""", """for""", """men""", """and""", """women""", """in""", """the""", """lobby""", """adjacent""", """to""", """exhibit""", """area.""", """Please""", """move""", """into""", """exhibit""", """area.""", """(Exhibits""", """Open)""", """TRRF""", """GENERAL""", """SESSION""", """(PART""", """|)""", """Presiding:""", """Lee""", """A.""", """Waller""", """TRRF""", """Vice""", """President""", """“Introductory""", """Remarks”""", """Lee""", """A.""", """Waller,""", """TRRF""", """Vice""", """Presi-""", """dent""", """Individual""", """Interviews""", """with""", """TRRF""", """Public""", """Board""", """Members""", """and""", """Sci-""", """entific""", """Advisory""", """Council""", """Mem-""", """bers""", """Conducted""", """by""", """TRRF""", """Treasurer""", """Philip""", """G.""", """Kuehn""", """to""", """get""", """answers""", """which""", """the""", """public""", """refrigerated""", """warehousing""", """industry""", """is""", """looking""", """for.""", """Plus""", """questions""", """from""", """the""", """floor.""", """Dr.""", """Emil""", """M.""", """Mrak,""", """University""", """of""", """Cal-""", """ifornia,""", """Chairman,""", """TRRF""", """Board;""", """Sam""", """R.""", """Cecil,""", """University""", """of""", """Georgia""", """College""", """of""", """Agriculture;""", """Dr.""", """Stanley""", """Charm,""", """Tufts""", """University""", """School""", """of""", """Medicine;""", """Dr.""", """Robert""", """H.""", """Cotton,""", """ITT""", """Continental""", """Baking""", """Company;""", """Dr.""", """Owen""", """Fennema,""", """University""", """of""", """Wis-""", """consin;""", """Dr.""", """Robert""", """E.""", """Hardenburg,""", """USDA.""", """Questions""", """and""", """Answers""", """Exhibits""", """Open""", """Capt.""", """Jack""", """Stoney""", """Room""", """TRRF""", """Scientific""", """Advisory""", """Council""", """Meeting""", """Ballroom""", """Foyer"""]] # noqa: E231 snake_case = [[[1_41, 57, 2_14, 69], [2_28, 58, 2_52, 69], [1_41, 75, 2_16, 88], [2_30, 79, 2_80, 88], [1_42, 2_60, 2_18, 2_73], [2_30, 2_61, 2_55, 2_73], [1_43, 2_79, 2_18, 2_90], [2_31, 2_82, 2_90, 2_91], [1_43, 3_42, 2_18, 3_54], [2_31, 3_45, 2_89, 3_55], [2_02, 3_62, 2_27, 3_73], [1_43, 3_79, 2_20, 3_92], [2_31, 3_82, 2_91, 3_94], [1_44, 7_14, 2_20, 7_26], [2_31, 7_15, 2_56, 7_26], [1_44, 7_32, 2_20, 7_45], [2_32, 7_36, 2_91, 7_47], [1_44, 7_69, 2_18, 7_82], [2_31, 7_70, 2_56, 7_82], [1_41, 7_88, 2_02, 8_01], [2_15, 7_91, 2_74, 8_04], [1_43, 8_26, 2_04, 8_38], [2_15, 8_26, 2_40, 8_38], [1_42, 8_44, 2_02, 8_57], [2_15, 8_47, 2_74, 8_59], [3_34, 57, 4_27, 69], [4_40, 57, 5_22, 69], [3_69, 75, 4_61, 88], [4_69, 75, 5_16, 88], [5_28, 76, 5_62, 88], [5_70, 76, 6_67, 88], [6_75, 75, 7_11, 87], [7_21, 79, 7_78, 88], [7_89, 75, 8_40, 88], [3_69, 97, 4_70, 1_07], [4_84, 94, 5_07, 1_06], [5_18, 94, 5_62, 1_07], [5_76, 94, 6_55, 1_10], [6_68, 94, 7_92, 1_09], [8_04, 95, 8_29, 1_07], [3_69, 1_13, 4_65, 1_25], [4_77, 1_16, 5_47, 1_25], [5_62, 1_13, 6_58, 1_25], [6_71, 1_16, 7_48, 1_25], [7_61, 1_13, 8_11, 1_25], [3_69, 1_31, 4_65, 1_43], [4_77, 1_33, 5_48, 1_43], [5_63, 1_30, 6_98, 1_45], [7_10, 1_30, 8_02, 1_46], [3_36, 1_71, 4_12, 1_83], [4_23, 1_71, 5_72, 1_83], [5_82, 1_70, 7_16, 1_84], [7_28, 1_71, 8_17, 1_87], [8_29, 1_71, 8_44, 1_86], [3_38, 1_97, 4_82, 2_12], [5_07, 1_96, 5_57, 2_09], [5_69, 1_96, 5_95, 2_08], [6_10, 1_96, 7_02, 2_09], [5_05, 2_14, 5_83, 2_26], [5_95, 2_14, 6_56, 2_27], [6_70, 2_15, 8_07, 2_27], [3_35, 2_59, 5_43, 2_74], [5_56, 2_59, 7_08, 2_72], [3_72, 2_79, 4_22, 2_91], [4_35, 2_79, 4_60, 2_91], [4_74, 2_79, 5_74, 2_92], [5_87, 2_78, 6_64, 2_91], [6_76, 2_78, 7_38, 2_91], [7_51, 2_79, 8_34, 2_91], [3_72, 2_98, 4_34, 3_10], [3_35, 3_41, 4_83, 3_54], [4_97, 3_41, 6_55, 3_54], [6_67, 3_41, 7_28, 3_54], [7_40, 3_41, 8_25, 3_54], [3_35, 3_60, 4_30, 3_72], [4_42, 3_60, 5_34, 3_72], [5_45, 3_59, 6_87, 3_72], [6_97, 3_60, 7_54, 3_72], [7_65, 3_60, 8_23, 3_73], [3_34, 3_78, 4_28, 3_91], [4_40, 3_78, 5_77, 3_94], [5_90, 3_78, 7_05, 3_91], [7_20, 3_78, 8_01, 3_91], [3_34, 3_97, 4_00, 4_09], [3_70, 4_16, 5_29, 4_29], [5_44, 4_16, 5_76, 4_32], [5_87, 4_16, 6_65, 4_28], [6_77, 4_16, 8_14, 4_29], [3_72, 4_35, 4_52, 4_50], [4_65, 4_34, 4_95, 4_47], [5_11, 4_34, 6_00, 4_47], [6_11, 4_36, 6_37, 4_47], [6_49, 4_36, 6_94, 4_51], [7_05, 4_38, 8_24, 4_47], [3_69, 4_53, 4_52, 4_66], [4_64, 4_54, 5_09, 4_66], [5_22, 4_53, 6_11, 4_69], [6_25, 4_53, 7_92, 4_69], [3_70, 4_72, 5_56, 4_88], [5_70, 4_72, 6_84, 4_87], [6_97, 4_72, 7_18, 4_85], [7_32, 4_72, 8_35, 4_88], [3_69, 4_90, 4_11, 5_03], [4_25, 4_90, 4_84, 5_03], [4_96, 4_90, 6_35, 5_06], [6_45, 4_90, 7_07, 5_03], [7_18, 4_91, 7_61, 5_03], [7_71, 4_90, 8_40, 5_03], [3_36, 5_10, 3_74, 5_21], [3_88, 5_10, 4_47, 5_22], [4_60, 5_10, 4_89, 5_21], [5_03, 5_10, 5_80, 5_22], [5_92, 5_09, 7_36, 5_25], [7_45, 5_09, 7_70, 5_22], [7_81, 5_09, 8_40, 5_22], [3_38, 5_28, 4_34, 5_41], [4_48, 5_28, 5_96, 5_41], [6_09, 5_27, 6_87, 5_40], [7_00, 5_28, 7_92, 5_41], [3_36, 5_46, 3_97, 5_59], [4_07, 5_46, 4_31, 5_59], [4_43, 5_46, 5_25, 5_60], [5_37, 5_46, 6_80, 5_62], [6_88, 5_46, 7_14, 5_59], [7_22, 5_46, 8_37, 5_62], [3_36, 5_65, 4_49, 5_81], [4_61, 5_65, 4_85, 5_77], [4_97, 5_65, 6_65, 5_81], [6_81, 5_65, 7_18, 5_77], [7_32, 5_65, 8_37, 5_80], [3_37, 5_84, 4_38, 5_97], [4_52, 5_83, 5_21, 5_96], [5_35, 5_84, 6_77, 5_99], [6_90, 5_83, 7_87, 5_96], [8_01, 5_83, 8_25, 5_96], [3_38, 6_02, 4_78, 6_15], [4_92, 6_02, 5_30, 6_14], [5_43, 6_02, 6_38, 6_15], [6_50, 6_02, 6_76, 6_14], [6_88, 6_02, 7_88, 6_15], [8_02, 6_02, 8_43, 6_14], [3_37, 6_21, 5_02, 6_33], [5_16, 6_21, 6_15, 6_37], [6_29, 6_21, 7_74, 6_36], [7_89, 6_21, 8_27, 6_33], [3_37, 6_39, 4_18, 6_52], [4_32, 6_40, 5_71, 6_53], [5_87, 6_39, 7_31, 6_55], [7_43, 6_39, 7_69, 6_52], [7_80, 6_39, 8_41, 6_52], [3_38, 6_58, 4_40, 6_73], [4_55, 6_58, 4_91, 6_70], [5_08, 6_58, 6_02, 6_71], [6_16, 6_58, 6_38, 6_70], [6_54, 6_58, 8_35, 6_74], [3_37, 6_77, 4_29, 6_89], [3_37, 7_14, 4_82, 7_26], [4_95, 7_14, 5_48, 7_26], [5_61, 7_14, 6_83, 7_26], [3_38, 7_70, 4_61, 7_82], [4_74, 7_69, 5_54, 7_85], [4_89, 7_88, 5_62, 8_03], [5_76, 7_88, 6_43, 8_01], [6_56, 7_87, 7_51, 8_04], [7_64, 7_88, 8_44, 8_01], [3_34, 8_25, 4_21, 8_38], [4_30, 8_24, 5_74, 8_38], [5_84, 8_24, 7_23, 8_41], [3_35, 8_44, 4_50, 8_57], [4_64, 8_43, 5_83, 8_60], [6_28, 8_62, 7_55, 8_75], [7_69, 8_61, 8_48, 8_78]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , __snake_case ) self.assertListEqual(encoding.boxes , __snake_case ) # with apply_OCR = False snake_case = LayoutLMvaImageProcessor(apply_ocr=__snake_case ) snake_case = image_processing(__snake_case , return_tensors="""pt""" ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 2_24, 2_24) )
3
1
def A_ ( _lowerCAmelCase ) -> Tuple: UpperCamelCase : Dict = 0 UpperCamelCase : Any = len(_lowerCAmelCase ) for i in range(n - 1 ): for j in range(i + 1 , _lowerCAmelCase ): if arr[i] > arr[j]: num_inversions += 1 return num_inversions def A_ ( _lowerCAmelCase ) -> Optional[int]: if len(_lowerCAmelCase ) <= 1: return arr, 0 UpperCamelCase : int = len(_lowerCAmelCase ) // 2 UpperCamelCase : Union[str, Any] = arr[0:mid] UpperCamelCase : Optional[Any] = arr[mid:] UpperCamelCase , UpperCamelCase : Optional[int] = count_inversions_recursive(_lowerCAmelCase ) UpperCamelCase , UpperCamelCase : Optional[Any] = count_inversions_recursive(_lowerCAmelCase ) UpperCamelCase , UpperCamelCase : List[str] = _count_cross_inversions(_lowerCAmelCase , _lowerCAmelCase ) UpperCamelCase : Optional[int] = inversion_p + inversions_q + cross_inversions return c, num_inversions def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> int: UpperCamelCase : Dict = [] UpperCamelCase : int = 0 while i < len(_lowerCAmelCase ) and j < len(_lowerCAmelCase ): if p[i] > q[j]: # if P[1] > Q[j], then P[k] > Q[k] for all i < k <= len(P) # These are all inversions. The claim emerges from the # property that P is sorted. num_inversion += len(_lowerCAmelCase ) - i r.append(q[j] ) j += 1 else: r.append(p[i] ) i += 1 if i < len(_lowerCAmelCase ): r.extend(p[i:] ) else: r.extend(q[j:] ) return r, num_inversion def A_ ( ) -> Dict: UpperCamelCase : Tuple = [10, 2, 1, 5, 5, 2, 11] # this arr has 8 inversions: # (10, 2), (10, 1), (10, 5), (10, 5), (10, 2), (2, 1), (5, 2), (5, 2) UpperCamelCase : Optional[int] = count_inversions_bf(_lowerCAmelCase ) UpperCamelCase , UpperCamelCase : str = count_inversions_recursive(_lowerCAmelCase ) assert num_inversions_bf == num_inversions_recursive == 8 print("number of inversions = " , _lowerCAmelCase ) # testing an array with zero inversion (a sorted arr_1) arr_a.sort() UpperCamelCase : int = count_inversions_bf(_lowerCAmelCase ) UpperCamelCase , UpperCamelCase : List[Any] = count_inversions_recursive(_lowerCAmelCase ) assert num_inversions_bf == num_inversions_recursive == 0 print("number of inversions = " , _lowerCAmelCase ) # an empty list should also have zero inversions UpperCamelCase : List[str] = [] UpperCamelCase : Any = count_inversions_bf(_lowerCAmelCase ) UpperCamelCase , UpperCamelCase : Tuple = count_inversions_recursive(_lowerCAmelCase ) assert num_inversions_bf == num_inversions_recursive == 0 print("number of inversions = " , _lowerCAmelCase ) if __name__ == "__main__": main()
52
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCamelCase : Optional[int] = {"""configuration_mmbt""": ["""MMBTConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : List[Any] = ["""MMBTForClassification""", """MMBTModel""", """ModalEmbeddings"""] if TYPE_CHECKING: from .configuration_mmbt import MMBTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mmbt import MMBTForClassification, MMBTModel, ModalEmbeddings else: import sys __lowerCamelCase : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
52
1
"""simple docstring""" from __future__ import annotations def snake_case__ ( __lowerCamelCase : int ): """simple docstring""" lowerCamelCase__ : List[Any] =2 lowerCamelCase__ : Any =[] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(__lowerCamelCase ) if n > 1: factors.append(__lowerCamelCase ) return factors if __name__ == "__main__": import doctest doctest.testmod()
272
"""simple docstring""" from ..utils import DummyObject, requires_backends class __SCREAMING_SNAKE_CASE ( metaclass=lowerCAmelCase_ ): '''simple docstring''' _a = ['torch', 'torchsde'] def __init__( self : Union[str, Any], *lowerCamelCase : str, **lowerCamelCase : int )-> Tuple: requires_backends(self, ['''torch''', '''torchsde'''] ) @classmethod def snake_case ( cls : List[str], *lowerCamelCase : Optional[Any], **lowerCamelCase : Dict )-> str: requires_backends(cls, ['''torch''', '''torchsde'''] ) @classmethod def snake_case ( cls : Tuple, *lowerCamelCase : Dict, **lowerCamelCase : Tuple )-> List[str]: requires_backends(cls, ['''torch''', '''torchsde'''] )
272
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase_ : Dict = { '''configuration_rembert''': ['''REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''RemBertConfig''', '''RemBertOnnxConfig'''] } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Dict = ['''RemBertTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : int = ['''RemBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Union[str, Any] = [ '''REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''RemBertForCausalLM''', '''RemBertForMaskedLM''', '''RemBertForMultipleChoice''', '''RemBertForQuestionAnswering''', '''RemBertForSequenceClassification''', '''RemBertForTokenClassification''', '''RemBertLayer''', '''RemBertModel''', '''RemBertPreTrainedModel''', '''load_tf_weights_in_rembert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Optional[int] = [ '''TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFRemBertForCausalLM''', '''TFRemBertForMaskedLM''', '''TFRemBertForMultipleChoice''', '''TFRemBertForQuestionAnswering''', '''TFRemBertForSequenceClassification''', '''TFRemBertForTokenClassification''', '''TFRemBertLayer''', '''TFRemBertModel''', '''TFRemBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_rembert import REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RemBertConfig, RemBertOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert import RemBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert_fast import RemBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rembert import ( REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, RemBertForCausalLM, RemBertForMaskedLM, RemBertForMultipleChoice, RemBertForQuestionAnswering, RemBertForSequenceClassification, RemBertForTokenClassification, RemBertLayer, RemBertModel, RemBertPreTrainedModel, load_tf_weights_in_rembert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rembert import ( TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFRemBertForCausalLM, TFRemBertForMaskedLM, TFRemBertForMultipleChoice, TFRemBertForQuestionAnswering, TFRemBertForSequenceClassification, TFRemBertForTokenClassification, TFRemBertLayer, TFRemBertModel, TFRemBertPreTrainedModel, ) else: import sys UpperCAmelCase_ : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
38
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class A__ : def __init__( self : Any , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any]=13 , _UpperCAmelCase : List[str]=7 , _UpperCAmelCase : List[str]=True , _UpperCAmelCase : str=True , _UpperCAmelCase : Union[str, Any]=True , _UpperCAmelCase : List[Any]=True , _UpperCAmelCase : Optional[Any]=99 , _UpperCAmelCase : Dict=32 , _UpperCAmelCase : List[str]=2 , _UpperCAmelCase : Union[str, Any]=4 , _UpperCAmelCase : Optional[int]=37 , _UpperCAmelCase : Union[str, Any]="gelu" , _UpperCAmelCase : str=0.1 , _UpperCAmelCase : Dict=0.1 , _UpperCAmelCase : str=5_12 , _UpperCAmelCase : Optional[int]=16 , _UpperCAmelCase : Optional[int]=2 , _UpperCAmelCase : List[str]=0.02 , _UpperCAmelCase : Optional[int]=3 , _UpperCAmelCase : Any=4 , _UpperCAmelCase : List[Any]=None , ) -> Union[str, Any]: """simple docstring""" __lowercase = parent __lowercase = 13 __lowercase = 7 __lowercase = True __lowercase = True __lowercase = True __lowercase = True __lowercase = 99 __lowercase = 3_84 __lowercase = 2 __lowercase = 4 __lowercase = 37 __lowercase = 'gelu' __lowercase = 0.1 __lowercase = 0.1 __lowercase = 5_12 __lowercase = 16 __lowercase = 2 __lowercase = 0.02 __lowercase = 3 __lowercase = 4 __lowercase = 1_28 __lowercase = 2 __lowercase = 9 __lowercase = 1 __lowercase = None def a__ ( self : Dict ) -> List[Any]: """simple docstring""" __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase = ids_tensor([self.batch_size] , self.num_choices ) __lowercase = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=_UpperCAmelCase , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def a__ ( self : Any , _UpperCAmelCase : Dict , _UpperCAmelCase : str , _UpperCAmelCase : Tuple , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Any , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : int ) -> List[Any]: """simple docstring""" __lowercase = TFConvBertModel(config=_UpperCAmelCase ) __lowercase = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} __lowercase = [input_ids, input_mask] __lowercase = model(_UpperCAmelCase ) __lowercase = model(_UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self : int , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Tuple , _UpperCAmelCase : int , _UpperCAmelCase : Union[str, Any] ) -> str: """simple docstring""" __lowercase = TFConvBertForMaskedLM(config=_UpperCAmelCase ) __lowercase = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } __lowercase = model(_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a__ ( self : str , _UpperCAmelCase : int , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Any , _UpperCAmelCase : Any , _UpperCAmelCase : str ) -> Dict: """simple docstring""" __lowercase = self.num_labels __lowercase = TFConvBertForSequenceClassification(config=_UpperCAmelCase ) __lowercase = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } __lowercase = model(_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a__ ( self : Union[str, Any] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Any , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : int , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : List[str] ) -> Union[str, Any]: """simple docstring""" __lowercase = self.num_choices __lowercase = TFConvBertForMultipleChoice(config=_UpperCAmelCase ) __lowercase = tf.tile(tf.expand_dims(_UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(_UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(_UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) __lowercase = { 'input_ids': multiple_choice_inputs_ids, 'attention_mask': multiple_choice_input_mask, 'token_type_ids': multiple_choice_token_type_ids, } __lowercase = model(_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def a__ ( self : Dict , _UpperCAmelCase : str , _UpperCAmelCase : int , _UpperCAmelCase : Tuple , _UpperCAmelCase : int , _UpperCAmelCase : str , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Optional[Any] ) -> int: """simple docstring""" __lowercase = self.num_labels __lowercase = TFConvBertForTokenClassification(config=_UpperCAmelCase ) __lowercase = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } __lowercase = model(_UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def a__ ( self : Tuple , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : List[str] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Any , _UpperCAmelCase : str , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : List[Any] ) -> Any: """simple docstring""" __lowercase = TFConvBertForQuestionAnswering(config=_UpperCAmelCase ) __lowercase = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } __lowercase = model(_UpperCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def a__ ( self : int ) -> Optional[int]: """simple docstring""" __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class A__ ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): lowerCAmelCase__ : List[str] = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) lowerCAmelCase__ : List[str] = ( { "feature-extraction": TFConvBertModel, "fill-mask": TFConvBertForMaskedLM, "question-answering": TFConvBertForQuestionAnswering, "text-classification": TFConvBertForSequenceClassification, "token-classification": TFConvBertForTokenClassification, "zero-shot": TFConvBertForSequenceClassification, } if is_tf_available() else {} ) lowerCAmelCase__ : List[str] = False lowerCAmelCase__ : int = False lowerCAmelCase__ : List[str] = False def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" __lowercase = TFConvBertModelTester(self ) __lowercase = ConfigTester(self , config_class=_UpperCAmelCase , hidden_size=37 ) def a__ ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" self.config_tester.run_common_tests() def a__ ( self : Any ) -> Dict: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_UpperCAmelCase ) def a__ ( self : int ) -> str: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_UpperCAmelCase ) def a__ ( self : List[str] ) -> int: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*_UpperCAmelCase ) def a__ ( self : Any ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_UpperCAmelCase ) def a__ ( self : List[str] ) -> List[str]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_UpperCAmelCase ) def a__ ( self : List[Any] ) -> Optional[int]: """simple docstring""" __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_UpperCAmelCase ) @slow def a__ ( self : List[str] ) -> Optional[int]: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = True __lowercase = True if hasattr(_UpperCAmelCase , 'use_cache' ): __lowercase = True __lowercase = getattr(self.model_tester , 'encoder_seq_length' , self.model_tester.seq_length ) __lowercase = getattr(self.model_tester , 'key_length' , _UpperCAmelCase ) for model_class in self.all_model_classes: __lowercase = self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) __lowercase = model_class(_UpperCAmelCase ) __lowercase = len(model(_UpperCAmelCase ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_UpperCAmelCase , saved_model=_UpperCAmelCase ) __lowercase = os.path.join(_UpperCAmelCase , 'saved_model' , '1' ) __lowercase = tf.keras.models.load_model(_UpperCAmelCase ) __lowercase = model(_UpperCAmelCase ) if self.is_encoder_decoder: __lowercase = outputs['encoder_hidden_states'] __lowercase = outputs['encoder_attentions'] else: __lowercase = outputs['hidden_states'] __lowercase = outputs['attentions'] self.assertEqual(len(_UpperCAmelCase ) , _UpperCAmelCase ) __lowercase = getattr( self.model_tester , 'expected_num_hidden_layers' , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(_UpperCAmelCase ) , _UpperCAmelCase ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(_UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def a__ ( self : List[str] ) -> Dict: """simple docstring""" __lowercase = TFConvBertModel.from_pretrained('YituTech/conv-bert-base' ) self.assertIsNotNone(_UpperCAmelCase ) def a__ ( self : Tuple ) -> Tuple: """simple docstring""" __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = True __lowercase = getattr(self.model_tester , 'decoder_seq_length' , self.model_tester.seq_length ) __lowercase = getattr(self.model_tester , 'encoder_seq_length' , self.model_tester.seq_length ) __lowercase = getattr(self.model_tester , 'key_length' , _UpperCAmelCase ) __lowercase = getattr(self.model_tester , 'key_length' , _UpperCAmelCase ) def check_decoder_attentions_output(_UpperCAmelCase : int ): __lowercase = len(_UpperCAmelCase ) self.assertEqual(out_len % 2 , 0 ) __lowercase = outputs.decoder_attentions self.assertEqual(len(_UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(_UpperCAmelCase : Union[str, Any] ): __lowercase = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(_UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: __lowercase = True __lowercase = False __lowercase = model_class(_UpperCAmelCase ) __lowercase = model(self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) ) __lowercase = len(_UpperCAmelCase ) self.assertEqual(config.output_hidden_states , _UpperCAmelCase ) check_encoder_attentions_output(_UpperCAmelCase ) if self.is_encoder_decoder: __lowercase = model_class(_UpperCAmelCase ) __lowercase = model(self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) ) self.assertEqual(config.output_hidden_states , _UpperCAmelCase ) check_decoder_attentions_output(_UpperCAmelCase ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] __lowercase = True __lowercase = model_class(_UpperCAmelCase ) __lowercase = model(self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) ) self.assertEqual(config.output_hidden_states , _UpperCAmelCase ) check_encoder_attentions_output(_UpperCAmelCase ) # Check attention is always last and order is fine __lowercase = True __lowercase = True __lowercase = model_class(_UpperCAmelCase ) __lowercase = model(self._prepare_for_class(_UpperCAmelCase , _UpperCAmelCase ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(_UpperCAmelCase ) ) self.assertEqual(model.config.output_hidden_states , _UpperCAmelCase ) check_encoder_attentions_output(_UpperCAmelCase ) @require_tf class A__ ( unittest.TestCase ): @slow def a__ ( self : Dict ) -> Union[str, Any]: """simple docstring""" __lowercase = TFConvBertModel.from_pretrained('YituTech/conv-bert-base' ) __lowercase = tf.constant([[0, 1, 2, 3, 4, 5]] ) __lowercase = model(_UpperCAmelCase )[0] __lowercase = [1, 6, 7_68] self.assertEqual(output.shape , _UpperCAmelCase ) __lowercase = tf.constant( [ [ [-0.03_475_493, -0.4_686_034, -0.30_638_832], [0.22_637_248, -0.26_988_646, -0.7_423_424], [0.10_324_868, -0.45_013_508, -0.58_280_784], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , _UpperCAmelCase , atol=1e-4 )
325
0
'''simple docstring''' from copy import deepcopy class SCREAMING_SNAKE_CASE__ : """simple docstring""" def __init__( self : Tuple , UpperCAmelCase_ : list[int] | None = None , UpperCAmelCase_ : int | None = None ): """simple docstring""" if arr is None and size is not None: __UpperCAmelCase : Tuple = size __UpperCAmelCase : Optional[Any] = [0] * size elif arr is not None: self.init(UpperCAmelCase_ ) else: raise ValueError("Either arr or size must be specified" ) def lowerCamelCase_ ( self : List[Any] , UpperCAmelCase_ : list[int] ): """simple docstring""" __UpperCAmelCase : Optional[Any] = len(UpperCAmelCase_ ) __UpperCAmelCase : Tuple = deepcopy(UpperCAmelCase_ ) for i in range(1 , self.size ): __UpperCAmelCase : List[Any] = self.next_(UpperCAmelCase_ ) if j < self.size: self.tree[j] += self.tree[i] def lowerCamelCase_ ( self : Tuple ): """simple docstring""" __UpperCAmelCase : List[Any] = self.tree[:] for i in range(self.size - 1 , 0 , -1 ): __UpperCAmelCase : str = self.next_(UpperCAmelCase_ ) if j < self.size: arr[j] -= arr[i] return arr @staticmethod def lowerCamelCase_ ( UpperCAmelCase_ : int ): """simple docstring""" return index + (index & (-index)) @staticmethod def lowerCamelCase_ ( UpperCAmelCase_ : int ): """simple docstring""" return index - (index & (-index)) def lowerCamelCase_ ( self : str , UpperCAmelCase_ : int , UpperCAmelCase_ : int ): """simple docstring""" if index == 0: self.tree[0] += value return while index < self.size: self.tree[index] += value __UpperCAmelCase : Optional[int] = self.next_(UpperCAmelCase_ ) def lowerCamelCase_ ( self : int , UpperCAmelCase_ : int , UpperCAmelCase_ : int ): """simple docstring""" self.add(UpperCAmelCase_ , value - self.get(UpperCAmelCase_ ) ) def lowerCamelCase_ ( self : List[Any] , UpperCAmelCase_ : int ): """simple docstring""" if right == 0: return 0 __UpperCAmelCase : Tuple = self.tree[0] right -= 1 # make right inclusive while right > 0: result += self.tree[right] __UpperCAmelCase : str = self.prev(UpperCAmelCase_ ) return result def lowerCamelCase_ ( self : Union[str, Any] , UpperCAmelCase_ : int , UpperCAmelCase_ : int ): """simple docstring""" return self.prefix(UpperCAmelCase_ ) - self.prefix(UpperCAmelCase_ ) def lowerCamelCase_ ( self : Optional[Any] , UpperCAmelCase_ : int ): """simple docstring""" return self.query(UpperCAmelCase_ , index + 1 ) def lowerCamelCase_ ( self : Any , UpperCAmelCase_ : int ): """simple docstring""" value -= self.tree[0] if value < 0: return -1 __UpperCAmelCase : int = 1 # Largest power of 2 <= size while j * 2 < self.size: j *= 2 __UpperCAmelCase : List[str] = 0 while j > 0: if i + j < self.size and self.tree[i + j] <= value: value -= self.tree[i + j] i += j j //= 2 return i if __name__ == "__main__": import doctest doctest.testmod()
37
'''simple docstring''' from __future__ import annotations def __UpperCamelCase ( _UpperCAmelCase ): if not nums: raise ValueError("List is empty" ) return sum(_UpperCAmelCase ) / len(_UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
37
1