code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase : int = logging.get_logger(__name__) UpperCamelCase : Optional[int] = { '''funnel-transformer/small''': '''https://huggingface.co/funnel-transformer/small/resolve/main/config.json''', '''funnel-transformer/small-base''': '''https://huggingface.co/funnel-transformer/small-base/resolve/main/config.json''', '''funnel-transformer/medium''': '''https://huggingface.co/funnel-transformer/medium/resolve/main/config.json''', '''funnel-transformer/medium-base''': '''https://huggingface.co/funnel-transformer/medium-base/resolve/main/config.json''', '''funnel-transformer/intermediate''': ( '''https://huggingface.co/funnel-transformer/intermediate/resolve/main/config.json''' ), '''funnel-transformer/intermediate-base''': ( '''https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/config.json''' ), '''funnel-transformer/large''': '''https://huggingface.co/funnel-transformer/large/resolve/main/config.json''', '''funnel-transformer/large-base''': '''https://huggingface.co/funnel-transformer/large-base/resolve/main/config.json''', '''funnel-transformer/xlarge''': '''https://huggingface.co/funnel-transformer/xlarge/resolve/main/config.json''', '''funnel-transformer/xlarge-base''': '''https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/config.json''', } class __lowerCAmelCase ( UpperCAmelCase__ ): lowercase = "funnel" lowercase = { "hidden_size": "d_model", "num_attention_heads": "n_head", } def __init__( self , __UpperCAmelCase=3_0522 , __UpperCAmelCase=[4, 4, 4] , __UpperCAmelCase=None , __UpperCAmelCase=2 , __UpperCAmelCase=768 , __UpperCAmelCase=12 , __UpperCAmelCase=64 , __UpperCAmelCase=3072 , __UpperCAmelCase="gelu_new" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.1 , __UpperCAmelCase=None , __UpperCAmelCase=1E-9 , __UpperCAmelCase="mean" , __UpperCAmelCase="relative_shift" , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=True , **__UpperCAmelCase , ): '''simple docstring''' __UpperCamelCase = vocab_size __UpperCamelCase = block_sizes __UpperCamelCase = [1] * len(__UpperCAmelCase ) if block_repeats is None else block_repeats assert len(__UpperCAmelCase ) == len( self.block_repeats ), "`block_sizes` and `block_repeats` should have the same length." __UpperCamelCase = num_decoder_layers __UpperCamelCase = d_model __UpperCamelCase = n_head __UpperCamelCase = d_head __UpperCamelCase = d_inner __UpperCamelCase = hidden_act __UpperCamelCase = hidden_dropout __UpperCamelCase = attention_dropout __UpperCamelCase = activation_dropout __UpperCamelCase = initializer_range __UpperCamelCase = initializer_std __UpperCamelCase = layer_norm_eps assert pooling_type in [ "mean", "max", ], F'Got {pooling_type} for `pooling_type` but only \'mean\' and \'max\' are supported.' __UpperCamelCase = pooling_type assert attention_type in [ "relative_shift", "factorized", ], F'Got {attention_type} for `attention_type` but only \'relative_shift\' and \'factorized\' are supported.' __UpperCamelCase = attention_type __UpperCamelCase = separate_cls __UpperCamelCase = truncate_seq __UpperCamelCase = pool_q_only super().__init__(**__UpperCAmelCase ) @property def UpperCAmelCase ( self ): '''simple docstring''' return sum(self.block_sizes ) @num_hidden_layers.setter def UpperCAmelCase ( self , __UpperCAmelCase ): '''simple docstring''' raise NotImplementedError( 'This model does not support the setting of `num_hidden_layers`. Please set `block_sizes`.' ) @property def UpperCAmelCase ( self ): '''simple docstring''' return len(self.block_sizes ) @num_blocks.setter def UpperCAmelCase ( self , __UpperCAmelCase ): '''simple docstring''' raise NotImplementedError('This model does not support the setting of `num_blocks`. Please set `block_sizes`.' )
316
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A__: List[str] = logging.get_logger(__name__) A__: Union[str, Any] = { '''facebook/data2vec-text-base''': '''https://huggingface.co/data2vec/resolve/main/config.json''', } class A__ ( UpperCAmelCase__ ): __UpperCamelCase : int = "data2vec-text" def __init__( self :str , SCREAMING_SNAKE_CASE :Optional[Any]=3_0_5_2_2 , SCREAMING_SNAKE_CASE :Any=7_6_8 , SCREAMING_SNAKE_CASE :List[Any]=1_2 , SCREAMING_SNAKE_CASE :List[str]=1_2 , SCREAMING_SNAKE_CASE :Dict=3_0_7_2 , SCREAMING_SNAKE_CASE :List[str]="gelu" , SCREAMING_SNAKE_CASE :Any=0.1 , SCREAMING_SNAKE_CASE :List[str]=0.1 , SCREAMING_SNAKE_CASE :int=5_1_2 , SCREAMING_SNAKE_CASE :int=2 , SCREAMING_SNAKE_CASE :Union[str, Any]=0.02 , SCREAMING_SNAKE_CASE :Dict=1e-12 , SCREAMING_SNAKE_CASE :int=1 , SCREAMING_SNAKE_CASE :Dict=0 , SCREAMING_SNAKE_CASE :List[Any]=2 , SCREAMING_SNAKE_CASE :str="absolute" , SCREAMING_SNAKE_CASE :Tuple=True , SCREAMING_SNAKE_CASE :Union[str, Any]=None , **SCREAMING_SNAKE_CASE :Union[str, Any] , ) -> List[str]: '''simple docstring''' super().__init__(pad_token_id=SCREAMING_SNAKE_CASE , bos_token_id=SCREAMING_SNAKE_CASE , eos_token_id=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) _a : Optional[Any] =vocab_size _a : Optional[Any] =hidden_size _a : Any =num_hidden_layers _a : List[str] =num_attention_heads _a : Union[str, Any] =hidden_act _a : Any =intermediate_size _a : str =hidden_dropout_prob _a : Optional[Any] =attention_probs_dropout_prob _a : Optional[Any] =max_position_embeddings _a : Union[str, Any] =type_vocab_size _a : Tuple =initializer_range _a : Optional[int] =layer_norm_eps _a : Tuple =position_embedding_type _a : int =use_cache _a : List[str] =classifier_dropout class A__ ( UpperCAmelCase__ ): @property def __UpperCAmelCase ( self :int ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": _a : Tuple ={0: """batch""", 1: """choice""", 2: """sequence"""} else: _a : List[Any] ={0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
276
0
"""simple docstring""" import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCAmelCase_ : def __init__( self : Union[str, Any] , snake_case_ : Union[str, Any] , snake_case_ : Optional[Any]=13 , snake_case_ : Tuple=30 , snake_case_ : Optional[int]=2 , snake_case_ : List[str]=3 , snake_case_ : Optional[Any]=True , snake_case_ : Union[str, Any]=True , snake_case_ : Optional[Any]=32 , snake_case_ : Union[str, Any]=5 , snake_case_ : List[Any]=4 , snake_case_ : str=37 , snake_case_ : List[Any]="gelu" , snake_case_ : str=0.1 , snake_case_ : int=0.1 , snake_case_ : Union[str, Any]=10 , snake_case_ : int=0.02 , snake_case_ : Union[str, Any]=None , snake_case_ : Union[str, Any]=2 , ) -> Optional[Any]: '''simple docstring''' A__ = parent A__ = batch_size A__ = image_size A__ = patch_size A__ = num_channels A__ = is_training A__ = use_labels A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = hidden_act A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = type_sequence_label_size A__ = initializer_range A__ = scope A__ = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) A__ = (image_size // patch_size) ** 2 A__ = num_patches + 1 def __magic_name__ ( self : str ) -> Any: '''simple docstring''' A__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A__ = None if self.use_labels: A__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ = self.get_config() return config, pixel_values, labels def __magic_name__ ( self : Dict ) -> Optional[int]: '''simple docstring''' return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case_ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __magic_name__ ( self : int , snake_case_ : Any , snake_case_ : Tuple , snake_case_ : Optional[int] ) -> List[Any]: '''simple docstring''' A__ = ViTModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() A__ = model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __magic_name__ ( self : str , snake_case_ : List[str] , snake_case_ : int , snake_case_ : List[Any] ) -> str: '''simple docstring''' A__ = ViTForMaskedImageModeling(config=snake_case_ ) model.to(snake_case_ ) model.eval() A__ = model(snake_case_ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images A__ = 1 A__ = ViTForMaskedImageModeling(snake_case_ ) model.to(snake_case_ ) model.eval() A__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A__ = model(snake_case_ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __magic_name__ ( self : Any , snake_case_ : Any , snake_case_ : Any , snake_case_ : str ) -> int: '''simple docstring''' A__ = self.type_sequence_label_size A__ = ViTForImageClassification(snake_case_ ) model.to(snake_case_ ) model.eval() A__ = model(snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images A__ = 1 A__ = ViTForImageClassification(snake_case_ ) model.to(snake_case_ ) model.eval() A__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A__ = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __magic_name__ ( self : int ) -> List[str]: '''simple docstring''' A__ = self.prepare_config_and_inputs() ( ( A__ ), ( A__ ), ( A__ ), ) = config_and_inputs A__ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class UpperCAmelCase_ ( A_, A_, unittest.TestCase ): lowercase__ = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) lowercase__ = ( {'''feature-extraction''': ViTModel, '''image-classification''': ViTForImageClassification} if is_torch_available() else {} ) lowercase__ = True lowercase__ = False lowercase__ = False lowercase__ = False def __magic_name__ ( self : List[str] ) -> Optional[Any]: '''simple docstring''' A__ = ViTModelTester(self ) A__ = ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_ , hidden_size=37 ) def __magic_name__ ( self : Optional[int] ) -> Dict: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="ViT does not use inputs_embeds" ) def __magic_name__ ( self : List[Any] ) -> Any: '''simple docstring''' pass def __magic_name__ ( self : Tuple ) -> Any: '''simple docstring''' A__, A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ = model_class(snake_case_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) A__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case_ , nn.Linear ) ) def __magic_name__ ( self : int ) -> Any: '''simple docstring''' A__, A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ = model_class(snake_case_ ) A__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ = [*signature.parameters.keys()] A__ = ["pixel_values"] self.assertListEqual(arg_names[:1] , snake_case_ ) def __magic_name__ ( self : List[str] ) -> List[Any]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def __magic_name__ ( self : Optional[Any] ) -> Tuple: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*snake_case_ ) def __magic_name__ ( self : List[str] ) -> List[Any]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case_ ) @slow def __magic_name__ ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ = ViTModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _SCREAMING_SNAKE_CASE ( ) -> Any: A__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class UpperCAmelCase_ ( unittest.TestCase ): @cached_property def __magic_name__ ( self : Any ) -> Dict: '''simple docstring''' return ViTImageProcessor.from_pretrained("google/vit-base-patch16-224" ) if is_vision_available() else None @slow def __magic_name__ ( self : str ) -> int: '''simple docstring''' A__ = ViTForImageClassification.from_pretrained("google/vit-base-patch16-224" ).to(snake_case_ ) A__ = self.default_image_processor A__ = prepare_img() A__ = image_processor(images=snake_case_ , return_tensors="pt" ).to(snake_case_ ) # forward pass with torch.no_grad(): A__ = model(**snake_case_ ) # verify the logits A__ = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , snake_case_ ) A__ = torch.tensor([-0.2744, 0.8215, -0.0836] ).to(snake_case_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , snake_case_ , atol=1e-4 ) ) @slow def __magic_name__ ( self : Dict ) -> Any: '''simple docstring''' A__ = ViTModel.from_pretrained("facebook/dino-vits8" ).to(snake_case_ ) A__ = ViTImageProcessor.from_pretrained("facebook/dino-vits8" , size=480 ) A__ = prepare_img() A__ = image_processor(images=snake_case_ , return_tensors="pt" ) A__ = inputs.pixel_values.to(snake_case_ ) # forward pass with torch.no_grad(): A__ = model(snake_case_ , interpolate_pos_encoding=snake_case_ ) # verify the logits A__ = torch.Size((1, 3_601, 384) ) self.assertEqual(outputs.last_hidden_state.shape , snake_case_ ) A__ = torch.tensor( [[4.2340, 4.3906, -6.6692], [4.5463, 1.8928, -6.7257], [4.4429, 0.8496, -5.8585]] ).to(snake_case_ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , snake_case_ , atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def __magic_name__ ( self : str ) -> Any: '''simple docstring''' A__ = ViTModel.from_pretrained("facebook/dino-vits8" , torch_dtype=torch.floataa , device_map="auto" ) A__ = self.default_image_processor A__ = prepare_img() A__ = image_processor(images=snake_case_ , return_tensors="pt" ) A__ = inputs.pixel_values.to(snake_case_ ) # forward pass to make sure inference works in fp16 with torch.no_grad(): A__ = model(snake_case_ )
230
"""simple docstring""" from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class UpperCAmelCase_ ( A_ ): lowercase__ = ['''image_processor''', '''tokenizer'''] lowercase__ = '''AutoImageProcessor''' lowercase__ = '''AutoTokenizer''' def __init__( self : str , snake_case_ : Dict , snake_case_ : List[str] ) -> str: '''simple docstring''' super().__init__(snake_case_ , snake_case_ ) A__ = self.image_processor def __call__( self : int , snake_case_ : Any=None , snake_case_ : Any=None , snake_case_ : Union[str, Any]=None , **snake_case_ : Optional[int] ) -> Optional[int]: '''simple docstring''' if text is None and images is None: raise ValueError("You have to specify either text or images. Both cannot be none." ) if text is not None: A__ = self.tokenizer(snake_case_ , return_tensors=snake_case_ , **snake_case_ ) if images is not None: A__ = self.image_processor(snake_case_ , return_tensors=snake_case_ , **snake_case_ ) if text is not None and images is not None: A__ = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**snake_case_ ) , tensor_type=snake_case_ ) def __magic_name__ ( self : Optional[int] , *snake_case_ : Union[str, Any] , **snake_case_ : List[Any] ) -> int: '''simple docstring''' return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def __magic_name__ ( self : List[str] , *snake_case_ : List[str] , **snake_case_ : Optional[int] ) -> Tuple: '''simple docstring''' return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property def __magic_name__ ( self : List[Any] ) -> List[Any]: '''simple docstring''' return ["input_ids", "attention_mask", "pixel_values"]
230
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def _A ( self : List[str] ): UpperCamelCase :int = tempfile.mkdtemp() # fmt: off UpperCamelCase :str = ["""""", """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """lo""", """l</w>""", """w</w>""", """r</w>""", """t</w>""", """low</w>""", """er</w>""", """lowest</w>""", """newer</w>""", """wider""", """<unk>""", """<|startoftext|>""", """<|endoftext|>"""] # fmt: on UpperCamelCase :Optional[int] = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) UpperCamelCase :List[str] = ["""#version: 0.2""", """l o""", """lo w</w>""", """e r</w>""", """"""] UpperCamelCase :List[str] = {"""unk_token""": """<unk>"""} UpperCamelCase :List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCamelCase :Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCamelCase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(__lowerCamelCase ) ) UpperCamelCase :List[Any] = { """do_resize""": True, """size""": 20, """do_center_crop""": True, """crop_size""": 18, """do_normalize""": True, """image_mean""": [0.48145466, 0.4578275, 0.40821073], """image_std""": [0.26862954, 0.26130258, 0.27577711], } UpperCamelCase :Optional[Any] = os.path.join(self.tmpdirname , __lowerCamelCase ) with open(self.image_processor_file , """w""" , encoding="""utf-8""" ) as fp: json.dump(__lowerCamelCase , __lowerCamelCase ) def _A ( self : Dict , **__lowerCamelCase : Dict ): return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _A ( self : str , **__lowerCamelCase : Any ): return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token="""!""" , **__lowerCamelCase ) def _A ( self : Tuple , **__lowerCamelCase : List[Any] ): return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCamelCase ) def _A ( self : str ): shutil.rmtree(self.tmpdirname ) def _A ( self : str ): UpperCamelCase :Any = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] UpperCamelCase :Any = [Image.fromarray(np.moveaxis(__lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def _A ( self : str ): UpperCamelCase :List[Any] = self.get_tokenizer() UpperCamelCase :List[Any] = self.get_rust_tokenizer() UpperCamelCase :Tuple = self.get_image_processor() UpperCamelCase :Any = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_slow.save_pretrained(self.tmpdirname ) UpperCamelCase :int = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCamelCase ) UpperCamelCase :Union[str, Any] = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) processor_fast.save_pretrained(self.tmpdirname ) UpperCamelCase :str = OwlViTProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , __lowerCamelCase ) self.assertIsInstance(processor_fast.tokenizer , __lowerCamelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , __lowerCamelCase ) self.assertIsInstance(processor_fast.image_processor , __lowerCamelCase ) def _A ( self : List[str] ): UpperCamelCase :Optional[int] = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCamelCase :Any = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCamelCase :str = self.get_image_processor(do_normalize=__lowerCamelCase ) UpperCamelCase :str = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCamelCase ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __lowerCamelCase ) def _A ( self : Union[str, Any] ): UpperCamelCase :Union[str, Any] = self.get_image_processor() UpperCamelCase :Dict = self.get_tokenizer() UpperCamelCase :Optional[int] = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCamelCase :Dict = self.prepare_image_inputs() UpperCamelCase :List[Any] = image_processor(__lowerCamelCase , return_tensors="""np""" ) UpperCamelCase :Dict = processor(images=__lowerCamelCase , return_tensors="""np""" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1E-2 ) def _A ( self : Dict ): UpperCamelCase :List[Any] = self.get_image_processor() UpperCamelCase :str = self.get_tokenizer() UpperCamelCase :List[Any] = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCamelCase :Optional[Any] = """lower newer""" UpperCamelCase :List[str] = processor(text=__lowerCamelCase , return_tensors="""np""" ) UpperCamelCase :List[str] = tokenizer(__lowerCamelCase , return_tensors="""np""" ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist() ) def _A ( self : Optional[int] ): UpperCamelCase :List[str] = self.get_image_processor() UpperCamelCase :Union[str, Any] = self.get_tokenizer() UpperCamelCase :Any = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCamelCase :Optional[int] = """lower newer""" UpperCamelCase :Tuple = self.prepare_image_inputs() UpperCamelCase :int = processor(text=__lowerCamelCase , images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _A ( self : Any ): UpperCamelCase :Optional[Any] = """google/owlvit-base-patch32""" UpperCamelCase :List[Any] = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCamelCase :Union[str, Any] = ["""cat""", """nasa badge"""] UpperCamelCase :str = processor(text=__lowerCamelCase ) UpperCamelCase :Optional[Any] = 16 self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _A ( self : List[str] ): UpperCamelCase :Dict = """google/owlvit-base-patch32""" UpperCamelCase :Dict = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCamelCase :Tuple = [["""cat""", """nasa badge"""], ["""person"""]] UpperCamelCase :Union[str, Any] = processor(text=__lowerCamelCase ) UpperCamelCase :int = 16 UpperCamelCase :int = len(__lowerCamelCase ) UpperCamelCase :Dict = max([len(__lowerCamelCase ) for texts in input_texts] ) self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (batch_size * num_max_text_queries, seq_length) ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _A ( self : Any ): UpperCamelCase :Optional[Any] = """google/owlvit-base-patch32""" UpperCamelCase :List[Any] = OwlViTProcessor.from_pretrained(__lowerCamelCase ) UpperCamelCase :Optional[int] = ["""cat""", """nasa badge"""] UpperCamelCase :Union[str, Any] = processor(text=__lowerCamelCase ) UpperCamelCase :str = 16 UpperCamelCase :Optional[int] = inputs["""input_ids"""] UpperCamelCase :str = [ [49_406, 2_368, 49_407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [49_406, 6_841, 11_301, 49_407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys() ) , ["""input_ids""", """attention_mask"""] ) self.assertEqual(inputs["""input_ids"""].shape , (2, seq_length) ) self.assertListEqual(list(input_ids[0] ) , predicted_ids[0] ) self.assertListEqual(list(input_ids[1] ) , predicted_ids[1] ) def _A ( self : str ): UpperCamelCase :Union[str, Any] = self.get_image_processor() UpperCamelCase :str = self.get_tokenizer() UpperCamelCase :Optional[int] = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCamelCase :List[Any] = self.prepare_image_inputs() UpperCamelCase :List[str] = self.prepare_image_inputs() UpperCamelCase :Union[str, Any] = processor(images=__lowerCamelCase , query_images=__lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["""query_pixel_values""", """pixel_values"""] ) # test if it raises when no input is passed with pytest.raises(__lowerCamelCase ): processor() def _A ( self : Dict ): UpperCamelCase :Dict = self.get_image_processor() UpperCamelCase :str = self.get_tokenizer() UpperCamelCase :str = OwlViTProcessor(tokenizer=__lowerCamelCase , image_processor=__lowerCamelCase ) UpperCamelCase :str = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCamelCase :Any = processor.batch_decode(__lowerCamelCase ) UpperCamelCase :Optional[Any] = tokenizer.batch_decode(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , __lowerCamelCase )
38
import os import tempfile import unittest from transformers import FlaubertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FlaubertForMultipleChoice, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertModel, FlaubertWithLMHeadModel, ) from transformers.models.flaubert.modeling_flaubert import FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST class _SCREAMING_SNAKE_CASE ( _a ): def __init__( self : List[str] , __lowerCamelCase : List[str] , __lowerCamelCase : Union[str, Any]=13 , __lowerCamelCase : str=7 , __lowerCamelCase : Tuple=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Any=True , __lowerCamelCase : List[Any]=True , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : Optional[int]=False , __lowerCamelCase : str=False , __lowerCamelCase : List[Any]=False , __lowerCamelCase : List[Any]=2 , __lowerCamelCase : Union[str, Any]=99 , __lowerCamelCase : Optional[Any]=0 , __lowerCamelCase : Tuple=32 , __lowerCamelCase : Any=5 , __lowerCamelCase : Optional[Any]=4 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : List[str]=512 , __lowerCamelCase : List[Any]=12 , __lowerCamelCase : int=2 , __lowerCamelCase : List[str]=0.02 , __lowerCamelCase : Union[str, Any]=3 , __lowerCamelCase : Tuple=4 , __lowerCamelCase : Optional[int]="last" , __lowerCamelCase : Optional[Any]=None , __lowerCamelCase : List[str]=None , ): UpperCamelCase :int = parent UpperCamelCase :Optional[int] = batch_size UpperCamelCase :str = seq_length UpperCamelCase :Optional[int] = is_training UpperCamelCase :Optional[int] = use_input_lengths UpperCamelCase :Union[str, Any] = use_token_type_ids UpperCamelCase :List[str] = use_labels UpperCamelCase :Dict = gelu_activation UpperCamelCase :Optional[int] = sinusoidal_embeddings UpperCamelCase :List[Any] = causal UpperCamelCase :Optional[int] = asm UpperCamelCase :List[str] = n_langs UpperCamelCase :int = vocab_size UpperCamelCase :List[Any] = n_special UpperCamelCase :List[Any] = hidden_size UpperCamelCase :List[str] = num_hidden_layers UpperCamelCase :List[Any] = num_attention_heads UpperCamelCase :Tuple = hidden_dropout_prob UpperCamelCase :List[str] = attention_probs_dropout_prob UpperCamelCase :Tuple = max_position_embeddings UpperCamelCase :List[str] = type_vocab_size UpperCamelCase :Union[str, Any] = type_sequence_label_size UpperCamelCase :int = initializer_range UpperCamelCase :List[str] = num_labels UpperCamelCase :Optional[int] = num_choices UpperCamelCase :Optional[Any] = summary_type UpperCamelCase :Tuple = use_proj UpperCamelCase :Optional[Any] = scope def _A ( self : List[str] ): UpperCamelCase :Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase :Any = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase :List[Any] = None if self.use_input_lengths: UpperCamelCase :Dict = ( ids_tensor([self.batch_size] , vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length UpperCamelCase :str = None if self.use_token_type_ids: UpperCamelCase :int = ids_tensor([self.batch_size, self.seq_length] , self.n_langs ) UpperCamelCase :Optional[int] = None UpperCamelCase :int = None UpperCamelCase :List[Any] = None if self.use_labels: UpperCamelCase :Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase :List[str] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCamelCase :List[str] = ids_tensor([self.batch_size] , 2 ).float() UpperCamelCase :List[str] = ids_tensor([self.batch_size] , self.num_choices ) UpperCamelCase :Union[str, Any] = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def _A ( self : List[Any] ): return FlaubertConfig( vocab_size=self.vocab_size , n_special=self.n_special , emb_dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , gelu_activation=self.gelu_activation , sinusoidal_embeddings=self.sinusoidal_embeddings , asm=self.asm , causal=self.causal , n_langs=self.n_langs , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , summary_type=self.summary_type , use_proj=self.use_proj , ) def _A ( self : Union[str, Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] , __lowerCamelCase : Any , __lowerCamelCase : int , ): UpperCamelCase :Tuple = FlaubertModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :int = model(__lowerCamelCase , lengths=__lowerCamelCase , langs=__lowerCamelCase ) UpperCamelCase :List[Any] = model(__lowerCamelCase , langs=__lowerCamelCase ) UpperCamelCase :int = model(__lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _A ( self : Optional[int] , __lowerCamelCase : List[str] , __lowerCamelCase : Any , __lowerCamelCase : Tuple , __lowerCamelCase : int , __lowerCamelCase : List[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] , __lowerCamelCase : Any , __lowerCamelCase : Dict , ): UpperCamelCase :Any = FlaubertWithLMHeadModel(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Dict = model(__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _A ( self : List[Any] , __lowerCamelCase : str , __lowerCamelCase : Tuple , __lowerCamelCase : List[str] , __lowerCamelCase : int , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : str , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] , ): UpperCamelCase :Any = FlaubertForQuestionAnsweringSimple(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Any = model(__lowerCamelCase ) UpperCamelCase :int = model(__lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _A ( self : str , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : Optional[Any] , __lowerCamelCase : int , __lowerCamelCase : Tuple , __lowerCamelCase : Any , __lowerCamelCase : List[str] , __lowerCamelCase : Dict , __lowerCamelCase : str , ): UpperCamelCase :str = FlaubertForQuestionAnswering(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Any = model(__lowerCamelCase ) UpperCamelCase :Optional[int] = model( __lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase , cls_index=__lowerCamelCase , is_impossible=__lowerCamelCase , p_mask=__lowerCamelCase , ) UpperCamelCase :Union[str, Any] = model( __lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase , cls_index=__lowerCamelCase , is_impossible=__lowerCamelCase , ) ((UpperCamelCase) , ) :int = result_with_labels.to_tuple() UpperCamelCase :int = model(__lowerCamelCase , start_positions=__lowerCamelCase , end_positions=__lowerCamelCase ) ((UpperCamelCase) , ) :List[Any] = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape , () ) self.parent.assertEqual(result.start_top_log_probs.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual(result.start_top_index.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual( result.end_top_log_probs.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual( result.end_top_index.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual(result.cls_logits.shape , (self.batch_size,) ) def _A ( self : List[Any] , __lowerCamelCase : Any , __lowerCamelCase : Any , __lowerCamelCase : int , __lowerCamelCase : Dict , __lowerCamelCase : Any , __lowerCamelCase : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Tuple , ): UpperCamelCase :Optional[int] = FlaubertForSequenceClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Tuple = model(__lowerCamelCase ) UpperCamelCase :List[str] = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _A ( self : Any , __lowerCamelCase : int , __lowerCamelCase : Optional[int] , __lowerCamelCase : str , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : int , __lowerCamelCase : Optional[int] , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] , ): UpperCamelCase :Dict = self.num_labels UpperCamelCase :Tuple = FlaubertForTokenClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Optional[Any] = model(__lowerCamelCase , attention_mask=__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _A ( self : Tuple , __lowerCamelCase : Tuple , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[Any] , __lowerCamelCase : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : Tuple , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] , ): UpperCamelCase :Union[str, Any] = self.num_choices UpperCamelCase :List[Any] = FlaubertForMultipleChoice(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() UpperCamelCase :Optional[Any] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase :Optional[int] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase :int = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase :Union[str, Any] = model( __lowerCamelCase , attention_mask=__lowerCamelCase , token_type_ids=__lowerCamelCase , labels=__lowerCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _A ( self : str ): UpperCamelCase :List[str] = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) :List[Any] = config_and_inputs UpperCamelCase :Union[str, Any] = { """input_ids""": input_ids, """token_type_ids""": token_type_ids, """lengths""": input_lengths, """attention_mask""": input_mask, } return config, inputs_dict @require_torch class _SCREAMING_SNAKE_CASE ( _a , _a , unittest.TestCase ): snake_case__ : Optional[int] = ( ( FlaubertModel, FlaubertWithLMHeadModel, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertForMultipleChoice, ) if is_torch_available() else () ) snake_case__ : Tuple = ( { """feature-extraction""": FlaubertModel, """fill-mask""": FlaubertWithLMHeadModel, """question-answering""": FlaubertForQuestionAnsweringSimple, """text-classification""": FlaubertForSequenceClassification, """token-classification""": FlaubertForTokenClassification, """zero-shot""": FlaubertForSequenceClassification, } if is_torch_available() else {} ) def _A ( self : int , __lowerCamelCase : List[str] , __lowerCamelCase : Tuple , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[int] ): if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith("""Fast""" ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def _A ( self : Optional[int] , __lowerCamelCase : Dict , __lowerCamelCase : Tuple , __lowerCamelCase : Tuple=False ): UpperCamelCase :Tuple = super()._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) if return_labels: if model_class.__name__ == "FlaubertForQuestionAnswering": UpperCamelCase :Tuple = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__lowerCamelCase ) UpperCamelCase :List[Any] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__lowerCamelCase ) return inputs_dict def _A ( self : str ): UpperCamelCase :List[Any] = FlaubertModelTester(self ) UpperCamelCase :Any = ConfigTester(self , config_class=__lowerCamelCase , emb_dim=37 ) def _A ( self : Optional[int] ): self.config_tester.run_common_tests() def _A ( self : List[Any] ): UpperCamelCase :List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_model(*__lowerCamelCase ) def _A ( self : Optional[int] ): UpperCamelCase :Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_lm_head(*__lowerCamelCase ) def _A ( self : List[Any] ): UpperCamelCase :Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_simple_qa(*__lowerCamelCase ) def _A ( self : Union[str, Any] ): UpperCamelCase :int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_qa(*__lowerCamelCase ) def _A ( self : Optional[Any] ): UpperCamelCase :List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_sequence_classif(*__lowerCamelCase ) def _A ( self : Tuple ): UpperCamelCase :Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_token_classif(*__lowerCamelCase ) def _A ( self : int ): UpperCamelCase :Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_multiple_choice(*__lowerCamelCase ) @slow def _A ( self : Any ): for model_name in FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase :Optional[int] = FlaubertModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) @slow @require_torch_gpu def _A ( self : Tuple ): UpperCamelCase , UpperCamelCase :Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # FlauBertForMultipleChoice behaves incorrectly in JIT environments. if model_class == FlaubertForMultipleChoice: return UpperCamelCase :Optional[Any] = True UpperCamelCase :Optional[Any] = model_class(config=__lowerCamelCase ) UpperCamelCase :str = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) UpperCamelCase :str = torch.jit.trace( __lowerCamelCase , (inputs_dict["""input_ids"""].to("""cpu""" ), inputs_dict["""attention_mask"""].to("""cpu""" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(__lowerCamelCase , os.path.join(__lowerCamelCase , """traced_model.pt""" ) ) UpperCamelCase :int = torch.jit.load(os.path.join(__lowerCamelCase , """traced_model.pt""" ) , map_location=__lowerCamelCase ) loaded(inputs_dict["""input_ids"""].to(__lowerCamelCase ) , inputs_dict["""attention_mask"""].to(__lowerCamelCase ) ) @require_torch class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _A ( self : Optional[Any] ): UpperCamelCase :Union[str, Any] = FlaubertModel.from_pretrained("""flaubert/flaubert_base_cased""" ) UpperCamelCase :Optional[Any] = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) with torch.no_grad(): UpperCamelCase :Tuple = model(__lowerCamelCase )[0] UpperCamelCase :Union[str, Any] = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , __lowerCamelCase ) UpperCamelCase :int = torch.tensor( [[[-2.6251, -1.4298, -0.0227], [-2.8510, -1.6387, 0.2258], [-2.8114, -1.1832, -0.3066]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowerCamelCase , atol=1E-4 ) )
38
1
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : Dict, UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : List[Any]=7, UpperCAmelCase__ : Any=3, UpperCAmelCase__ : int=1_8, UpperCAmelCase__ : Tuple=3_0, UpperCAmelCase__ : Optional[int]=4_0_0, UpperCAmelCase__ : Union[str, Any]=True, UpperCAmelCase__ : Optional[int]=None, UpperCAmelCase__ : Any=True, ): __lowercase = size if size is not None else {"height": 1_8, "width": 1_8} __lowercase = parent __lowercase = batch_size __lowercase = num_channels __lowercase = image_size __lowercase = min_resolution __lowercase = max_resolution __lowercase = do_resize __lowercase = size __lowercase = apply_ocr def _lowercase ( self : List[Any] ): return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class _lowerCAmelCase ( lowercase ,unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = LayoutLMvaImageProcessor if is_pytesseract_available() else None def _lowercase ( self : int ): __lowercase = LayoutLMvaImageProcessingTester(self ) @property def _lowercase ( self : Union[str, Any] ): return self.image_processor_tester.prepare_image_processor_dict() def _lowercase ( self : Optional[int] ): __lowercase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(UpperCAmelCase__, "do_resize" ) ) self.assertTrue(hasattr(UpperCAmelCase__, "size" ) ) self.assertTrue(hasattr(UpperCAmelCase__, "apply_ocr" ) ) def _lowercase ( self : List[Any] ): __lowercase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size, {"height": 1_8, "width": 1_8} ) __lowercase = self.image_processing_class.from_dict(self.image_processor_dict, size=4_2 ) self.assertEqual(image_processor.size, {"height": 4_2, "width": 4_2} ) def _lowercase ( self : Tuple ): pass def _lowercase ( self : int ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowercase = prepare_image_inputs(self.image_processor_tester, equal_resolution=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__, Image.Image ) # Test not batched input __lowercase = image_processing(image_inputs[0], return_tensors="pt" ) self.assertEqual( encoding.pixel_values.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) self.assertIsInstance(encoding.words, UpperCAmelCase__ ) self.assertIsInstance(encoding.boxes, UpperCAmelCase__ ) # Test batched __lowercase = image_processing(UpperCAmelCase__, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) def _lowercase ( self : str ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowercase = prepare_image_inputs(self.image_processor_tester, equal_resolution=UpperCAmelCase__, numpify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__, np.ndarray ) # Test not batched input __lowercase = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) # Test batched __lowercase = image_processing(UpperCAmelCase__, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) def _lowercase ( self : int ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowercase = prepare_image_inputs(self.image_processor_tester, equal_resolution=UpperCAmelCase__, torchify=UpperCAmelCase__ ) for image in image_inputs: self.assertIsInstance(UpperCAmelCase__, torch.Tensor ) # Test not batched input __lowercase = image_processing(image_inputs[0], return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) # Test batched __lowercase = image_processing(UpperCAmelCase__, return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ), ) def _lowercase ( self : Optional[int] ): # with apply_OCR = True __lowercase = LayoutLMvaImageProcessor() from datasets import load_dataset __lowercase = load_dataset("hf-internal-testing/fixtures_docvqa", split="test" ) __lowercase = Image.open(ds[0]["file"] ).convert("RGB" ) __lowercase = image_processing(UpperCAmelCase__, return_tensors="pt" ) self.assertEqual(encoding.pixel_values.shape, (1, 3, 2_2_4, 2_2_4) ) self.assertEqual(len(encoding.words ), len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 __lowercase = [["11:14", "to", "11:39", "a.m", "11:39", "to", "11:44", "a.m.", "11:44", "a.m.", "to", "12:25", "p.m.", "12:25", "to", "12:58", "p.m.", "12:58", "to", "4:00", "p.m.", "2:00", "to", "5:00", "p.m.", "Coffee", "Break", "Coffee", "will", "be", "served", "for", "men", "and", "women", "in", "the", "lobby", "adjacent", "to", "exhibit", "area.", "Please", "move", "into", "exhibit", "area.", "(Exhibits", "Open)", "TRRF", "GENERAL", "SESSION", "(PART", "|)", "Presiding:", "Lee", "A.", "Waller", "TRRF", "Vice", "President", "“Introductory", "Remarks”", "Lee", "A.", "Waller,", "TRRF", "Vice", "Presi-", "dent", "Individual", "Interviews", "with", "TRRF", "Public", "Board", "Members", "and", "Sci-", "entific", "Advisory", "Council", "Mem-", "bers", "Conducted", "by", "TRRF", "Treasurer", "Philip", "G.", "Kuehn", "to", "get", "answers", "which", "the", "public", "refrigerated", "warehousing", "industry", "is", "looking", "for.", "Plus", "questions", "from", "the", "floor.", "Dr.", "Emil", "M.", "Mrak,", "University", "of", "Cal-", "ifornia,", "Chairman,", "TRRF", "Board;", "Sam", "R.", "Cecil,", "University", "of", "Georgia", "College", "of", "Agriculture;", "Dr.", "Stanley", "Charm,", "Tufts", "University", "School", "of", "Medicine;", "Dr.", "Robert", "H.", "Cotton,", "ITT", "Continental", "Baking", "Company;", "Dr.", "Owen", "Fennema,", "University", "of", "Wis-", "consin;", "Dr.", "Robert", "E.", "Hardenburg,", "USDA.", "Questions", "and", "Answers", "Exhibits", "Open", "Capt.", "Jack", "Stoney", "Room", "TRRF", "Scientific", "Advisory", "Council", "Meeting", "Ballroom", "Foyer"]] # noqa: E231 __lowercase = [[[1_4_1, 5_7, 2_1_4, 6_9], [2_2_8, 5_8, 2_5_2, 6_9], [1_4_1, 7_5, 2_1_6, 8_8], [2_3_0, 7_9, 2_8_0, 8_8], [1_4_2, 2_6_0, 2_1_8, 2_7_3], [2_3_0, 2_6_1, 2_5_5, 2_7_3], [1_4_3, 2_7_9, 2_1_8, 2_9_0], [2_3_1, 2_8_2, 2_9_0, 2_9_1], [1_4_3, 3_4_2, 2_1_8, 3_5_4], [2_3_1, 3_4_5, 2_8_9, 3_5_5], [2_0_2, 3_6_2, 2_2_7, 3_7_3], [1_4_3, 3_7_9, 2_2_0, 3_9_2], [2_3_1, 3_8_2, 2_9_1, 3_9_4], [1_4_4, 7_1_4, 2_2_0, 7_2_6], [2_3_1, 7_1_5, 2_5_6, 7_2_6], [1_4_4, 7_3_2, 2_2_0, 7_4_5], [2_3_2, 7_3_6, 2_9_1, 7_4_7], [1_4_4, 7_6_9, 2_1_8, 7_8_2], [2_3_1, 7_7_0, 2_5_6, 7_8_2], [1_4_1, 7_8_8, 2_0_2, 8_0_1], [2_1_5, 7_9_1, 2_7_4, 8_0_4], [1_4_3, 8_2_6, 2_0_4, 8_3_8], [2_1_5, 8_2_6, 2_4_0, 8_3_8], [1_4_2, 8_4_4, 2_0_2, 8_5_7], [2_1_5, 8_4_7, 2_7_4, 8_5_9], [3_3_4, 5_7, 4_2_7, 6_9], [4_4_0, 5_7, 5_2_2, 6_9], [3_6_9, 7_5, 4_6_1, 8_8], [4_6_9, 7_5, 5_1_6, 8_8], [5_2_8, 7_6, 5_6_2, 8_8], [5_7_0, 7_6, 6_6_7, 8_8], [6_7_5, 7_5, 7_1_1, 8_7], [7_2_1, 7_9, 7_7_8, 8_8], [7_8_9, 7_5, 8_4_0, 8_8], [3_6_9, 9_7, 4_7_0, 1_0_7], [4_8_4, 9_4, 5_0_7, 1_0_6], [5_1_8, 9_4, 5_6_2, 1_0_7], [5_7_6, 9_4, 6_5_5, 1_1_0], [6_6_8, 9_4, 7_9_2, 1_0_9], [8_0_4, 9_5, 8_2_9, 1_0_7], [3_6_9, 1_1_3, 4_6_5, 1_2_5], [4_7_7, 1_1_6, 5_4_7, 1_2_5], [5_6_2, 1_1_3, 6_5_8, 1_2_5], [6_7_1, 1_1_6, 7_4_8, 1_2_5], [7_6_1, 1_1_3, 8_1_1, 1_2_5], [3_6_9, 1_3_1, 4_6_5, 1_4_3], [4_7_7, 1_3_3, 5_4_8, 1_4_3], [5_6_3, 1_3_0, 6_9_8, 1_4_5], [7_1_0, 1_3_0, 8_0_2, 1_4_6], [3_3_6, 1_7_1, 4_1_2, 1_8_3], [4_2_3, 1_7_1, 5_7_2, 1_8_3], [5_8_2, 1_7_0, 7_1_6, 1_8_4], [7_2_8, 1_7_1, 8_1_7, 1_8_7], [8_2_9, 1_7_1, 8_4_4, 1_8_6], [3_3_8, 1_9_7, 4_8_2, 2_1_2], [5_0_7, 1_9_6, 5_5_7, 2_0_9], [5_6_9, 1_9_6, 5_9_5, 2_0_8], [6_1_0, 1_9_6, 7_0_2, 2_0_9], [5_0_5, 2_1_4, 5_8_3, 2_2_6], [5_9_5, 2_1_4, 6_5_6, 2_2_7], [6_7_0, 2_1_5, 8_0_7, 2_2_7], [3_3_5, 2_5_9, 5_4_3, 2_7_4], [5_5_6, 2_5_9, 7_0_8, 2_7_2], [3_7_2, 2_7_9, 4_2_2, 2_9_1], [4_3_5, 2_7_9, 4_6_0, 2_9_1], [4_7_4, 2_7_9, 5_7_4, 2_9_2], [5_8_7, 2_7_8, 6_6_4, 2_9_1], [6_7_6, 2_7_8, 7_3_8, 2_9_1], [7_5_1, 2_7_9, 8_3_4, 2_9_1], [3_7_2, 2_9_8, 4_3_4, 3_1_0], [3_3_5, 3_4_1, 4_8_3, 3_5_4], [4_9_7, 3_4_1, 6_5_5, 3_5_4], [6_6_7, 3_4_1, 7_2_8, 3_5_4], [7_4_0, 3_4_1, 8_2_5, 3_5_4], [3_3_5, 3_6_0, 4_3_0, 3_7_2], [4_4_2, 3_6_0, 5_3_4, 3_7_2], [5_4_5, 3_5_9, 6_8_7, 3_7_2], [6_9_7, 3_6_0, 7_5_4, 3_7_2], [7_6_5, 3_6_0, 8_2_3, 3_7_3], [3_3_4, 3_7_8, 4_2_8, 3_9_1], [4_4_0, 3_7_8, 5_7_7, 3_9_4], [5_9_0, 3_7_8, 7_0_5, 3_9_1], [7_2_0, 3_7_8, 8_0_1, 3_9_1], [3_3_4, 3_9_7, 4_0_0, 4_0_9], [3_7_0, 4_1_6, 5_2_9, 4_2_9], [5_4_4, 4_1_6, 5_7_6, 4_3_2], [5_8_7, 4_1_6, 6_6_5, 4_2_8], [6_7_7, 4_1_6, 8_1_4, 4_2_9], [3_7_2, 4_3_5, 4_5_2, 4_5_0], [4_6_5, 4_3_4, 4_9_5, 4_4_7], [5_1_1, 4_3_4, 6_0_0, 4_4_7], [6_1_1, 4_3_6, 6_3_7, 4_4_7], [6_4_9, 4_3_6, 6_9_4, 4_5_1], [7_0_5, 4_3_8, 8_2_4, 4_4_7], [3_6_9, 4_5_3, 4_5_2, 4_6_6], [4_6_4, 4_5_4, 5_0_9, 4_6_6], [5_2_2, 4_5_3, 6_1_1, 4_6_9], [6_2_5, 4_5_3, 7_9_2, 4_6_9], [3_7_0, 4_7_2, 5_5_6, 4_8_8], [5_7_0, 4_7_2, 6_8_4, 4_8_7], [6_9_7, 4_7_2, 7_1_8, 4_8_5], [7_3_2, 4_7_2, 8_3_5, 4_8_8], [3_6_9, 4_9_0, 4_1_1, 5_0_3], [4_2_5, 4_9_0, 4_8_4, 5_0_3], [4_9_6, 4_9_0, 6_3_5, 5_0_6], [6_4_5, 4_9_0, 7_0_7, 5_0_3], [7_1_8, 4_9_1, 7_6_1, 5_0_3], [7_7_1, 4_9_0, 8_4_0, 5_0_3], [3_3_6, 5_1_0, 3_7_4, 5_2_1], [3_8_8, 5_1_0, 4_4_7, 5_2_2], [4_6_0, 5_1_0, 4_8_9, 5_2_1], [5_0_3, 5_1_0, 5_8_0, 5_2_2], [5_9_2, 5_0_9, 7_3_6, 5_2_5], [7_4_5, 5_0_9, 7_7_0, 5_2_2], [7_8_1, 5_0_9, 8_4_0, 5_2_2], [3_3_8, 5_2_8, 4_3_4, 5_4_1], [4_4_8, 5_2_8, 5_9_6, 5_4_1], [6_0_9, 5_2_7, 6_8_7, 5_4_0], [7_0_0, 5_2_8, 7_9_2, 5_4_1], [3_3_6, 5_4_6, 3_9_7, 5_5_9], [4_0_7, 5_4_6, 4_3_1, 5_5_9], [4_4_3, 5_4_6, 5_2_5, 5_6_0], [5_3_7, 5_4_6, 6_8_0, 5_6_2], [6_8_8, 5_4_6, 7_1_4, 5_5_9], [7_2_2, 5_4_6, 8_3_7, 5_6_2], [3_3_6, 5_6_5, 4_4_9, 5_8_1], [4_6_1, 5_6_5, 4_8_5, 5_7_7], [4_9_7, 5_6_5, 6_6_5, 5_8_1], [6_8_1, 5_6_5, 7_1_8, 5_7_7], [7_3_2, 5_6_5, 8_3_7, 5_8_0], [3_3_7, 5_8_4, 4_3_8, 5_9_7], [4_5_2, 5_8_3, 5_2_1, 5_9_6], [5_3_5, 5_8_4, 6_7_7, 5_9_9], [6_9_0, 5_8_3, 7_8_7, 5_9_6], [8_0_1, 5_8_3, 8_2_5, 5_9_6], [3_3_8, 6_0_2, 4_7_8, 6_1_5], [4_9_2, 6_0_2, 5_3_0, 6_1_4], [5_4_3, 6_0_2, 6_3_8, 6_1_5], [6_5_0, 6_0_2, 6_7_6, 6_1_4], [6_8_8, 6_0_2, 7_8_8, 6_1_5], [8_0_2, 6_0_2, 8_4_3, 6_1_4], [3_3_7, 6_2_1, 5_0_2, 6_3_3], [5_1_6, 6_2_1, 6_1_5, 6_3_7], [6_2_9, 6_2_1, 7_7_4, 6_3_6], [7_8_9, 6_2_1, 8_2_7, 6_3_3], [3_3_7, 6_3_9, 4_1_8, 6_5_2], [4_3_2, 6_4_0, 5_7_1, 6_5_3], [5_8_7, 6_3_9, 7_3_1, 6_5_5], [7_4_3, 6_3_9, 7_6_9, 6_5_2], [7_8_0, 6_3_9, 8_4_1, 6_5_2], [3_3_8, 6_5_8, 4_4_0, 6_7_3], [4_5_5, 6_5_8, 4_9_1, 6_7_0], [5_0_8, 6_5_8, 6_0_2, 6_7_1], [6_1_6, 6_5_8, 6_3_8, 6_7_0], [6_5_4, 6_5_8, 8_3_5, 6_7_4], [3_3_7, 6_7_7, 4_2_9, 6_8_9], [3_3_7, 7_1_4, 4_8_2, 7_2_6], [4_9_5, 7_1_4, 5_4_8, 7_2_6], [5_6_1, 7_1_4, 6_8_3, 7_2_6], [3_3_8, 7_7_0, 4_6_1, 7_8_2], [4_7_4, 7_6_9, 5_5_4, 7_8_5], [4_8_9, 7_8_8, 5_6_2, 8_0_3], [5_7_6, 7_8_8, 6_4_3, 8_0_1], [6_5_6, 7_8_7, 7_5_1, 8_0_4], [7_6_4, 7_8_8, 8_4_4, 8_0_1], [3_3_4, 8_2_5, 4_2_1, 8_3_8], [4_3_0, 8_2_4, 5_7_4, 8_3_8], [5_8_4, 8_2_4, 7_2_3, 8_4_1], [3_3_5, 8_4_4, 4_5_0, 8_5_7], [4_6_4, 8_4_3, 5_8_3, 8_6_0], [6_2_8, 8_6_2, 7_5_5, 8_7_5], [7_6_9, 8_6_1, 8_4_8, 8_7_8]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words, UpperCAmelCase__ ) self.assertListEqual(encoding.boxes, UpperCAmelCase__ ) # with apply_OCR = False __lowercase = LayoutLMvaImageProcessor(apply_ocr=UpperCAmelCase__ ) __lowercase = image_processing(UpperCAmelCase__, return_tensors="pt" ) self.assertEqual(encoding.pixel_values.shape, (1, 3, 2_2_4, 2_2_4) )
144
"""simple docstring""" import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration _a = 5_00_00 _a = 50_00 _a , _a = os.path.split(__file__) _a = os.path.join(RESULTS_BASEPATH, 'results', RESULTS_FILENAME.replace('.py', '.json')) @get_duration def _A ( UpperCamelCase_ : datasets.Dataset, UpperCamelCase_ : List[str]) -> List[str]: '''simple docstring''' for i in range(UpperCamelCase_): __lowercase = dataset[i] @get_duration def _A ( UpperCamelCase_ : datasets.Dataset, UpperCamelCase_ : List[Any], UpperCamelCase_ : int) -> Dict: '''simple docstring''' for i in range(0, len(UpperCamelCase_), UpperCamelCase_): __lowercase = dataset[i : i + batch_size] @get_duration def _A ( UpperCamelCase_ : datasets.Dataset, UpperCamelCase_ : Any, UpperCamelCase_ : Optional[int]) -> List[str]: '''simple docstring''' with dataset.formatted_as(type=UpperCamelCase_): for i in range(UpperCamelCase_): __lowercase = dataset[i] @get_duration def _A ( UpperCamelCase_ : datasets.Dataset, UpperCamelCase_ : str, UpperCamelCase_ : Union[str, Any], UpperCamelCase_ : Union[str, Any]) -> Dict: '''simple docstring''' with dataset.formatted_as(type=UpperCamelCase_): for i in range(0, UpperCamelCase_, UpperCamelCase_): __lowercase = dataset[i : i + batch_size] def _A ( ) -> List[str]: '''simple docstring''' __lowercase = {"num examples": SPEED_TEST_N_EXAMPLES} __lowercase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted, {"type": "pandas", "length": SMALL_TEST}), (read_formatted, {"type": "torch", "length": SMALL_TEST}), (read_formatted, {"type": "tensorflow", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1000}), ] __lowercase = [ (read, {"length": SMALL_TEST}), (read, {"length": SPEED_TEST_N_EXAMPLES}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 10}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 100}), (read_batch, {"length": SPEED_TEST_N_EXAMPLES, "batch_size": 1000}), (read_formatted, {"type": "numpy", "length": SMALL_TEST}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 10}), (read_formatted_batch, {"type": "numpy", "length": SMALL_TEST, "batch_size": 1000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print("generating dataset") __lowercase = datasets.Features( {"list": datasets.Sequence(datasets.Value("float32")), "numbers": datasets.Value("float32")}) __lowercase = generate_example_dataset( os.path.join(UpperCamelCase_, "dataset.arrow"), UpperCamelCase_, num_examples=UpperCamelCase_, seq_shapes={"list": (100,)}, ) print("first set of iterations") for func, kwargs in functions: print(func.__name__, str(UpperCamelCase_)) __lowercase = func(UpperCamelCase_, **UpperCamelCase_) print("shuffling dataset") __lowercase = dataset.shuffle() print("Second set of iterations (after shuffling") for func, kwargs in functions_shuffled: print("shuffled ", func.__name__, str(UpperCamelCase_)) __lowercase = func( UpperCamelCase_, **UpperCamelCase_) with open(UpperCamelCase_, "wb") as f: f.write(json.dumps(UpperCamelCase_).encode("utf-8")) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
144
1
'''simple docstring''' from __future__ import annotations import math def __lowerCAmelCase ( snake_case__ ): if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(snake_case__ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def __lowerCAmelCase ( snake_case__ ): __UpperCamelCase : Any = str(snake_case__ ) __UpperCamelCase : str = [n] for i in range(1 , len(snake_case__ ) ): list_nums.append(int(str_num[i:] ) ) list_nums.append(int(str_num[:-i] ) ) return list_nums def __lowerCAmelCase ( snake_case__ ): if len(str(snake_case__ ) ) > 3: if not is_prime(int(str(snake_case__ )[-3:] ) ) or not is_prime(int(str(snake_case__ )[:3] ) ): return False return True def __lowerCAmelCase ( snake_case__ = 11 ): __UpperCamelCase : list[int] = [] __UpperCamelCase : Union[str, Any] = 13 while len(snake_case__ ) != count: if validate(snake_case__ ): __UpperCamelCase : Any = list_truncated_nums(snake_case__ ) if all(is_prime(snake_case__ ) for i in list_nums ): list_truncated_primes.append(snake_case__ ) num += 2 return list_truncated_primes def __lowerCAmelCase ( ): return sum(compute_truncated_primes(11 ) ) if __name__ == "__main__": print(f'{sum(compute_truncated_primes(11)) = }')
298
'''simple docstring''' import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEmbeddings, BertLayer, BertPooler, BertPreTrainedModel, ) def __lowerCAmelCase ( snake_case__ ): __UpperCamelCase : Tuple = torch.exp(snake_case__ ) __UpperCamelCase : str = torch.sum(snake_case__ , dim=1 ) # sum of exp(x_i) __UpperCamelCase : int = torch.sum(x * exp_x , dim=1 ) # sum of x_i * exp(x_i) return torch.log(snake_case__ ) - B / A class A ( nn.Module ): '''simple docstring''' def __init__(self , _UpperCAmelCase ) -> Union[str, Any]: super().__init__() __UpperCamelCase : Any = config.output_attentions __UpperCamelCase : Dict = config.output_hidden_states __UpperCamelCase : Union[str, Any] = nn.ModuleList([BertLayer(_UpperCAmelCase ) for _ in range(config.num_hidden_layers )] ) __UpperCamelCase : Tuple = nn.ModuleList([BertHighway(_UpperCAmelCase ) for _ in range(config.num_hidden_layers )] ) __UpperCamelCase : Optional[int] = [-1 for _ in range(config.num_hidden_layers )] def a_ (self , _UpperCAmelCase ) -> int: if (type(_UpperCAmelCase ) is float) or (type(_UpperCAmelCase ) is int): for i in range(len(self.early_exit_entropy ) ): __UpperCamelCase : str = x else: __UpperCamelCase : List[Any] = x def a_ (self , _UpperCAmelCase ) -> str: __UpperCamelCase : Tuple = pooler.state_dict() for highway in self.highway: for name, param in highway.pooler.state_dict().items(): param.copy_(loaded_model[name] ) def a_ (self , _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , ) -> List[Any]: __UpperCamelCase : Optional[Any] = () __UpperCamelCase : Tuple = () __UpperCamelCase : Dict = () for i, layer_module in enumerate(self.layer ): if self.output_hidden_states: __UpperCamelCase : Tuple = all_hidden_states + (hidden_states,) __UpperCamelCase : Optional[int] = layer_module( _UpperCAmelCase , _UpperCAmelCase , head_mask[i] , _UpperCAmelCase , _UpperCAmelCase ) __UpperCamelCase : Tuple = layer_outputs[0] if self.output_attentions: __UpperCamelCase : Optional[Any] = all_attentions + (layer_outputs[1],) __UpperCamelCase : Any = (hidden_states,) if self.output_hidden_states: __UpperCamelCase : Any = current_outputs + (all_hidden_states,) if self.output_attentions: __UpperCamelCase : int = current_outputs + (all_attentions,) __UpperCamelCase : Optional[int] = self.highway[i](_UpperCAmelCase ) # logits, pooled_output if not self.training: __UpperCamelCase : Dict = highway_exit[0] __UpperCamelCase : Any = entropy(_UpperCAmelCase ) __UpperCamelCase : str = highway_exit + (highway_entropy,) # logits, hidden_states(?), entropy __UpperCamelCase : Optional[Any] = all_highway_exits + (highway_exit,) if highway_entropy < self.early_exit_entropy[i]: __UpperCamelCase : str = (highway_logits,) + current_outputs[1:] + (all_highway_exits,) raise HighwayException(_UpperCAmelCase , i + 1 ) else: __UpperCamelCase : Optional[int] = all_highway_exits + (highway_exit,) # Add last layer if self.output_hidden_states: __UpperCamelCase : int = all_hidden_states + (hidden_states,) __UpperCamelCase : Dict = (hidden_states,) if self.output_hidden_states: __UpperCamelCase : Union[str, Any] = outputs + (all_hidden_states,) if self.output_attentions: __UpperCamelCase : Optional[int] = outputs + (all_attentions,) __UpperCamelCase : List[Any] = outputs + (all_highway_exits,) return outputs # last-layer hidden state, (all hidden states), (all attentions), all highway exits @add_start_docstrings( "The Bert Model transformer with early exiting (DeeBERT). " , SCREAMING_SNAKE_CASE__ , ) class A ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def __init__(self , _UpperCAmelCase ) -> Dict: super().__init__(_UpperCAmelCase ) __UpperCamelCase : Union[str, Any] = config __UpperCamelCase : Dict = BertEmbeddings(_UpperCAmelCase ) __UpperCamelCase : Optional[Any] = DeeBertEncoder(_UpperCAmelCase ) __UpperCamelCase : str = BertPooler(_UpperCAmelCase ) self.init_weights() def a_ (self ) -> Any: self.encoder.init_highway_pooler(self.pooler ) def a_ (self ) -> Optional[int]: return self.embeddings.word_embeddings def a_ (self , _UpperCAmelCase ) -> Dict: __UpperCamelCase : int = value def a_ (self , _UpperCAmelCase ) -> Tuple: for layer, heads in heads_to_prune.items(): self.encoder.layer[layer].attention.prune_heads(_UpperCAmelCase ) @add_start_docstrings_to_model_forward(_UpperCAmelCase ) def a_ (self , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , ) -> Union[str, Any]: if input_ids is not None and inputs_embeds is not None: raise ValueError("You cannot specify both input_ids and inputs_embeds at the same time" ) elif input_ids is not None: __UpperCamelCase : Tuple = input_ids.size() elif inputs_embeds is not None: __UpperCamelCase : Optional[int] = inputs_embeds.size()[:-1] else: raise ValueError("You have to specify either input_ids or inputs_embeds" ) __UpperCamelCase : List[str] = input_ids.device if input_ids is not None else inputs_embeds.device if attention_mask is None: __UpperCamelCase : int = torch.ones(_UpperCAmelCase , device=_UpperCAmelCase ) if encoder_attention_mask is None: __UpperCamelCase : Tuple = torch.ones(_UpperCAmelCase , device=_UpperCAmelCase ) if token_type_ids is None: __UpperCamelCase : Optional[Any] = torch.zeros(_UpperCAmelCase , dtype=torch.long , device=_UpperCAmelCase ) # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] # ourselves in which case we just need to make it broadcastable to all heads. __UpperCamelCase : torch.Tensor = self.get_extended_attention_mask(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # If a 2D ou 3D attention mask is provided for the cross-attention # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] if encoder_attention_mask.dim() == 3: __UpperCamelCase : Tuple = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.dim() == 2: __UpperCamelCase : Any = encoder_attention_mask[:, None, None, :] __UpperCamelCase : List[Any] = encoder_extended_attention_mask.to( dtype=next(self.parameters() ).dtype ) # fp16 compatibility __UpperCamelCase : Dict = (1.0 - encoder_extended_attention_mask) * -10_000.0 # Prepare head mask if needed # 1.0 in head_mask indicate we keep the head # attention_probs has shape bsz x n_heads x N x N # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] __UpperCamelCase : Dict = self.get_head_mask(_UpperCAmelCase , self.config.num_hidden_layers ) __UpperCamelCase : Optional[int] = self.embeddings( input_ids=_UpperCAmelCase , position_ids=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , inputs_embeds=_UpperCAmelCase ) __UpperCamelCase : List[Any] = self.encoder( _UpperCAmelCase , attention_mask=_UpperCAmelCase , head_mask=_UpperCAmelCase , encoder_hidden_states=_UpperCAmelCase , encoder_attention_mask=_UpperCAmelCase , ) __UpperCamelCase : Union[str, Any] = encoder_outputs[0] __UpperCamelCase : Any = self.pooler(_UpperCAmelCase ) __UpperCamelCase : Union[str, Any] = ( sequence_output, pooled_output, ) + encoder_outputs[ 1: ] # add hidden_states and attentions if they are here return outputs # sequence_output, pooled_output, (hidden_states), (attentions), highway exits class A ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def __init__(self , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[Any]: __UpperCamelCase : Tuple = message __UpperCamelCase : Union[str, Any] = exit_layer # start from 1! class A ( nn.Module ): '''simple docstring''' def __init__(self , _UpperCAmelCase ) -> Dict: super().__init__() __UpperCamelCase : Union[str, Any] = BertPooler(_UpperCAmelCase ) __UpperCamelCase : int = nn.Dropout(config.hidden_dropout_prob ) __UpperCamelCase : Union[str, Any] = nn.Linear(config.hidden_size , config.num_labels ) def a_ (self , _UpperCAmelCase ) -> Any: # Pooler __UpperCamelCase : Optional[int] = encoder_outputs[0] __UpperCamelCase : str = self.pooler(_UpperCAmelCase ) # "return" pooler_output # BertModel __UpperCamelCase : Tuple = (pooler_input, pooler_output) + encoder_outputs[1:] # "return" bmodel_output # Dropout and classification __UpperCamelCase : Dict = bmodel_output[1] __UpperCamelCase : List[Any] = self.dropout(_UpperCAmelCase ) __UpperCamelCase : Any = self.classifier(_UpperCAmelCase ) return logits, pooled_output @add_start_docstrings( "Bert Model (with early exiting - DeeBERT) with a classifier on top,\n also takes care of multi-layer training. " , SCREAMING_SNAKE_CASE__ , ) class A ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def __init__(self , _UpperCAmelCase ) -> Any: super().__init__(_UpperCAmelCase ) __UpperCamelCase : List[Any] = config.num_labels __UpperCamelCase : List[Any] = config.num_hidden_layers __UpperCamelCase : Optional[int] = DeeBertModel(_UpperCAmelCase ) __UpperCamelCase : List[str] = nn.Dropout(config.hidden_dropout_prob ) __UpperCamelCase : str = nn.Linear(config.hidden_size , self.config.num_labels ) self.init_weights() @add_start_docstrings_to_model_forward(_UpperCAmelCase ) def a_ (self , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=-1 , _UpperCAmelCase=False , ) -> int: __UpperCamelCase : int = self.num_layers try: __UpperCamelCase : Tuple = self.bert( _UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , position_ids=_UpperCAmelCase , head_mask=_UpperCAmelCase , inputs_embeds=_UpperCAmelCase , ) # sequence_output, pooled_output, (hidden_states), (attentions), highway exits __UpperCamelCase : str = outputs[1] __UpperCamelCase : List[Any] = self.dropout(_UpperCAmelCase ) __UpperCamelCase : Dict = self.classifier(_UpperCAmelCase ) __UpperCamelCase : Tuple = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: __UpperCamelCase : int = e.message __UpperCamelCase : Optional[Any] = e.exit_layer __UpperCamelCase : Optional[int] = outputs[0] if not self.training: __UpperCamelCase : Optional[int] = entropy(_UpperCAmelCase ) __UpperCamelCase : Optional[Any] = [] __UpperCamelCase : Any = [] if labels is not None: if self.num_labels == 1: # We are doing regression __UpperCamelCase : List[str] = MSELoss() __UpperCamelCase : Tuple = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: __UpperCamelCase : Dict = CrossEntropyLoss() __UpperCamelCase : Any = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits __UpperCamelCase : List[Any] = [] for highway_exit in outputs[-1]: __UpperCamelCase : Union[str, Any] = highway_exit[0] if not self.training: highway_logits_all.append(_UpperCAmelCase ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression __UpperCamelCase : Union[str, Any] = MSELoss() __UpperCamelCase : str = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: __UpperCamelCase : Optional[Any] = CrossEntropyLoss() __UpperCamelCase : List[str] = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(_UpperCAmelCase ) if train_highway: __UpperCamelCase : int = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: __UpperCamelCase : Dict = (loss,) + outputs if not self.training: __UpperCamelCase : Optional[int] = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: __UpperCamelCase : int = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), (highway_exits)
298
1
"""simple docstring""" import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sentencepiece @require_tokenizers @require_flax class _lowerCamelCase ( unittest.TestCase ): @slow def snake_case_ (self ) -> Optional[Any]: UpperCamelCase = FlaxXLMRobertaModel.from_pretrained("xlm-roberta-base" ) UpperCamelCase = AutoTokenizer.from_pretrained("xlm-roberta-base" ) UpperCamelCase = "The dog is cute and lives in the garden house" UpperCamelCase = jnp.array([tokenizer.encode(__a )] ) UpperCamelCase = (1, 12, 7_68) # batch_size, sequence_length, embedding_vector_dim UpperCamelCase = jnp.array( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) UpperCamelCase = model(__a )["last_hidden_state"] self.assertEqual(output.shape , __a ) # compare the actual values for a slice of last dim self.assertTrue(jnp.allclose(output[:, :, -1] , __a , atol=1e-3 ) )
244
"""simple docstring""" import doctest from collections import deque import numpy as np class _lowerCamelCase : def __init__(self ) -> None: UpperCamelCase = [2, 1, 2, -1] UpperCamelCase = [1, 2, 3, 4] def snake_case_ (self ) -> list[float]: UpperCamelCase = len(self.first_signal ) UpperCamelCase = len(self.second_signal ) UpperCamelCase = max(__a , __a ) # create a zero matrix of max_length x max_length UpperCamelCase = [[0] * max_length for i in range(__a )] # fills the smaller signal with zeros to make both signals of same length if length_first_signal < length_second_signal: self.first_signal += [0] * (max_length - length_first_signal) elif length_first_signal > length_second_signal: self.second_signal += [0] * (max_length - length_second_signal) for i in range(__a ): UpperCamelCase = deque(self.second_signal ) rotated_signal.rotate(__a ) for j, item in enumerate(__a ): matrix[i][j] += item # multiply the matrix with the first signal UpperCamelCase = np.matmul(np.transpose(__a ) , np.transpose(self.first_signal ) ) # rounding-off to two decimal places return [round(__a , 2 ) for i in final_signal] if __name__ == "__main__": doctest.testmod()
244
1
"""simple docstring""" import os import re import shutil import sys import tempfile import unittest import black _A = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If DDPMSchedulerOutput is changed in scheduling_ddpm.py, this code needs to be manually updated. _A = """ \"\"\" Output class for the scheduler's step function output. Args: prev_sample (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)` for images): Computed sample (x_{t-1}) of previous timestep. `prev_sample` should be used as next model input in the denoising loop. pred_original_sample (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)` for images): The predicted denoised sample (x_{0}) based on the model output from the current timestep. `pred_original_sample` can be used to preview progress or for guidance. \"\"\" prev_sample: torch.FloatTensor pred_original_sample: Optional[torch.FloatTensor] = None """ class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def _a (self ): """simple docstring""" UpperCAmelCase__ : int = tempfile.mkdtemp() os.makedirs(os.path.join(self.diffusers_dir , """schedulers/""" ) ) UpperCAmelCase__ : List[Any] = self.diffusers_dir shutil.copy( os.path.join(_lowerCamelCase , """src/diffusers/schedulers/scheduling_ddpm.py""" ) , os.path.join(self.diffusers_dir , """schedulers/scheduling_ddpm.py""" ) , ) def _a (self ): """simple docstring""" UpperCAmelCase__ : List[str] = """src/diffusers""" shutil.rmtree(self.diffusers_dir ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=None ): """simple docstring""" UpperCAmelCase__ : Dict = comment + F"""\nclass {class_name}(nn.Module):\n""" + class_code if overwrite_result is not None: UpperCAmelCase__ : List[Any] = comment + F"""\nclass {class_name}(nn.Module):\n""" + overwrite_result UpperCAmelCase__ : str = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 ) UpperCAmelCase__ : Dict = black.format_str(_lowerCamelCase , mode=_lowerCamelCase ) UpperCAmelCase__ : Union[str, Any] = os.path.join(self.diffusers_dir , """new_code.py""" ) with open(_lowerCamelCase , """w""" , newline="""\n""" ) as f: f.write(_lowerCamelCase ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(_lowerCamelCase ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=_lowerCamelCase ) with open(_lowerCamelCase , """r""" ) as f: self.assertTrue(f.read() , _lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Tuple = check_copies.find_code_in_diffusers("""schedulers.scheduling_ddpm.DDPMSchedulerOutput""" ) self.assertEqual(_lowerCamelCase , _lowerCamelCase ) def _a (self ): """simple docstring""" self.check_copy_consistency( """# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput""" , """DDPMSchedulerOutput""" , REFERENCE_CODE + """\n""" , ) # With no empty line at the end self.check_copy_consistency( """# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput""" , """DDPMSchedulerOutput""" , _lowerCamelCase , ) # Copy consistency with rename self.check_copy_consistency( """# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->Test""" , """TestSchedulerOutput""" , re.sub("""DDPM""" , """Test""" , _lowerCamelCase ) , ) # Copy consistency with a really long name UpperCAmelCase__ : Any = """TestClassWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason""" self.check_copy_consistency( F"""# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->{long_class_name}""" , F"""{long_class_name}SchedulerOutput""" , re.sub("""Bert""" , _lowerCamelCase , _lowerCamelCase ) , ) # Copy consistency with overwrite self.check_copy_consistency( """# Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->Test""" , """TestSchedulerOutput""" , _lowerCamelCase , overwrite_result=re.sub("""DDPM""" , """Test""" , _lowerCamelCase ) , )
171
"""simple docstring""" from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_torch_tensor, is_vision_available, logging if is_vision_available(): import PIL if is_torch_available(): import torch _A = logging.get_logger(__name__) class lowerCamelCase ( lowerCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE = ['pixel_values'] def __init__(self , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = PILImageResampling.BILINEAR , _lowerCamelCase = True , _lowerCamelCase = 1 / 255 , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = True , **_lowerCamelCase , ): """simple docstring""" super().__init__(**_lowerCamelCase ) UpperCAmelCase__ : Optional[Any] = size if size is not None else {"""shortest_edge""": 224} UpperCAmelCase__ : List[Any] = get_size_dict(_lowerCamelCase , default_to_square=_lowerCamelCase ) UpperCAmelCase__ : str = crop_size if crop_size is not None else {"""height""": 256, """width""": 256} UpperCAmelCase__ : str = get_size_dict(_lowerCamelCase , param_name="""crop_size""" ) UpperCAmelCase__ : int = do_resize UpperCAmelCase__ : Any = size UpperCAmelCase__ : int = resample UpperCAmelCase__ : Union[str, Any] = do_rescale UpperCAmelCase__ : Union[str, Any] = rescale_factor UpperCAmelCase__ : str = do_center_crop UpperCAmelCase__ : Dict = crop_size UpperCAmelCase__ : List[str] = do_flip_channel_order def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = PIL.Image.BILINEAR , _lowerCamelCase = None , **_lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : List[str] = get_size_dict(_lowerCamelCase , default_to_square=_lowerCamelCase ) if "shortest_edge" not in size: raise ValueError(F"""The `size` dictionary must contain the key `shortest_edge`. Got {size.keys()}""" ) UpperCAmelCase__ : Union[str, Any] = get_resize_output_image_size(_lowerCamelCase , size=size["""shortest_edge"""] , default_to_square=_lowerCamelCase ) return resize(_lowerCamelCase , size=_lowerCamelCase , resample=_lowerCamelCase , data_format=_lowerCamelCase , **_lowerCamelCase ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = None , **_lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = get_size_dict(_lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""" ) return center_crop(_lowerCamelCase , size=(size["""height"""], size["""width"""]) , data_format=_lowerCamelCase , **_lowerCamelCase ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = None , **_lowerCamelCase , ): """simple docstring""" return rescale(_lowerCamelCase , scale=_lowerCamelCase , data_format=_lowerCamelCase , **_lowerCamelCase ) def _a (self , _lowerCamelCase , _lowerCamelCase = None ): """simple docstring""" return flip_channel_order(_lowerCamelCase , data_format=_lowerCamelCase ) def _a (self , _lowerCamelCase , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = ChannelDimension.FIRST , **_lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : Any = do_resize if do_resize is not None else self.do_resize UpperCAmelCase__ : List[str] = resample if resample is not None else self.resample UpperCAmelCase__ : Tuple = do_rescale if do_rescale is not None else self.do_rescale UpperCAmelCase__ : Optional[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCAmelCase__ : List[Any] = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCAmelCase__ : int = ( do_flip_channel_order if do_flip_channel_order is not None else self.do_flip_channel_order ) UpperCAmelCase__ : List[str] = size if size is not None else self.size UpperCAmelCase__ : Tuple = get_size_dict(_lowerCamelCase , default_to_square=_lowerCamelCase ) UpperCAmelCase__ : Optional[int] = crop_size if crop_size is not None else self.crop_size UpperCAmelCase__ : str = get_size_dict(_lowerCamelCase , param_name="""crop_size""" ) UpperCAmelCase__ : List[str] = make_list_of_images(_lowerCamelCase ) if not valid_images(_lowerCamelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) # All transformations expect numpy arrays. UpperCAmelCase__ : Union[str, Any] = [to_numpy_array(_lowerCamelCase ) for image in images] if do_resize: UpperCAmelCase__ : Tuple = [self.resize(image=_lowerCamelCase , size=_lowerCamelCase , resample=_lowerCamelCase ) for image in images] if do_center_crop: UpperCAmelCase__ : Optional[Any] = [self.center_crop(image=_lowerCamelCase , size=_lowerCamelCase ) for image in images] if do_rescale: UpperCAmelCase__ : List[Any] = [self.rescale(image=_lowerCamelCase , scale=_lowerCamelCase ) for image in images] # the pretrained checkpoints assume images are BGR, not RGB if do_flip_channel_order: UpperCAmelCase__ : Any = [self.flip_channel_order(image=_lowerCamelCase ) for image in images] UpperCAmelCase__ : int = [to_channel_dimension_format(_lowerCamelCase , _lowerCamelCase ) for image in images] UpperCAmelCase__ : Optional[Any] = {"""pixel_values""": images} return BatchFeature(data=_lowerCamelCase , tensor_type=_lowerCamelCase ) def _a (self , _lowerCamelCase , _lowerCamelCase = None ): """simple docstring""" UpperCAmelCase__ : List[str] = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(_lowerCamelCase ) != len(_lowerCamelCase ): raise ValueError( """Make sure that you pass in as many target sizes as the batch dimension of the logits""" ) if is_torch_tensor(_lowerCamelCase ): UpperCAmelCase__ : Optional[int] = target_sizes.numpy() UpperCAmelCase__ : Tuple = [] for idx in range(len(_lowerCamelCase ) ): UpperCAmelCase__ : Union[str, Any] = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode="""bilinear""" , align_corners=_lowerCamelCase ) UpperCAmelCase__ : str = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(_lowerCamelCase ) else: UpperCAmelCase__ : str = logits.argmax(dim=1 ) UpperCAmelCase__ : int = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
171
1
import unittest from transformers import AutoConfig, AutoTokenizer, BertConfig, TensorType, is_flax_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, require_flax, slow if is_flax_available(): import jax from transformers.models.auto.modeling_flax_auto import FlaxAutoModel from transformers.models.bert.modeling_flax_bert import FlaxBertModel from transformers.models.roberta.modeling_flax_roberta import FlaxRobertaModel @require_flax class snake_case__ ( unittest.TestCase ): @slow def __magic_name__ ( self ) -> Tuple: for model_name in ["bert-base-cased", "bert-large-uncased"]: with self.subTest(lowerCAmelCase__ ): __magic_name__ : Dict = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) __magic_name__ : List[Any] = FlaxAutoModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) @slow def __magic_name__ ( self ) -> Optional[Any]: for model_name in ["roberta-base", "roberta-large"]: with self.subTest(lowerCAmelCase__ ): __magic_name__ : Dict = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) __magic_name__ : List[str] = FlaxAutoModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) @slow def __magic_name__ ( self ) -> Optional[Any]: for model_name in ["bert-base-cased", "bert-large-uncased"]: __magic_name__ : Tuple = AutoTokenizer.from_pretrained(lowerCAmelCase__ ) __magic_name__ : Any = FlaxBertModel.from_pretrained(lowerCAmelCase__ ) __magic_name__ : str = tokenizer("""Do you support jax jitted function?""" , return_tensors=TensorType.JAX ) @jax.jit def eval(**lowerCAmelCase__ ): return model(**lowerCAmelCase__ ) eval(**lowerCAmelCase__ ).block_until_ready() @slow def __magic_name__ ( self ) -> Dict: for model_name in ["roberta-base", "roberta-large"]: __magic_name__ : Dict = AutoTokenizer.from_pretrained(lowerCAmelCase__ ) __magic_name__ : List[str] = FlaxRobertaModel.from_pretrained(lowerCAmelCase__ ) __magic_name__ : Tuple = tokenizer("""Do you support jax jitted function?""" , return_tensors=TensorType.JAX ) @jax.jit def eval(**lowerCAmelCase__ ): return model(**lowerCAmelCase__ ) eval(**lowerCAmelCase__ ).block_until_ready() def __magic_name__ ( self ) -> List[Any]: with self.assertRaisesRegex( lowerCAmelCase__ , """bert-base is not a local folder and is not a valid model identifier""" ): __magic_name__ : List[str] = FlaxAutoModel.from_pretrained("""bert-base""" ) def __magic_name__ ( self ) -> Tuple: with self.assertRaisesRegex( lowerCAmelCase__ , R"""aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)""" ): __magic_name__ : Tuple = FlaxAutoModel.from_pretrained(lowerCAmelCase__ , revision="""aaaaaa""" ) def __magic_name__ ( self ) -> Any: with self.assertRaisesRegex( lowerCAmelCase__ , """hf-internal-testing/config-no-model does not appear to have a file named flax_model.msgpack""" , ): __magic_name__ : Any = FlaxAutoModel.from_pretrained("""hf-internal-testing/config-no-model""" ) def __magic_name__ ( self ) -> Union[str, Any]: with self.assertRaisesRegex(lowerCAmelCase__ , """Use `from_pt=True` to load this model""" ): __magic_name__ : Tuple = FlaxAutoModel.from_pretrained("""hf-internal-testing/tiny-bert-pt-only""" )
138
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_xlnet import XLNetTokenizer else: __magic_name__: Any = None __magic_name__: Dict = logging.get_logger(__name__) __magic_name__: Any = {"vocab_file": "spiece.model", "tokenizer_file": "tokenizer.json"} __magic_name__: str = { "vocab_file": { "xlnet-base-cased": "https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model", "xlnet-large-cased": "https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model", }, "tokenizer_file": { "xlnet-base-cased": "https://huggingface.co/xlnet-base-cased/resolve/main/tokenizer.json", "xlnet-large-cased": "https://huggingface.co/xlnet-large-cased/resolve/main/tokenizer.json", }, } __magic_name__: Optional[Any] = { "xlnet-base-cased": None, "xlnet-large-cased": None, } __magic_name__: Optional[Any] = "▁" # Segments (not really needed) __magic_name__: List[Any] = 0 __magic_name__: Dict = 1 __magic_name__: List[str] = 2 __magic_name__: List[Any] = 3 __magic_name__: Optional[int] = 4 class snake_case__ ( _lowerCAmelCase ): lowercase__ : Dict = VOCAB_FILES_NAMES lowercase__ : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP lowercase__ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : List[str] = '''left''' lowercase__ : List[str] = XLNetTokenizer def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="<sep>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<cls>" , lowerCAmelCase__="<mask>" , lowerCAmelCase__=["<eop>", "<eod>"] , **lowerCAmelCase__ , ) -> Tuple: # Mask token behave like a normal word, i.e. include the space before it __magic_name__ : Optional[int] = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token super().__init__( vocab_file=lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , remove_space=lowerCAmelCase__ , keep_accents=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , additional_special_tokens=lowerCAmelCase__ , **lowerCAmelCase__ , ) __magic_name__ : List[str] = 3 __magic_name__ : str = do_lower_case __magic_name__ : Union[str, Any] = remove_space __magic_name__ : str = keep_accents __magic_name__ : Tuple = vocab_file __magic_name__ : List[Any] = False if not self.vocab_file else True def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: __magic_name__ : Any = [self.sep_token_id] __magic_name__ : Any = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: __magic_name__ : List[str] = [self.sep_token_id] __magic_name__ : Optional[Any] = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __magic_name__ : List[str] = os.path.join( lowerCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
138
1
'''simple docstring''' import os from typing import List, Optional, Union from ...tokenization_utils import PreTrainedTokenizer from ...tokenization_utils_base import AddedToken from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = {'''vocab_file''': '''vocab.txt'''} UpperCamelCase__ = { '''vocab_file''': { '''facebook/esm2_t6_8M_UR50D''': '''https://huggingface.co/facebook/esm2_t6_8M_UR50D/resolve/main/vocab.txt''', '''facebook/esm2_t12_35M_UR50D''': '''https://huggingface.co/facebook/esm2_t12_35M_UR50D/resolve/main/vocab.txt''', }, } UpperCamelCase__ = { '''facebook/esm2_t6_8M_UR50D''': 1_0_2_4, '''facebook/esm2_t12_35M_UR50D''': 1_0_2_4, } def a__ ( lowerCAmelCase__ ) -> Union[str, Any]: with open(lowerCAmelCase__ , '''r''' ) as f: UpperCAmelCase__ : List[Any] = f.read().splitlines() return [l.strip() for l in lines] class lowerCamelCase_ ( __a ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = ['input_ids', 'attention_mask'] def __init__( self : Optional[Any] , _A : str , _A : List[Any]="<unk>" , _A : str="<cls>" , _A : Tuple="<pad>" , _A : Any="<mask>" , _A : Any="<eos>" , **_A : List[Any] , ): '''simple docstring''' super().__init__(**_A ) UpperCAmelCase__ : List[Any] = load_vocab_file(_A ) UpperCAmelCase__ : Tuple = dict(enumerate(self.all_tokens ) ) UpperCAmelCase__ : Optional[int] = {tok: ind for ind, tok in enumerate(self.all_tokens )} UpperCAmelCase__ : Tuple = unk_token UpperCAmelCase__ : Optional[Any] = cls_token UpperCAmelCase__ : Any = pad_token UpperCAmelCase__ : str = mask_token UpperCAmelCase__ : Tuple = eos_token UpperCAmelCase__ : List[str] = self.all_tokens self._create_trie(self.unique_no_split_tokens ) def lowercase_ ( self : Any , _A : int ): '''simple docstring''' return self._id_to_token.get(_A , self.unk_token ) def lowercase_ ( self : List[Any] , _A : str ): '''simple docstring''' return self._token_to_id.get(_A , self._token_to_id.get(self.unk_token ) ) def lowercase_ ( self : Dict , _A : Optional[int] , **_A : Optional[Any] ): '''simple docstring''' return text.split() def lowercase_ ( self : Optional[int] , _A : Union[str, Any]=False ): '''simple docstring''' return len(self._id_to_token ) def lowercase_ ( self : str ): '''simple docstring''' return {token: i for i, token in enumerate(self.all_tokens )} def lowercase_ ( self : Optional[Any] , _A : str ): '''simple docstring''' return self._token_to_id.get(_A , self._token_to_id.get(self.unk_token ) ) def lowercase_ ( self : Dict , _A : int ): '''simple docstring''' return self._id_to_token.get(_A , self.unk_token ) def lowercase_ ( self : int , _A : List[int] , _A : Optional[List[int]] = None ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = [self.cls_token_id] UpperCAmelCase__ : List[str] = [self.eos_token_id] # No sep token in ESM vocabulary if token_ids_a is None: if self.eos_token_id is None: return cls + token_ids_a else: return cls + token_ids_a + sep elif self.eos_token_id is None: raise ValueError('''Cannot tokenize multiple sequences when EOS token is not set!''' ) return cls + token_ids_a + sep + token_ids_a + sep # Multiple inputs always have an EOS token def lowercase_ ( self : List[Any] , _A : List , _A : Optional[List] = None , _A : bool = False ): '''simple docstring''' if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if token in self.all_special_ids else 0 for token in token_ids_a] UpperCAmelCase__ : List[Any] = [1] + ([0] * len(_A )) + [1] if token_ids_a is not None: mask += [0] * len(_A ) + [1] return mask def lowercase_ ( self : List[str] , _A : str , _A : int ): '''simple docstring''' UpperCAmelCase__ : int = os.path.join(_A , (filename_prefix + '''-''' if filename_prefix else '''''') + '''vocab.txt''' ) with open(_A , '''w''' ) as f: f.write('''\n'''.join(self.all_tokens ) ) return (vocab_file,) @property def lowercase_ ( self : str ): '''simple docstring''' return self.get_vocab_size(with_added_tokens=_A ) def lowercase_ ( self : List[str] , _A : Union[List[str], List[AddedToken]] , _A : bool = False ): '''simple docstring''' return super()._add_tokens(_A , special_tokens=_A )
181
'''simple docstring''' import math class lowerCamelCase_ : def lowercase_ ( self : Optional[Any] , _A : list[list[float]] , _A : list[int] ): '''simple docstring''' UpperCAmelCase__ : List[str] = 0.0 UpperCAmelCase__ : Union[str, Any] = 0.0 for i in range(len(_A ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def lowercase_ ( self : Dict , _A : list[list[int | float]] , _A : list[int] , _A : int , _A : float ): '''simple docstring''' for i in range(len(_A ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def a__ ( ) -> None: # Training Examples ( m, n ) UpperCAmelCase__ : Optional[Any] = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) UpperCAmelCase__ : str = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training UpperCAmelCase__ : Optional[Any] = SelfOrganizingMap() UpperCAmelCase__ : Tuple = 3 UpperCAmelCase__ : Dict = 0.5 for _ in range(lowerCAmelCase__ ): for j in range(len(lowerCAmelCase__ ) ): # training sample UpperCAmelCase__ : Optional[int] = training_samples[j] # Compute the winning vector UpperCAmelCase__ : str = self_organizing_map.get_winner(lowerCAmelCase__ , lowerCAmelCase__ ) # Update the winning vector UpperCAmelCase__ : int = self_organizing_map.update(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # classify test sample UpperCAmelCase__ : str = [0, 0, 0, 1] UpperCAmelCase__ : int = self_organizing_map.get_winner(lowerCAmelCase__ , lowerCAmelCase__ ) # results print(F"""Clusters that the test sample belongs to : {winner}""" ) print(F"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
181
1
"""simple docstring""" import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __A = logging.get_logger(__name__) __A = { "vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_config_file": "tokenizer_config.json", } __A = { "vocab_file": {"facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json"}, "merges_file": {"facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt"}, "tokenizer_config_file": { "facebook/blenderbot-3B": "https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json" }, } __A = {"facebook/blenderbot-3B": 128} @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def a__ ( ) -> Dict: __lowerCAmelCase: str = ( list(range(ord("!" ) , ord("~" ) + 1 ) ) + list(range(ord("¡" ) , ord("¬" ) + 1 ) ) + list(range(ord("®" ) , ord("ÿ" ) + 1 ) ) ) __lowerCAmelCase: Dict = bs[:] __lowerCAmelCase: Dict = 0 for b in range(2**8 ): if b not in bs: bs.append(snake_case__ ) cs.append(2**8 + n ) n += 1 __lowerCAmelCase: List[str] = [chr(snake_case__ ) for n in cs] return dict(zip(snake_case__ , snake_case__ ) ) def a__ ( __SCREAMING_SNAKE_CASE ) -> str: __lowerCAmelCase: Dict = set() __lowerCAmelCase: Union[str, Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __lowerCAmelCase: Optional[Any] = char return pairs class snake_case ( __lowerCAmelCase ): SCREAMING_SNAKE_CASE_ : List[str] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE_ : List[str] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE_ : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE_ : Optional[int] = ['''input_ids''', '''attention_mask'''] def __init__( self : Optional[Any] , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : List[str] , UpperCamelCase__ : int="replace" , UpperCamelCase__ : List[str]="<s>" , UpperCamelCase__ : Union[str, Any]="</s>" , UpperCamelCase__ : Any="</s>" , UpperCamelCase__ : Any="<s>" , UpperCamelCase__ : Tuple="<unk>" , UpperCamelCase__ : Any="<pad>" , UpperCamelCase__ : Tuple="<mask>" , UpperCamelCase__ : int=False , **UpperCamelCase__ : Any , )-> int: '''simple docstring''' __lowerCAmelCase: Any = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_) if isinstance(lowerCAmelCase_ , lowerCAmelCase_) else bos_token __lowerCAmelCase: Optional[Any] = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_) if isinstance(lowerCAmelCase_ , lowerCAmelCase_) else eos_token __lowerCAmelCase: Any = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_) if isinstance(lowerCAmelCase_ , lowerCAmelCase_) else sep_token __lowerCAmelCase: List[str] = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_) if isinstance(lowerCAmelCase_ , lowerCAmelCase_) else cls_token __lowerCAmelCase: Optional[int] = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_) if isinstance(lowerCAmelCase_ , lowerCAmelCase_) else unk_token __lowerCAmelCase: List[str] = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_) if isinstance(lowerCAmelCase_ , lowerCAmelCase_) else pad_token # Mask token behave like a normal word, i.e. include the space before it __lowerCAmelCase: Tuple = AddedToken(lowerCAmelCase_ , lstrip=lowerCAmelCase_ , rstrip=lowerCAmelCase_) if isinstance(lowerCAmelCase_ , lowerCAmelCase_) else mask_token super().__init__( errors=lowerCAmelCase_ , bos_token=lowerCAmelCase_ , eos_token=lowerCAmelCase_ , unk_token=lowerCAmelCase_ , sep_token=lowerCAmelCase_ , cls_token=lowerCAmelCase_ , pad_token=lowerCAmelCase_ , mask_token=lowerCAmelCase_ , add_prefix_space=lowerCAmelCase_ , **lowerCAmelCase_ , ) with open(lowerCAmelCase_ , encoding="utf-8") as vocab_handle: __lowerCAmelCase: Optional[Any] = json.load(lowerCAmelCase_) __lowerCAmelCase: Optional[Any] = {v: k for k, v in self.encoder.items()} __lowerCAmelCase: int = errors # how to handle errors in decoding __lowerCAmelCase: Tuple = bytes_to_unicode() __lowerCAmelCase: str = {v: k for k, v in self.byte_encoder.items()} with open(lowerCAmelCase_ , encoding="utf-8") as merges_handle: __lowerCAmelCase: Union[str, Any] = merges_handle.read().split("\n")[1:-1] __lowerCAmelCase: List[Any] = [tuple(merge.split()) for merge in bpe_merges] __lowerCAmelCase: Optional[Any] = dict(zip(lowerCAmelCase_ , range(len(lowerCAmelCase_)))) __lowerCAmelCase: Union[str, Any] = {} __lowerCAmelCase: Optional[int] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions __lowerCAmelCase: List[Any] = re.compile(R"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+") @property # Copied from transformers.models.roberta.tokenization_roberta.RobertaTokenizer.vocab_size with Roberta->Blenderbot, RoBERTa->Blenderbot def lowercase_ ( self : Dict)-> str: '''simple docstring''' return len(self.encoder) def lowercase_ ( self : Optional[Any])-> Any: '''simple docstring''' return dict(self.encoder , **self.added_tokens_encoder) def lowercase_ ( self : Optional[Any] , UpperCamelCase__ : Union[str, Any])-> Tuple: '''simple docstring''' if token in self.cache: return self.cache[token] __lowerCAmelCase: Optional[int] = tuple(lowerCAmelCase_) __lowerCAmelCase: Any = get_pairs(lowerCAmelCase_) if not pairs: return token while True: __lowerCAmelCase: int = min(lowerCAmelCase_ , key=lambda UpperCamelCase__: self.bpe_ranks.get(lowerCAmelCase_ , float("inf"))) if bigram not in self.bpe_ranks: break __lowerCAmelCase , __lowerCAmelCase: Optional[int] = bigram __lowerCAmelCase: Union[str, Any] = [] __lowerCAmelCase: Tuple = 0 while i < len(lowerCAmelCase_): try: __lowerCAmelCase: Optional[Any] = word.index(lowerCAmelCase_ , lowerCAmelCase_) except ValueError: new_word.extend(word[i:]) break else: new_word.extend(word[i:j]) __lowerCAmelCase: Tuple = j if word[i] == first and i < len(lowerCAmelCase_) - 1 and word[i + 1] == second: new_word.append(first + second) i += 2 else: new_word.append(word[i]) i += 1 __lowerCAmelCase: Tuple = tuple(lowerCAmelCase_) __lowerCAmelCase: Union[str, Any] = new_word if len(lowerCAmelCase_) == 1: break else: __lowerCAmelCase: Optional[Any] = get_pairs(lowerCAmelCase_) __lowerCAmelCase: Optional[Any] = " ".join(lowerCAmelCase_) __lowerCAmelCase: Optional[int] = word return word def lowercase_ ( self : Dict , UpperCamelCase__ : Optional[int])-> Union[str, Any]: '''simple docstring''' __lowerCAmelCase: Tuple = [] for token in re.findall(self.pat , lowerCAmelCase_): __lowerCAmelCase: Tuple = "".join( self.byte_encoder[b] for b in token.encode("utf-8")) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCAmelCase_).split(" ")) return bpe_tokens def lowercase_ ( self : Optional[int] , UpperCamelCase__ : Dict)-> List[str]: '''simple docstring''' return self.encoder.get(lowerCAmelCase_ , self.encoder.get(self.unk_token)) def lowercase_ ( self : Optional[Any] , UpperCamelCase__ : List[str])-> Union[str, Any]: '''simple docstring''' return self.decoder.get(lowerCAmelCase_) def lowercase_ ( self : Union[str, Any] , UpperCamelCase__ : int)-> int: '''simple docstring''' __lowerCAmelCase: Any = "".join(lowerCAmelCase_) __lowerCAmelCase: List[Any] = bytearray([self.byte_decoder[c] for c in text]).decode("utf-8" , errors=self.errors) return text def lowercase_ ( self : str , UpperCamelCase__ : List[str] , UpperCamelCase__ : List[str] = None)-> Tuple[str]: '''simple docstring''' if not os.path.isdir(lowerCAmelCase_): logger.error(f"Vocabulary path ({save_directory}) should be a directory") return __lowerCAmelCase: int = os.path.join( lowerCAmelCase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) __lowerCAmelCase: List[str] = os.path.join( lowerCAmelCase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"]) with open(lowerCAmelCase_ , "w" , encoding="utf-8") as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCAmelCase_ , ensure_ascii=lowerCAmelCase_) + "\n") __lowerCAmelCase: List[str] = 0 with open(lowerCAmelCase_ , "w" , encoding="utf-8") as writer: writer.write("#version: 0.2\n") for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda UpperCamelCase__: kv[1]): if index != token_index: logger.warning( f"Saving vocabulary to {merge_file}: BPE merge indices are not consecutive." " Please check that the tokenizer is not corrupted!") __lowerCAmelCase: Optional[int] = token_index writer.write(" ".join(lowerCAmelCase_) + "\n") index += 1 return vocab_file, merge_file def lowercase_ ( self : Dict , UpperCamelCase__ : int , UpperCamelCase__ : List[Any] = None , UpperCamelCase__ : Tuple = False)-> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase_ , token_ids_a=lowerCAmelCase_ , already_has_special_tokens=lowerCAmelCase_) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase_)) + [1] return [1] + ([0] * len(lowerCAmelCase_)) + [1, 1] + ([0] * len(lowerCAmelCase_)) + [1] def lowercase_ ( self : Optional[int] , UpperCamelCase__ : int , UpperCamelCase__ : Optional[int] = None)-> List[int]: '''simple docstring''' __lowerCAmelCase: List[str] = [self.sep_token_id] __lowerCAmelCase: int = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def lowercase_ ( self : Dict , UpperCamelCase__ : Dict , UpperCamelCase__ : int=False , **UpperCamelCase__ : str)-> List[Any]: '''simple docstring''' __lowerCAmelCase: Dict = kwargs.pop("add_prefix_space" , self.add_prefix_space) if (is_split_into_words or add_prefix_space) and (len(lowerCAmelCase_) > 0 and not text[0].isspace()): __lowerCAmelCase: List[str] = " " + text return (text, kwargs) def lowercase_ ( self : Union[str, Any] , UpperCamelCase__ : int , UpperCamelCase__ : Dict = None)-> List[str]: '''simple docstring''' return token_ids_a + [self.eos_token_id] def lowercase_ ( self : Optional[Any] , UpperCamelCase__ : Tuple)-> List[int]: '''simple docstring''' __lowerCAmelCase: Dict = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(" " + text) else: # Generated responses should contain them already. inputs.append(lowerCAmelCase_) __lowerCAmelCase: Optional[int] = " ".join(lowerCAmelCase_) __lowerCAmelCase: List[Any] = self.encode(lowerCAmelCase_) if len(lowerCAmelCase_) > self.model_max_length: __lowerCAmelCase: Dict = input_ids[-self.model_max_length :] logger.warning(f"Trimmed input from conversation as it was longer than {self.model_max_length} tokens.") return input_ids
365
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __A = logging.get_logger(__name__) __A = { "facebook/data2vec-vision-base-ft": ( "https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json" ), } class snake_case ( __snake_case ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = """data2vec-vision""" def __init__( self : Optional[int] , UpperCamelCase__ : str=7_6_8 , UpperCamelCase__ : Any=1_2 , UpperCamelCase__ : str=1_2 , UpperCamelCase__ : Optional[Any]=3_0_7_2 , UpperCamelCase__ : Union[str, Any]="gelu" , UpperCamelCase__ : List[Any]=0.0 , UpperCamelCase__ : Dict=0.0 , UpperCamelCase__ : Union[str, Any]=0.02 , UpperCamelCase__ : str=1e-12 , UpperCamelCase__ : Dict=2_2_4 , UpperCamelCase__ : List[Any]=1_6 , UpperCamelCase__ : Optional[Any]=3 , UpperCamelCase__ : List[str]=False , UpperCamelCase__ : List[Any]=False , UpperCamelCase__ : str=False , UpperCamelCase__ : List[str]=False , UpperCamelCase__ : int=0.1 , UpperCamelCase__ : List[Any]=0.1 , UpperCamelCase__ : Optional[Any]=True , UpperCamelCase__ : List[Any]=[3, 5, 7, 1_1] , UpperCamelCase__ : List[str]=[1, 2, 3, 6] , UpperCamelCase__ : Union[str, Any]=True , UpperCamelCase__ : Any=0.4 , UpperCamelCase__ : Union[str, Any]=2_5_6 , UpperCamelCase__ : List[Any]=1 , UpperCamelCase__ : str=False , UpperCamelCase__ : Optional[int]=2_5_5 , **UpperCamelCase__ : Dict , )-> List[str]: '''simple docstring''' super().__init__(**UpperCamelCase__) __lowerCAmelCase: List[str] = hidden_size __lowerCAmelCase: Union[str, Any] = num_hidden_layers __lowerCAmelCase: Dict = num_attention_heads __lowerCAmelCase: Optional[int] = intermediate_size __lowerCAmelCase: int = hidden_act __lowerCAmelCase: Union[str, Any] = hidden_dropout_prob __lowerCAmelCase: Any = attention_probs_dropout_prob __lowerCAmelCase: Dict = initializer_range __lowerCAmelCase: Any = layer_norm_eps __lowerCAmelCase: Union[str, Any] = image_size __lowerCAmelCase: Tuple = patch_size __lowerCAmelCase: List[str] = num_channels __lowerCAmelCase: Optional[Any] = use_mask_token __lowerCAmelCase: str = use_absolute_position_embeddings __lowerCAmelCase: Optional[int] = use_relative_position_bias __lowerCAmelCase: str = use_shared_relative_position_bias __lowerCAmelCase: Union[str, Any] = layer_scale_init_value __lowerCAmelCase: Any = drop_path_rate __lowerCAmelCase: Dict = use_mean_pooling # decode head attributes (semantic segmentation) __lowerCAmelCase: int = out_indices __lowerCAmelCase: Any = pool_scales # auxiliary head attributes (semantic segmentation) __lowerCAmelCase: List[Any] = use_auxiliary_head __lowerCAmelCase: int = auxiliary_loss_weight __lowerCAmelCase: Dict = auxiliary_channels __lowerCAmelCase: Any = auxiliary_num_convs __lowerCAmelCase: Any = auxiliary_concat_input __lowerCAmelCase: Any = semantic_loss_ignore_index class snake_case ( __snake_case ): SCREAMING_SNAKE_CASE_ : Dict = version.parse("""1.11""" ) @property def lowercase_ ( self : Optional[Any])-> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ]) @property def lowercase_ ( self : Any)-> float: '''simple docstring''' return 1e-4
108
0
'''simple docstring''' import json import multiprocessing as mp import re from collections import defaultdict from functools import partial from typing import Dict, List, Optional, Set, Tuple, Type from datasets import Dataset from datasketch import MinHash, MinHashLSH from dpu_utils.utils.iterators import ThreadedIterator from tqdm import tqdm UpperCAmelCase = re.compile('''[^A-Za-z_0-9]''') # parameters used in DuplicationIndex UpperCAmelCase = 10 UpperCAmelCase = 256 def __UpperCamelCase ( lowercase__ : List[str] ): '''simple docstring''' if len(lowercase__ ) < MIN_NUM_TOKENS: return None __lowercase =MinHash(num_perm=lowercase__ ) for token in set(lowercase__ ): min_hash.update(token.encode() ) return min_hash def __UpperCamelCase ( lowercase__ : str ): '''simple docstring''' return {t for t in NON_ALPHA.split(lowercase__ ) if len(t.strip() ) > 0} class lowerCAmelCase : def __init__( self : Optional[int] , *, __lowercase : float = 0.8_5 , ): """simple docstring""" __lowercase =duplication_jaccard_threshold __lowercase =NUM_PERM __lowercase =MinHashLSH(threshold=self._duplication_jaccard_threshold , num_perm=self._num_perm ) __lowercase =defaultdict(__lowercase ) def snake_case ( self : Optional[int] , __lowercase : Tuple , __lowercase : MinHash ): """simple docstring""" __lowercase =self._index.query(__lowercase ) if code_key in self._index.keys: print(f'''Duplicate key {code_key}''' ) return self._index.insert(__lowercase , __lowercase ) if len(__lowercase ) > 0: for base_duplicate in close_duplicates: if base_duplicate in self._duplicate_clusters: self._duplicate_clusters[base_duplicate].add(__lowercase ) break else: self._duplicate_clusters[close_duplicates[0]].add(__lowercase ) def snake_case ( self : int ): """simple docstring""" __lowercase =[] for base, duplicates in self._duplicate_clusters.items(): __lowercase =[base] + list(__lowercase ) # reformat the cluster to be a list of dict __lowercase =[{'base_index': el[0], 'repo_name': el[1], 'path': el[2]} for el in cluster] duplicate_clusters.append(__lowercase ) return duplicate_clusters def snake_case ( self : int , __lowercase : Tuple ): """simple docstring""" __lowercase =self.get_duplicate_clusters() with open(__lowercase , 'w' ) as f: json.dump(__lowercase , __lowercase ) def __UpperCamelCase ( lowercase__ : Any ): '''simple docstring''' __lowercase , __lowercase =element __lowercase =get_min_hash([t for t in NON_ALPHA.split(data['content'] ) if len(t.strip() ) > 0] ) if min_hash is not None: return (index, data["repo_name"], data["path"]), min_hash def __UpperCamelCase ( lowercase__ : Type[Dataset] ): '''simple docstring''' with mp.Pool() as pool: for data in pool.imap_unordered( _compute_min_hash, ThreadedIterator(lowercase__, max_queue_size=1_00_00 ), chunksize=1_00, ): if data is not None: yield data def __UpperCamelCase ( lowercase__ : Type[Dataset], lowercase__ : float ): '''simple docstring''' __lowercase =DuplicationIndex(duplication_jaccard_threshold=lowercase__ ) for filename, min_hash in tqdm(ThreadedIterator(minhash_iter(enumerate(lowercase__ ) ), max_queue_size=1_00 ) ): di.add(lowercase__, lowercase__ ) # Returns a List[Cluster] where Cluster is List[str] with the filenames. return di.get_duplicate_clusters() def __UpperCamelCase ( lowercase__ : str, lowercase__ : str ): '''simple docstring''' __lowercase =get_tokens(lowercase__ ) __lowercase =get_tokens(lowercase__ ) return len(tokensa & tokensa ) / len(tokensa | tokensa ) UpperCAmelCase = None def __UpperCamelCase ( lowercase__ : Union[str, Any], lowercase__ : Dict ): '''simple docstring''' __lowercase =[] for elementa in cluster: __lowercase =_shared_dataset[elementa['base_index']]['content'] for elementa in extremes: __lowercase =_shared_dataset[elementa['base_index']]['content'] if jaccard_similarity(lowercase__, lowercase__ ) >= jaccard_threshold: elementa["copies"] += 1 break else: __lowercase =1 extremes.append(lowercase__ ) return extremes def __UpperCamelCase ( lowercase__ : List[str], lowercase__ : Tuple, lowercase__ : Optional[int] ): '''simple docstring''' global _shared_dataset __lowercase =dataset __lowercase =[] __lowercase =partial(_find_cluster_extremes_shared, jaccard_threshold=lowercase__ ) with mp.Pool() as pool: for extremes in tqdm( pool.imap_unordered( lowercase__, lowercase__, ), total=len(lowercase__ ), ): extremes_list.append(lowercase__ ) return extremes_list def __UpperCamelCase ( lowercase__ : Type[Dataset], lowercase__ : float = 0.85 ): '''simple docstring''' __lowercase =make_duplicate_clusters(lowercase__, lowercase__ ) __lowercase ={x['base_index'] for cluster in duplicate_clusters for x in cluster} __lowercase ={} __lowercase =find_extremes(lowercase__, lowercase__, lowercase__ ) for extremes in extremes_clusters: for element in extremes: __lowercase =element __lowercase =duplicate_indices - set(extreme_dict.keys() ) __lowercase =dataset.filter(lambda lowercase__, lowercase__ : idx not in remove_indices, with_indices=lowercase__ ) # update duplicate_clusters for cluster in duplicate_clusters: for element in cluster: __lowercase =element['base_index'] in extreme_dict if element["is_extreme"]: __lowercase =extreme_dict[element['base_index']]['copies'] print(F'''Original dataset size: {len(lowercase__ )}''' ) print(F'''Number of duplicate clusters: {len(lowercase__ )}''' ) print(F'''Files in duplicate cluster: {len(lowercase__ )}''' ) print(F'''Unique files in duplicate cluster: {len(lowercase__ )}''' ) print(F'''Filtered dataset size: {len(lowercase__ )}''' ) return ds_filter, duplicate_clusters
141
'''simple docstring''' from __future__ import annotations import math import random from collections.abc import Collection from typing import overload class lowerCAmelCase : def __init__( self : List[str] , __lowercase : Collection[float] | None = None ): """simple docstring""" if components is None: __lowercase =[] __lowercase =list(__lowercase ) def __len__( self : Union[str, Any] ): """simple docstring""" return len(self.__components ) def __str__( self : int ): """simple docstring""" return "(" + ",".join(map(__lowercase , self.__components ) ) + ")" def __add__( self : List[Any] , __lowercase : Vector ): """simple docstring""" __lowercase =len(self ) if size == len(__lowercase ): __lowercase =[self.__components[i] + other.component(__lowercase ) for i in range(__lowercase )] return Vector(__lowercase ) else: raise Exception('must have the same size' ) def __sub__( self : str , __lowercase : Vector ): """simple docstring""" __lowercase =len(self ) if size == len(__lowercase ): __lowercase =[self.__components[i] - other.component(__lowercase ) for i in range(__lowercase )] return Vector(__lowercase ) else: # error case raise Exception('must have the same size' ) @overload def __mul__( self : Tuple , __lowercase : float ): """simple docstring""" ... @overload def __mul__( self : Tuple , __lowercase : Vector ): """simple docstring""" ... def __mul__( self : int , __lowercase : float | Vector ): """simple docstring""" if isinstance(__lowercase , (float, int) ): __lowercase =[c * other for c in self.__components] return Vector(__lowercase ) elif isinstance(__lowercase , __lowercase ) and len(self ) == len(__lowercase ): __lowercase =len(self ) __lowercase =[self.__components[i] * other.component(__lowercase ) for i in range(__lowercase )] return sum(__lowercase ) else: # error case raise Exception('invalid operand!' ) def snake_case ( self : Optional[Any] ): """simple docstring""" return Vector(self.__components ) def snake_case ( self : str , __lowercase : int ): """simple docstring""" if isinstance(__lowercase , __lowercase ) and -len(self.__components ) <= i < len(self.__components ): return self.__components[i] else: raise Exception('index out of range' ) def snake_case ( self : List[str] , __lowercase : int , __lowercase : float ): """simple docstring""" assert -len(self.__components ) <= pos < len(self.__components ) __lowercase =value def snake_case ( self : Tuple ): """simple docstring""" if len(self.__components ) == 0: raise Exception('Vector is empty' ) __lowercase =[c**2 for c in self.__components] return math.sqrt(sum(__lowercase ) ) def snake_case ( self : List[Any] , __lowercase : Vector , __lowercase : bool = False ): """simple docstring""" __lowercase =self * other __lowercase =self.euclidean_length() * other.euclidean_length() if deg: return math.degrees(math.acos(num / den ) ) else: return math.acos(num / den ) def __UpperCamelCase ( lowercase__ : int ): '''simple docstring''' assert isinstance(lowercase__, lowercase__ ) return Vector([0] * dimension ) def __UpperCamelCase ( lowercase__ : int, lowercase__ : int ): '''simple docstring''' assert isinstance(lowercase__, lowercase__ ) and (isinstance(lowercase__, lowercase__ )) __lowercase =[0] * dimension __lowercase =1 return Vector(lowercase__ ) def __UpperCamelCase ( lowercase__ : float, lowercase__ : Vector, lowercase__ : Vector ): '''simple docstring''' assert ( isinstance(lowercase__, lowercase__ ) and isinstance(lowercase__, lowercase__ ) and (isinstance(lowercase__, (int, float) )) ) return x * scalar + y def __UpperCamelCase ( lowercase__ : int, lowercase__ : int, lowercase__ : int ): '''simple docstring''' random.seed(lowercase__ ) __lowercase =[random.randint(lowercase__, lowercase__ ) for _ in range(lowercase__ )] return Vector(lowercase__ ) class lowerCAmelCase : def __init__( self : Dict , __lowercase : list[list[float]] , __lowercase : int , __lowercase : int ): """simple docstring""" __lowercase =matrix __lowercase =w __lowercase =h def __str__( self : Optional[Any] ): """simple docstring""" __lowercase ='' for i in range(self.__height ): ans += "|" for j in range(self.__width ): if j < self.__width - 1: ans += str(self.__matrix[i][j] ) + "," else: ans += str(self.__matrix[i][j] ) + "|\n" return ans def __add__( self : Union[str, Any] , __lowercase : Matrix ): """simple docstring""" if self.__width == other.width() and self.__height == other.height(): __lowercase =[] for i in range(self.__height ): __lowercase =[ self.__matrix[i][j] + other.component(__lowercase , __lowercase ) for j in range(self.__width ) ] matrix.append(__lowercase ) return Matrix(__lowercase , self.__width , self.__height ) else: raise Exception('matrix must have the same dimension!' ) def __sub__( self : int , __lowercase : Matrix ): """simple docstring""" if self.__width == other.width() and self.__height == other.height(): __lowercase =[] for i in range(self.__height ): __lowercase =[ self.__matrix[i][j] - other.component(__lowercase , __lowercase ) for j in range(self.__width ) ] matrix.append(__lowercase ) return Matrix(__lowercase , self.__width , self.__height ) else: raise Exception('matrices must have the same dimension!' ) @overload def __mul__( self : int , __lowercase : float ): """simple docstring""" ... @overload def __mul__( self : str , __lowercase : Vector ): """simple docstring""" ... def __mul__( self : Tuple , __lowercase : float | Vector ): """simple docstring""" if isinstance(__lowercase , __lowercase ): # matrix-vector if len(__lowercase ) == self.__width: __lowercase =zero_vector(self.__height ) for i in range(self.__height ): __lowercase =[ self.__matrix[i][j] * other.component(__lowercase ) for j in range(self.__width ) ] ans.change_component(__lowercase , sum(__lowercase ) ) return ans else: raise Exception( 'vector must have the same size as the ' 'number of columns of the matrix!' ) elif isinstance(__lowercase , (int, float) ): # matrix-scalar __lowercase =[ [self.__matrix[i][j] * other for j in range(self.__width )] for i in range(self.__height ) ] return Matrix(__lowercase , self.__width , self.__height ) return None def snake_case ( self : int ): """simple docstring""" return self.__height def snake_case ( self : List[str] ): """simple docstring""" return self.__width def snake_case ( self : Dict , __lowercase : int , __lowercase : int ): """simple docstring""" if 0 <= x < self.__height and 0 <= y < self.__width: return self.__matrix[x][y] else: raise Exception('change_component: indices out of bounds' ) def snake_case ( self : Dict , __lowercase : int , __lowercase : int , __lowercase : float ): """simple docstring""" if 0 <= x < self.__height and 0 <= y < self.__width: __lowercase =value else: raise Exception('change_component: indices out of bounds' ) def snake_case ( self : Dict , __lowercase : int , __lowercase : int ): """simple docstring""" if self.__height != self.__width: raise Exception('Matrix is not square' ) __lowercase =self.__matrix[:x] + self.__matrix[x + 1 :] for i in range(len(__lowercase ) ): __lowercase =minor[i][:y] + minor[i][y + 1 :] return Matrix(__lowercase , self.__width - 1 , self.__height - 1 ).determinant() def snake_case ( self : Union[str, Any] , __lowercase : int , __lowercase : int ): """simple docstring""" if self.__height != self.__width: raise Exception('Matrix is not square' ) if 0 <= x < self.__height and 0 <= y < self.__width: return (-1) ** (x + y) * self.minor(__lowercase , __lowercase ) else: raise Exception('Indices out of bounds' ) def snake_case ( self : Tuple ): """simple docstring""" if self.__height != self.__width: raise Exception('Matrix is not square' ) if self.__height < 1: raise Exception('Matrix has no element' ) elif self.__height == 1: return self.__matrix[0][0] elif self.__height == 2: return ( self.__matrix[0][0] * self.__matrix[1][1] - self.__matrix[0][1] * self.__matrix[1][0] ) else: __lowercase =[ self.__matrix[0][y] * self.cofactor(0 , __lowercase ) for y in range(self.__width ) ] return sum(__lowercase ) def __UpperCamelCase ( lowercase__ : int ): '''simple docstring''' __lowercase =[[0] * n for _ in range(lowercase__ )] return Matrix(lowercase__, lowercase__, lowercase__ ) def __UpperCamelCase ( lowercase__ : int, lowercase__ : int, lowercase__ : int, lowercase__ : int ): '''simple docstring''' random.seed(lowercase__ ) __lowercase =[ [random.randint(lowercase__, lowercase__ ) for _ in range(lowercase__ )] for _ in range(lowercase__ ) ] return Matrix(lowercase__, lowercase__, lowercase__ )
141
1
"""simple docstring""" import os from distutils.util import strtobool def __UpperCAmelCase ( UpperCAmelCase_ : int , UpperCAmelCase_ : Union[str, Any] ) -> List[Any]: '''simple docstring''' for e in env_keys: __snake_case : int = int(os.environ.get(UpperCAmelCase_ , -1 ) ) if val >= 0: return val return default def __UpperCAmelCase ( UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : int=False ) -> str: '''simple docstring''' __snake_case : Dict = os.environ.get(UpperCAmelCase_ , str(UpperCAmelCase_ ) ) return strtobool(UpperCAmelCase_ ) == 1 # As its name indicates `strtobool` actually returns an int... def __UpperCAmelCase ( UpperCAmelCase_ : Dict , UpperCAmelCase_ : Tuple="no" ) -> Dict: '''simple docstring''' __snake_case : Optional[Any] = os.environ.get(UpperCAmelCase_ , str(UpperCAmelCase_ ) ) return value
95
"""simple docstring""" from __future__ import annotations from math import pow, sqrt def __UpperCAmelCase ( UpperCAmelCase_ : float , UpperCAmelCase_ : float , UpperCAmelCase_ : float ) -> dict[str, float]: '''simple docstring''' if (resistance, reactance, impedance).count(0 ) != 1: raise ValueError('One and only one argument must be 0' ) if resistance == 0: return {"resistance": sqrt(pow(UpperCAmelCase_ , 2 ) - pow(UpperCAmelCase_ , 2 ) )} elif reactance == 0: return {"reactance": sqrt(pow(UpperCAmelCase_ , 2 ) - pow(UpperCAmelCase_ , 2 ) )} elif impedance == 0: return {"impedance": sqrt(pow(UpperCAmelCase_ , 2 ) + pow(UpperCAmelCase_ , 2 ) )} else: raise ValueError('Exactly one argument must be 0' ) if __name__ == "__main__": import doctest doctest.testmod()
95
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available a_ = { 'configuration_nezha': ['NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'NezhaConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ 'NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST', 'NezhaForNextSentencePrediction', 'NezhaForMaskedLM', 'NezhaForPreTraining', 'NezhaForMultipleChoice', 'NezhaForQuestionAnswering', 'NezhaForSequenceClassification', 'NezhaForTokenClassification', 'NezhaModel', 'NezhaPreTrainedModel', ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys a_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
175
import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMInverseScheduler, DDIMScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, StableDiffusionDiffEditPipeline, UNetaDConditionModel, ) from diffusers.utils import load_image, slow from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, require_torch_gpu, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _lowercase ( snake_case_ , snake_case_ , unittest.TestCase ): lowercase = StableDiffusionDiffEditPipeline lowercase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'height', 'width', 'image'} | {'image_latents'} lowercase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS - {'image'} | {'image_latents'} lowercase = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess lowercase = frozenset([] ) def SCREAMING_SNAKE_CASE__ ( self : List[Any] ) -> Dict: """simple docstring""" torch.manual_seed(0 ) UpperCamelCase_ : Dict = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=3_2 , attention_head_dim=(2, 4) , use_linear_projection=snake_case , ) UpperCamelCase_ : Optional[int] = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=snake_case , set_alpha_to_one=snake_case , ) UpperCamelCase_ : Tuple = DDIMInverseScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=snake_case , set_alpha_to_zero=snake_case , ) torch.manual_seed(0 ) UpperCamelCase_ : List[Any] = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0 ) UpperCamelCase_ : List[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='gelu' , projection_dim=5_1_2 , ) UpperCamelCase_ : List[str] = CLIPTextModel(snake_case ) UpperCamelCase_ : List[str] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) UpperCamelCase_ : Any = { 'unet': unet, 'scheduler': scheduler, 'inverse_scheduler': inverse_scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , snake_case : Dict , snake_case : Tuple=0 ) -> int: """simple docstring""" UpperCamelCase_ : Optional[Any] = floats_tensor((1, 1_6, 1_6) , rng=random.Random(snake_case ) ).to(snake_case ) UpperCamelCase_ : Optional[int] = floats_tensor((1, 2, 4, 1_6, 1_6) , rng=random.Random(snake_case ) ).to(snake_case ) if str(snake_case ).startswith('mps' ): UpperCamelCase_ : int = torch.manual_seed(snake_case ) else: UpperCamelCase_ : Any = torch.Generator(device=snake_case ).manual_seed(snake_case ) UpperCamelCase_ : Union[str, Any] = { 'prompt': 'a dog and a newt', 'mask_image': mask, 'image_latents': latents, 'generator': generator, 'num_inference_steps': 2, 'inpaint_strength': 1.0, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , snake_case : Union[str, Any] , snake_case : Dict=0 ) -> str: """simple docstring""" UpperCamelCase_ : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(snake_case ) ).to(snake_case ) UpperCamelCase_ : int = image.cpu().permute(0 , 2 , 3 , 1 )[0] UpperCamelCase_ : Optional[int] = Image.fromarray(np.uinta(snake_case ) ).convert('RGB' ) if str(snake_case ).startswith('mps' ): UpperCamelCase_ : int = torch.manual_seed(snake_case ) else: UpperCamelCase_ : Optional[int] = torch.Generator(device=snake_case ).manual_seed(snake_case ) UpperCamelCase_ : List[str] = { 'image': image, 'source_prompt': 'a cat and a frog', 'target_prompt': 'a dog and a newt', 'generator': generator, 'num_inference_steps': 2, 'num_maps_per_mask': 2, 'mask_encode_strength': 1.0, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def SCREAMING_SNAKE_CASE__ ( self : List[str] , snake_case : str , snake_case : Dict=0 ) -> int: """simple docstring""" UpperCamelCase_ : List[Any] = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(snake_case ) ).to(snake_case ) UpperCamelCase_ : Any = image.cpu().permute(0 , 2 , 3 , 1 )[0] UpperCamelCase_ : List[str] = Image.fromarray(np.uinta(snake_case ) ).convert('RGB' ) if str(snake_case ).startswith('mps' ): UpperCamelCase_ : Any = torch.manual_seed(snake_case ) else: UpperCamelCase_ : int = torch.Generator(device=snake_case ).manual_seed(snake_case ) UpperCamelCase_ : int = { 'image': image, 'prompt': 'a cat and a frog', 'generator': generator, 'num_inference_steps': 2, 'inpaint_strength': 1.0, 'guidance_scale': 6.0, 'decode_latents': True, 'output_type': 'numpy', } return inputs def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" if not hasattr(self.pipeline_class , '_optional_components' ): return UpperCamelCase_ : int = self.get_dummy_components() UpperCamelCase_ : str = self.pipeline_class(**snake_case ) pipe.to(snake_case ) pipe.set_progress_bar_config(disable=snake_case ) # set all optional components to None and update pipeline config accordingly for optional_component in pipe._optional_components: setattr(snake_case , snake_case , snake_case ) pipe.register_modules(**{optional_component: None for optional_component in pipe._optional_components} ) UpperCamelCase_ : Any = self.get_dummy_inputs(snake_case ) UpperCamelCase_ : Any = pipe(**snake_case )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(snake_case ) UpperCamelCase_ : int = self.pipeline_class.from_pretrained(snake_case ) pipe_loaded.to(snake_case ) pipe_loaded.set_progress_bar_config(disable=snake_case ) for optional_component in pipe._optional_components: self.assertTrue( getattr(snake_case , snake_case ) is None , f"`{optional_component}` did not stay set to None after loading." , ) UpperCamelCase_ : Dict = self.get_dummy_inputs(snake_case ) UpperCamelCase_ : int = pipe_loaded(**snake_case )[0] UpperCamelCase_ : int = np.abs(output - output_loaded ).max() self.assertLess(snake_case , 1e-4 ) def SCREAMING_SNAKE_CASE__ ( self : str ) -> Any: """simple docstring""" UpperCamelCase_ : Optional[int] = 'cpu' UpperCamelCase_ : Any = self.get_dummy_components() UpperCamelCase_ : int = self.pipeline_class(**snake_case ) pipe.to(snake_case ) pipe.set_progress_bar_config(disable=snake_case ) UpperCamelCase_ : Union[str, Any] = self.get_dummy_mask_inputs(snake_case ) UpperCamelCase_ : Optional[int] = pipe.generate_mask(**snake_case ) UpperCamelCase_ : List[str] = mask[0, -3:, -3:] self.assertEqual(mask.shape , (1, 1_6, 1_6) ) UpperCamelCase_ : Tuple = np.array([0] * 9 ) UpperCamelCase_ : int = np.abs(mask_slice.flatten() - expected_slice ).max() self.assertLessEqual(snake_case , 1e-3 ) self.assertEqual(mask[0, -3, -4] , 0 ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> List[Any]: """simple docstring""" UpperCamelCase_ : Dict = 'cpu' UpperCamelCase_ : List[str] = self.get_dummy_components() UpperCamelCase_ : List[Any] = self.pipeline_class(**snake_case ) pipe.to(snake_case ) pipe.set_progress_bar_config(disable=snake_case ) UpperCamelCase_ : Optional[int] = self.get_dummy_inversion_inputs(snake_case ) UpperCamelCase_ : Optional[Any] = pipe.invert(**snake_case ).images UpperCamelCase_ : Tuple = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 3_2, 3_2, 3) ) UpperCamelCase_ : str = np.array( [0.5150, 0.5134, 0.5043, 0.5376, 0.4694, 0.51050, 0.5015, 0.4407, 0.4799] , ) UpperCamelCase_ : Any = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(snake_case , 1e-3 ) def SCREAMING_SNAKE_CASE__ ( self : Any ) -> Any: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=5e-3 ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> str: """simple docstring""" UpperCamelCase_ : Optional[Any] = 'cpu' UpperCamelCase_ : Tuple = self.get_dummy_components() UpperCamelCase_ : int = {'beta_start': 0.00085, 'beta_end': 0.012, 'beta_schedule': 'scaled_linear'} UpperCamelCase_ : str = DPMSolverMultistepScheduler(**snake_case ) UpperCamelCase_ : Union[str, Any] = DPMSolverMultistepInverseScheduler(**snake_case ) UpperCamelCase_ : Optional[int] = self.pipeline_class(**snake_case ) pipe.to(snake_case ) pipe.set_progress_bar_config(disable=snake_case ) UpperCamelCase_ : Any = self.get_dummy_inversion_inputs(snake_case ) UpperCamelCase_ : Union[str, Any] = pipe.invert(**snake_case ).images UpperCamelCase_ : Any = image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 3_2, 3_2, 3) ) UpperCamelCase_ : int = np.array( [0.5150, 0.5134, 0.5043, 0.5376, 0.4694, 0.51050, 0.5015, 0.4407, 0.4799] , ) UpperCamelCase_ : Union[str, Any] = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(snake_case , 1e-3 ) @require_torch_gpu @slow class _lowercase ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self : int ) -> Any: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @classmethod def SCREAMING_SNAKE_CASE__ ( cls : Any ) -> Any: """simple docstring""" UpperCamelCase_ : Optional[int] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/diffedit/fruit.png' ) UpperCamelCase_ : Any = raw_image.convert('RGB' ).resize((7_6_8, 7_6_8) ) UpperCamelCase_ : Optional[Any] = raw_image def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" UpperCamelCase_ : Dict = torch.manual_seed(0 ) UpperCamelCase_ : List[str] = StableDiffusionDiffEditPipeline.from_pretrained( 'stabilityai/stable-diffusion-2-1' , safety_checker=snake_case , torch_dtype=torch.floataa ) UpperCamelCase_ : int = DDIMScheduler.from_config(pipe.scheduler.config ) UpperCamelCase_ : Tuple = DDIMInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=snake_case ) UpperCamelCase_ : Optional[int] = 'a bowl of fruit' UpperCamelCase_ : int = 'a bowl of pears' UpperCamelCase_ : Any = pipe.generate_mask( image=self.raw_image , source_prompt=snake_case , target_prompt=snake_case , generator=snake_case , ) UpperCamelCase_ : int = pipe.invert( prompt=snake_case , image=self.raw_image , inpaint_strength=0.7 , generator=snake_case ).latents UpperCamelCase_ : Optional[int] = pipe( prompt=snake_case , mask_image=snake_case , image_latents=snake_case , generator=snake_case , negative_prompt=snake_case , inpaint_strength=0.7 , output_type='numpy' , ).images[0] UpperCamelCase_ : Any = ( np.array( load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/diffedit/pears.png' ).resize((7_6_8, 7_6_8) ) ) / 2_5_5 ) assert np.abs((expected_image - image).max() ) < 5e-1 def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> int: """simple docstring""" UpperCamelCase_ : Union[str, Any] = torch.manual_seed(0 ) UpperCamelCase_ : List[Any] = StableDiffusionDiffEditPipeline.from_pretrained( 'stabilityai/stable-diffusion-2-1' , safety_checker=snake_case , torch_dtype=torch.floataa ) UpperCamelCase_ : Optional[int] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) UpperCamelCase_ : Union[str, Any] = DPMSolverMultistepInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=snake_case ) UpperCamelCase_ : str = 'a bowl of fruit' UpperCamelCase_ : int = 'a bowl of pears' UpperCamelCase_ : Union[str, Any] = pipe.generate_mask( image=self.raw_image , source_prompt=snake_case , target_prompt=snake_case , generator=snake_case , ) UpperCamelCase_ : Any = pipe.invert( prompt=snake_case , image=self.raw_image , inpaint_strength=0.7 , generator=snake_case , num_inference_steps=2_5 , ).latents UpperCamelCase_ : int = pipe( prompt=snake_case , mask_image=snake_case , image_latents=snake_case , generator=snake_case , negative_prompt=snake_case , inpaint_strength=0.7 , num_inference_steps=2_5 , output_type='numpy' , ).images[0] UpperCamelCase_ : str = ( np.array( load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/diffedit/pears.png' ).resize((7_6_8, 7_6_8) ) ) / 2_5_5 ) assert np.abs((expected_image - image).max() ) < 5e-1
175
1
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class __A ( unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: a =[ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertTrue(is_safetensors_compatible(__A ) ) def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: a =[ '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertTrue(is_safetensors_compatible(__A ) ) def SCREAMING_SNAKE_CASE ( self ) -> Any: a =[ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', '''unet/diffusion_pytorch_model.bin''', # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(__A ) ) def SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: a =[ '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', ] self.assertTrue(is_safetensors_compatible(__A ) ) def SCREAMING_SNAKE_CASE ( self ) -> Any: a =[ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', # Removed: 'text_encoder/model.safetensors', '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertFalse(is_safetensors_compatible(__A ) ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: a =[ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] a ='''fp16''' self.assertTrue(is_safetensors_compatible(__A , variant=__A ) ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: a =[ '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] a ='''fp16''' self.assertTrue(is_safetensors_compatible(__A , variant=__A ) ) def SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: # pass variant but use the non-variant filenames a =[ '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] a ='''fp16''' self.assertTrue(is_safetensors_compatible(__A , variant=__A ) ) def SCREAMING_SNAKE_CASE ( self ) -> List[Any]: a =[ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', '''unet/diffusion_pytorch_model.fp16.bin''', # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] a ='''fp16''' self.assertFalse(is_safetensors_compatible(__A , variant=__A ) ) def SCREAMING_SNAKE_CASE ( self ) -> Any: a =[ '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', ] a ='''fp16''' self.assertTrue(is_safetensors_compatible(__A , variant=__A ) ) def SCREAMING_SNAKE_CASE ( self ) -> str: # pass variant but use the non-variant filenames a =[ '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', ] a ='''fp16''' self.assertTrue(is_safetensors_compatible(__A , variant=__A ) ) def SCREAMING_SNAKE_CASE ( self ) -> Any: a =[ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', # 'text_encoder/model.fp16.safetensors', '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] a ='''fp16''' self.assertFalse(is_safetensors_compatible(__A , variant=__A ) )
215
"""simple docstring""" def _A ( lowercase = 2_00_00_00 ): """simple docstring""" a =[0 for i in range(n + 1 )] a =1 a =1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , lowercase ): a =1 a =0 for i in range(lowercase ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(F'{solution() = }')
215
1
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL UpperCAmelCase_ : int = logging.get_logger(__name__) def _A (__a ) -> str: """simple docstring""" if isinstance(__lowerCamelCase , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(__lowerCamelCase , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(__lowerCamelCase ): return [[videos]] raise ValueError(f'Could not make batched video from {videos}' ) class lowerCAmelCase__ ( A_ ): '''simple docstring''' __UpperCamelCase = ["pixel_values"] def __init__( self : Any , lowercase_ : bool = True , lowercase_ : Dict[str, int] = None , lowercase_ : PILImageResampling = PILImageResampling.BILINEAR , lowercase_ : bool = True , lowercase_ : Dict[str, int] = None , lowercase_ : bool = True , lowercase_ : Union[int, float] = 1 / 255 , lowercase_ : bool = True , lowercase_ : Optional[Union[float, List[float]]] = None , lowercase_ : Optional[Union[float, List[float]]] = None , **lowercase_ : int , ): '''simple docstring''' super().__init__(**snake_case__) SCREAMING_SNAKE_CASE_ : Dict = size if size is not None else {"shortest_edge": 224} SCREAMING_SNAKE_CASE_ : Optional[Any] = get_size_dict(snake_case__ , default_to_square=snake_case__) SCREAMING_SNAKE_CASE_ : Optional[Any] = crop_size if crop_size is not None else {"height": 224, "width": 224} SCREAMING_SNAKE_CASE_ : str = get_size_dict(snake_case__ , param_name='''crop_size''') SCREAMING_SNAKE_CASE_ : Any = do_resize SCREAMING_SNAKE_CASE_ : Optional[Any] = size SCREAMING_SNAKE_CASE_ : Tuple = do_center_crop SCREAMING_SNAKE_CASE_ : Dict = crop_size SCREAMING_SNAKE_CASE_ : Tuple = resample SCREAMING_SNAKE_CASE_ : Dict = do_rescale SCREAMING_SNAKE_CASE_ : Any = rescale_factor SCREAMING_SNAKE_CASE_ : Optional[int] = do_normalize SCREAMING_SNAKE_CASE_ : List[str] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN SCREAMING_SNAKE_CASE_ : List[str] = image_std if image_std is not None else IMAGENET_STANDARD_STD def _SCREAMING_SNAKE_CASE ( self : Tuple , lowercase_ : np.ndarray , lowercase_ : Dict[str, int] , lowercase_ : PILImageResampling = PILImageResampling.BILINEAR , lowercase_ : Optional[Union[str, ChannelDimension]] = None , **lowercase_ : Union[str, Any] , ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Optional[int] = get_size_dict(snake_case__ , default_to_square=snake_case__) if "shortest_edge" in size: SCREAMING_SNAKE_CASE_ : Union[str, Any] = get_resize_output_image_size(snake_case__ , size['''shortest_edge'''] , default_to_square=snake_case__) elif "height" in size and "width" in size: SCREAMING_SNAKE_CASE_ : Optional[int] = (size["height"], size["width"]) else: raise ValueError(F'Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}') return resize(snake_case__ , size=snake_case__ , resample=snake_case__ , data_format=snake_case__ , **snake_case__) def _SCREAMING_SNAKE_CASE ( self : Tuple , lowercase_ : np.ndarray , lowercase_ : Dict[str, int] , lowercase_ : Optional[Union[str, ChannelDimension]] = None , **lowercase_ : Any , ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Optional[Any] = get_size_dict(snake_case__) if "height" not in size or "width" not in size: raise ValueError(F'Size must have \'height\' and \'width\' as keys. Got {size.keys()}') return center_crop(snake_case__ , size=(size['''height'''], size['''width''']) , data_format=snake_case__ , **snake_case__) def _SCREAMING_SNAKE_CASE ( self : int , lowercase_ : np.ndarray , lowercase_ : Union[int, float] , lowercase_ : Optional[Union[str, ChannelDimension]] = None , **lowercase_ : List[str] , ): '''simple docstring''' return rescale(snake_case__ , scale=snake_case__ , data_format=snake_case__ , **snake_case__) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowercase_ : np.ndarray , lowercase_ : Union[float, List[float]] , lowercase_ : Union[float, List[float]] , lowercase_ : Optional[Union[str, ChannelDimension]] = None , **lowercase_ : List[str] , ): '''simple docstring''' return normalize(snake_case__ , mean=snake_case__ , std=snake_case__ , data_format=snake_case__ , **snake_case__) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowercase_ : ImageInput , lowercase_ : bool = None , lowercase_ : Dict[str, int] = None , lowercase_ : PILImageResampling = None , lowercase_ : bool = None , lowercase_ : Dict[str, int] = None , lowercase_ : bool = None , lowercase_ : float = None , lowercase_ : bool = None , lowercase_ : Optional[Union[float, List[float]]] = None , lowercase_ : Optional[Union[float, List[float]]] = None , lowercase_ : Optional[ChannelDimension] = ChannelDimension.FIRST , ): '''simple docstring''' if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''') if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''') if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''') if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''') # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE_ : List[Any] = to_numpy_array(snake_case__) if do_resize: SCREAMING_SNAKE_CASE_ : int = self.resize(image=snake_case__ , size=snake_case__ , resample=snake_case__) if do_center_crop: SCREAMING_SNAKE_CASE_ : List[str] = self.center_crop(snake_case__ , size=snake_case__) if do_rescale: SCREAMING_SNAKE_CASE_ : int = self.rescale(image=snake_case__ , scale=snake_case__) if do_normalize: SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.normalize(image=snake_case__ , mean=snake_case__ , std=snake_case__) SCREAMING_SNAKE_CASE_ : List[str] = to_channel_dimension_format(snake_case__ , snake_case__) return image def _SCREAMING_SNAKE_CASE ( self : Dict , lowercase_ : ImageInput , lowercase_ : bool = None , lowercase_ : Dict[str, int] = None , lowercase_ : PILImageResampling = None , lowercase_ : bool = None , lowercase_ : Dict[str, int] = None , lowercase_ : bool = None , lowercase_ : float = None , lowercase_ : bool = None , lowercase_ : Optional[Union[float, List[float]]] = None , lowercase_ : Optional[Union[float, List[float]]] = None , lowercase_ : Optional[Union[str, TensorType]] = None , lowercase_ : ChannelDimension = ChannelDimension.FIRST , **lowercase_ : List[Any] , ): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Union[str, Any] = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE_ : Union[str, Any] = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE_ : List[str] = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE_ : Dict = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE_ : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE_ : Any = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE_ : Tuple = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE_ : Tuple = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE_ : Dict = size if size is not None else self.size SCREAMING_SNAKE_CASE_ : Any = get_size_dict(snake_case__ , default_to_square=snake_case__) SCREAMING_SNAKE_CASE_ : Union[str, Any] = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE_ : List[Any] = get_size_dict(snake_case__ , param_name='''crop_size''') if not valid_images(snake_case__): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''') SCREAMING_SNAKE_CASE_ : Tuple = make_batched(snake_case__) SCREAMING_SNAKE_CASE_ : Optional[Any] = [ [ self._preprocess_image( image=snake_case__ , do_resize=snake_case__ , size=snake_case__ , resample=snake_case__ , do_center_crop=snake_case__ , crop_size=snake_case__ , do_rescale=snake_case__ , rescale_factor=snake_case__ , do_normalize=snake_case__ , image_mean=snake_case__ , image_std=snake_case__ , data_format=snake_case__ , ) for img in video ] for video in videos ] SCREAMING_SNAKE_CASE_ : List[str] = {"pixel_values": videos} return BatchFeature(data=snake_case__ , tensor_type=snake_case__)
91
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer __lowerCamelCase = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast __lowerCamelCase = TaTokenizerFast __lowerCamelCase = {"""configuration_mt5""": ["""MT5Config""", """MT5OnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = [ """MT5EncoderModel""", """MT5ForConditionalGeneration""", """MT5ForQuestionAnswering""", """MT5Model""", """MT5PreTrainedModel""", """MT5Stack""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = ["""TFMT5EncoderModel""", """TFMT5ForConditionalGeneration""", """TFMT5Model"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = ["""FlaxMT5EncoderModel""", """FlaxMT5ForConditionalGeneration""", """FlaxMT5Model"""] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys __lowerCamelCase = _LazyModule( __name__, globals()["""__file__"""], _import_structure, extra_objects={"""MT5Tokenizer""": MTaTokenizer, """MT5TokenizerFast""": MTaTokenizerFast}, module_spec=__spec__, )
59
0
import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py UpperCAmelCase__ : Any = '.' # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) UpperCAmelCase__ : Optional[Any] = [ 'Assert', 'AssignVariableOp', 'EmptyTensorList', 'MergeV2Checkpoints', 'ReadVariableOp', 'ResourceGather', 'RestoreV2', 'SaveV2', 'ShardedFilename', 'StatefulPartitionedCall', 'StaticRegexFullMatch', 'VarHandleOp', ] def lowerCamelCase__ ( a , a , a ) -> Union[str, Any]: _A: Optional[int] = SavedModel() _A: int = [] with open(os.path.join(a , '''utils''' , '''tf_ops''' , '''onnx.json''' ) ) as f: _A: List[Any] = json.load(a )['''opsets'''] for i in range(1 , opset + 1 ): onnx_ops.extend(onnx_opsets[str(a )] ) with open(a , '''rb''' ) as f: saved_model.ParseFromString(f.read() ) _A: Optional[Any] = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want _A: Optional[int] = sorted(a ) _A: Tuple = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(a ) if strict and len(a ) > 0: raise Exception(f"""Found the following incompatible ops for the opset {opset}:\n""" + incompatible_ops ) elif len(a ) > 0: print(f"""Found the following incompatible ops for the opset {opset}:""" ) print(*a , sep='''\n''' ) else: print(f"""The saved model {saved_model_path} can properly be converted with ONNX.""" ) if __name__ == "__main__": UpperCAmelCase__ : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('--saved_model_path', help='Path of the saved model to check (the .pb file).') parser.add_argument( '--opset', default=12, type=int, help='The ONNX opset against which the model has to be tested.' ) parser.add_argument( '--framework', choices=['onnx'], default='onnx', help='Frameworks against which to test the saved model.' ) parser.add_argument( '--strict', action='store_true', help='Whether make the checking strict (raise errors) or not (raise warnings)' ) UpperCAmelCase__ : int = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
301
import argparse import json import os from tensorflow.core.protobuf.saved_model_pba import SavedModel # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py UpperCAmelCase__ : Any = '.' # Internal TensorFlow ops that can be safely ignored (mostly specific to a saved model) UpperCAmelCase__ : Optional[Any] = [ 'Assert', 'AssignVariableOp', 'EmptyTensorList', 'MergeV2Checkpoints', 'ReadVariableOp', 'ResourceGather', 'RestoreV2', 'SaveV2', 'ShardedFilename', 'StatefulPartitionedCall', 'StaticRegexFullMatch', 'VarHandleOp', ] def lowerCamelCase__ ( a , a , a ) -> Union[str, Any]: _A: Optional[int] = SavedModel() _A: int = [] with open(os.path.join(a , '''utils''' , '''tf_ops''' , '''onnx.json''' ) ) as f: _A: List[Any] = json.load(a )['''opsets'''] for i in range(1 , opset + 1 ): onnx_ops.extend(onnx_opsets[str(a )] ) with open(a , '''rb''' ) as f: saved_model.ParseFromString(f.read() ) _A: Optional[Any] = set() # Iterate over every metagraph in case there is more than one (a saved model can contain multiple graphs) for meta_graph in saved_model.meta_graphs: # Add operations in the graph definition model_op_names.update(node.op for node in meta_graph.graph_def.node ) # Go through the functions in the graph definition for func in meta_graph.graph_def.library.function: # Add operations in each function model_op_names.update(node.op for node in func.node_def ) # Convert to list, sorted if you want _A: Optional[int] = sorted(a ) _A: Tuple = [] for op in model_op_names: if op not in onnx_ops and op not in INTERNAL_OPS: incompatible_ops.append(a ) if strict and len(a ) > 0: raise Exception(f"""Found the following incompatible ops for the opset {opset}:\n""" + incompatible_ops ) elif len(a ) > 0: print(f"""Found the following incompatible ops for the opset {opset}:""" ) print(*a , sep='''\n''' ) else: print(f"""The saved model {saved_model_path} can properly be converted with ONNX.""" ) if __name__ == "__main__": UpperCAmelCase__ : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('--saved_model_path', help='Path of the saved model to check (the .pb file).') parser.add_argument( '--opset', default=12, type=int, help='The ONNX opset against which the model has to be tested.' ) parser.add_argument( '--framework', choices=['onnx'], default='onnx', help='Frameworks against which to test the saved model.' ) parser.add_argument( '--strict', action='store_true', help='Whether make the checking strict (raise errors) or not (raise warnings)' ) UpperCAmelCase__ : int = parser.parse_args() if args.framework == "onnx": onnx_compliancy(args.saved_model_path, args.strict, args.opset)
301
1
'''simple docstring''' import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class __SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self : Tuple , __a : Dict , __a : Optional[int]=13 , __a : Union[str, Any]=3 , __a : List[str]=True , __a : Dict=True , __a : str=0.1 , __a : Union[str, Any]=0.1 , __a : Optional[Any]=2_24 , __a : List[Any]=10_00 , __a : Dict=[3, 3, 6, 4] , __a : Optional[int]=[48, 56, 1_12, 2_20] , ): _a = parent _a = batch_size _a = num_channels _a = is_training _a = use_labels _a = hidden_dropout_prob _a = attention_probs_dropout_prob _a = num_labels _a = image_size _a = layer_depths _a = embed_dims def UpperCamelCase__ ( self : List[str] ): _a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _a = None if self.use_labels: _a = ids_tensor([self.batch_size] , self.num_labels ) _a = self.get_config() return config, pixel_values, labels def UpperCamelCase__ ( self : Optional[int] ): return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act="gelu" , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=lowerCamelCase_ , layer_scale_init_value=1e-5 , ) def UpperCamelCase__ ( self : str , __a : Tuple , __a : Union[str, Any] , __a : Tuple ): _a = SwiftFormerModel(config=lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() _a = model(lowerCamelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7) ) def UpperCamelCase__ ( self : Optional[Any] , __a : List[str] , __a : List[Any] , __a : Dict ): _a = self.num_labels _a = SwiftFormerForImageClassification(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() _a = model(lowerCamelCase_ , labels=lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) _a = SwiftFormerForImageClassification(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() _a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _a = model(lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase__ ( self : Dict ): ((_a) , (_a) , (_a)) = self.prepare_config_and_inputs() _a = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): """simple docstring""" __a =(SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () __a =( {'feature-extraction': SwiftFormerModel, 'image-classification': SwiftFormerForImageClassification} if is_torch_available() else {} ) __a =False __a =False __a =False __a =False __a =False def UpperCamelCase__ ( self : Union[str, Any] ): _a = SwiftFormerModelTester(self ) _a = ConfigTester( self , config_class=lowerCamelCase_ , has_text_modality=lowerCamelCase_ , hidden_size=37 , num_attention_heads=12 , num_hidden_layers=12 , ) def UpperCamelCase__ ( self : Optional[Any] ): self.config_tester.run_common_tests() @unittest.skip(reason="SwiftFormer does not use inputs_embeds" ) def UpperCamelCase__ ( self : Tuple ): pass def UpperCamelCase__ ( self : Dict ): _a , _a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a = model_class(lowerCamelCase_ ) _a = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase_ , nn.Linear ) ) def UpperCamelCase__ ( self : Optional[int] ): _a , _a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a = model_class(lowerCamelCase_ ) _a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a = [*signature.parameters.keys()] _a = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase_ ) def UpperCamelCase__ ( self : Dict ): _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase_ ) def UpperCamelCase__ ( self : int ): _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase_ ) @slow def UpperCamelCase__ ( self : Tuple ): for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a = SwiftFormerModel.from_pretrained(lowerCamelCase_ ) self.assertIsNotNone(lowerCamelCase_ ) @unittest.skip(reason="SwiftFormer does not output attentions" ) def UpperCamelCase__ ( self : Union[str, Any] ): pass def UpperCamelCase__ ( self : List[str] ): def check_hidden_states_output(__a : int , __a : List[str] , __a : int ): _a = model_class(lowerCamelCase_ ) model.to(lowerCamelCase_ ) model.eval() with torch.no_grad(): _a = model(**self._prepare_for_class(lowerCamelCase_ , lowerCamelCase_ ) ) _a = outputs.hidden_states _a = 8 self.assertEqual(len(lowerCamelCase_ ) , lowerCamelCase_ ) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(lowerCamelCase_ ) ): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ] ) , ) _a , _a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a = True check_hidden_states_output(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _a = True check_hidden_states_output(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) def UpperCamelCase__ ( self : str ): def _config_zero_init(__a : List[Any] ): _a = copy.deepcopy(lowerCamelCase_ ) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(lowerCamelCase_ , lowerCamelCase_ , 1e-1_0 ) if isinstance(getattr(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) , lowerCamelCase_ ): _a = _config_zero_init(getattr(lowerCamelCase_ , lowerCamelCase_ ) ) setattr(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) return configs_no_init _a , _a = self.model_tester.prepare_config_and_inputs_for_common() _a = _config_zero_init(lowerCamelCase_ ) for model_class in self.all_model_classes: _a = model_class(config=lowerCamelCase_ ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9) / 1e9).round().item() , [0.0, 1.0] , msg=f'Parameter {name} of model {model_class} seems not properly initialized' , ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def UpperCamelCase__ ( self : Any ): pass def _lowerCamelCase ( ) -> Any: _a = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class __SCREAMING_SNAKE_CASE (unittest.TestCase ): """simple docstring""" @cached_property def UpperCamelCase__ ( self : Union[str, Any] ): return ViTImageProcessor.from_pretrained("MBZUAI/swiftformer-xs" ) if is_vision_available() else None @slow def UpperCamelCase__ ( self : Optional[int] ): _a = SwiftFormerForImageClassification.from_pretrained("MBZUAI/swiftformer-xs" ).to(lowerCamelCase_ ) _a = self.default_image_processor _a = prepare_img() _a = image_processor(images=lowerCamelCase_ , return_tensors="pt" ).to(lowerCamelCase_ ) # forward pass with torch.no_grad(): _a = model(**lowerCamelCase_ ) # verify the logits _a = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , lowerCamelCase_ ) _a = torch.tensor([[-2.1_7_0_3e0_0, 2.1_1_0_7e0_0, -2.0_8_1_1e0_0]] ).to(lowerCamelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase_ , atol=1e-4 ) )
63
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE = { """xlnet-base-cased""": """https://huggingface.co/xlnet-base-cased/resolve/main/config.json""", """xlnet-large-cased""": """https://huggingface.co/xlnet-large-cased/resolve/main/config.json""", } class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): __lowerCAmelCase = """xlnet""" __lowerCAmelCase = ["""mems"""] __lowerCAmelCase = { """n_token""": """vocab_size""", # Backward compatibility """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : Any , lowerCamelCase_ : Any=3_2000 , lowerCamelCase_ : Dict=1024 , lowerCamelCase_ : List[str]=24 , lowerCamelCase_ : Dict=16 , lowerCamelCase_ : List[Any]=4096 , lowerCamelCase_ : List[Any]="gelu" , lowerCamelCase_ : List[Any]=True , lowerCamelCase_ : Union[str, Any]="bi" , lowerCamelCase_ : Optional[Any]=0.0_2 , lowerCamelCase_ : Optional[int]=1E-12 , lowerCamelCase_ : List[Any]=0.1 , lowerCamelCase_ : Union[str, Any]=512 , lowerCamelCase_ : Any=None , lowerCamelCase_ : str=True , lowerCamelCase_ : List[str]=False , lowerCamelCase_ : List[str]=False , lowerCamelCase_ : Optional[int]=-1 , lowerCamelCase_ : List[str]=False , lowerCamelCase_ : Union[str, Any]="last" , lowerCamelCase_ : List[str]=True , lowerCamelCase_ : str="tanh" , lowerCamelCase_ : Any=0.1 , lowerCamelCase_ : Dict=5 , lowerCamelCase_ : str=5 , lowerCamelCase_ : Optional[int]=5 , lowerCamelCase_ : Any=1 , lowerCamelCase_ : int=2 , **lowerCamelCase_ : List[Any] , ): """simple docstring""" UpperCamelCase = vocab_size UpperCamelCase = d_model UpperCamelCase = n_layer UpperCamelCase = n_head if d_model % n_head != 0: raise ValueError(f"""'d_model % n_head' ({d_model % n_head}) should be equal to 0""" ) if "d_head" in kwargs: if kwargs["d_head"] != d_model // n_head: raise ValueError( f"""`d_head` ({kwargs["d_head"]}) should be equal to `d_model // n_head` ({d_model // n_head})""" ) UpperCamelCase = d_model // n_head UpperCamelCase = ff_activation UpperCamelCase = d_inner UpperCamelCase = untie_r UpperCamelCase = attn_type UpperCamelCase = initializer_range UpperCamelCase = layer_norm_eps UpperCamelCase = dropout UpperCamelCase = mem_len UpperCamelCase = reuse_len UpperCamelCase = bi_data UpperCamelCase = clamp_len UpperCamelCase = same_length UpperCamelCase = summary_type UpperCamelCase = summary_use_proj UpperCamelCase = summary_activation UpperCamelCase = summary_last_dropout UpperCamelCase = start_n_top UpperCamelCase = end_n_top UpperCamelCase = bos_token_id UpperCamelCase = pad_token_id UpperCamelCase = eos_token_id if "use_cache" in kwargs: warnings.warn( """The `use_cache` argument is deprecated and will be removed in a future version, use `use_mems_eval`""" """ instead.""" , lowerCamelCase_ , ) UpperCamelCase = kwargs["""use_cache"""] UpperCamelCase = use_mems_eval UpperCamelCase = use_mems_train super().__init__(pad_token_id=lowerCamelCase_ , bos_token_id=lowerCamelCase_ , eos_token_id=lowerCamelCase_ , **lowerCamelCase_ ) @property def lowerCamelCase_ ( self : Dict ): """simple docstring""" logger.info(f"""The model {self.model_type} is one of the few models that has no sequence length limit.""" ) return -1 @max_position_embeddings.setter def lowerCamelCase_ ( self : Optional[int] , lowerCamelCase_ : Optional[int] ): """simple docstring""" raise NotImplementedError( f"""The model {self.model_type} is one of the few models that has no sequence length limit.""" )
343
0
from __future__ import annotations from collections.abc import Sequence from typing import Literal def A__ ( __lowerCamelCase, __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = list(__lowerCamelCase ) SCREAMING_SNAKE_CASE_ = list(__lowerCamelCase ) SCREAMING_SNAKE_CASE_ = 0 for i in range(len(__lowerCamelCase ) ): if lista[i] != lista[i]: count += 1 SCREAMING_SNAKE_CASE_ = '''_''' if count > 1: return False else: return "".join(__lowerCamelCase ) def A__ ( __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = [] while True: SCREAMING_SNAKE_CASE_ = ['''$'''] * len(__lowerCamelCase ) SCREAMING_SNAKE_CASE_ = [] for i in range(len(__lowerCamelCase ) ): for j in range(i + 1, len(__lowerCamelCase ) ): SCREAMING_SNAKE_CASE_ = compare_string(binary[i], binary[j] ) if k is False: SCREAMING_SNAKE_CASE_ = '''*''' SCREAMING_SNAKE_CASE_ = '''*''' temp.append('''X''' ) for i in range(len(__lowerCamelCase ) ): if checka[i] == "$": pi.append(binary[i] ) if len(__lowerCamelCase ) == 0: return pi SCREAMING_SNAKE_CASE_ = list(set(__lowerCamelCase ) ) def A__ ( __lowerCamelCase, __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = [] for minterm in minterms: SCREAMING_SNAKE_CASE_ = '''''' for _ in range(__lowerCamelCase ): SCREAMING_SNAKE_CASE_ = str(minterm % 2 ) + string minterm //= 2 temp.append(__lowerCamelCase ) return temp def A__ ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = list(__lowerCamelCase ) SCREAMING_SNAKE_CASE_ = list(__lowerCamelCase ) SCREAMING_SNAKE_CASE_ = 0 for i in range(len(__lowerCamelCase ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def A__ ( __lowerCamelCase, __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = [0] * len(__lowerCamelCase ) for i in range(len(chart[0] ) ): SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = -1 for j in range(len(__lowerCamelCase ) ): if chart[j][i] == 1: count += 1 SCREAMING_SNAKE_CASE_ = j if count == 1: SCREAMING_SNAKE_CASE_ = 1 for i in range(len(__lowerCamelCase ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(__lowerCamelCase ) ): SCREAMING_SNAKE_CASE_ = 0 temp.append(prime_implicants[i] ) while True: SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = -1 SCREAMING_SNAKE_CASE_ = 0 for i in range(len(__lowerCamelCase ) ): SCREAMING_SNAKE_CASE_ = chart[i].count(1 ) if count_n > max_n: SCREAMING_SNAKE_CASE_ = count_n SCREAMING_SNAKE_CASE_ = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(__lowerCamelCase ) ): SCREAMING_SNAKE_CASE_ = 0 def A__ ( __lowerCamelCase, __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = [[0 for x in range(len(__lowerCamelCase ) )] for x in range(len(__lowerCamelCase ) )] for i in range(len(__lowerCamelCase ) ): SCREAMING_SNAKE_CASE_ = prime_implicants[i].count('''_''' ) for j in range(len(__lowerCamelCase ) ): if is_for_table(prime_implicants[i], binary[j], __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = 1 return chart def A__ ( ): SCREAMING_SNAKE_CASE_ = int(input('''Enter the no. of variables\n''' ) ) SCREAMING_SNAKE_CASE_ = [ float(__lowerCamelCase ) for x in input( '''Enter the decimal representation of Minterms \'Spaces Separated\'\n''' ).split() ] SCREAMING_SNAKE_CASE_ = decimal_to_binary(__lowerCamelCase, __lowerCamelCase ) SCREAMING_SNAKE_CASE_ = check(__lowerCamelCase ) print('''Prime Implicants are:''' ) print(__lowerCamelCase ) SCREAMING_SNAKE_CASE_ = prime_implicant_chart(__lowerCamelCase, __lowerCamelCase ) SCREAMING_SNAKE_CASE_ = selection(__lowerCamelCase, __lowerCamelCase ) print('''Essential Prime Implicants are:''' ) print(__lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
257
import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, require_tf, slow if is_tf_available(): import tensorflow as tf if is_tensorflow_text_available(): from transformers.models.bert import TFBertTokenizer __UpperCAmelCase = ["bert-base-uncased", "bert-base-cased"] __UpperCAmelCase = "hf-internal-testing/tiny-bert-tf-only" if is_tf_available(): class UpperCamelCase__ ( tf.keras.Model ): """simple docstring""" def __init__( self , _A ) -> int: super().__init__() SCREAMING_SNAKE_CASE_ = tokenizer SCREAMING_SNAKE_CASE_ = AutoConfig.from_pretrained(_A ) SCREAMING_SNAKE_CASE_ = TFAutoModel.from_config(_A ) def _UpperCamelCase ( self , _A ) -> Optional[Any]: SCREAMING_SNAKE_CASE_ = self.tokenizer(_A ) SCREAMING_SNAKE_CASE_ = self.bert(**_A ) return out["pooler_output"] @require_tf @require_tensorflow_text class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def _UpperCamelCase ( self ) -> Optional[Any]: super().setUp() SCREAMING_SNAKE_CASE_ = [ BertTokenizer.from_pretrained(_A ) for checkpoint in (TOKENIZER_CHECKPOINTS * 2) ] # repeat for when fast_bert_tokenizer=false SCREAMING_SNAKE_CASE_ = [TFBertTokenizer.from_pretrained(_A ) for checkpoint in TOKENIZER_CHECKPOINTS] + [ TFBertTokenizer.from_pretrained(_A , use_fast_bert_tokenizer=_A ) for checkpoint in TOKENIZER_CHECKPOINTS ] assert len(self.tokenizers ) == len(self.tf_tokenizers ) SCREAMING_SNAKE_CASE_ = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] SCREAMING_SNAKE_CASE_ = list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def _UpperCamelCase ( self ) -> str: for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in (self.test_sentences, self.paired_sentences): SCREAMING_SNAKE_CASE_ = tokenizer(_A , return_tensors='''tf''' , padding='''longest''' ) SCREAMING_SNAKE_CASE_ = tf_tokenizer(_A ) for key in python_outputs.keys(): self.assertTrue(tf.reduce_all(python_outputs[key].shape == tf_outputs[key].shape ) ) self.assertTrue(tf.reduce_all(tf.cast(python_outputs[key] , tf.intaa ) == tf_outputs[key] ) ) @slow def _UpperCamelCase ( self ) -> Dict: for tf_tokenizer in self.tf_tokenizers: SCREAMING_SNAKE_CASE_ = tf_tokenizer(self.paired_sentences ) SCREAMING_SNAKE_CASE_ = tf_tokenizer( text=[sentence[0] for sentence in self.paired_sentences] , text_pair=[sentence[1] for sentence in self.paired_sentences] , ) for key in merged_outputs.keys(): self.assertTrue(tf.reduce_all(tf.cast(merged_outputs[key] , tf.intaa ) == separated_outputs[key] ) ) @slow def _UpperCamelCase ( self ) -> int: for tf_tokenizer in self.tf_tokenizers: SCREAMING_SNAKE_CASE_ = tf.function(_A ) for test_inputs in (self.test_sentences, self.paired_sentences): SCREAMING_SNAKE_CASE_ = tf.constant(_A ) SCREAMING_SNAKE_CASE_ = compiled_tokenizer(_A ) SCREAMING_SNAKE_CASE_ = tf_tokenizer(_A ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def _UpperCamelCase ( self ) -> List[str]: for tf_tokenizer in self.tf_tokenizers: SCREAMING_SNAKE_CASE_ = ModelToSave(tokenizer=_A ) SCREAMING_SNAKE_CASE_ = tf.convert_to_tensor(self.test_sentences ) SCREAMING_SNAKE_CASE_ = model(_A ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: SCREAMING_SNAKE_CASE_ = Path(_A ) / '''saved.model''' model.save(_A ) SCREAMING_SNAKE_CASE_ = tf.keras.models.load_model(_A ) SCREAMING_SNAKE_CASE_ = loaded_model(_A ) # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertLessEqual(tf.reduce_max(tf.abs(out - loaded_output ) ) , 1E-5 )
257
1
'''simple docstring''' from __future__ import annotations from statistics import mean def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : List[Any] = [0] * no_of_processes _UpperCamelCase : Union[str, Any] = [0] * no_of_processes # Initialize remaining_time to waiting_time. for i in range(UpperCAmelCase_ ): _UpperCamelCase : List[Any] = burst_time[i] _UpperCamelCase : list[int] = [] _UpperCamelCase : List[str] = 0 _UpperCamelCase : Optional[Any] = 0 # When processes are not completed, # A process whose arrival time has passed \ # and has remaining execution time is put into the ready_process. # The shortest process in the ready_process, target_process is executed. while completed != no_of_processes: _UpperCamelCase : str = [] _UpperCamelCase : str = -1 for i in range(UpperCAmelCase_ ): if (arrival_time[i] <= total_time) and (remaining_time[i] > 0): ready_process.append(UpperCAmelCase_ ) if len(UpperCAmelCase_ ) > 0: _UpperCamelCase : Optional[Any] = ready_process[0] for i in ready_process: if remaining_time[i] < remaining_time[target_process]: _UpperCamelCase : Optional[int] = i total_time += burst_time[target_process] completed += 1 _UpperCamelCase : List[Any] = 0 _UpperCamelCase : int = ( total_time - arrival_time[target_process] - burst_time[target_process] ) else: total_time += 1 return waiting_time def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : Optional[int] = [0] * no_of_processes for i in range(UpperCAmelCase_ ): _UpperCamelCase : Tuple = burst_time[i] + waiting_time[i] return turn_around_time if __name__ == "__main__": print('[TEST CASE 01]') snake_case_ : List[Any] = 4 snake_case_ : List[str] = [2, 5, 3, 7] snake_case_ : List[Any] = [0, 0, 0, 0] snake_case_ : Tuple = calculate_waitingtime(arrival_time, burst_time, no_of_processes) snake_case_ : List[Any] = calculate_turnaroundtime( burst_time, no_of_processes, waiting_time ) # Printing the Result print('PID\tBurst Time\tArrival Time\tWaiting Time\tTurnaround Time') for i, process_id in enumerate(list(range(1, 5))): print( F"""{process_id}\t{burst_time[i]}\t\t\t{arrival_time[i]}\t\t\t\t""" F"""{waiting_time[i]}\t\t\t\t{turn_around_time[i]}""" ) print(F"""\nAverage waiting time = {mean(waiting_time):.5f}""") print(F"""Average turnaround time = {mean(turn_around_time):.5f}""")
83
'''simple docstring''' def A__ ( UpperCAmelCase_ = 1_0_0_0 ): _UpperCamelCase : Dict = 3 _UpperCamelCase : Any = 0 while a < n: if a % 3 == 0 or a % 5 == 0: result += a elif a % 1_5 == 0: result -= a a += 1 return result if __name__ == "__main__": print(F"""{solution() = }""")
83
1
"""simple docstring""" import os import unittest from huggingface_hub.utils import are_progress_bars_disabled import transformers.models.bart.tokenization_bart from transformers import logging from transformers.testing_utils import CaptureLogger, mockenv, mockenv_context from transformers.utils.logging import disable_progress_bar, enable_progress_bar class lowercase__ ( unittest.TestCase ): '''simple docstring''' def lowercase__ ( self : List[str] ) -> int: '''simple docstring''' UpperCAmelCase_ = logging.get_logger() # the current default level is logging.WARNING UpperCAmelCase_ = logging.get_verbosity() logging.set_verbosity_error() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_warning() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_info() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_debug() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) # restore to the original level logging.set_verbosity(_UpperCAmelCase ) def lowercase__ ( self : List[Any] ) -> List[Any]: '''simple docstring''' UpperCAmelCase_ = logging.get_verbosity() UpperCAmelCase_ = logging.get_logger("transformers.models.bart.tokenization_bart" ) UpperCAmelCase_ = "Testing 1, 2, 3" # should be able to log warnings (if default settings weren't overridden by `pytest --log-level-all`) if level_origin <= logging.WARNING: with CaptureLogger(_UpperCAmelCase ) as cl: logger.warning(_UpperCAmelCase ) self.assertEqual(cl.out , msg + "\n" ) # this is setting the level for all of `transformers.*` loggers logging.set_verbosity_error() # should not be able to log warnings with CaptureLogger(_UpperCAmelCase ) as cl: logger.warning(_UpperCAmelCase ) self.assertEqual(cl.out , "" ) # should be able to log warnings again logging.set_verbosity_warning() with CaptureLogger(_UpperCAmelCase ) as cl: logger.warning(_UpperCAmelCase ) self.assertEqual(cl.out , msg + "\n" ) # restore to the original level logging.set_verbosity(_UpperCAmelCase ) @mockenv(TRANSFORMERS_VERBOSITY="error" ) def lowercase__ ( self : str ) -> Union[str, Any]: '''simple docstring''' transformers.utils.logging._reset_library_root_logger() # this action activates the env var UpperCAmelCase_ = logging.get_logger("transformers.models.bart.tokenization_bart" ) UpperCAmelCase_ = os.getenv("TRANSFORMERS_VERBOSITY" , _UpperCAmelCase ) UpperCAmelCase_ = logging.log_levels[env_level_str] UpperCAmelCase_ = logging.get_verbosity() self.assertEqual( _UpperCAmelCase , _UpperCAmelCase , F"""TRANSFORMERS_VERBOSITY={env_level_str}/{env_level}, but internal verbosity is {current_level}""" , ) # restore to the original level UpperCAmelCase_ = "" transformers.utils.logging._reset_library_root_logger() @mockenv(TRANSFORMERS_VERBOSITY="super-error" ) def lowercase__ ( self : List[str] ) -> int: '''simple docstring''' transformers.utils.logging._reset_library_root_logger() UpperCAmelCase_ = logging.logging.getLogger() with CaptureLogger(_UpperCAmelCase ) as cl: # this action activates the env var logging.get_logger("transformers.models.bart.tokenization_bart" ) self.assertIn("Unknown option TRANSFORMERS_VERBOSITY=super-error" , cl.out ) # no need to restore as nothing was changed def lowercase__ ( self : List[Any] ) -> List[str]: '''simple docstring''' transformers.utils.logging._reset_library_root_logger() UpperCAmelCase_ = logging.get_logger("transformers.models.bart.tokenization_bart" ) UpperCAmelCase_ = "Testing 1, 2, 3" with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS="1" ): # nothing should be logged as env var disables this method with CaptureLogger(_UpperCAmelCase ) as cl: logger.warning_advice(_UpperCAmelCase ) self.assertEqual(cl.out , "" ) with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS="" ): # should log normally as TRANSFORMERS_NO_ADVISORY_WARNINGS is unset with CaptureLogger(_UpperCAmelCase ) as cl: logger.warning_advice(_UpperCAmelCase ) self.assertEqual(cl.out , msg + "\n" ) def a__ ( ): disable_progress_bar() assert are_progress_bars_disabled() enable_progress_bar() assert not are_progress_bars_disabled()
241
"""simple docstring""" import glob import os import random from string import ascii_lowercase, digits import cva lowerCamelCase = """""" lowerCamelCase = """""" lowerCamelCase = """""" lowerCamelCase = 1 # (0 is vertical, 1 is horizontal) def a__ ( ): UpperCAmelCase_ , UpperCAmelCase_ = get_dataset(lowerCAmelCase__ , lowerCAmelCase__ ) print("Processing..." ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = update_image_and_anno(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) for index, image in enumerate(lowerCAmelCase__ ): # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' UpperCAmelCase_ = random_chars(32 ) UpperCAmelCase_ = paths[index].split(os.sep )[-1].rsplit("." , 1 )[0] UpperCAmelCase_ = f"""{OUTPUT_DIR}/{file_name}_FLIP_{letter_code}""" cva.imwrite(f"""/{file_root}.jpg""" , lowerCAmelCase__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(f"""Success {index+1}/{len(lowerCAmelCase__ )} with {file_name}""" ) UpperCAmelCase_ = [] for anno in new_annos[index]: UpperCAmelCase_ = f"""{anno[0]} {anno[1]} {anno[2]} {anno[3]} {anno[4]}""" annos_list.append(lowerCAmelCase__ ) with open(f"""/{file_root}.txt""" , "w" ) as outfile: outfile.write("\n".join(line for line in annos_list ) ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = [] UpperCAmelCase_ = [] for label_file in glob.glob(os.path.join(lowerCAmelCase__ , "*.txt" ) ): UpperCAmelCase_ = label_file.split(os.sep )[-1].rsplit("." , 1 )[0] with open(lowerCAmelCase__ ) as in_file: UpperCAmelCase_ = in_file.readlines() UpperCAmelCase_ = os.path.join(lowerCAmelCase__ , f"""{label_name}.jpg""" ) UpperCAmelCase_ = [] for obj_list in obj_lists: UpperCAmelCase_ = obj_list.rstrip("\n" ).split(" " ) boxes.append( [ int(obj[0] ), float(obj[1] ), float(obj[2] ), float(obj[3] ), float(obj[4] ), ] ) if not boxes: continue img_paths.append(lowerCAmelCase__ ) labels.append(lowerCAmelCase__ ) return img_paths, labels def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = 1 ): UpperCAmelCase_ = [] UpperCAmelCase_ = [] UpperCAmelCase_ = [] for idx in range(len(lowerCAmelCase__ ) ): UpperCAmelCase_ = [] UpperCAmelCase_ = img_list[idx] path_list.append(lowerCAmelCase__ ) UpperCAmelCase_ = anno_list[idx] UpperCAmelCase_ = cva.imread(lowerCAmelCase__ ) if flip_type == 1: UpperCAmelCase_ = cva.flip(lowerCAmelCase__ , lowerCAmelCase__ ) for bbox in img_annos: UpperCAmelCase_ = 1 - bbox[1] new_annos.append([bbox[0], x_center_new, bbox[2], bbox[3], bbox[4]] ) elif flip_type == 0: UpperCAmelCase_ = cva.flip(lowerCAmelCase__ , lowerCAmelCase__ ) for bbox in img_annos: UpperCAmelCase_ = 1 - bbox[2] new_annos.append([bbox[0], bbox[1], y_center_new, bbox[3], bbox[4]] ) new_annos_lists.append(lowerCAmelCase__ ) new_imgs_list.append(lowerCAmelCase__ ) return new_imgs_list, new_annos_lists, path_list def a__ ( lowerCAmelCase__ = 32 ): assert number_char > 1, "The number of character should greater than 1" UpperCAmelCase_ = ascii_lowercase + digits return "".join(random.choice(lowerCAmelCase__ ) for _ in range(lowerCAmelCase__ ) ) if __name__ == "__main__": main() print("""DONE ✅""")
241
1
from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def UpperCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ): '''simple docstring''' lowercase , lowercase = coefficient_matrix.shape lowercase , lowercase = constant_matrix.shape if rowsa != colsa: lowercase = f'Coefficient matrix dimensions must be nxn but received {rowsa}x{colsa}' raise ValueError(lowerCAmelCase__ ) if colsa != 1: lowercase = f'Constant matrix must be nx1 but received {rowsa}x{colsa}' raise ValueError(lowerCAmelCase__ ) if rowsa != rowsa: lowercase = ( '''Coefficient and constant matrices dimensions must be nxn and nx1 but ''' f'received {rowsa}x{colsa} and {rowsa}x{colsa}' ) raise ValueError(lowerCAmelCase__ ) if len(lowerCAmelCase__ ) != rowsa: lowercase = ( '''Number of initial values must be equal to number of rows in coefficient ''' f'matrix but received {len(lowerCAmelCase__ )} and {rowsa}' ) raise ValueError(lowerCAmelCase__ ) if iterations <= 0: raise ValueError('''Iterations must be at least 1''' ) lowercase = np.concatenate( (coefficient_matrix, constant_matrix) , axis=1 ) lowercase , lowercase = table.shape strictly_diagonally_dominant(lowerCAmelCase__ ) # Iterates the whole matrix for given number of times for _ in range(lowerCAmelCase__ ): lowercase = [] for row in range(lowerCAmelCase__ ): lowercase = 0 for col in range(lowerCAmelCase__ ): if col == row: lowercase = table[row][col] elif col == cols - 1: lowercase = table[row][col] else: temp += (-1) * table[row][col] * init_val[col] lowercase = (temp + val) / denom new_val.append(lowerCAmelCase__ ) lowercase = new_val return [float(lowerCAmelCase__ ) for i in new_val] def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase , lowercase = table.shape lowercase = True for i in range(0 , lowerCAmelCase__ ): lowercase = 0 for j in range(0 , cols - 1 ): if i == j: continue else: total += table[i][j] if table[i][i] <= total: raise ValueError('''Coefficient matrix is not strictly diagonally dominant''' ) return is_diagonally_dominant # Test Cases if __name__ == "__main__": import doctest doctest.testmod()
101
def lowerCAmelCase_ ( __A, __A ) -> None: '''simple docstring''' UpperCAmelCase__ = len(__A ) print("The following activities are selected:" ) # The first activity is always selected UpperCAmelCase__ = 0 print(__A, end="," ) # Consider rest of the activities for j in range(__A ): # If this activity has start time greater than # or equal to the finish time of previously # selected activity, then select it if start[j] >= finish[i]: print(__A, end="," ) UpperCAmelCase__ = j if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ = [1, 3, 0, 5, 8, 5] UpperCamelCase__ = [2, 4, 6, 7, 9, 9] print_max_activities(start, finish)
65
0
'''simple docstring''' import numpy class A : def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> None: """simple docstring""" A : str = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. A : Optional[Any] = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. A : List[str] = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. A : str = numpy.random.rand(3 , 1 ) # Real output values provided. A : Dict = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. A : Union[str, Any] = numpy.zeros(output_array.shape ) def __lowerCAmelCase ( self ) -> numpy.ndarray: """simple docstring""" A : str = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. A : int = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. A : List[Any] = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def __lowerCAmelCase ( self ) -> None: """simple docstring""" A : Any = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) A : Optional[Any] = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) A : Optional[Any] = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> None: """simple docstring""" for iteration in range(1 , iterations + 1 ): A : str = self.feedforward() self.back_propagation() if give_loss: A : Optional[Any] = numpy.mean(numpy.square(output - self.feedforward() ) ) print(F'Iteration {iteration} Loss: {loss}' ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE ) -> int: """simple docstring""" A : str = input_arr A : Union[str, Any] = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) A : List[Any] = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) A : Optional[Any] = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' return 1 / (1 + numpy.exp(-value )) def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' return (value) * (1 - (value)) def lowerCAmelCase_ ( ): '''simple docstring''' A : Dict = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. A : List[str] = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. A : int = TwoHiddenLayerNeuralNetwork( input_array=snake_case__ , output_array=snake_case__ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=snake_case__ , iterations=10 , give_loss=snake_case__ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
311
'''simple docstring''' import warnings from typing import Dict import numpy as np from ..utils import ExplicitEnum, add_end_docstrings, is_tf_available, is_torch_available from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' return 1.0 / (1.0 + np.exp(-_outputs )) def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' A : Optional[int] = np.max(_outputs , axis=-1 , keepdims=snake_case__ ) A : Any = np.exp(_outputs - maxes ) return shifted_exp / shifted_exp.sum(axis=-1 , keepdims=snake_case__ ) class A ( __snake_case ): __magic_name__ = '''sigmoid''' __magic_name__ = '''softmax''' __magic_name__ = '''none''' @add_end_docstrings( __snake_case , R''' return_all_scores (`bool`, *optional*, defaults to `False`): Whether to return all prediction scores or just the one of the predicted class. function_to_apply (`str`, *optional*, defaults to `"default"`): The function to apply to the model outputs in order to retrieve the scores. Accepts four different values: - `"default"`: if the model has a single label, will apply the sigmoid function on the output. If the model has several labels, will apply the softmax function on the output. - `"sigmoid"`: Applies the sigmoid function on the output. - `"softmax"`: Applies the softmax function on the output. - `"none"`: Does not apply any function on the output. ''' , ) class A ( __snake_case ): __magic_name__ = False __magic_name__ = ClassificationFunction.NONE def __init__( self , **SCREAMING_SNAKE_CASE ) -> Optional[Any]: """simple docstring""" super().__init__(**SCREAMING_SNAKE_CASE ) self.check_model_type( TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if self.framework == '''tf''' else MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE="" , **SCREAMING_SNAKE_CASE ) -> Dict: """simple docstring""" A : Optional[Any] = tokenizer_kwargs A : int = {} if hasattr(self.model.config , '''return_all_scores''' ) and return_all_scores is None: A : int = self.model.config.return_all_scores if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) or top_k is None: A : Union[str, Any] = top_k A : Dict = False elif return_all_scores is not None: warnings.warn( '''`return_all_scores` is now deprecated, if want a similar functionality use `top_k=None` instead of''' ''' `return_all_scores=True` or `top_k=1` instead of `return_all_scores=False`.''' , SCREAMING_SNAKE_CASE , ) if return_all_scores: A : Optional[int] = None else: A : Dict = 1 if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): A : Dict = ClassificationFunction[function_to_apply.upper()] if function_to_apply is not None: A : int = function_to_apply return preprocess_params, {}, postprocess_params def __call__( self , *SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) -> Tuple: """simple docstring""" A : str = super().__call__(*SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) # TODO try and retrieve it in a nicer way from _sanitize_parameters. A : Any = '''top_k''' not in kwargs if isinstance(args[0] , SCREAMING_SNAKE_CASE ) and _legacy: # This pipeline is odd, and return a list when single item is run return [result] else: return result def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) -> Dict[str, GenericTensor]: """simple docstring""" A : List[Any] = self.framework if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): return self.tokenizer(**SCREAMING_SNAKE_CASE , return_tensors=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) elif isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) and len(SCREAMING_SNAKE_CASE ) == 1 and isinstance(inputs[0] , SCREAMING_SNAKE_CASE ) and len(inputs[0] ) == 2: # It used to be valid to use a list of list of list for text pairs, keeping this path for BC return self.tokenizer( text=inputs[0][0] , text_pair=inputs[0][1] , return_tensors=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) elif isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): # This is likely an invalid usage of the pipeline attempting to pass text pairs. raise ValueError( '''The pipeline received invalid inputs, if you are trying to send text pairs, you can try to send a''' ''' dictionary `{"text": "My text", "text_pair": "My pair"}` in order to send a text pair.''' ) return self.tokenizer(SCREAMING_SNAKE_CASE , return_tensors=SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE ) -> Union[str, Any]: """simple docstring""" return self.model(**SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=1 , SCREAMING_SNAKE_CASE=True ) -> List[str]: """simple docstring""" if function_to_apply is None: if self.model.config.problem_type == "multi_label_classification" or self.model.config.num_labels == 1: A : Optional[int] = ClassificationFunction.SIGMOID elif self.model.config.problem_type == "single_label_classification" or self.model.config.num_labels > 1: A : Any = ClassificationFunction.SOFTMAX elif hasattr(self.model.config , '''function_to_apply''' ) and function_to_apply is None: A : Optional[int] = self.model.config.function_to_apply else: A : Optional[int] = ClassificationFunction.NONE A : Any = model_outputs['''logits'''][0] A : List[Any] = outputs.numpy() if function_to_apply == ClassificationFunction.SIGMOID: A : int = sigmoid(SCREAMING_SNAKE_CASE ) elif function_to_apply == ClassificationFunction.SOFTMAX: A : Any = softmax(SCREAMING_SNAKE_CASE ) elif function_to_apply == ClassificationFunction.NONE: A : int = outputs else: raise ValueError(F'Unrecognized `function_to_apply` argument: {function_to_apply}' ) if top_k == 1 and _legacy: return {"label": self.model.config.idalabel[scores.argmax().item()], "score": scores.max().item()} A : int = [ {'''label''': self.model.config.idalabel[i], '''score''': score.item()} for i, score in enumerate(SCREAMING_SNAKE_CASE ) ] if not _legacy: dict_scores.sort(key=lambda SCREAMING_SNAKE_CASE : x["score"] , reverse=SCREAMING_SNAKE_CASE ) if top_k is not None: A : Union[str, Any] = dict_scores[:top_k] return dict_scores
311
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { '''google/canine-s''': '''https://huggingface.co/google/canine-s/resolve/main/config.json''', # See all CANINE models at https://huggingface.co/models?filter=canine } class lowerCAmelCase_ ( __A ): UpperCAmelCase__ : List[str] = "canine" def __init__( self, SCREAMING_SNAKE_CASE_=768, SCREAMING_SNAKE_CASE_=12, SCREAMING_SNAKE_CASE_=12, SCREAMING_SNAKE_CASE_=3072, SCREAMING_SNAKE_CASE_="gelu", SCREAMING_SNAKE_CASE_=0.1, SCREAMING_SNAKE_CASE_=0.1, SCREAMING_SNAKE_CASE_=1_6384, SCREAMING_SNAKE_CASE_=16, SCREAMING_SNAKE_CASE_=0.02, SCREAMING_SNAKE_CASE_=1e-12, SCREAMING_SNAKE_CASE_=0, SCREAMING_SNAKE_CASE_=0XE000, SCREAMING_SNAKE_CASE_=0XE001, SCREAMING_SNAKE_CASE_=4, SCREAMING_SNAKE_CASE_=4, SCREAMING_SNAKE_CASE_=8, SCREAMING_SNAKE_CASE_=1_6384, SCREAMING_SNAKE_CASE_=128, **SCREAMING_SNAKE_CASE_, ) -> List[Any]: super().__init__(pad_token_id=_UpperCamelCase, bos_token_id=_UpperCamelCase, eos_token_id=_UpperCamelCase, **_UpperCamelCase ) UpperCamelCase : int = max_position_embeddings UpperCamelCase : Dict = hidden_size UpperCamelCase : Union[str, Any] = num_hidden_layers UpperCamelCase : Any = num_attention_heads UpperCamelCase : str = intermediate_size UpperCamelCase : Union[str, Any] = hidden_act UpperCamelCase : Optional[int] = hidden_dropout_prob UpperCamelCase : Any = attention_probs_dropout_prob UpperCamelCase : int = initializer_range UpperCamelCase : Optional[int] = type_vocab_size UpperCamelCase : List[Any] = layer_norm_eps # Character config: UpperCamelCase : Dict = downsampling_rate UpperCamelCase : List[Any] = upsampling_kernel_size UpperCamelCase : Dict = num_hash_functions UpperCamelCase : Any = num_hash_buckets UpperCamelCase : str = local_transformer_stride
119
from collections import deque from .hash_table import HashTable class snake_case_ ( __A ): '''simple docstring''' def __init__( self : int , *_UpperCamelCase : int , **_UpperCamelCase : Tuple ) ->Tuple: super().__init__(*_UpperCamelCase , **_UpperCamelCase ) def snake_case__( self : Dict , _UpperCamelCase : List[str] , _UpperCamelCase : Dict ) ->Tuple: snake_case_ = deque([] ) if self.values[key] is None else self.values[key] self.values[key].appendleft(_UpperCamelCase ) snake_case_ = self.values[key] def snake_case__( self : List[Any] ) ->str: return ( sum(self.charge_factor - len(_UpperCamelCase ) for slot in self.values ) / self.size_table * self.charge_factor ) def snake_case__( self : Dict , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Optional[int]=None ) ->str: if not ( len(self.values[key] ) == self.charge_factor and self.values.count(_UpperCamelCase ) == 0 ): return key return super()._collision_resolution(_UpperCamelCase , _UpperCamelCase )
8
0
import unittest from transformers import PegasusTokenizer, PegasusTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin SCREAMING_SNAKE_CASE :int = get_tests_dir('fixtures/test_sentencepiece_no_bos.model') @require_sentencepiece @require_tokenizers class UpperCAmelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' snake_case_ = PegasusTokenizer snake_case_ = PegasusTokenizerFast snake_case_ = True snake_case_ = True def UpperCamelCase_ ( self : Union[str, Any] ): super().setUp() # We have a SentencePiece fixture for testing __A = PegasusTokenizer(A ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def UpperCamelCase_ ( self : Tuple ): return PegasusTokenizer.from_pretrained("google/pegasus-large" ) def UpperCamelCase_ ( self : str ,**A : Any ): return PegasusTokenizer.from_pretrained(self.tmpdirname ,**A ) def UpperCamelCase_ ( self : Optional[Any] ,A : Dict ): return ("This is a test", "This is a test") def UpperCamelCase_ ( self : str ): __A = "</s>" __A = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A ) ,A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A ) ,A ) def UpperCamelCase_ ( self : Union[str, Any] ): __A = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,"<pad>" ) self.assertEqual(vocab_keys[1] ,"</s>" ) self.assertEqual(vocab_keys[-1] ,"v" ) self.assertEqual(len(A ) ,11_03 ) def UpperCamelCase_ ( self : int ): self.assertEqual(self.get_tokenizer().vocab_size ,11_03 ) def UpperCamelCase_ ( self : str ): __A = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) __A = self.tokenizer_class.from_pretrained(self.tmpdirname ) __A = ( "Let's see which <unk> is the better <unk_token_11> one <mask_1> It seems like this <mask_2> was important" " </s> <pad> <pad> <pad>" ) __A = rust_tokenizer([raw_input_str] ,return_tensors=A ,add_special_tokens=A ).input_ids[0] __A = py_tokenizer([raw_input_str] ,return_tensors=A ,add_special_tokens=A ).input_ids[0] self.assertListEqual(A ,A ) def UpperCamelCase_ ( self : Tuple ): __A = self._large_tokenizer # <mask_1> masks whole sentence while <mask_2> masks single word __A = "<mask_1> To ensure a <mask_2> flow of bank resolutions." __A = [2, 4_13, 6_15, 1_14, 3, 19_71, 1_13, 16_79, 1_07_10, 1_07, 1] __A = tokenizer([raw_input_str] ,return_tensors=A ).input_ids[0] self.assertListEqual(A ,A ) def UpperCamelCase_ ( self : List[Any] ): __A = self._large_tokenizer # The tracebacks for the following asserts are **better** without messages or self.assertEqual assert tokenizer.vocab_size == 9_61_03 assert tokenizer.pad_token_id == 0 assert tokenizer.eos_token_id == 1 assert tokenizer.offset == 1_03 assert tokenizer.unk_token_id == tokenizer.offset + 2 == 1_05 assert tokenizer.unk_token == "<unk>" assert tokenizer.model_max_length == 10_24 __A = "To ensure a smooth flow of bank resolutions." __A = [4_13, 6_15, 1_14, 22_91, 19_71, 1_13, 16_79, 1_07_10, 1_07, 1] __A = tokenizer([raw_input_str] ,return_tensors=A ).input_ids[0] self.assertListEqual(A ,A ) assert tokenizer.convert_ids_to_tokens([0, 1, 2, 3] ) == ["<pad>", "</s>", "<mask_1>", "<mask_2>"] @require_torch def UpperCamelCase_ ( self : Optional[Any] ): __A = ["This is going to be way too long." * 1_50, "short example"] __A = ["not super long but more than 5 tokens", "tiny"] __A = self._large_tokenizer(A ,padding=A ,truncation=A ,return_tensors="pt" ) __A = self._large_tokenizer( text_target=A ,max_length=5 ,padding=A ,truncation=A ,return_tensors="pt" ) assert batch.input_ids.shape == (2, 10_24) assert batch.attention_mask.shape == (2, 10_24) assert targets["input_ids"].shape == (2, 5) assert len(A ) == 2 # input_ids, attention_mask. @slow def UpperCamelCase_ ( self : Dict ): # fmt: off __A = {"input_ids": [[3_89_79, 1_43, 1_84_85, 6_06, 1_30, 2_66_69, 8_76_86, 1_21, 5_41_89, 11_29, 1_11, 2_66_69, 8_76_86, 1_21, 91_14, 1_47_87, 1_21, 1_32_49, 1_58, 5_92, 9_56, 1_21, 1_46_21, 3_15_76, 1_43, 6_26_13, 1_08, 96_88, 9_30, 4_34_30, 1_15_62, 6_26_13, 3_04, 1_08, 1_14_43, 8_97, 1_08, 93_14, 1_74_15, 6_33_99, 1_08, 1_14_43, 76_14, 1_83_16, 1_18, 42_84, 71_48, 1_24_30, 1_43, 14_00, 2_57_03, 1_58, 1_11, 42_84, 71_48, 1_17_72, 1_43, 2_12_97, 10_64, 1_58, 1_22, 2_04, 35_06, 17_54, 11_33, 1_47_87, 15_81, 1_15, 3_32_24, 44_82, 1_11, 13_55, 1_10, 2_91_73, 3_17, 5_08_33, 1_08, 2_01_47, 9_46_65, 1_11, 7_71_98, 1_07, 1], [1_10, 6_26_13, 1_17, 6_38, 1_12, 11_33, 1_21, 2_00_98, 13_55, 7_90_50, 1_38_72, 1_35, 15_96, 5_35_41, 13_52, 1_41, 1_30_39, 55_42, 1_24, 3_02, 5_18, 1_11, 2_68, 29_56, 1_15, 1_49, 44_27, 1_07, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_39, 12_35, 27_99, 1_82_89, 1_77_80, 2_04, 1_09, 94_74, 12_96, 1_07, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=A ,model_name="google/bigbird-pegasus-large-arxiv" ,revision="ba85d0851d708441f91440d509690f1ab6353415" ,) @require_sentencepiece @require_tokenizers class UpperCAmelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' snake_case_ = PegasusTokenizer snake_case_ = PegasusTokenizerFast snake_case_ = True snake_case_ = True def UpperCamelCase_ ( self : Union[str, Any] ): super().setUp() # We have a SentencePiece fixture for testing __A = PegasusTokenizer(A ,offset=0 ,mask_token_sent=A ,mask_token="[MASK]" ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def UpperCamelCase_ ( self : Any ): return PegasusTokenizer.from_pretrained("google/bigbird-pegasus-large-arxiv" ) def UpperCamelCase_ ( self : int ,**A : str ): return PegasusTokenizer.from_pretrained(self.tmpdirname ,**A ) def UpperCamelCase_ ( self : List[str] ,A : str ): return ("This is a test", "This is a test") def UpperCamelCase_ ( self : Dict ): __A = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) __A = self.tokenizer_class.from_pretrained(self.tmpdirname ) __A = ( "Let's see which <unk> is the better <unk_token> one [MASK] It seems like this [MASK] was important </s>" " <pad> <pad> <pad>" ) __A = rust_tokenizer([raw_input_str] ,return_tensors=A ,add_special_tokens=A ).input_ids[0] __A = py_tokenizer([raw_input_str] ,return_tensors=A ,add_special_tokens=A ).input_ids[0] self.assertListEqual(A ,A ) @require_torch def UpperCamelCase_ ( self : Dict ): __A = ["This is going to be way too long." * 10_00, "short example"] __A = ["not super long but more than 5 tokens", "tiny"] __A = self._large_tokenizer(A ,padding=A ,truncation=A ,return_tensors="pt" ) __A = self._large_tokenizer( text_target=A ,max_length=5 ,padding=A ,truncation=A ,return_tensors="pt" ) assert batch.input_ids.shape == (2, 40_96) assert batch.attention_mask.shape == (2, 40_96) assert targets["input_ids"].shape == (2, 5) assert len(A ) == 2 # input_ids, attention_mask. def UpperCamelCase_ ( self : List[Any] ): __A = ( "This is an example string that is used to test the original TF implementation against the HF" " implementation" ) __A = self._large_tokenizer(A ).input_ids self.assertListEqual( A ,[1_82, 1_17, 1_42, 5_87, 42_11, 1_20, 1_17, 2_63, 1_12, 8_04, 1_09, 8_56, 2_50_16, 31_37, 4_64, 1_09, 2_69_55, 31_37, 1] ,)
124
import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor SCREAMING_SNAKE_CASE :Any = logging.get_logger(__name__) class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : int ,*A : Dict ,**A : Dict ): warnings.warn( "The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use CLIPImageProcessor instead." ,A ,) super().__init__(*A ,**A )
124
1
from __future__ import annotations import queue class a : def __init__( self :Dict ,__lowercase :Optional[Any] ): snake_case__ : Union[str, Any] = data snake_case__ : Any = None snake_case__ : List[Any] = None def _lowerCAmelCase ( ) -> TreeNode: """simple docstring""" print('''\n********Press N to stop entering at any point of time********\n''' ) snake_case__ : Dict = input('''Enter the value of the root node: ''' ).strip().lower() snake_case__ : queue.Queue = queue.Queue() snake_case__ : Any = TreeNode(int(__lowerCAmelCase ) ) q.put(__lowerCAmelCase ) while not q.empty(): snake_case__ : Any = q.get() snake_case__ : Optional[Any] = f"""Enter the left node of {node_found.data}: """ snake_case__ : Any = input(__lowerCAmelCase ).strip().lower() or '''n''' if check == "n": return tree_node snake_case__ : str = TreeNode(int(__lowerCAmelCase ) ) snake_case__ : Any = left_node q.put(__lowerCAmelCase ) snake_case__ : str = f"""Enter the right node of {node_found.data}: """ snake_case__ : List[Any] = input(__lowerCAmelCase ).strip().lower() or '''n''' if check == "n": return tree_node snake_case__ : str = TreeNode(int(__lowerCAmelCase ) ) snake_case__ : str = right_node q.put(__lowerCAmelCase ) raise def _lowerCAmelCase ( __lowerCAmelCase ) -> None: """simple docstring""" if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return print(node.data , end=''',''' ) pre_order(node.left ) pre_order(node.right ) def _lowerCAmelCase ( __lowerCAmelCase ) -> None: """simple docstring""" if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return in_order(node.left ) print(node.data , end=''',''' ) in_order(node.right ) def _lowerCAmelCase ( __lowerCAmelCase ) -> None: """simple docstring""" if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return post_order(node.left ) post_order(node.right ) print(node.data , end=''',''' ) def _lowerCAmelCase ( __lowerCAmelCase ) -> None: """simple docstring""" if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return snake_case__ : queue.Queue = queue.Queue() q.put(__lowerCAmelCase ) while not q.empty(): snake_case__ : int = q.get() print(node_dequeued.data , end=''',''' ) if node_dequeued.left: q.put(node_dequeued.left ) if node_dequeued.right: q.put(node_dequeued.right ) def _lowerCAmelCase ( __lowerCAmelCase ) -> None: """simple docstring""" if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return snake_case__ : queue.Queue = queue.Queue() q.put(__lowerCAmelCase ) while not q.empty(): snake_case__ : Optional[Any] = [] while not q.empty(): snake_case__ : Dict = q.get() print(node_dequeued.data , end=''',''' ) if node_dequeued.left: list_.append(node_dequeued.left ) if node_dequeued.right: list_.append(node_dequeued.right ) print() for node in list_: q.put(__lowerCAmelCase ) def _lowerCAmelCase ( __lowerCAmelCase ) -> None: """simple docstring""" if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return snake_case__ : list[TreeNode] = [] snake_case__ : Tuple = node while n or stack: while n: # start from root node, find its left child print(n.data , end=''',''' ) stack.append(__lowerCAmelCase ) snake_case__ : List[Any] = n.left # end of while means current node doesn't have left child snake_case__ : Union[str, Any] = stack.pop() # start to traverse its right child snake_case__ : str = n.right def _lowerCAmelCase ( __lowerCAmelCase ) -> None: """simple docstring""" if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return snake_case__ : list[TreeNode] = [] snake_case__ : Optional[Any] = node while n or stack: while n: stack.append(__lowerCAmelCase ) snake_case__ : Dict = n.left snake_case__ : Union[str, Any] = stack.pop() print(n.data , end=''',''' ) snake_case__ : Any = n.right def _lowerCAmelCase ( __lowerCAmelCase ) -> None: """simple docstring""" if not isinstance(__lowerCAmelCase , __lowerCAmelCase ) or not node: return snake_case__ , snake_case__ : Optional[int] = [], [] snake_case__ : Optional[Any] = node stacka.append(__lowerCAmelCase ) while stacka: # to find the reversed order of post order, store it in stack2 snake_case__ : List[str] = stacka.pop() if n.left: stacka.append(n.left ) if n.right: stacka.append(n.right ) stacka.append(__lowerCAmelCase ) while stacka: # pop up from stack2 will be the post order print(stacka.pop().data , end=''',''' ) def _lowerCAmelCase ( __lowerCAmelCase = "" , __lowerCAmelCase=50 , __lowerCAmelCase="*" ) -> str: """simple docstring""" if not s: return "\n" + width * char snake_case__ , snake_case__ : Optional[Any] = divmod(width - len(__lowerCAmelCase ) - 2 , 2 ) return f"""{left * char} {s} {(left + extra) * char}""" if __name__ == "__main__": import doctest doctest.testmod() print(prompt('''Binary Tree Traversals''')) A__ = build_tree() print(prompt('''Pre Order Traversal''')) pre_order(node) print(prompt() + '''\n''') print(prompt('''In Order Traversal''')) in_order(node) print(prompt() + '''\n''') print(prompt('''Post Order Traversal''')) post_order(node) print(prompt() + '''\n''') print(prompt('''Level Order Traversal''')) level_order(node) print(prompt() + '''\n''') print(prompt('''Actual Level Order Traversal''')) level_order_actual(node) print('''*''' * 50 + '''\n''') print(prompt('''Pre Order Traversal - Iteration Version''')) pre_order_iter(node) print(prompt() + '''\n''') print(prompt('''In Order Traversal - Iteration Version''')) in_order_iter(node) print(prompt() + '''\n''') print(prompt('''Post Order Traversal - Iteration Version''')) post_order_iter(node) print(prompt())
230
from importlib import import_module from .logging import get_logger A__ = get_logger(__name__) class a : def __init__( self :Optional[int] ,__lowercase :List[str] ,__lowercase :Any=None ): snake_case__ : List[str] = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith('''__''' ): setattr(self ,__lowercase ,getattr(__lowercase ,__lowercase ) ) snake_case__ : Optional[Any] = module._original_module if isinstance(__lowercase ,_PatchedModuleObj ) else module class a : __lowerCAmelCase : Any = [] def __init__( self :List[str] ,__lowercase :Optional[Any] ,__lowercase :str ,__lowercase :Dict ,__lowercase :Any=None ): snake_case__ : Dict = obj snake_case__ : Dict = target snake_case__ : List[str] = new snake_case__ : int = target.split('''.''' )[0] snake_case__ : List[str] = {} snake_case__ : Any = attrs or [] def __enter__( self :Tuple ): *snake_case__ , snake_case__ : str = self.target.split('''.''' ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(__lowercase ) ): try: snake_case__ : Any = import_module('''.'''.join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): snake_case__ : Optional[int] = getattr(self.obj ,__lowercase ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(__lowercase ,_PatchedModuleObj ) and obj_attr._original_module is submodule) ): snake_case__ : List[Any] = obj_attr # patch at top level setattr(self.obj ,__lowercase ,_PatchedModuleObj(__lowercase ,attrs=self.attrs ) ) snake_case__ : List[Any] = getattr(self.obj ,__lowercase ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(__lowercase ,__lowercase ,_PatchedModuleObj(getattr(__lowercase ,__lowercase ,__lowercase ) ,attrs=self.attrs ) ) snake_case__ : List[Any] = getattr(__lowercase ,__lowercase ) # finally set the target attribute setattr(__lowercase ,__lowercase ,self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: snake_case__ : int = getattr(import_module('''.'''.join(__lowercase ) ) ,__lowercase ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj ,__lowercase ) is attr_value: snake_case__ : str = getattr(self.obj ,__lowercase ) setattr(self.obj ,__lowercase ,self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" snake_case__ : str = globals()['''__builtins__'''][target_attr] setattr(self.obj ,__lowercase ,self.new ) else: raise RuntimeError(F"""Tried to patch attribute {target_attr} instead of a submodule.""" ) def __exit__( self :Tuple ,*__lowercase :Optional[int] ): for attr in list(self.original ): setattr(self.obj ,__lowercase ,self.original.pop(__lowercase ) ) def __lowerCamelCase ( self :Tuple ): self.__enter__() self._active_patches.append(self ) def __lowerCamelCase ( self :Dict ): try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
230
1
def snake_case (__lowercase , __lowercase ) -> Any: '''simple docstring''' _snake_case : Optional[int] = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def snake_case (__lowercase , __lowercase , __lowercase ) -> Union[str, Any]: '''simple docstring''' _snake_case : Dict = 0 while b > 0: if b & 1: _snake_case : List[Any] = ((res % c) + (a % c)) % c a += a b >>= 1 return res
352
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_tf_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_tf_available(): import tensorflow as tf __SCREAMING_SNAKE_CASE : str = logging.get_logger(__name__) @dataclass class lowercase_ ( __snake_case ): _lowerCamelCase = [ 'no_inference', 'no_cuda', 'no_tpu', 'no_speed', 'no_memory', 'no_env_print', 'no_multi_process', ] def __init__( self , **lowercase_ ): for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: _snake_case : List[str] = deprecated_arg[3:] _snake_case : Optional[int] = not kwargs.pop(lowercase_ ) logger.warning( f"""{deprecated_arg} is depreciated. Please use --no-{positive_arg} or""" f""" {positive_arg}={kwargs[positive_arg]}""" ) _snake_case : Tuple = kwargs.pop("tpu_name" , self.tpu_name ) _snake_case : Any = kwargs.pop("device_idx" , self.device_idx ) _snake_case : List[str] = kwargs.pop("eager_mode" , self.eager_mode ) _snake_case : List[str] = kwargs.pop("use_xla" , self.use_xla ) super().__init__(**lowercase_ ) _lowerCamelCase = field( default=__snake_case , metadata={'help': 'Name of TPU'} , ) _lowerCamelCase = field( default=0 , metadata={'help': 'CPU / GPU device index. Defaults to 0.'} , ) _lowerCamelCase = field(default=__snake_case , metadata={'help': 'Benchmark models in eager model.'} ) _lowerCamelCase = field( default=__snake_case , metadata={ 'help': 'Benchmark models using XLA JIT compilation. Note that `eager_model` has to be set to `False`.' } , ) @cached_property def UpperCamelCase ( self ): requires_backends(self , ["tf"] ) _snake_case : str = None if self.tpu: try: if self.tpu_name: _snake_case : Optional[Any] = tf.distribute.cluster_resolver.TPUClusterResolver(self.tpu_name ) else: _snake_case : List[str] = tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: _snake_case : Union[str, Any] = None return tpu @cached_property def UpperCamelCase ( self ): requires_backends(self , ["tf"] ) if self.is_tpu: tf.config.experimental_connect_to_cluster(self._setup_tpu ) tf.tpu.experimental.initialize_tpu_system(self._setup_tpu ) _snake_case : List[str] = tf.distribute.TPUStrategy(self._setup_tpu ) else: # currently no multi gpu is allowed if self.is_gpu: # TODO: Currently only single GPU is supported tf.config.set_visible_devices(self.gpu_list[self.device_idx] , "GPU" ) _snake_case : Any = tf.distribute.OneDeviceStrategy(device=f"""/gpu:{self.device_idx}""" ) else: tf.config.set_visible_devices([] , "GPU" ) # disable GPU _snake_case : Any = tf.distribute.OneDeviceStrategy(device=f"""/cpu:{self.device_idx}""" ) return strategy @property def UpperCamelCase ( self ): requires_backends(self , ["tf"] ) return self._setup_tpu is not None @property def UpperCamelCase ( self ): requires_backends(self , ["tf"] ) return self._setup_strategy @property def UpperCamelCase ( self ): requires_backends(self , ["tf"] ) return tf.config.list_physical_devices("GPU" ) @property def UpperCamelCase ( self ): requires_backends(self , ["tf"] ) if self.cuda: return len(self.gpu_list ) return 0 @property def UpperCamelCase ( self ): return self.n_gpu > 0
284
0
'''simple docstring''' from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class __A : def __init__(self : Any , __a : str , __a : List[str]=13 , __a : Optional[int]=30 , __a : Tuple=2 , __a : str=3 , __a : Tuple=True , __a : List[Any]=True , __a : Optional[int]=32 , __a : Optional[int]=2 , __a : int=4 , __a : Optional[Any]=37 , __a : Optional[Any]="gelu" , __a : Optional[Any]=0.1 , __a : int=0.1 , __a : int=10 , __a : Optional[int]=0.02 , __a : Dict=3 , __a : Optional[int]=None , __a : List[str]=2 , ): UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = image_size UpperCAmelCase_ = patch_size UpperCAmelCase_ = num_channels UpperCAmelCase_ = is_training UpperCAmelCase_ = use_labels UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = type_sequence_label_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = scope UpperCAmelCase_ = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase_ = (image_size // patch_size) ** 2 UpperCAmelCase_ = num_patches + 2 def _lowercase (self : Optional[Any] ): UpperCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase_ = self.get_config() return config, pixel_values, labels def _lowercase (self : List[str] ): return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__a , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowercase (self : Union[str, Any] , __a : str , __a : Any , __a : Dict ): UpperCAmelCase_ = TFDeiTModel(config=__a ) UpperCAmelCase_ = model(__a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase (self : str , __a : Optional[int] , __a : List[str] , __a : Tuple ): UpperCAmelCase_ = TFDeiTForMaskedImageModeling(config=__a ) UpperCAmelCase_ = model(__a ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase_ = 1 UpperCAmelCase_ = TFDeiTForMaskedImageModeling(__a ) UpperCAmelCase_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase_ = model(__a ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowercase (self : List[Any] , __a : List[str] , __a : int , __a : int ): UpperCAmelCase_ = self.type_sequence_label_size UpperCAmelCase_ = TFDeiTForImageClassification(__a ) UpperCAmelCase_ = model(__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase_ = 1 UpperCAmelCase_ = TFDeiTForImageClassification(__a ) UpperCAmelCase_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase_ = model(__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowercase (self : Dict ): UpperCAmelCase_ = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = config_and_inputs UpperCAmelCase_ = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class __A ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): a__ : Tuple = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) a__ : Union[str, Any] = ( { """feature-extraction""": TFDeiTModel, """image-classification""": (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) a__ : Tuple = False a__ : List[Any] = False a__ : Any = False a__ : Dict = False def _lowercase (self : List[Any] ): UpperCAmelCase_ = TFDeiTModelTester(self ) UpperCAmelCase_ = ConfigTester(self , config_class=__a , has_text_modality=__a , hidden_size=37 ) def _lowercase (self : Union[str, Any] ): self.config_tester.run_common_tests() @unittest.skip(reason="DeiT does not use inputs_embeds" ) def _lowercase (self : List[str] ): pass def _lowercase (self : List[Any] ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(__a ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) UpperCAmelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__a , tf.keras.layers.Dense ) ) def _lowercase (self : Optional[Any] ): UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(__a ) UpperCAmelCase_ = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ = [*signature.parameters.keys()] UpperCAmelCase_ = ["pixel_values"] self.assertListEqual(arg_names[:1] , __a ) def _lowercase (self : Dict ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a ) def _lowercase (self : Any ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__a ) def _lowercase (self : Any ): UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__a ) def _lowercase (self : Optional[Any] , __a : Tuple , __a : Tuple , __a : Union[str, Any]=False ): UpperCAmelCase_ = super()._prepare_for_class(__a , __a , return_labels=__a ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def _lowercase (self : int ): for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ = TFDeiTModel.from_pretrained(__a ) self.assertIsNotNone(__a ) def lowerCAmelCase_ ( ) -> Dict: '''simple docstring''' UpperCAmelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf @require_vision class __A ( unittest.TestCase ): @cached_property def _lowercase (self : Union[str, Any] ): return ( DeiTImageProcessor.from_pretrained("facebook/deit-base-distilled-patch16-224" ) if is_vision_available() else None ) @slow def _lowercase (self : Optional[Any] ): UpperCAmelCase_ = TFDeiTForImageClassificationWithTeacher.from_pretrained("facebook/deit-base-distilled-patch16-224" ) UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = image_processor(images=__a , return_tensors="tf" ) # forward pass UpperCAmelCase_ = model(**__a ) # verify the logits UpperCAmelCase_ = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , __a ) UpperCAmelCase_ = tf.constant([-1.02_66, 0.19_12, -1.28_61] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , __a , atol=1E-4 ) )
1
"""simple docstring""" import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint _snake_case = { '169M': 12, '430M': 24, '1B5': 24, '3B': 32, '7B': 32, '14B': 40, } _snake_case = { '169M': 768, '430M': 1024, '1B5': 2048, '3B': 2560, '7B': 4096, '14B': 5120, } def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : int = list(state_dict.keys() ) for name in state_dict_keys: _a : str = state_dict.pop(UpperCamelCase__ ) # emb -> embedding if name.startswith("""emb.""" ): _a : Dict = name.replace("""emb.""" , """embeddings.""" ) # ln_0 -> pre_ln (only present at block 0) if name.startswith("""blocks.0.ln0""" ): _a : Dict = name.replace("""blocks.0.ln0""" , """blocks.0.pre_ln""" ) # att -> attention _a : Any = re.sub(R"""blocks\.(\d+)\.att""" , R"""blocks.\1.attention""" , UpperCamelCase__ ) # ffn -> feed_forward _a : int = re.sub(R"""blocks\.(\d+)\.ffn""" , R"""blocks.\1.feed_forward""" , UpperCamelCase__ ) # time_mix_k -> time_mix_key and reshape if name.endswith(""".time_mix_k""" ): _a : List[str] = name.replace(""".time_mix_k""" , """.time_mix_key""" ) # time_mix_v -> time_mix_value and reshape if name.endswith(""".time_mix_v""" ): _a : Tuple = name.replace(""".time_mix_v""" , """.time_mix_value""" ) # time_mix_r -> time_mix_key and reshape if name.endswith(""".time_mix_r""" ): _a : Dict = name.replace(""".time_mix_r""" , """.time_mix_receptance""" ) if name != "head.weight": _a : Optional[int] = """rwkv.""" + name _a : Any = weight return state_dict def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=False , UpperCamelCase__=None ): '''simple docstring''' # 1. If possible, build the tokenizer. if tokenizer_file is None: print("""No `--tokenizer_file` provided, we will use the default tokenizer.""" ) _a : Tuple = 5_0_2_7_7 _a : str = AutoTokenizer.from_pretrained("""EleutherAI/gpt-neox-20b""" ) else: _a : int = PreTrainedTokenizerFast(tokenizer_file=UpperCamelCase__ ) _a : int = len(UpperCamelCase__ ) tokenizer.save_pretrained(UpperCamelCase__ ) # 2. Build the config _a : Optional[Any] = list(NUM_HIDDEN_LAYERS_MAPPING.keys() ) if size is None: # Try to infer size from the checkpoint name for candidate in possible_sizes: if candidate in checkpoint_file: _a : Tuple = candidate break if size is None: raise ValueError("""Could not infer the size, please provide it with the `--size` argument.""" ) if size not in possible_sizes: raise ValueError(F"""`size` should be one of {possible_sizes}, got {size}.""" ) _a : List[Any] = RwkvConfig( vocab_size=UpperCamelCase__ , num_hidden_layers=NUM_HIDDEN_LAYERS_MAPPING[size] , hidden_size=HIDEN_SIZE_MAPPING[size] , ) config.save_pretrained(UpperCamelCase__ ) # 3. Download model file then convert state_dict _a : str = hf_hub_download(UpperCamelCase__ , UpperCamelCase__ ) _a : int = torch.load(UpperCamelCase__ , map_location="""cpu""" ) _a : List[str] = convert_state_dict(UpperCamelCase__ ) # 4. Split in shards and save _a , _a : List[str] = shard_checkpoint(UpperCamelCase__ ) for shard_file, shard in shards.items(): torch.save(UpperCamelCase__ , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) if index is not None: _a : Optional[Any] = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) # Save the index as well with open(UpperCamelCase__ , """w""" , encoding="""utf-8""" ) as f: _a : Dict = json.dumps(UpperCamelCase__ , indent=2 , sort_keys=UpperCamelCase__ ) + """\n""" f.write(UpperCamelCase__ ) # 5. Clean up shards (for some reason the file PyTorch saves take the same space as the whole state_dict print( """Cleaning up shards. This may error with an OOM error, it this is the case don't worry you still have converted the model.""" ) _a : List[Any] = list(shards.keys() ) del state_dict del shards gc.collect() for shard_file in shard_files: _a : Any = torch.load(os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) torch.save({k: v.cpu().clone() for k, v in state_dict.items()} , os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) del state_dict gc.collect() if push_to_hub: if model_name is None: raise ValueError("""Please provide a `model_name` to push the model to the Hub.""" ) _a : Dict = AutoModelForCausalLM.from_pretrained(UpperCamelCase__ ) model.push_to_hub(UpperCamelCase__ , max_shard_size="""2GB""" ) tokenizer.push_to_hub(UpperCamelCase__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '--repo_id', default=None, type=str, required=True, help='Repo ID from which to pull the checkpoint.' ) parser.add_argument( '--checkpoint_file', default=None, type=str, required=True, help='Name of the checkpoint file in the repo.' ) parser.add_argument( '--output_dir', default=None, type=str, required=True, help='Where to save the converted model.' ) parser.add_argument( '--tokenizer_file', default=None, type=str, help='Path to the tokenizer file to use (if not provided, only the model is converted).', ) parser.add_argument( '--size', default=None, type=str, help='Size of the model. Will be inferred from the `checkpoint_file` if not passed.', ) parser.add_argument( '--push_to_hub', action='store_true', help='Push to the Hub the converted model.', ) parser.add_argument( '--model_name', default=None, type=str, help='Name of the pushed model on the Hub, including the username / organization.', ) _snake_case = parser.parse_args() convert_rmkv_checkpoint_to_hf_format( args.repo_id, args.checkpoint_file, args.output_dir, size=args.size, tokenizer_file=args.tokenizer_file, push_to_hub=args.push_to_hub, model_name=args.model_name, )
294
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __SCREAMING_SNAKE_CASE ={ "configuration_chinese_clip": [ "CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "ChineseCLIPConfig", "ChineseCLIPOnnxConfig", "ChineseCLIPTextConfig", "ChineseCLIPVisionConfig", ], "processing_chinese_clip": ["ChineseCLIPProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE =["ChineseCLIPFeatureExtractor"] __SCREAMING_SNAKE_CASE =["ChineseCLIPImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE =[ "CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "ChineseCLIPModel", "ChineseCLIPPreTrainedModel", "ChineseCLIPTextModel", "ChineseCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys __SCREAMING_SNAKE_CASE =_LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
366
"""simple docstring""" import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class UpperCamelCase : def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> int: '''simple docstring''' return None class UpperCamelCase : def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> str: '''simple docstring''' return None class UpperCamelCase ( unittest.TestCase ): lowercase = [ # (model_name, model_kwargs) ('bert-base-cased', {}), ('gpt2', {'use_cache': False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def _UpperCAmelCase ( self ) -> str: '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(__UpperCamelCase ,'tf' ,12 ,**__UpperCamelCase ) @require_torch @slow def _UpperCAmelCase ( self ) -> List[Any]: '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(__UpperCamelCase ,'pt' ,12 ,**__UpperCamelCase ) @require_torch @slow def _UpperCAmelCase ( self ) -> Optional[int]: '''simple docstring''' from transformers import BertModel lowercase_ : Union[str, Any] = ['[UNK]', '[SEP]', '[CLS]', '[PAD]', '[MASK]', 'some', 'other', 'words'] with NamedTemporaryFile(mode='w+t' ) as vocab_file: vocab_file.write('\n'.join(__UpperCamelCase ) ) vocab_file.flush() lowercase_ : List[str] = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: lowercase_ : Optional[Any] = BertModel(BertConfig(vocab_size=len(__UpperCamelCase ) ) ) model.save_pretrained(__UpperCamelCase ) self._test_export(__UpperCamelCase ,'pt' ,12 ,__UpperCamelCase ) @require_tf @slow def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: lowercase_ : Optional[int] = self._test_export(__UpperCamelCase ,'tf' ,12 ,**__UpperCamelCase ) lowercase_ : int = quantize(Path(__UpperCamelCase ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(__UpperCamelCase ).stat().st_size: self.fail('Quantized model is bigger than initial ONNX model' ) @require_torch @slow def _UpperCAmelCase ( self ) -> List[str]: '''simple docstring''' for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: lowercase_ : Tuple = self._test_export(__UpperCamelCase ,'pt' ,12 ,**__UpperCamelCase ) lowercase_ : Tuple = quantize(__UpperCamelCase ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(__UpperCamelCase ).stat().st_size: self.fail('Quantized model is bigger than initial ONNX model' ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase=None ,**__UpperCamelCase ) -> Optional[int]: '''simple docstring''' try: # Compute path with TemporaryDirectory() as tempdir: lowercase_ : Dict = Path(__UpperCamelCase ).joinpath('model.onnx' ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,**__UpperCamelCase ) return path except Exception as e: self.fail(__UpperCamelCase ) @require_torch @require_tokenizers @slow def _UpperCAmelCase ( self ) -> Any: '''simple docstring''' from transformers import BertModel lowercase_ : List[Any] = BertModel(BertConfig.from_pretrained('lysandre/tiny-bert-random' ) ) lowercase_ : Union[str, Any] = BertTokenizerFast.from_pretrained('lysandre/tiny-bert-random' ) self._test_infer_dynamic_axis(__UpperCamelCase ,__UpperCamelCase ,'pt' ) @require_tf @require_tokenizers @slow def _UpperCAmelCase ( self ) -> str: '''simple docstring''' from transformers import TFBertModel lowercase_ : Optional[Any] = TFBertModel(BertConfig.from_pretrained('lysandre/tiny-bert-random' ) ) lowercase_ : Any = BertTokenizerFast.from_pretrained('lysandre/tiny-bert-random' ) self._test_infer_dynamic_axis(__UpperCamelCase ,__UpperCamelCase ,'tf' ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> Dict: '''simple docstring''' lowercase_ : Tuple = FeatureExtractionPipeline(__UpperCamelCase ,__UpperCamelCase ) lowercase_ : Dict = ['input_ids', 'token_type_ids', 'attention_mask', 'output_0', 'output_1'] lowercase_ , lowercase_ , lowercase_ , lowercase_ : Any = infer_shapes(__UpperCamelCase ,__UpperCamelCase ) # Assert all variables are present self.assertEqual(len(__UpperCamelCase ) ,len(__UpperCamelCase ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] ,__UpperCamelCase ) self.assertSequenceEqual(variable_names[3:] ,__UpperCamelCase ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] ,{0: 'batch', 1: 'sequence'} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes['output_0'] ,{0: 'batch', 1: 'sequence'} ) self.assertDictEqual(shapes['output_1'] ,{0: 'batch'} ) def _UpperCAmelCase ( self ) -> Any: '''simple docstring''' lowercase_ : Any = ['input_ids', 'attention_mask', 'token_type_ids'] lowercase_ : List[Any] = {'input_ids': [1, 2, 3, 4], 'attention_mask': [0, 0, 0, 0], 'token_type_ids': [1, 1, 1, 1]} lowercase_ , lowercase_ : int = ensure_valid_input(FuncContiguousArgs() ,__UpperCamelCase ,__UpperCamelCase ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(__UpperCamelCase ) ,3 ) # Should have exactly the same input names self.assertEqual(set(__UpperCamelCase ) ,set(__UpperCamelCase ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(__UpperCamelCase ,(tokens['input_ids'], tokens['token_type_ids'], tokens['attention_mask']) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) lowercase_ , lowercase_ : Optional[int] = ensure_valid_input(FuncNonContiguousArgs() ,__UpperCamelCase ,__UpperCamelCase ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(__UpperCamelCase ) ,1 ) self.assertEqual(len(__UpperCamelCase ) ,1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] ,tokens['input_ids'] ) self.assertEqual(ordered_input_names[0] ,'input_ids' ) def _UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' lowercase_ : Dict = generate_identified_filename(Path('/home/something/my_fake_model.onnx' ) ,'-test' ) self.assertEqual('/home/something/my_fake_model-test.onnx' ,generated.as_posix() )
321
0
from __future__ import annotations def a__ ( snake_case , snake_case , snake_case ): """simple docstring""" if days_between_payments <= 0: raise ValueError('''days_between_payments must be > 0''' ) if daily_interest_rate < 0: raise ValueError('''daily_interest_rate must be >= 0''' ) if principal <= 0: raise ValueError('''principal must be > 0''' ) return principal * daily_interest_rate * days_between_payments def a__ ( snake_case , snake_case , snake_case , ): """simple docstring""" if number_of_compounding_periods <= 0: raise ValueError('''number_of_compounding_periods must be > 0''' ) if nominal_annual_interest_rate_percentage < 0: raise ValueError('''nominal_annual_interest_rate_percentage must be >= 0''' ) if principal <= 0: raise ValueError('''principal must be > 0''' ) return principal * ( (1 + nominal_annual_interest_rate_percentage) ** number_of_compounding_periods - 1 ) def a__ ( snake_case , snake_case , snake_case , ): """simple docstring""" if number_of_years <= 0: raise ValueError('''number_of_years must be > 0''' ) if nominal_annual_percentage_rate < 0: raise ValueError('''nominal_annual_percentage_rate must be >= 0''' ) if principal <= 0: raise ValueError('''principal must be > 0''' ) return compound_interest( snake_case , nominal_annual_percentage_rate / 365 , number_of_years * 365 ) if __name__ == "__main__": import doctest doctest.testmod()
303
from __future__ import annotations def _a ( SCREAMING_SNAKE_CASE : list[int] , SCREAMING_SNAKE_CASE : int ) -> list[int]: """simple docstring""" __lowerCAmelCase: int = 0 __lowerCAmelCase: Tuple = len(SCREAMING_SNAKE_CASE ) - 1 while i < j: if nums[i] + nums[j] == target: return [i, j] elif nums[i] + nums[j] < target: __lowerCAmelCase: Tuple = i + 1 else: __lowerCAmelCase: List[str] = j - 1 return [] if __name__ == "__main__": import doctest doctest.testmod() print(f"{two_pointer([2, 7, 1_1, 1_5], 9) = }")
322
0
"""simple docstring""" import argparse import random import joblib import numpy as np import torch from igf.igf import ( SecondaryLearner, collect_objective_set, compute_perplexity, generate_datasets, load_gpta, recopy_gpta, set_seed, train_secondary_learner, ) from torch.utils.data import DataLoader, RandomSampler from transformers import GPTaLMHeadModel def __lowerCAmelCase (_UpperCamelCase=32 , _UpperCamelCase=10 , _UpperCamelCase=100 , _UpperCamelCase=1026 , _UpperCamelCase=True , _UpperCamelCase="data/tokenized_stories_train_wikitext103.jbl" , _UpperCamelCase="igf_context_pairs.jbl" , ): set_seed(3 ) # generate train_data and objective_set __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = generate_datasets( _UpperCamelCase , _UpperCamelCase , number=_UpperCamelCase , min_len=1026 , trim=_UpperCamelCase ) # keeps model same across runs set_seed(4 ) # model, lm_optimizer, lm_scheduler = recopy_gpt2(model, device, max_steps) # store original model weights # can we train on GPU? __lowerCAmelCase : Union[str, Any] = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu' ) # load pretrained model __lowerCAmelCase : List[str] = load_gpta('gpt2' ).to(_UpperCamelCase ) print('computing perplexity on objective set' ) __lowerCAmelCase : Union[str, Any] = compute_perplexity(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ).item() print('perplexity on objective set:' , _UpperCamelCase ) # collect igf pairs and save to file demo.jbl collect_objective_set(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) # clean up, delete model and data we don't need anymore del model, train_data, objective_set torch.cuda.empty_cache() def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase=15 , _UpperCamelCase=128 , _UpperCamelCase=100 , _UpperCamelCase="igf_model.pt" , ): set_seed(42 ) # Load pre-trained model __lowerCAmelCase : int = GPTaLMHeadModel.from_pretrained('gpt2' ) # Initialize secondary learner to use embedding weights of model __lowerCAmelCase : List[str] = SecondaryLearner(_UpperCamelCase ) # Train secondary learner __lowerCAmelCase : Dict = train_secondary_learner( _UpperCamelCase , _UpperCamelCase , max_epochs=_UpperCamelCase , batch_size=_UpperCamelCase , eval_freq=100 , igf_model_path=_UpperCamelCase , ) del model, secondary_learner_train_data torch.cuda.empty_cache() return secondary_learner def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase=32 , _UpperCamelCase=1000 , _UpperCamelCase=16 , _UpperCamelCase=1.0 , _UpperCamelCase=recopy_gpta , _UpperCamelCase=None , _UpperCamelCase=10 , _UpperCamelCase="gpt2_finetuned.pt" , ): __lowerCAmelCase : Any = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu' ) __lowerCAmelCase : int = RandomSampler(_UpperCamelCase ) __lowerCAmelCase : Optional[int] = DataLoader(_UpperCamelCase , sampler=_UpperCamelCase ) __lowerCAmelCase : Union[str, Any] = max_steps // (len(_UpperCamelCase )) + 1 __lowerCAmelCase : int = 0 __lowerCAmelCase : Optional[int] = torch.zeros((1, context_len) , dtype=torch.long , device=_UpperCamelCase ) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Optional[int] = recopy_model(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) model.train() if secondary_learner is not None: secondary_learner.to(_UpperCamelCase ) secondary_learner.eval() __lowerCAmelCase : Tuple = [] __lowerCAmelCase : Union[str, Any] = 0 __lowerCAmelCase : Tuple = [] __lowerCAmelCase : List[Any] = [] # Compute the performance of the transformer model at the beginning __lowerCAmelCase : List[str] = compute_perplexity(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) test_perps.append(_UpperCamelCase ) print('Test perplexity, step' , _UpperCamelCase , ':' , _UpperCamelCase ) for epoch in range(int(_UpperCamelCase ) ): for step, example in enumerate(_UpperCamelCase ): torch.cuda.empty_cache() __lowerCAmelCase : List[Any] = random.randint(0 , example.size(2 ) - context_len - 1 ) __lowerCAmelCase : Any = example[0, 0, start : start + context_len] lm_optimizer.zero_grad() __lowerCAmelCase : str = model(_UpperCamelCase , labels=_UpperCamelCase ) __lowerCAmelCase : Tuple = True if secondary_learner is not None: __lowerCAmelCase : Any = secondary_learner.forward( torch.tensor(_UpperCamelCase , dtype=torch.long , device=_UpperCamelCase ).unsqueeze(0 ) )[0].item() observed_qs.append(float(_UpperCamelCase ) ) # Here we implement the simple non-constant threshold for the predicted IG(X) value # We will decay the selectivity of our secondary learner filter from # 1 standard deviation above average to 1 below average after 10 batches. if global_step == 10: __lowerCAmelCase : Tuple = -1 if predicted_q < threshold: __lowerCAmelCase : int = False # If we passed the filter, add the context to the batch! if do_backprop: contexts.append(np.array(context.cpu() ) ) __lowerCAmelCase : List[str] = outputs[0] lm_loss.backward() examples += 1 del outputs # Once the batch is filled with enough contexts, backprop on the batch. if examples == batch_size: torch.cuda.empty_cache() __lowerCAmelCase : Union[str, Any] = 0 # Do LM backprop torch.nn.utils.clip_grad_norm_(model.parameters() , 3.0 ) lm_optimizer.step() lm_scheduler.step() # Update learning rate schedule global_step += 1 # Compute the performance of the transformer model at this batch if global_step % eval_interval == 0: __lowerCAmelCase : Tuple = compute_perplexity(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) test_perps.append(_UpperCamelCase ) print('Test perplexity, step' , _UpperCamelCase , ':' , _UpperCamelCase ) # Break out of the loop after 60 batches if max_steps > 0 and global_step > 60: break if max_steps > 0 and global_step > 60: break # save finetuned transformer model torch.save(model.state_dict() , _UpperCamelCase ) torch.cuda.empty_cache() # Do some cleaning up so we can reinitialize for the next run of this function del lm_optimizer del lm_scheduler return model def __lowerCAmelCase (): __lowerCAmelCase : Tuple = argparse.ArgumentParser(description='Fine-tune a transformer model with IGF on a language modeling task' ) # Required parameters parser.add_argument( '--data_dir' , default=_UpperCamelCase , type=_UpperCamelCase , required=_UpperCamelCase , help='The input data dir. Should contain data files for WikiText.' , ) parser.add_argument( '--model_name_or_path' , default=_UpperCamelCase , type=_UpperCamelCase , required=_UpperCamelCase , help='Path to pretrained model or model identifier from huggingface.co/models' , ) parser.add_argument( '--data_file' , type=_UpperCamelCase , default=_UpperCamelCase , help=( 'A jbl file containing tokenized data which can be split as objective dataset, ' 'train_dataset and test_dataset.' ) , ) parser.add_argument( '--igf_data_file' , type=_UpperCamelCase , default=_UpperCamelCase , help='A jbl file containing the context and information gain pairs to train secondary learner.' , ) parser.add_argument( '--output_dir' , default=_UpperCamelCase , type=_UpperCamelCase , required=_UpperCamelCase , help='The output directory where the final fine-tuned model is stored.' , ) parser.add_argument( '--tokenizer_name' , default=_UpperCamelCase , type=_UpperCamelCase , help='Pretrained tokenizer name or path if not the same as model_name' , ) parser.add_argument('--seed' , type=_UpperCamelCase , default=_UpperCamelCase , help='A seed for reproducible training.' ) parser.add_argument( '--context_len' , default=32 , type=_UpperCamelCase , help=( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) , ) parser.add_argument( '--size_objective_set' , default=100 , type=_UpperCamelCase , help='number of articles that are long enough to be used as our objective set' , ) parser.add_argument( '--eval_freq' , default=100 , type=_UpperCamelCase , help='secondary model evaluation is triggered at eval_freq' ) parser.add_argument('--max_steps' , default=1000 , type=_UpperCamelCase , help='To calculate training epochs' ) parser.add_argument( '--secondary_learner_batch_size' , default=128 , type=_UpperCamelCase , help='batch size of training data for secondary learner' , ) parser.add_argument( '--batch_size' , default=16 , type=_UpperCamelCase , help='batch size of training data of language model(gpt2) ' ) parser.add_argument( '--eval_interval' , default=10 , type=_UpperCamelCase , help=( 'decay the selectivity of our secondary learner filter from' '1 standard deviation above average to 1 below average after 10 batches' ) , ) parser.add_argument( '--number' , default=100 , type=_UpperCamelCase , help='The number of examples split to be used as objective_set/test_data' ) parser.add_argument( '--min_len' , default=1026 , type=_UpperCamelCase , help='The minimum length of the article to be used as objective set' ) parser.add_argument( '--secondary_learner_max_epochs' , default=15 , type=_UpperCamelCase , help='number of epochs to train secondary learner' ) parser.add_argument('--trim' , default=_UpperCamelCase , type=_UpperCamelCase , help='truncate the example if it exceeds context length' ) parser.add_argument( '--threshold' , default=1.0 , type=_UpperCamelCase , help=( 'The threshold value used by secondary learner to filter the train_data and allow only' ' informative data as input to the model' ) , ) parser.add_argument('--finetuned_model_name' , default='gpt2_finetuned.pt' , type=_UpperCamelCase , help='finetuned_model_name' ) parser.add_argument( '--recopy_model' , default=_UpperCamelCase , type=_UpperCamelCase , help='Reset the model to the original pretrained GPT-2 weights after each iteration' , ) # function calls # Collecting *n* pairs of context and information gain(X, IG(X)) for training the secondary learner generate_n_pairs( context_len=32 , max_steps=10 , size_objective_set=100 , min_len=1026 , trim=_UpperCamelCase , data_file='data/tokenized_stories_train_wikitext103.jbl' , igf_data_file='igf_context_pairs.jbl' , ) # Load train data for secondary learner __lowerCAmelCase : int = joblib.load('data/IGF_values.jbl' ) # Train secondary learner __lowerCAmelCase : List[Any] = training_secondary_learner( _UpperCamelCase , secondary_learner_max_epochs=15 , secondary_learner_batch_size=128 , eval_freq=100 , igf_model_path='igf_model.pt' , ) # load pretrained gpt2 model __lowerCAmelCase : List[str] = GPTaLMHeadModel.from_pretrained('gpt2' ) set_seed(42 ) # Generate train and test data to train and evaluate gpt2 model __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = generate_datasets( context_len=32 , file='data/tokenized_stories_train_wikitext103.jbl' , number=100 , min_len=1026 , trim=_UpperCamelCase ) # fine-tuning of the gpt2 model using igf (Information Gain Filtration) finetune( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , context_len=32 , max_steps=1000 , batch_size=16 , threshold=1.0 , recopy_model=_UpperCamelCase , secondary_learner=_UpperCamelCase , eval_interval=10 , finetuned_model_name='gpt2_finetuned.pt' , ) if __name__ == "__main__": main()
182
"""simple docstring""" import pytest import datasets.config from datasets.utils.info_utils import is_small_dataset @pytest.mark.parametrize('dataset_size' , [None, 400 * 2**20, 600 * 2**20] ) @pytest.mark.parametrize('input_in_memory_max_size' , ['default', 0, 100 * 2**20, 900 * 2**20] ) def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): if input_in_memory_max_size != "default": monkeypatch.setattr(datasets.config , 'IN_MEMORY_MAX_SIZE' , _UpperCamelCase ) __lowerCAmelCase : Union[str, Any] = datasets.config.IN_MEMORY_MAX_SIZE if input_in_memory_max_size == "default": assert in_memory_max_size == 0 else: assert in_memory_max_size == input_in_memory_max_size if dataset_size and in_memory_max_size: __lowerCAmelCase : Tuple = dataset_size < in_memory_max_size else: __lowerCAmelCase : str = False __lowerCAmelCase : Optional[int] = is_small_dataset(_UpperCamelCase ) assert result == expected
182
1
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed __UpperCAmelCase = "true" def A__ ( __lowerCamelCase, __lowerCamelCase=82, __lowerCamelCase=16 ): set_seed(42 ) SCREAMING_SNAKE_CASE_ = RegressionModel() SCREAMING_SNAKE_CASE_ = deepcopy(UpperCamelCase_ ) SCREAMING_SNAKE_CASE_ = RegressionDataset(length=UpperCamelCase_ ) SCREAMING_SNAKE_CASE_ = DataLoader(UpperCamelCase_, batch_size=UpperCamelCase_ ) model.to(accelerator.device ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = accelerator.prepare(UpperCamelCase_, UpperCamelCase_ ) return model, ddp_model, dataloader def A__ ( __lowerCamelCase, __lowerCamelCase=False ): SCREAMING_SNAKE_CASE_ = AutoTokenizer.from_pretrained('''hf-internal-testing/mrpc-bert-base-cased''' ) SCREAMING_SNAKE_CASE_ = load_dataset('''glue''', '''mrpc''', split='''validation''' ) def tokenize_function(__lowerCamelCase ): SCREAMING_SNAKE_CASE_ = tokenizer(examples['''sentence1'''], examples['''sentence2'''], truncation=UpperCamelCase_, max_length=UpperCamelCase_ ) return outputs with accelerator.main_process_first(): SCREAMING_SNAKE_CASE_ = dataset.map( UpperCamelCase_, batched=UpperCamelCase_, remove_columns=['''idx''', '''sentence1''', '''sentence2'''], ) SCREAMING_SNAKE_CASE_ = tokenized_datasets.rename_column('''label''', '''labels''' ) def collate_fn(__lowerCamelCase ): if use_longest: return tokenizer.pad(UpperCamelCase_, padding='''longest''', return_tensors='''pt''' ) return tokenizer.pad(UpperCamelCase_, padding='''max_length''', max_length=1_28, return_tensors='''pt''' ) return DataLoader(UpperCamelCase_, shuffle=UpperCamelCase_, collate_fn=UpperCamelCase_, batch_size=16 ) def A__ ( __lowerCamelCase, __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = Accelerator(dispatch_batches=UpperCamelCase_, split_batches=UpperCamelCase_ ) SCREAMING_SNAKE_CASE_ = get_dataloader(UpperCamelCase_, not dispatch_batches ) SCREAMING_SNAKE_CASE_ = AutoModelForSequenceClassification.from_pretrained( '''hf-internal-testing/mrpc-bert-base-cased''', return_dict=UpperCamelCase_ ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = accelerator.prepare(UpperCamelCase_, UpperCamelCase_ ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def A__ ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): SCREAMING_SNAKE_CASE_ = [] for batch in dataloader: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = batch.values() with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(UpperCamelCase_ ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = [], [] for logit, targ in logits_and_targets: logits.append(UpperCamelCase_ ) targs.append(UpperCamelCase_ ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = torch.cat(UpperCamelCase_ ), torch.cat(UpperCamelCase_ ) return logits, targs def A__ ( __lowerCamelCase, __lowerCamelCase=82, __lowerCamelCase=False, __lowerCamelCase=False, __lowerCamelCase=16 ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = get_basic_setup(UpperCamelCase_, UpperCamelCase_, UpperCamelCase_ ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = generate_predictions(UpperCamelCase_, UpperCamelCase_, UpperCamelCase_ ) assert ( len(UpperCamelCase_ ) == num_samples ), F'''Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(UpperCamelCase_ )}''' def A__ ( __lowerCamelCase = False, __lowerCamelCase = False ): SCREAMING_SNAKE_CASE_ = evaluate.load('''glue''', '''mrpc''' ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = get_mrpc_setup(UpperCamelCase_, UpperCamelCase_ ) # First do baseline SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = setup['''no'''] model.to(UpperCamelCase_ ) model.eval() for batch in dataloader: batch.to(UpperCamelCase_ ) with torch.inference_mode(): SCREAMING_SNAKE_CASE_ = model(**UpperCamelCase_ ) SCREAMING_SNAKE_CASE_ = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=UpperCamelCase_, references=batch['''labels'''] ) SCREAMING_SNAKE_CASE_ = metric.compute() # Then do distributed SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = setup['''ddp'''] model.eval() for batch in dataloader: with torch.inference_mode(): SCREAMING_SNAKE_CASE_ = model(**UpperCamelCase_ ) SCREAMING_SNAKE_CASE_ = outputs.logits.argmax(dim=-1 ) SCREAMING_SNAKE_CASE_ = batch['''labels'''] SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=UpperCamelCase_, references=UpperCamelCase_ ) SCREAMING_SNAKE_CASE_ = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key], distributed[key] ), F'''Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n''' def A__ ( ): SCREAMING_SNAKE_CASE_ = Accelerator(split_batches=UpperCamelCase_, dispatch_batches=UpperCamelCase_ ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print('''**Testing gather_for_metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(F'''With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`''' ) test_mrpc(UpperCamelCase_, UpperCamelCase_ ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test torch metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: SCREAMING_SNAKE_CASE_ = Accelerator(split_batches=UpperCamelCase_, dispatch_batches=UpperCamelCase_ ) if accelerator.is_local_main_process: print(F'''With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99''' ) test_torch_metrics(UpperCamelCase_, 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test last batch is not dropped when perfectly divisible**''' ) SCREAMING_SNAKE_CASE_ = Accelerator() test_torch_metrics(UpperCamelCase_, 5_12 ) accelerator.state._reset_state() def A__ ( __lowerCamelCase ): main() if __name__ == "__main__": main()
299
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm _SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) @dataclass class A__ ( snake_case__ ): """simple docstring""" __magic_name__ = [ 'no_inference', 'no_cuda', 'no_tpu', 'no_speed', 'no_memory', 'no_env_print', 'no_multi_process', ] def __init__( self , **__snake_case ): for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: snake_case = deprecated_arg[3:] setattr(self , __snake_case , not kwargs.pop(__snake_case ) ) logger.warning( F'''{deprecated_arg} is depreciated. Please use --no_{positive_arg} or''' F''' {positive_arg}={kwargs[positive_arg]}''' ) snake_case = kwargs.pop('''torchscript''' , self.torchscript ) snake_case = kwargs.pop('''torch_xla_tpu_print_metrics''' , self.torch_xla_tpu_print_metrics ) snake_case = kwargs.pop('''fp16_opt_level''' , self.fpaa_opt_level ) super().__init__(**__snake_case ) __magic_name__ = field(default=snake_case__ , metadata={'help': 'Trace the models using torchscript'} ) __magic_name__ = field(default=snake_case__ , metadata={'help': 'Print Xla/PyTorch tpu metrics'} ) __magic_name__ = field( default='O1' , metadata={ 'help': ( 'For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\']. ' 'See details at https://nvidia.github.io/apex/amp.html' ) } , ) @cached_property def a_ ( self ): requires_backends(self , ['''torch'''] ) logger.info('''PyTorch: setting up devices''' ) if not self.cuda: snake_case = torch.device('''cpu''' ) snake_case = 0 elif is_torch_tpu_available(): snake_case = xm.xla_device() snake_case = 0 else: snake_case = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) snake_case = torch.cuda.device_count() return device, n_gpu @property def a_ ( self ): return is_torch_tpu_available() and self.tpu @property def a_ ( self ): requires_backends(self , ['''torch'''] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def a_ ( self ): requires_backends(self , ['''torch'''] ) return self._setup_devices[0] @property def a_ ( self ): requires_backends(self , ['''torch'''] ) return self._setup_devices[1] @property def a_ ( self ): return self.n_gpu > 0
127
0
"""simple docstring""" from math import log from scipy.constants import Boltzmann, physical_constants A_ : Tuple =3_0_0 # TEMPERATURE (unit = K) def SCREAMING_SNAKE_CASE_ ( snake_case : float , snake_case : float , snake_case : float , )-> float: if donor_conc <= 0: raise ValueError('Donor concentration should be positive' ) elif acceptor_conc <= 0: raise ValueError('Acceptor concentration should be positive' ) elif intrinsic_conc <= 0: raise ValueError('Intrinsic concentration should be positive' ) elif donor_conc <= intrinsic_conc: raise ValueError( 'Donor concentration should be greater than intrinsic concentration' ) elif acceptor_conc <= intrinsic_conc: raise ValueError( 'Acceptor concentration should be greater than intrinsic concentration' ) else: return ( Boltzmann * T * log((donor_conc * acceptor_conc) / intrinsic_conc**2 ) / physical_constants["electron volt"][0] ) if __name__ == "__main__": import doctest doctest.testmod()
368
"""simple docstring""" import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor A_ : Union[str, Any] =logging.get_logger(__name__) class __a ( lowerCAmelCase__ ): def __init__( self , *a__ , **a__ ): warnings.warn( 'The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use DPTImageProcessor instead.' , a__ , ) super().__init__(*a__ , **a__ )
80
0
"""simple docstring""" import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () lowercase__ = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). lowercase__ = [0, 25, 50] lowercase__ = [25, 50, 75] lowercase__ = fuzz.membership.trimf(X, abca) lowercase__ = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. lowercase__ = np.ones(75) lowercase__ = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) lowercase__ = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) lowercase__ = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) lowercase__ = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) lowercase__ = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] lowercase__ = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) lowercase__ = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] lowercase__ = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] lowercase__ = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title("""Young""") plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title("""Middle aged""") plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title("""union""") plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title("""intersection""") plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title("""complement_a""") plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title("""difference a/b""") plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title("""alg_sum""") plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title("""alg_product""") plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title("""bdd_sum""") plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title("""bdd_difference""") plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
96
"""simple docstring""" def a__ ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' if height >= 1: move_tower(height - 1 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) move_disk(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) move_tower(height - 1 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def a__ ( SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' print("moving disk from" , SCREAMING_SNAKE_CASE , "to" , SCREAMING_SNAKE_CASE ) def a__ ( ): '''simple docstring''' lowerCAmelCase : Optional[int] = int(input("Height of hanoi: " ).strip() ) move_tower(SCREAMING_SNAKE_CASE , "A" , "B" , "C" ) if __name__ == "__main__": main()
108
0
from __future__ import annotations import inspect import unittest from math import floor import numpy as np from transformers import CvtConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFCvtForImageClassification, TFCvtModel from transformers.models.cvt.modeling_tf_cvt import TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def __magic_name__ ( self : List[str] ): """simple docstring""" _A: List[Any] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCAmelCase_ , '''embed_dim''' ) ) self.parent.assertTrue(hasattr(lowerCAmelCase_ , '''num_heads''' ) ) class UpperCAmelCase : '''simple docstring''' def __init__( self : Union[str, Any] , lowerCAmelCase_ : str , lowerCAmelCase_ : List[str]=1_3 , lowerCAmelCase_ : Tuple=6_4 , lowerCAmelCase_ : Tuple=3 , lowerCAmelCase_ : Any=[1_6, 4_8, 9_6] , lowerCAmelCase_ : Dict=[1, 3, 6] , lowerCAmelCase_ : List[str]=[1, 2, 1_0] , lowerCAmelCase_ : Tuple=[7, 3, 3] , lowerCAmelCase_ : Tuple=[4, 2, 2] , lowerCAmelCase_ : List[Any]=[2, 1, 1] , lowerCAmelCase_ : Any=[2, 2, 2] , lowerCAmelCase_ : Any=[False, False, True] , lowerCAmelCase_ : Optional[Any]=[0.0, 0.0, 0.0] , lowerCAmelCase_ : Optional[int]=0.02 , lowerCAmelCase_ : Dict=1e-12 , lowerCAmelCase_ : Any=True , lowerCAmelCase_ : int=True , lowerCAmelCase_ : List[str]=2 , ): """simple docstring""" _A: Tuple = parent _A: Optional[int] = batch_size _A: Tuple = image_size _A: str = patch_sizes _A: int = patch_stride _A: List[str] = patch_padding _A: Tuple = is_training _A: Tuple = use_labels _A: Union[str, Any] = num_labels _A: List[str] = num_channels _A: Tuple = embed_dim _A: int = num_heads _A: Optional[Any] = stride_kv _A: int = depth _A: Optional[int] = cls_token _A: List[str] = attention_drop_rate _A: Tuple = initializer_range _A: Optional[Any] = layer_norm_eps def __magic_name__ ( self : Optional[int] ): """simple docstring""" _A: Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _A: str = None if self.use_labels: # create a random int32 tensor of given shape _A: Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) _A: Dict = self.get_config() return config, pixel_values, labels def __magic_name__ ( self : Dict ): """simple docstring""" return CvtConfig( image_size=self.image_size , num_labels=self.num_labels , num_channels=self.num_channels , embed_dim=self.embed_dim , num_heads=self.num_heads , patch_sizes=self.patch_sizes , patch_padding=self.patch_padding , patch_stride=self.patch_stride , stride_kv=self.stride_kv , depth=self.depth , cls_token=self.cls_token , attention_drop_rate=self.attention_drop_rate , initializer_range=self.initializer_range , ) def __magic_name__ ( self : Union[str, Any] , lowerCAmelCase_ : int , lowerCAmelCase_ : str , lowerCAmelCase_ : List[str] ): """simple docstring""" _A: str = TFCvtModel(config=lowerCAmelCase_ ) _A: int = model(lowerCAmelCase_ , training=lowerCAmelCase_ ) _A: int = (self.image_size, self.image_size) _A , _A: List[str] = image_size[0], image_size[1] for i in range(len(self.depth ) ): _A: List[str] = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) _A: Optional[int] = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dim[-1], height, width) ) def __magic_name__ ( self : Union[str, Any] , lowerCAmelCase_ : Any , lowerCAmelCase_ : Dict , lowerCAmelCase_ : Union[str, Any] ): """simple docstring""" _A: List[str] = self.num_labels _A: Union[str, Any] = TFCvtForImageClassification(lowerCAmelCase_ ) _A: Tuple = model(lowerCAmelCase_ , labels=lowerCAmelCase_ , training=lowerCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __magic_name__ ( self : Optional[int] ): """simple docstring""" _A: List[Any] = self.prepare_config_and_inputs() _A , _A , _A: List[str] = config_and_inputs _A: Any = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): '''simple docstring''' __UpperCamelCase : Any = (TFCvtModel, TFCvtForImageClassification) if is_tf_available() else () __UpperCamelCase : Optional[Any] = ( {'''feature-extraction''': TFCvtModel, '''image-classification''': TFCvtForImageClassification} if is_tf_available() else {} ) __UpperCamelCase : Dict = False __UpperCamelCase : Tuple = False __UpperCamelCase : int = False __UpperCamelCase : Optional[Any] = False __UpperCamelCase : Any = False def __magic_name__ ( self : str ): """simple docstring""" _A: str = TFCvtModelTester(self ) _A: Any = TFCvtConfigTester(self , config_class=lowerCAmelCase_ , has_text_modality=lowerCAmelCase_ , hidden_size=3_7 ) def __magic_name__ ( self : int ): """simple docstring""" self.config_tester.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() @unittest.skip(reason='''Cvt does not output attentions''' ) def __magic_name__ ( self : Union[str, Any] ): """simple docstring""" pass @unittest.skip(reason='''Cvt does not use inputs_embeds''' ) def __magic_name__ ( self : Optional[int] ): """simple docstring""" pass @unittest.skip(reason='''Cvt does not support input and output embeddings''' ) def __magic_name__ ( self : Optional[int] ): """simple docstring""" pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices('''GPU''' ) ) == 0 , reason='''TF does not support backprop for grouped convolutions on CPU.''' , ) def __magic_name__ ( self : Any ): """simple docstring""" super().test_dataset_conversion() @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices('''GPU''' ) ) == 0 , reason='''TF does not support backprop for grouped convolutions on CPU.''' , ) @slow def __magic_name__ ( self : str ): """simple docstring""" super().test_keras_fit() @unittest.skip(reason='''Get `Failed to determine best cudnn convolution algo.` error after using TF 2.12+cuda 11.8''' ) def __magic_name__ ( self : Optional[int] ): """simple docstring""" _A: Tuple = tf.keras.mixed_precision.Policy('''mixed_float16''' ) tf.keras.mixed_precision.set_global_policy(lowerCAmelCase_ ) super().test_keras_fit() tf.keras.mixed_precision.set_global_policy('''float32''' ) def __magic_name__ ( self : Any ): """simple docstring""" _A , _A: List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A: List[str] = model_class(lowerCAmelCase_ ) _A: Optional[Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _A: int = [*signature.parameters.keys()] _A: str = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCAmelCase_ ) def __magic_name__ ( self : str ): """simple docstring""" def check_hidden_states_output(lowerCAmelCase_ : Dict , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : str ): _A: Any = model_class(lowerCAmelCase_ ) _A: Optional[Any] = model(**self._prepare_for_class(lowerCAmelCase_ , lowerCAmelCase_ ) ) _A: Any = outputs.hidden_states _A: Tuple = len(self.model_tester.depth ) self.assertEqual(len(lowerCAmelCase_ ) , lowerCAmelCase_ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) _A , _A: Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _A: List[str] = True check_hidden_states_output(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _A: List[Any] = True check_hidden_states_output(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def __magic_name__ ( self : List[Any] ): """simple docstring""" _A: List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase_ ) def __magic_name__ ( self : int ): """simple docstring""" _A: int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase_ ) @slow def __magic_name__ ( self : Optional[int] ): """simple docstring""" for model_name in TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _A: Any = TFCvtModel.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) def lowerCamelCase__ ( ) -> Any: _A: int = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def __magic_name__ ( self : List[Any] ): """simple docstring""" return AutoImageProcessor.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def __magic_name__ ( self : Any ): """simple docstring""" _A: Dict = TFCvtForImageClassification.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) _A: Optional[Any] = self.default_image_processor _A: Tuple = prepare_img() _A: Optional[Any] = image_processor(images=lowerCAmelCase_ , return_tensors='''tf''' ) # forward pass _A: List[str] = model(**lowerCAmelCase_ ) # verify the logits _A: List[str] = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase_ ) _A: List[str] = tf.constant([0.9285, 0.9015, -0.3150] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , lowerCAmelCase_ , atol=1e-4 ) )
301
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase__ : Union[str, Any] = logging.get_logger(__name__) UpperCAmelCase__ : List[Any] = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } UpperCAmelCase__ : Tuple = { 'vocab_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json' }, 'merges_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt' }, 'tokenizer_config_file': { 'facebook/blenderbot_small-90M': ( 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json' ) }, } UpperCAmelCase__ : Optional[int] = {'facebook/blenderbot_small-90M': 512} def lowerCamelCase__ ( a ) -> Optional[Any]: _A: List[Any] = set() _A: List[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _A: List[Any] = char _A: Union[str, Any] = set(a ) return pairs class UpperCAmelCase ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' __UpperCamelCase : str = VOCAB_FILES_NAMES __UpperCamelCase : List[Any] = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase : Union[str, Any] = ['''input_ids''', '''attention_mask'''] def __init__( self : int , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : List[str]="__start__" , lowerCAmelCase_ : Any="__end__" , lowerCAmelCase_ : Any="__unk__" , lowerCAmelCase_ : Any="__null__" , **lowerCAmelCase_ : int , ): """simple docstring""" super().__init__(unk_token=lowerCAmelCase_ , bos_token=lowerCAmelCase_ , eos_token=lowerCAmelCase_ , pad_token=lowerCAmelCase_ , **lowerCAmelCase_ ) with open(lowerCAmelCase_ , encoding='''utf-8''' ) as vocab_handle: _A: Optional[int] = json.load(lowerCAmelCase_ ) _A: int = {v: k for k, v in self.encoder.items()} with open(lowerCAmelCase_ , encoding='''utf-8''' ) as merges_handle: _A: Dict = merges_handle.read().split('''\n''' )[1:-1] _A: int = [tuple(merge.split() ) for merge in merges] _A: Dict = dict(zip(lowerCAmelCase_ , range(len(lowerCAmelCase_ ) ) ) ) _A: Union[str, Any] = {} @property def __magic_name__ ( self : Optional[int] ): """simple docstring""" return len(self.encoder ) def __magic_name__ ( self : Optional[int] ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def __magic_name__ ( self : str , lowerCAmelCase_ : str ): """simple docstring""" if token in self.cache: return self.cache[token] _A: List[Any] = re.sub('''([.,!?()])''' , R''' \1''' , lowerCAmelCase_ ) _A: List[Any] = re.sub('''(\')''' , R''' \1 ''' , lowerCAmelCase_ ) _A: List[Any] = re.sub(R'''\s{2,}''' , ''' ''' , lowerCAmelCase_ ) if "\n" in token: _A: Dict = token.replace('''\n''' , ''' __newln__''' ) _A: Any = token.split(''' ''' ) _A: Optional[Any] = [] for token in tokens: if not len(lowerCAmelCase_ ): continue _A: str = token.lower() _A: List[str] = tuple(lowerCAmelCase_ ) _A: str = tuple(list(word[:-1] ) + [word[-1] + '''</w>'''] ) _A: Dict = get_pairs(lowerCAmelCase_ ) if not pairs: words.append(lowerCAmelCase_ ) continue while True: _A: str = min(lowerCAmelCase_ , key=lambda lowerCAmelCase_ : self.bpe_ranks.get(lowerCAmelCase_ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break _A , _A: Optional[int] = bigram _A: str = [] _A: Dict = 0 while i < len(lowerCAmelCase_ ): try: _A: List[Any] = word.index(lowerCAmelCase_ , lowerCAmelCase_ ) new_word.extend(word[i:j] ) _A: Optional[int] = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(lowerCAmelCase_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _A: Union[str, Any] = tuple(lowerCAmelCase_ ) _A: Tuple = new_word if len(lowerCAmelCase_ ) == 1: break else: _A: Optional[int] = get_pairs(lowerCAmelCase_ ) _A: str = '''@@ '''.join(lowerCAmelCase_ ) _A: Tuple = word[:-4] _A: List[Any] = word words.append(lowerCAmelCase_ ) return " ".join(lowerCAmelCase_ ) def __magic_name__ ( self : str , lowerCAmelCase_ : str ): """simple docstring""" _A: List[Any] = [] _A: List[Any] = re.findall(R'''\S+\n?''' , lowerCAmelCase_ ) for token in words: split_tokens.extend(list(self.bpe(lowerCAmelCase_ ).split(''' ''' ) ) ) return split_tokens def __magic_name__ ( self : str , lowerCAmelCase_ : str ): """simple docstring""" _A: List[str] = token.lower() return self.encoder.get(lowerCAmelCase_ , self.encoder.get(self.unk_token ) ) def __magic_name__ ( self : int , lowerCAmelCase_ : int ): """simple docstring""" return self.decoder.get(lowerCAmelCase_ , self.unk_token ) def __magic_name__ ( self : List[str] , lowerCAmelCase_ : List[str] ): """simple docstring""" _A: List[str] = ''' '''.join(lowerCAmelCase_ ).replace('''@@ ''' , '''''' ).strip() return out_string def __magic_name__ ( self : List[Any] , lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[str] = None ): """simple docstring""" if not os.path.isdir(lowerCAmelCase_ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _A: Dict = os.path.join( lowerCAmelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) _A: Any = os.path.join( lowerCAmelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(lowerCAmelCase_ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCAmelCase_ , ensure_ascii=lowerCAmelCase_ ) + '''\n''' ) _A: List[str] = 0 with open(lowerCAmelCase_ , '''w''' , encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCAmelCase_ : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ''' Please check that the tokenizer is not corrupted!''' ) _A: Optional[int] = token_index writer.write(''' '''.join(lowerCAmelCase_ ) + '''\n''' ) index += 1 return vocab_file, merge_file
301
1
from __future__ import annotations import unittest from transformers import DistilBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.distilbert.modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertModel, ) class __lowerCAmelCase : def __init__( self , lowerCAmelCase__ , ) -> int: '''simple docstring''' a__ : Optional[int] =parent a__ : Union[str, Any] =1_3 a__ : Tuple =7 a__ : List[str] =True a__ : Union[str, Any] =True a__ : Optional[Any] =False a__ : List[Any] =True a__ : Optional[int] =9_9 a__ : Any =3_2 a__ : List[str] =2 a__ : List[Any] =4 a__ : List[Any] =3_7 a__ : Tuple ="gelu" a__ : Optional[Any] =0.1 a__ : Dict =0.1 a__ : int =5_1_2 a__ : str =1_6 a__ : List[Any] =2 a__ : int =0.02 a__ : List[Any] =3 a__ : Any =4 a__ : Optional[int] =None def _lowercase ( self ) -> Optional[Any]: '''simple docstring''' a__ : Optional[int] =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a__ : str =None if self.use_input_mask: a__ : Optional[Any] =random_attention_mask([self.batch_size, self.seq_length] ) a__ : List[str] =None a__ : str =None a__ : Dict =None if self.use_labels: a__ : Optional[int] =ids_tensor([self.batch_size] , self.type_sequence_label_size ) a__ : List[Any] =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a__ : Any =ids_tensor([self.batch_size] , self.num_choices ) a__ : List[str] =DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Union[str, Any]: '''simple docstring''' a__ : int =TFDistilBertModel(config=lowerCAmelCase__ ) a__ : Tuple ={"input_ids": input_ids, "attention_mask": input_mask} a__ : List[Any] =model(lowerCAmelCase__ ) a__ : Tuple =[input_ids, input_mask] a__ : Tuple =model(lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Union[str, Any]: '''simple docstring''' a__ : Union[str, Any] =TFDistilBertForMaskedLM(config=lowerCAmelCase__ ) a__ : Dict ={"input_ids": input_ids, "attention_mask": input_mask} a__ : List[str] =model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[Any]: '''simple docstring''' a__ : List[str] =TFDistilBertForQuestionAnswering(config=lowerCAmelCase__ ) a__ : Union[str, Any] ={ "input_ids": input_ids, "attention_mask": input_mask, } a__ : int =model(lowerCAmelCase__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: '''simple docstring''' a__ : Tuple =self.num_labels a__ : Optional[int] =TFDistilBertForSequenceClassification(lowerCAmelCase__ ) a__ : List[Any] ={"input_ids": input_ids, "attention_mask": input_mask} a__ : List[Any] =model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Union[str, Any]: '''simple docstring''' a__ : Union[str, Any] =self.num_choices a__ : Dict =TFDistilBertForMultipleChoice(lowerCAmelCase__ ) a__ : Dict =tf.tile(tf.expand_dims(lowerCAmelCase__ , 1 ) , (1, self.num_choices, 1) ) a__ : Optional[int] =tf.tile(tf.expand_dims(lowerCAmelCase__ , 1 ) , (1, self.num_choices, 1) ) a__ : Optional[int] ={ "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, } a__ : int =model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[Any]: '''simple docstring''' a__ : Optional[int] =self.num_labels a__ : Dict =TFDistilBertForTokenClassification(lowerCAmelCase__ ) a__ : str ={"input_ids": input_ids, "attention_mask": input_mask} a__ : Tuple =model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowercase ( self ) -> Dict: '''simple docstring''' a__ : Dict =self.prepare_config_and_inputs() ((a__) , (a__) , (a__) , (a__) , (a__) , (a__)) : Dict =config_and_inputs a__ : List[str] ={"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase): _lowercase : Any = ( ( TFDistilBertModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertForMultipleChoice, ) if is_tf_available() else None ) _lowercase : List[Any] = ( { """feature-extraction""": TFDistilBertModel, """fill-mask""": TFDistilBertForMaskedLM, """question-answering""": TFDistilBertForQuestionAnswering, """text-classification""": TFDistilBertForSequenceClassification, """token-classification""": TFDistilBertForTokenClassification, """zero-shot""": TFDistilBertForSequenceClassification, } if is_tf_available() else {} ) _lowercase : int = False _lowercase : List[str] = False def _lowercase ( self ) -> Tuple: '''simple docstring''' a__ : Dict =TFDistilBertModelTester(self ) a__ : Optional[int] =ConfigTester(self , config_class=lowerCAmelCase__ , dim=3_7 ) def _lowercase ( self ) -> int: '''simple docstring''' self.config_tester.run_common_tests() def _lowercase ( self ) -> Tuple: '''simple docstring''' a__ : int =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*lowerCAmelCase__ ) def _lowercase ( self ) -> Union[str, Any]: '''simple docstring''' a__ : Optional[int] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*lowerCAmelCase__ ) def _lowercase ( self ) -> Optional[int]: '''simple docstring''' a__ : Dict =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*lowerCAmelCase__ ) def _lowercase ( self ) -> List[str]: '''simple docstring''' a__ : List[str] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*lowerCAmelCase__ ) def _lowercase ( self ) -> str: '''simple docstring''' a__ : int =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*lowerCAmelCase__ ) def _lowercase ( self ) -> Union[str, Any]: '''simple docstring''' a__ : Union[str, Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*lowerCAmelCase__ ) @slow def _lowercase ( self ) -> Union[str, Any]: '''simple docstring''' for model_name in list(TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1] ): a__ : List[str] =TFDistilBertModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) @require_tf class __lowerCAmelCase ( unittest.TestCase): @slow def _lowercase ( self ) -> Optional[int]: '''simple docstring''' a__ : Dict =TFDistilBertModel.from_pretrained("distilbert-base-uncased" ) a__ : Optional[int] =tf.constant([[0, 1, 2, 3, 4, 5]] ) a__ : Tuple =model(lowerCAmelCase__ )[0] a__ : List[str] =[1, 6, 7_6_8] self.assertEqual(output.shape , lowerCAmelCase__ ) a__ : List[Any] =tf.constant( [ [ [0.19_26_18_85, -0.13_73_29_55, 0.4_11_97_99], [0.22_15_01_56, -0.07_42_26_61, 0.39_03_72_04], [0.22_75_60_18, -0.0_89_64_14, 0.3_70_14_67], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , lowerCAmelCase__ , atol=1E-4 )
95
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_funnel import FunnelTokenizer UpperCAmelCase : int = logging.get_logger(__name__) UpperCAmelCase : Optional[Any] = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} UpperCAmelCase : List[Any] = [ """small""", """small-base""", """medium""", """medium-base""", """intermediate""", """intermediate-base""", """large""", """large-base""", """xlarge""", """xlarge-base""", ] UpperCAmelCase : Optional[int] = { """vocab_file""": { """funnel-transformer/small""": """https://huggingface.co/funnel-transformer/small/resolve/main/vocab.txt""", """funnel-transformer/small-base""": """https://huggingface.co/funnel-transformer/small-base/resolve/main/vocab.txt""", """funnel-transformer/medium""": """https://huggingface.co/funnel-transformer/medium/resolve/main/vocab.txt""", """funnel-transformer/medium-base""": ( """https://huggingface.co/funnel-transformer/medium-base/resolve/main/vocab.txt""" ), """funnel-transformer/intermediate""": ( """https://huggingface.co/funnel-transformer/intermediate/resolve/main/vocab.txt""" ), """funnel-transformer/intermediate-base""": ( """https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/vocab.txt""" ), """funnel-transformer/large""": """https://huggingface.co/funnel-transformer/large/resolve/main/vocab.txt""", """funnel-transformer/large-base""": """https://huggingface.co/funnel-transformer/large-base/resolve/main/vocab.txt""", """funnel-transformer/xlarge""": """https://huggingface.co/funnel-transformer/xlarge/resolve/main/vocab.txt""", """funnel-transformer/xlarge-base""": ( """https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """funnel-transformer/small""": """https://huggingface.co/funnel-transformer/small/resolve/main/tokenizer.json""", """funnel-transformer/small-base""": ( """https://huggingface.co/funnel-transformer/small-base/resolve/main/tokenizer.json""" ), """funnel-transformer/medium""": """https://huggingface.co/funnel-transformer/medium/resolve/main/tokenizer.json""", """funnel-transformer/medium-base""": ( """https://huggingface.co/funnel-transformer/medium-base/resolve/main/tokenizer.json""" ), """funnel-transformer/intermediate""": ( """https://huggingface.co/funnel-transformer/intermediate/resolve/main/tokenizer.json""" ), """funnel-transformer/intermediate-base""": ( """https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/tokenizer.json""" ), """funnel-transformer/large""": """https://huggingface.co/funnel-transformer/large/resolve/main/tokenizer.json""", """funnel-transformer/large-base""": ( """https://huggingface.co/funnel-transformer/large-base/resolve/main/tokenizer.json""" ), """funnel-transformer/xlarge""": """https://huggingface.co/funnel-transformer/xlarge/resolve/main/tokenizer.json""", """funnel-transformer/xlarge-base""": ( """https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/tokenizer.json""" ), }, } UpperCAmelCase : Optional[int] = {F"""funnel-transformer/{name}""": 512 for name in _model_names} UpperCAmelCase : Optional[int] = {F"""funnel-transformer/{name}""": {"""do_lower_case""": True} for name in _model_names} class __lowerCAmelCase ( UpperCamelCase__): _lowercase : str = VOCAB_FILES_NAMES _lowercase : List[Any] = PRETRAINED_VOCAB_FILES_MAP _lowercase : Dict = PRETRAINED_INIT_CONFIGURATION _lowercase : Union[str, Any] = FunnelTokenizer _lowercase : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowercase : int = 2 def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__="<unk>" , lowerCAmelCase__="<sep>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<cls>" , lowerCAmelCase__="<mask>" , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__=True , lowerCAmelCase__=True , lowerCAmelCase__=None , lowerCAmelCase__="##" , **lowerCAmelCase__ , ) -> Optional[int]: '''simple docstring''' super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , clean_text=lowerCAmelCase__ , tokenize_chinese_chars=lowerCAmelCase__ , strip_accents=lowerCAmelCase__ , wordpieces_prefix=lowerCAmelCase__ , **lowerCAmelCase__ , ) a__ : Optional[Any] =json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowerCAmelCase__ ) != do_lower_case or normalizer_state.get("strip_accents" , lowerCAmelCase__ ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowerCAmelCase__ ) != tokenize_chinese_chars ): a__ : List[str] =getattr(lowerCAmelCase__ , normalizer_state.pop("type" ) ) a__ : Union[str, Any] =do_lower_case a__ : Any =strip_accents a__ : Optional[Any] =tokenize_chinese_chars a__ : Dict =normalizer_class(**lowerCAmelCase__ ) a__ : Any =do_lower_case def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__=None ) -> str: '''simple docstring''' a__ : Dict =[self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: '''simple docstring''' a__ : Optional[int] =[self.sep_token_id] a__ : Union[str, Any] =[self.cls_token_id] if token_ids_a is None: return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: '''simple docstring''' a__ : Tuple =self._tokenizer.model.save(lowerCAmelCase__ , name=lowerCAmelCase__ ) return tuple(lowerCAmelCase__ )
95
1
"""simple docstring""" import json import os import unittest from transformers import BatchEncoding, MvpTokenizer, MvpTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin, filter_roberta_detectors @require_tokenizers class lowerCAmelCase_ ( lowerCAmelCase , unittest.TestCase ): """simple docstring""" _lowerCAmelCase : Optional[int] = MvpTokenizer _lowerCAmelCase : int = MvpTokenizerFast _lowerCAmelCase : Optional[int] = True _lowerCAmelCase : str = filter_roberta_detectors def snake_case ( self ): """simple docstring""" super().setUp() snake_case = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', '\u0120', '\u0120l', '\u0120n', '\u0120lo', '\u0120low', 'er', '\u0120lowest', '\u0120newer', '\u0120wider', '<unk>', ] snake_case = dict(zip(lowerCAmelCase , range(len(lowerCAmelCase ) ) ) ) snake_case = ['#version: 0.2', '\u0120 l', '\u0120l o', '\u0120lo w', 'e r', ''] snake_case = {'unk_token': '<unk>'} snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(lowerCAmelCase ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(lowerCAmelCase ) ) def snake_case ( self , **lowerCAmelCase ): """simple docstring""" kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **lowerCAmelCase ) def snake_case ( self , **lowerCAmelCase ): """simple docstring""" kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **lowerCAmelCase ) def snake_case ( self , lowerCAmelCase ): """simple docstring""" return "lower newer", "lower newer" @cached_property def snake_case ( self ): """simple docstring""" return MvpTokenizer.from_pretrained('RUCAIBox/mvp' ) @cached_property def snake_case ( self ): """simple docstring""" return MvpTokenizerFast.from_pretrained('RUCAIBox/mvp' ) @require_torch def snake_case ( self ): """simple docstring""" snake_case = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] snake_case = [0, 2_50, 2_51, 1_78_18, 13, 3_91_86, 19_38, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: snake_case = tokenizer(lowerCAmelCase , max_length=len(lowerCAmelCase ) , padding=lowerCAmelCase , return_tensors='pt' ) self.assertIsInstance(lowerCAmelCase , lowerCAmelCase ) self.assertEqual((2, 9) , batch.input_ids.shape ) self.assertEqual((2, 9) , batch.attention_mask.shape ) snake_case = batch.input_ids.tolist()[0] self.assertListEqual(lowerCAmelCase , lowerCAmelCase ) # Test that special tokens are reset @require_torch def snake_case ( self ): """simple docstring""" snake_case = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: snake_case = tokenizer(lowerCAmelCase , padding=lowerCAmelCase , return_tensors='pt' ) # check if input_ids are returned and no labels self.assertIn('input_ids' , lowerCAmelCase ) self.assertIn('attention_mask' , lowerCAmelCase ) self.assertNotIn('labels' , lowerCAmelCase ) self.assertNotIn('decoder_attention_mask' , lowerCAmelCase ) @require_torch def snake_case ( self ): """simple docstring""" snake_case = [ 'Summary of the text.', 'Another summary.', ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: snake_case = tokenizer(text_target=lowerCAmelCase , max_length=32 , padding='max_length' , return_tensors='pt' ) self.assertEqual(32 , targets['input_ids'].shape[1] ) @require_torch def snake_case ( self ): """simple docstring""" for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: snake_case = tokenizer( ['I am a small frog' * 10_24, 'I am a small frog'] , padding=lowerCAmelCase , truncation=lowerCAmelCase , return_tensors='pt' ) self.assertIsInstance(lowerCAmelCase , lowerCAmelCase ) self.assertEqual(batch.input_ids.shape , (2, 10_24) ) @require_torch def snake_case ( self ): """simple docstring""" snake_case = ['A long paragraph for summarization.'] snake_case = [ 'Summary of the text.', ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: snake_case = tokenizer(lowerCAmelCase , text_target=lowerCAmelCase , return_tensors='pt' ) snake_case = inputs['input_ids'] snake_case = inputs['labels'] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) def snake_case ( self ): """simple docstring""" pass def snake_case ( self ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): snake_case = self.rust_tokenizer_class.from_pretrained(lowerCAmelCase , **lowerCAmelCase ) snake_case = self.tokenizer_class.from_pretrained(lowerCAmelCase , **lowerCAmelCase ) snake_case = 'A, <mask> AllenNLP sentence.' snake_case = tokenizer_r.encode_plus(lowerCAmelCase , add_special_tokens=lowerCAmelCase , return_token_type_ids=lowerCAmelCase ) snake_case = tokenizer_p.encode_plus(lowerCAmelCase , add_special_tokens=lowerCAmelCase , return_token_type_ids=lowerCAmelCase ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r['token_type_ids'] ) , sum(tokens_p['token_type_ids'] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r['attention_mask'] ) / len(tokens_r['attention_mask'] ) , sum(tokens_p['attention_mask'] ) / len(tokens_p['attention_mask'] ) , ) snake_case = tokenizer_r.convert_ids_to_tokens(tokens_r['input_ids'] ) snake_case = tokenizer_p.convert_ids_to_tokens(tokens_p['input_ids'] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p['input_ids'] , [0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual(tokens_r['input_ids'] , [0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual( lowerCAmelCase , ['<s>', 'A', ',', '<mask>', 'ĠAllen', 'N', 'LP', 'Ġsentence', '.', '</s>'] ) self.assertSequenceEqual( lowerCAmelCase , ['<s>', 'A', ',', '<mask>', 'ĠAllen', 'N', 'LP', 'Ġsentence', '.', '</s>'] )
149
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ = { "microsoft/swinv2-tiny-patch4-window8-256": ( "https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json" ), } class lowerCAmelCase_ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase : Tuple = """swinv2""" _lowerCAmelCase : Any = { """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , lowerCAmelCase=2_24 , lowerCAmelCase=4 , lowerCAmelCase=3 , lowerCAmelCase=96 , lowerCAmelCase=[2, 2, 6, 2] , lowerCAmelCase=[3, 6, 12, 24] , lowerCAmelCase=7 , lowerCAmelCase=4.0 , lowerCAmelCase=True , lowerCAmelCase=0.0 , lowerCAmelCase=0.0 , lowerCAmelCase=0.1 , lowerCAmelCase="gelu" , lowerCAmelCase=False , lowerCAmelCase=0.02 , lowerCAmelCase=1E-5 , lowerCAmelCase=32 , **lowerCAmelCase , ): """simple docstring""" super().__init__(**lowerCAmelCase ) snake_case = image_size snake_case = patch_size snake_case = num_channels snake_case = embed_dim snake_case = depths snake_case = len(lowerCAmelCase ) snake_case = num_heads snake_case = window_size snake_case = mlp_ratio snake_case = qkv_bias snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = drop_path_rate snake_case = hidden_act snake_case = use_absolute_embeddings snake_case = layer_norm_eps snake_case = initializer_range snake_case = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model snake_case = int(embed_dim * 2 ** (len(lowerCAmelCase ) - 1) ) snake_case = (0, 0, 0, 0)
149
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A_ : int = { """configuration_distilbert""": [ """DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """DistilBertConfig""", """DistilBertOnnxConfig""", ], """tokenization_distilbert""": ["""DistilBertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Tuple = ["""DistilBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Any = [ """DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """DistilBertForMaskedLM""", """DistilBertForMultipleChoice""", """DistilBertForQuestionAnswering""", """DistilBertForSequenceClassification""", """DistilBertForTokenClassification""", """DistilBertModel""", """DistilBertPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : Union[str, Any] = [ """TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFDistilBertForMaskedLM""", """TFDistilBertForMultipleChoice""", """TFDistilBertForQuestionAnswering""", """TFDistilBertForSequenceClassification""", """TFDistilBertForTokenClassification""", """TFDistilBertMainLayer""", """TFDistilBertModel""", """TFDistilBertPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ : List[Any] = [ """FlaxDistilBertForMaskedLM""", """FlaxDistilBertForMultipleChoice""", """FlaxDistilBertForQuestionAnswering""", """FlaxDistilBertForSequenceClassification""", """FlaxDistilBertForTokenClassification""", """FlaxDistilBertModel""", """FlaxDistilBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys A_ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
215
'''simple docstring''' import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class lowercase ( enum.Enum ): """simple docstring""" UpperCAmelCase = 0 UpperCAmelCase = 1 UpperCAmelCase = 2 @add_end_docstrings(_lowerCamelCase ) class lowercase ( _lowerCamelCase ): """simple docstring""" UpperCAmelCase = """ In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The voice of Nicholas's young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision and denounces one of the men as a horse thief. Although his father initially slaps him for making such an accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop, begging for his blessing. <eod> </s> <eos> """ def __init__( self ,*a_ ,**a_ ) -> Union[str, Any]: super().__init__(*a_ ,**a_ ) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == """tf""" else MODEL_FOR_CAUSAL_LM_MAPPING ) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. _UpperCAmelCase : List[str] = None if self.model.config.prefix is not None: _UpperCAmelCase : Any = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. _UpperCAmelCase : Union[str, Any] = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. _UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase : List[Any] = self._sanitize_parameters(prefix=a_ ,**self._forward_params ) _UpperCAmelCase : Optional[int] = {**self._preprocess_params, **preprocess_params} _UpperCAmelCase : List[Any] = {**self._forward_params, **forward_params} def _snake_case ( self ,a_=None ,a_=None ,a_=None ,a_=None ,a_=None ,a_=None ,a_=None ,a_=None ,**a_ ,) -> Dict: _UpperCAmelCase : int = {} if prefix is not None: _UpperCAmelCase : Union[str, Any] = prefix if prefix: _UpperCAmelCase : Union[str, Any] = self.tokenizer( a_ ,padding=a_ ,add_special_tokens=a_ ,return_tensors=self.framework ) _UpperCAmelCase : Optional[int] = prefix_inputs["""input_ids"""].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( f'''{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected''' """ [None, 'hole']""" ) _UpperCAmelCase : Optional[Any] = handle_long_generation preprocess_params.update(a_ ) _UpperCAmelCase : str = generate_kwargs _UpperCAmelCase : str = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError("""`return_text` is mutually exclusive with `return_full_text`""" ) if return_tensors is not None: raise ValueError("""`return_full_text` is mutually exclusive with `return_tensors`""" ) _UpperCAmelCase : Any = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError("""`return_text` is mutually exclusive with `return_tensors`""" ) _UpperCAmelCase : Tuple = ReturnType.TENSORS if return_type is not None: _UpperCAmelCase : int = return_type if clean_up_tokenization_spaces is not None: _UpperCAmelCase : Tuple = clean_up_tokenization_spaces if stop_sequence is not None: _UpperCAmelCase : str = self.tokenizer.encode(a_ ,add_special_tokens=a_ ) if len(a_ ) > 1: warnings.warn( """Stopping on a multiple token sequence is not yet supported on transformers. The first token of""" """ the stop sequence will be used as the stop sequence string in the interim.""" ) _UpperCAmelCase : str = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def _snake_case ( self ,*a_ ,**a_ ) -> Dict: # Parse arguments if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({"""add_space_before_punct_symbol""": True} ) return super()._parse_and_tokenize(*a_ ,**a_ ) def __call__( self ,a_ ,**a_ ) -> str: return super().__call__(a_ ,**a_ ) def _snake_case ( self ,a_ ,a_="" ,a_=None ,**a_ ) -> Optional[Any]: _UpperCAmelCase : str = self.tokenizer( prefix + prompt_text ,padding=a_ ,add_special_tokens=a_ ,return_tensors=self.framework ) _UpperCAmelCase : Optional[Any] = prompt_text if handle_long_generation == "hole": _UpperCAmelCase : Dict = inputs["""input_ids"""].shape[-1] if "max_new_tokens" in generate_kwargs: _UpperCAmelCase : str = generate_kwargs["""max_new_tokens"""] else: _UpperCAmelCase : Optional[int] = generate_kwargs.get("""max_length""" ,self.model.config.max_length ) - cur_len if new_tokens < 0: raise ValueError("""We cannot infer how many new tokens are expected""" ) if cur_len + new_tokens > self.tokenizer.model_max_length: _UpperCAmelCase : str = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( """We cannot use `hole` to handle this generation the number of desired tokens exceeds the""" """ models max length""" ) _UpperCAmelCase : Optional[Any] = inputs["""input_ids"""][:, -keep_length:] if "attention_mask" in inputs: _UpperCAmelCase : Optional[int] = inputs["""attention_mask"""][:, -keep_length:] return inputs def _snake_case ( self ,a_ ,**a_ ) -> Union[str, Any]: _UpperCAmelCase : Optional[Any] = model_inputs["""input_ids"""] _UpperCAmelCase : List[str] = model_inputs.get("""attention_mask""" ,a_ ) # Allow empty prompts if input_ids.shape[1] == 0: _UpperCAmelCase : Optional[int] = None _UpperCAmelCase : List[Any] = None _UpperCAmelCase : int = 1 else: _UpperCAmelCase : List[str] = input_ids.shape[0] _UpperCAmelCase : Any = model_inputs.pop("""prompt_text""" ) # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. _UpperCAmelCase : List[Any] = generate_kwargs.pop("""prefix_length""" ,0 ) if prefix_length > 0: _UpperCAmelCase : Optional[int] = """max_new_tokens""" in generate_kwargs or ( """generation_config""" in generate_kwargs and generate_kwargs["""generation_config"""].max_new_tokens is not None ) if not has_max_new_tokens: _UpperCAmelCase : Optional[int] = generate_kwargs.get("""max_length""" ) or self.model.config.max_length generate_kwargs["max_length"] += prefix_length _UpperCAmelCase : str = """min_new_tokens""" in generate_kwargs or ( """generation_config""" in generate_kwargs and generate_kwargs["""generation_config"""].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL _UpperCAmelCase : Optional[int] = self.model.generate(input_ids=a_ ,attention_mask=a_ ,**a_ ) _UpperCAmelCase : Dict = generated_sequence.shape[0] if self.framework == "pt": _UpperCAmelCase : Optional[int] = generated_sequence.reshape(a_ ,out_b // in_b ,*generated_sequence.shape[1:] ) elif self.framework == "tf": _UpperCAmelCase : Union[str, Any] = tf.reshape(a_ ,(in_b, out_b // in_b, *generated_sequence.shape[1:]) ) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def _snake_case ( self ,a_ ,a_=ReturnType.FULL_TEXT ,a_=True ) -> List[str]: _UpperCAmelCase : Optional[Any] = model_outputs["""generated_sequence"""][0] _UpperCAmelCase : Optional[int] = model_outputs["""input_ids"""] _UpperCAmelCase : List[str] = model_outputs["""prompt_text"""] _UpperCAmelCase : Optional[Any] = generated_sequence.numpy().tolist() _UpperCAmelCase : Dict = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: _UpperCAmelCase : str = {"""generated_token_ids""": sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text _UpperCAmelCase : Tuple = self.tokenizer.decode( a_ ,skip_special_tokens=a_ ,clean_up_tokenization_spaces=a_ ,) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: _UpperCAmelCase : Union[str, Any] = 0 else: _UpperCAmelCase : Tuple = len( self.tokenizer.decode( input_ids[0] ,skip_special_tokens=a_ ,clean_up_tokenization_spaces=a_ ,) ) if return_type == ReturnType.FULL_TEXT: _UpperCAmelCase : Any = prompt_text + text[prompt_length:] else: _UpperCAmelCase : Dict = text[prompt_length:] _UpperCAmelCase : Union[str, Any] = {"""generated_text""": all_text} records.append(a_ ) return records
215
1
"""simple docstring""" import pytest UpperCAmelCase = '''__dummy_dataset1__''' UpperCAmelCase = ''' import json import os import datasets REPO_URL = "https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/" URLS = {"train": REPO_URL + "wikiann-bn-train.jsonl", "validation": REPO_URL + "wikiann-bn-validation.jsonl"} class __DummyDataset1__(datasets.GeneratorBasedBuilder): def _info(self): features = datasets.Features( { "tokens": datasets.Sequence(datasets.Value("string")), "ner_tags": datasets.Sequence( datasets.features.ClassLabel( names=[ "O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC", ] ) ), "langs": datasets.Sequence(datasets.Value("string")), "spans": datasets.Sequence(datasets.Value("string")), } ) return datasets.DatasetInfo(features=features) def _split_generators(self, dl_manager): dl_path = dl_manager.download(URLS) return [ datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={"filepath": dl_path["train"]}), datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={"filepath": dl_path["validation"]}), ] def _generate_examples(self, filepath): with open(filepath, "r", encoding="utf-8") as f: for i, line in enumerate(f): yield i, json.loads(line) ''' @pytest.fixture def lowerCamelCase () -> Optional[Any]: return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def lowerCamelCase () -> List[str]: return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def lowerCamelCase (a_ :Optional[int] , a_ :List[Any] , a_ :Any) -> Optional[Any]: lowercase :Any = dataset_loading_script_name lowercase :Tuple = tmp_path / '''datasets''' / script_name script_dir.mkdir(parents=a_) lowercase :List[Any] = script_dir / F"""{script_name}.py""" with open(a_ , '''w''') as f: f.write(a_) return str(a_)
362
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCAmelCase = logging.get_logger(__name__) class __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ): __A : Dict = "maskformer-swin" __A : Union[str, Any] = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self : Dict , snake_case__ : Dict=2_2_4 , snake_case__ : Any=4 , snake_case__ : Dict=3 , snake_case__ : str=9_6 , snake_case__ : List[str]=[2, 2, 6, 2] , snake_case__ : Optional[int]=[3, 6, 1_2, 2_4] , snake_case__ : Optional[Any]=7 , snake_case__ : int=4.0 , snake_case__ : str=True , snake_case__ : Dict=0.0 , snake_case__ : List[Any]=0.0 , snake_case__ : Tuple=0.1 , snake_case__ : List[str]="gelu" , snake_case__ : Tuple=False , snake_case__ : int=0.02 , snake_case__ : Tuple=1e-5 , snake_case__ : Optional[int]=None , snake_case__ : Tuple=None , **snake_case__ : List[Any] , ): '''simple docstring''' super().__init__(**snake_case__ ) lowercase :Optional[int] = image_size lowercase :List[Any] = patch_size lowercase :Optional[Any] = num_channels lowercase :Union[str, Any] = embed_dim lowercase :Union[str, Any] = depths lowercase :List[Any] = len(snake_case__ ) lowercase :Optional[Any] = num_heads lowercase :Optional[Any] = window_size lowercase :Optional[int] = mlp_ratio lowercase :str = qkv_bias lowercase :int = hidden_dropout_prob lowercase :List[str] = attention_probs_dropout_prob lowercase :str = drop_path_rate lowercase :Optional[Any] = hidden_act lowercase :Tuple = use_absolute_embeddings lowercase :Union[str, Any] = layer_norm_eps lowercase :Any = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowercase :Optional[int] = int(embed_dim * 2 ** (len(snake_case__ ) - 1) ) lowercase :Optional[int] = ['''stem'''] + [f"""stage{idx}""" for idx in range(1 , len(snake_case__ ) + 1 )] lowercase , lowercase :List[str] = get_aligned_output_features_output_indices( out_features=snake_case__ , out_indices=snake_case__ , stage_names=self.stage_names )
172
0
"""simple docstring""" import os import random import sys from . import cryptomath_module as cryptomath from . import rabin_miller A_ : List[str] = 3 def A ( snake_case__ ): '''simple docstring''' print("""Generating primitive root of p""" ) while True: SCREAMING_SNAKE_CASE__ = random.randrange(3 , snake_case__ ) if pow(snake_case__ , 2 , snake_case__ ) == 1: continue if pow(snake_case__ , snake_case__ , snake_case__ ) == 1: continue return g def A ( snake_case__ ): '''simple docstring''' print("""Generating prime p...""" ) SCREAMING_SNAKE_CASE__ = rabin_miller.generate_large_prime(snake_case__ ) # select large prime number. SCREAMING_SNAKE_CASE__ = primitive_root(snake_case__ ) # one primitive root on modulo p. SCREAMING_SNAKE_CASE__ = random.randrange(3 , snake_case__ ) # private_key -> have to be greater than 2 for safety. SCREAMING_SNAKE_CASE__ = cryptomath.find_mod_inverse(pow(snake_case__ , snake_case__ , snake_case__ ) , snake_case__ ) SCREAMING_SNAKE_CASE__ = (key_size, e_a, e_a, p) SCREAMING_SNAKE_CASE__ = (key_size, d) return public_key, private_key def A ( snake_case__ , snake_case__ ): '''simple docstring''' if os.path.exists(f"""{name}_pubkey.txt""" ) or os.path.exists(f"""{name}_privkey.txt""" ): print("""\nWARNING:""" ) print( f"""\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n""" """Use a different name or delete these files and re-run this program.""" ) sys.exit() SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = generate_key(snake_case__ ) print(f"""\nWriting public key to file {name}_pubkey.txt...""" ) with open(f"""{name}_pubkey.txt""" , """w""" ) as fo: fo.write(f"""{public_key[0]},{public_key[1]},{public_key[2]},{public_key[3]}""" ) print(f"""Writing private key to file {name}_privkey.txt...""" ) with open(f"""{name}_privkey.txt""" , """w""" ) as fo: fo.write(f"""{private_key[0]},{private_key[1]}""" ) def A ( ): '''simple docstring''' print("""Making key files...""" ) make_key_files("""elgamal""" , 20_48 ) print("""Key files generation successful""" ) if __name__ == "__main__": main()
165
"""simple docstring""" import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class lowerCamelCase (unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE ( self : Dict ) -> Dict: SCREAMING_SNAKE_CASE__ = AutoImageProcessor.from_pretrained("""microsoft/dit-base-finetuned-rvlcdip""" ) SCREAMING_SNAKE_CASE__ = AutoModelForImageClassification.from_pretrained("""microsoft/dit-base-finetuned-rvlcdip""" ) model.to(__UpperCAmelCase ) from datasets import load_dataset SCREAMING_SNAKE_CASE__ = load_dataset("""nielsr/rvlcdip-demo""" ) SCREAMING_SNAKE_CASE__ = dataset["""train"""][0]["""image"""].convert("""RGB""" ) SCREAMING_SNAKE_CASE__ = image_processor(__UpperCAmelCase , return_tensors="""pt""" ).to(__UpperCAmelCase ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(**__UpperCAmelCase ) SCREAMING_SNAKE_CASE__ = outputs.logits SCREAMING_SNAKE_CASE__ = torch.Size((1, 1_6) ) self.assertEqual(logits.shape , __UpperCAmelCase ) SCREAMING_SNAKE_CASE__ = torch.tensor( [-0.4_158, -0.4_092, -0.4_347] , device=__UpperCAmelCase , dtype=torch.float , ) self.assertTrue(torch.allclose(logits[0, :3] , __UpperCAmelCase , atol=1e-4 ) )
165
1
from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup _UpperCamelCase = '''https://www.indeed.co.in/jobs?q=mobile+app+development&l=''' def UpperCamelCase_( snake_case__: str = "mumbai" ) -> Generator[tuple[str, str], None, None]: UpperCAmelCase__ = BeautifulSoup(requests.get(url + location ).content , 'html.parser' ) # This attribute finds out all the specifics listed in a job for job in soup.find_all('div' , attrs={'data-tn-component': 'organicJob'} ): UpperCAmelCase__ = job.find('a' , attrs={'data-tn-element': 'jobTitle'} ).text.strip() UpperCAmelCase__ = job.find('span' , {'class': 'company'} ).text.strip() yield job_title, company_name if __name__ == "__main__": for i, job in enumerate(fetch_jobs('''Bangalore'''), 1): print(F"""Job {i:>2} is {job[0]} at {job[1]}""")
335
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _UpperCamelCase = { '''configuration_squeezebert''': [ '''SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SqueezeBertConfig''', '''SqueezeBertOnnxConfig''', ], '''tokenization_squeezebert''': ['''SqueezeBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = ['''SqueezeBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ '''SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SqueezeBertForMaskedLM''', '''SqueezeBertForMultipleChoice''', '''SqueezeBertForQuestionAnswering''', '''SqueezeBertForSequenceClassification''', '''SqueezeBertForTokenClassification''', '''SqueezeBertModel''', '''SqueezeBertModule''', '''SqueezeBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_squeezebert import ( SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, SqueezeBertConfig, SqueezeBertOnnxConfig, ) from .tokenization_squeezebert import SqueezeBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_squeezebert_fast import SqueezeBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_squeezebert import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, SqueezeBertModule, SqueezeBertPreTrainedModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
335
1
from __future__ import annotations from math import pi def __lowercase ( a__ , a__ , a__ ) -> dict[str, float]: if (inductance, frequency, reactance).count(0 ) != 1: raise ValueError('One and only one argument must be 0' ) if inductance < 0: raise ValueError('Inductance cannot be negative' ) if frequency < 0: raise ValueError('Frequency cannot be negative' ) if reactance < 0: raise ValueError('Inductive reactance cannot be negative' ) if inductance == 0: return {"inductance": reactance / (2 * pi * frequency)} elif frequency == 0: return {"frequency": reactance / (2 * pi * inductance)} elif reactance == 0: return {"reactance": 2 * pi * frequency * inductance} else: raise ValueError('Exactly one argument must be 0' ) if __name__ == "__main__": import doctest doctest.testmod()
257
from __future__ import annotations import unittest from transformers import AutoTokenizer, MBartConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFMBartForConditionalGeneration, TFMBartModel @require_tf class UpperCAmelCase_ : '''simple docstring''' UpperCamelCase__ : int = MBartConfig UpperCamelCase__ : Optional[Any] = {} UpperCamelCase__ : Union[str, Any] = '''gelu''' def __init__( self , _A , _A=13 , _A=7 , _A=True , _A=False , _A=99 , _A=32 , _A=2 , _A=4 , _A=37 , _A=0.1 , _A=0.1 , _A=20 , _A=2 , _A=1 , _A=0 , ): '''simple docstring''' __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = seq_length __SCREAMING_SNAKE_CASE = is_training __SCREAMING_SNAKE_CASE = use_labels __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = max_position_embeddings __SCREAMING_SNAKE_CASE = eos_token_id __SCREAMING_SNAKE_CASE = pad_token_id __SCREAMING_SNAKE_CASE = bos_token_id def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __SCREAMING_SNAKE_CASE = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __SCREAMING_SNAKE_CASE = tf.concat([input_ids, eos_tensor] , axis=1 ) __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __SCREAMING_SNAKE_CASE = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) __SCREAMING_SNAKE_CASE = prepare_mbart_inputs_dict(_A , _A , _A ) return config, inputs_dict def _A ( self , _A , _A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = TFMBartModel(config=_A ).get_decoder() __SCREAMING_SNAKE_CASE = inputs_dict['input_ids'] __SCREAMING_SNAKE_CASE = input_ids[:1, :] __SCREAMING_SNAKE_CASE = inputs_dict['attention_mask'][:1, :] __SCREAMING_SNAKE_CASE = inputs_dict['head_mask'] __SCREAMING_SNAKE_CASE = 1 # first forward pass __SCREAMING_SNAKE_CASE = model(_A , attention_mask=_A , head_mask=_A , use_cache=_A ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = outputs.to_tuple() __SCREAMING_SNAKE_CASE = past_key_values[1] def __lowercase ( a__ , a__ , a__ , a__=None , a__=None , a__=None , a__=None , a__=None , ) -> str: if attention_mask is None: __SCREAMING_SNAKE_CASE = tf.cast(tf.math.not_equal(a__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: __SCREAMING_SNAKE_CASE = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: __SCREAMING_SNAKE_CASE = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: __SCREAMING_SNAKE_CASE = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: __SCREAMING_SNAKE_CASE = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class UpperCAmelCase_ ( UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase ): '''simple docstring''' UpperCamelCase__ : Any = (TFMBartForConditionalGeneration, TFMBartModel) if is_tf_available() else () UpperCamelCase__ : int = (TFMBartForConditionalGeneration,) if is_tf_available() else () UpperCamelCase__ : Optional[Any] = ( { '''conversational''': TFMBartForConditionalGeneration, '''feature-extraction''': TFMBartModel, '''summarization''': TFMBartForConditionalGeneration, '''text2text-generation''': TFMBartForConditionalGeneration, '''translation''': TFMBartForConditionalGeneration, } if is_tf_available() else {} ) UpperCamelCase__ : List[str] = True UpperCamelCase__ : Tuple = False UpperCamelCase__ : Union[str, Any] = False def _A ( self , _A , _A , _A , _A , _A ): '''simple docstring''' if pipeline_test_casse_name != "FeatureExtractionPipelineTests": # Exception encountered when calling layer '...' return True return False def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = TFMBartModelTester(self ) __SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=_A ) def _A ( self ): '''simple docstring''' self.config_tester.run_common_tests() def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_A ) @require_sentencepiece @require_tokenizers @require_tf class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' UpperCamelCase__ : Any = [ ''' UN Chief Says There Is No Military Solution in Syria''', ] UpperCamelCase__ : str = [ '''Şeful ONU declară că nu există o soluţie militară în Siria''', ] UpperCamelCase__ : List[str] = '''facebook/mbart-large-en-ro''' @cached_property def _A ( self ): '''simple docstring''' return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def _A ( self ): '''simple docstring''' __SCREAMING_SNAKE_CASE = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def _A ( self , **_A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.translate_src_text(**_A ) self.assertListEqual(self.expected_text , _A ) def _A ( self , **_A ): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.tokenizer(self.src_text , **_A , return_tensors='tf' ) __SCREAMING_SNAKE_CASE = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 ) __SCREAMING_SNAKE_CASE = self.tokenizer.batch_decode(_A , skip_special_tokens=_A ) return generated_words @slow def _A ( self ): '''simple docstring''' self._assert_generated_batch_equal_expected()
257
1
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase = None, ): lowercase :List[Any] = {} if train_file is not None: lowercase :Union[str, Any] = [train_file] if eval_file is not None: lowercase :str = [eval_file] if test_file is not None: lowercase :Optional[int] = [test_file] lowercase :Union[str, Any] = datasets.load_dataset("csv", data_files=lowerCamelCase ) lowercase :Optional[int] = list(ds[list(files.keys() )[0]].features.keys() ) lowercase :Union[str, Any] = features_name.pop(lowerCamelCase ) lowercase :int = list(set(ds[list(files.keys() )[0]][label_name] ) ) lowercase :Dict = {label: i for i, label in enumerate(lowerCamelCase )} lowercase :List[str] = tokenizer.model_input_names lowercase :int = {} if len(lowerCamelCase ) == 1: for k in files.keys(): lowercase :List[Any] = ds[k].map( lambda lowerCamelCase : tokenizer.batch_encode_plus( example[features_name[0]], truncation=lowerCamelCase, max_length=lowerCamelCase, padding="max_length" ), batched=lowerCamelCase, ) elif len(lowerCamelCase ) == 2: for k in files.keys(): lowercase :int = ds[k].map( lambda lowerCamelCase : tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]), truncation=lowerCamelCase, max_length=lowerCamelCase, padding="max_length", ), batched=lowerCamelCase, ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: lowercase :Dict = {k: v for k, v in ex.items() if k in input_names} lowercase :List[str] = labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: lowercase :Union[str, Any] = {k: v for k, v in ex.items() if k in input_names} lowercase :List[Any] = labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: lowercase :Optional[int] = {k: v for k, v in ex.items() if k in input_names} lowercase :int = labelaid[ex[label_name]] yield (d, label) lowercase :Optional[int] = ( tf.data.Dataset.from_generator( lowerCamelCase, ({k: tf.intaa for k in input_names}, tf.intaa), ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )), ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: lowercase :Optional[Any] = train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN] ) ) ) lowercase :str = ( tf.data.Dataset.from_generator( lowerCamelCase, ({k: tf.intaa for k in input_names}, tf.intaa), ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )), ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: lowercase :Optional[Any] = val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION] ) ) ) lowercase :List[str] = ( tf.data.Dataset.from_generator( lowerCamelCase, ({k: tf.intaa for k in input_names}, tf.intaa), ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )), ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: lowercase :List[str] = test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST] ) ) ) return train_ds, val_ds, test_ds, labelaid _UpperCAmelCase : Any = logging.getLogger(__name__) @dataclass class __lowerCAmelCase : _a = field(metadata={'''help''': '''Which column contains the label'''}) _a = field(default=lowerCAmelCase , metadata={'''help''': '''The path of the training file'''}) _a = field(default=lowerCAmelCase , metadata={'''help''': '''The path of the development file'''}) _a = field(default=lowerCAmelCase , metadata={'''help''': '''The path of the test file'''}) _a = field( default=128 , metadata={ '''help''': ( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) } , ) _a = field( default=lowerCAmelCase , metadata={'''help''': '''Overwrite the cached training and evaluation sets'''}) @dataclass class __lowerCAmelCase : _a = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''}) _a = field( default=lowerCAmelCase , metadata={'''help''': '''Pretrained config name or path if not the same as model_name'''}) _a = field( default=lowerCAmelCase , metadata={'''help''': '''Pretrained tokenizer name or path if not the same as model_name'''}) _a = field(default=lowerCAmelCase , metadata={'''help''': '''Set this flag to use fast tokenization.'''}) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. _a = field( default=lowerCAmelCase , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) def UpperCAmelCase__ ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowercase :Dict = HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments) ) lowercase , lowercase , lowercase :List[str] = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( F"Output directory ({training_args.output_dir}) already exists and is not empty. Use" " --overwrite_output_dir to overcome." ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s", datefmt="%m/%d/%Y %H:%M:%S", level=logging.INFO, ) logger.info( F"n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1 )}, " F"16-bits training: {training_args.fpaa}" ) logger.info(F"Training/evaluation parameters {training_args}" ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. lowercase :List[Any] = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path, cache_dir=model_args.cache_dir, ) lowercase , lowercase , lowercase , lowercase :int = get_tfds( train_file=data_args.train_file, eval_file=data_args.dev_file, test_file=data_args.test_file, tokenizer=lowerCamelCase, label_column_id=data_args.label_column_id, max_seq_length=data_args.max_seq_length, ) lowercase :Any = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path, num_labels=len(lowerCamelCase ), labelaid=lowerCamelCase, idalabel={id: label for label, id in labelaid.items()}, finetuning_task="text-classification", cache_dir=model_args.cache_dir, ) with training_args.strategy.scope(): lowercase :str = TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path, from_pt=bool(".bin" in model_args.model_name_or_path ), config=lowerCamelCase, cache_dir=model_args.cache_dir, ) def compute_metrics(lowerCamelCase ) -> Dict: lowercase :Optional[int] = np.argmax(p.predictions, axis=1 ) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer lowercase :Union[str, Any] = TFTrainer( model=lowerCamelCase, args=lowerCamelCase, train_dataset=lowerCamelCase, eval_dataset=lowerCamelCase, compute_metrics=lowerCamelCase, ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowercase :Dict = {} if training_args.do_eval: logger.info("*** Evaluate ***" ) lowercase :Optional[Any] = trainer.evaluate() lowercase :Tuple = os.path.join(training_args.output_dir, "eval_results.txt" ) with open(lowerCamelCase, "w" ) as writer: logger.info("***** Eval results *****" ) for key, value in result.items(): logger.info(F" {key} = {value}" ) writer.write(F"{key} = {value}\n" ) results.update(lowerCamelCase ) return results if __name__ == "__main__": main()
158
import pytest _UpperCAmelCase : List[Any] = "__dummy_dataset1__" _UpperCAmelCase : Union[str, Any] = "\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = \"https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/\"\nURLS = {\"train\": REPO_URL + \"wikiann-bn-train.jsonl\", \"validation\": REPO_URL + \"wikiann-bn-validation.jsonl\"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n \"tokens\": datasets.Sequence(datasets.Value(\"string\")),\n \"ner_tags\": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n \"O\",\n \"B-PER\",\n \"I-PER\",\n \"B-ORG\",\n \"I-ORG\",\n \"B-LOC\",\n \"I-LOC\",\n ]\n )\n ),\n \"langs\": datasets.Sequence(datasets.Value(\"string\")),\n \"spans\": datasets.Sequence(datasets.Value(\"string\")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={\"filepath\": dl_path[\"train\"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={\"filepath\": dl_path[\"validation\"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, \"r\", encoding=\"utf-8\") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n" @pytest.fixture def UpperCAmelCase__ ( ): return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def UpperCAmelCase__ ( ): return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): lowercase :Tuple = dataset_loading_script_name lowercase :Dict = tmp_path / "datasets" / script_name script_dir.mkdir(parents=lowerCamelCase ) lowercase :int = script_dir / F"{script_name}.py" with open(lowerCamelCase, "w" ) as f: f.write(lowerCamelCase ) return str(lowerCamelCase )
158
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { '''abeja/gpt-neox-japanese-2.7b''': '''https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/config.json''', } class __lowercase ( _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = '''gpt_neox_japanese''' def __init__( self , _UpperCAmelCase=32000 , _UpperCAmelCase=2560 , _UpperCAmelCase=32 , _UpperCAmelCase=32 , _UpperCAmelCase=4 , _UpperCAmelCase="gelu" , _UpperCAmelCase=1.0_0 , _UpperCAmelCase=10000 , _UpperCAmelCase=2048 , _UpperCAmelCase=0.0_2 , _UpperCAmelCase=1e-5 , _UpperCAmelCase=True , _UpperCAmelCase=31996 , _UpperCAmelCase=31999 , _UpperCAmelCase=0.1 , _UpperCAmelCase=0.0 , **_UpperCAmelCase , ): super().__init__(bos_token_id=_UpperCAmelCase , eos_token_id=_UpperCAmelCase , **_UpperCAmelCase ) __a : Optional[Any] = vocab_size __a : Dict = max_position_embeddings __a : List[Any] = hidden_size __a : Any = num_hidden_layers __a : List[Any] = num_attention_heads __a : int = intermediate_multiple_size __a : Optional[Any] = hidden_act __a : List[str] = rotary_pct __a : Any = rotary_emb_base __a : Dict = initializer_range __a : Union[str, Any] = layer_norm_eps __a : Dict = use_cache __a : Union[str, Any] = attention_dropout __a : List[str] = hidden_dropout
160
"""simple docstring""" from __future__ import annotations def __A ( a_ :str , a_ :str) -> bool: __a : Optional[Any] = get_failure_array(a_) # 2) Step through text searching for pattern __a , __a : Union[str, Any] = 0, 0 # index into text, pattern while i < len(a_): if pattern[j] == text[i]: if j == (len(a_) - 1): return True j += 1 # if this is a prefix in our pattern # just go back far enough to continue elif j > 0: __a : List[Any] = failure[j - 1] continue i += 1 return False def __A ( a_ :str) -> list[int]: __a : List[Any] = [0] __a : List[Any] = 0 __a : Any = 1 while j < len(a_): if pattern[i] == pattern[j]: i += 1 elif i > 0: __a : Any = failure[i - 1] continue j += 1 failure.append(a_) return failure if __name__ == "__main__": # Test 1) A = '''abc1abc12''' A = '''alskfjaldsabc1abc1abc12k23adsfabcabc''' A = '''alskfjaldsk23adsfabcabc''' assert kmp(pattern, texta) and not kmp(pattern, texta) # Test 2) A = '''ABABX''' A = '''ABABZABABYABABX''' assert kmp(pattern, text) # Test 3) A = '''AAAB''' A = '''ABAAAAAB''' assert kmp(pattern, text) # Test 4) A = '''abcdabcy''' A = '''abcxabcdabxabcdabcdabcy''' assert kmp(pattern, text) # Test 5) A = '''aabaabaaa''' assert get_failure_array(pattern) == [0, 1, 0, 1, 2, 3, 4, 5, 2]
160
1
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL import torch from transformers import CLIPImageProcessor, CLIPVisionModel from ...models import PriorTransformer from ...pipelines import DiffusionPipeline from ...schedulers import HeunDiscreteScheduler from ...utils import ( BaseOutput, is_accelerate_available, logging, randn_tensor, replace_example_docstring, ) from .renderer import ShapERenderer _lowerCamelCase : int = logging.get_logger(__name__) # pylint: disable=invalid-name _lowerCamelCase : Optional[int] = "\n Examples:\n ```py\n >>> from PIL import Image\n >>> import torch\n >>> from diffusers import DiffusionPipeline\n >>> from diffusers.utils import export_to_gif, load_image\n\n >>> device = torch.device(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n\n >>> repo = \"openai/shap-e-img2img\"\n >>> pipe = DiffusionPipeline.from_pretrained(repo, torch_dtype=torch.float16)\n >>> pipe = pipe.to(device)\n\n >>> guidance_scale = 3.0\n >>> image_url = \"https://hf.co/datasets/diffusers/docs-images/resolve/main/shap-e/corgi.png\"\n >>> image = load_image(image_url).convert(\"RGB\")\n\n >>> images = pipe(\n ... image,\n ... guidance_scale=guidance_scale,\n ... num_inference_steps=64,\n ... frame_size=256,\n ... ).images\n\n >>> gif_path = export_to_gif(images[0], \"corgi_3d.gif\")\n ```\n" @dataclass class __UpperCAmelCase ( lowerCamelCase__ ): UpperCamelCase = 4_2 class __UpperCAmelCase ( lowerCamelCase__ ): def __init__( self : Union[str, Any], __A : PriorTransformer, __A : CLIPVisionModel, __A : CLIPImageProcessor, __A : HeunDiscreteScheduler, __A : ShapERenderer, ): super().__init__() self.register_modules( prior=__A, image_encoder=__A, image_processor=__A, scheduler=__A, renderer=__A, ) def __magic_name__ ( self : Optional[int], __A : Optional[Any], __A : Any, __A : Union[str, Any], __A : List[Any], __A : Any, __A : Union[str, Any] ): if latents is None: UpperCAmelCase : Tuple = randn_tensor(__A, generator=__A, device=__A, dtype=__A ) else: if latents.shape != shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) UpperCAmelCase : List[str] = latents.to(__A ) UpperCAmelCase : Union[str, Any] = latents * scheduler.init_noise_sigma return latents def __magic_name__ ( self : Union[str, Any], __A : Optional[int]=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) UpperCAmelCase : int = torch.device(F'''cuda:{gpu_id}''' ) UpperCAmelCase : int = [self.image_encoder, self.prior] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(__A, __A ) @property def __magic_name__ ( self : Dict ): if self.device != torch.device('''meta''' ) or not hasattr(self.image_encoder, '''_hf_hook''' ): return self.device for module in self.image_encoder.modules(): if ( hasattr(__A, '''_hf_hook''' ) and hasattr(module._hf_hook, '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device def __magic_name__ ( self : Dict, __A : int, __A : List[str], __A : str, __A : List[Any], ): if isinstance(__A, __A ) and isinstance(image[0], torch.Tensor ): UpperCAmelCase : Dict = torch.cat(__A, axis=0 ) if image[0].ndim == 4 else torch.stack(__A, axis=0 ) if not isinstance(__A, torch.Tensor ): UpperCAmelCase : int = self.image_processor(__A, return_tensors='''pt''' ).pixel_values[0].unsqueeze(0 ) UpperCAmelCase : Dict = image.to(dtype=self.image_encoder.dtype, device=__A ) UpperCAmelCase : int = self.image_encoder(__A )['''last_hidden_state'''] UpperCAmelCase : Optional[Any] = image_embeds[:, 1:, :].contiguous() # batch_size, dim, 256 UpperCAmelCase : Union[str, Any] = image_embeds.repeat_interleave(__A, dim=0 ) if do_classifier_free_guidance: UpperCAmelCase : int = torch.zeros_like(__A ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes UpperCAmelCase : str = torch.cat([negative_image_embeds, image_embeds] ) return image_embeds @torch.no_grad() @replace_example_docstring(__A ) def __call__( self : List[Any], __A : Union[PIL.Image.Image, List[PIL.Image.Image]], __A : int = 1, __A : int = 2_5, __A : Optional[Union[torch.Generator, List[torch.Generator]]] = None, __A : Optional[torch.FloatTensor] = None, __A : float = 4.0, __A : int = 6_4, __A : Optional[str] = "pil", __A : bool = True, ): if isinstance(__A, PIL.Image.Image ): UpperCAmelCase : Tuple = 1 elif isinstance(__A, torch.Tensor ): UpperCAmelCase : Tuple = image.shape[0] elif isinstance(__A, __A ) and isinstance(image[0], (torch.Tensor, PIL.Image.Image) ): UpperCAmelCase : Dict = len(__A ) else: raise ValueError( F'''`image` has to be of type `PIL.Image.Image`, `torch.Tensor`, `List[PIL.Image.Image]` or `List[torch.Tensor]` but is {type(__A )}''' ) UpperCAmelCase : Tuple = self._execution_device UpperCAmelCase : int = batch_size * num_images_per_prompt UpperCAmelCase : str = guidance_scale > 1.0 UpperCAmelCase : List[Any] = self._encode_image(__A, __A, __A, __A ) # prior self.scheduler.set_timesteps(__A, device=__A ) UpperCAmelCase : Dict = self.scheduler.timesteps UpperCAmelCase : Union[str, Any] = self.prior.config.num_embeddings UpperCAmelCase : Optional[int] = self.prior.config.embedding_dim UpperCAmelCase : Any = self.prepare_latents( (batch_size, num_embeddings * embedding_dim), image_embeds.dtype, __A, __A, __A, self.scheduler, ) # YiYi notes: for testing only to match ldm, we can directly create a latents with desired shape: batch_size, num_embeddings, embedding_dim UpperCAmelCase : List[Any] = latents.reshape(latents.shape[0], __A, __A ) for i, t in enumerate(self.progress_bar(__A ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase : List[str] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase : Optional[Any] = self.scheduler.scale_model_input(__A, __A ) UpperCAmelCase : Any = self.prior( __A, timestep=__A, proj_embedding=__A, ).predicted_image_embedding # remove the variance UpperCAmelCase : Tuple = noise_pred.split( scaled_model_input.shape[2], dim=2 ) # batch_size, num_embeddings, embedding_dim if do_classifier_free_guidance is not None: UpperCAmelCase : List[str] = noise_pred.chunk(2 ) UpperCAmelCase : int = noise_pred_uncond + guidance_scale * (noise_pred - noise_pred_uncond) UpperCAmelCase : int = self.scheduler.step( __A, timestep=__A, sample=__A, ).prev_sample if output_type == "latent": return ShapEPipelineOutput(images=__A ) UpperCAmelCase : Union[str, Any] = [] for i, latent in enumerate(__A ): print() UpperCAmelCase : Any = self.renderer.decode( latent[None, :], __A, size=__A, ray_batch_size=4_0_9_6, n_coarse_samples=6_4, n_fine_samples=1_2_8, ) images.append(__A ) UpperCAmelCase : Optional[int] = torch.stack(__A ) if output_type not in ["np", "pil"]: raise ValueError(F'''Only the output types `pil` and `np` are supported not output_type={output_type}''' ) UpperCAmelCase : str = images.cpu().numpy() if output_type == "pil": UpperCAmelCase : str = [self.numpy_to_pil(__A ) for image in images] # Offload last model to CPU if hasattr(self, '''final_offload_hook''' ) and self.final_offload_hook is not None: self.final_offload_hook.offload() if not return_dict: return (images,) return ShapEPipelineOutput(images=__A )
357
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ChineseCLIPImageProcessor class __UpperCAmelCase ( unittest.TestCase ): def __init__( self : List[Any], __A : str, __A : List[str]=7, __A : List[str]=3, __A : Optional[int]=1_8, __A : List[Any]=3_0, __A : Tuple=4_0_0, __A : Tuple=True, __A : List[Any]=None, __A : str=True, __A : int=None, __A : Optional[Any]=True, __A : List[Any]=[0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3], __A : List[str]=[0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1], __A : Tuple=True, ): UpperCAmelCase : int = size if size is not None else {'''height''': 2_2_4, '''width''': 2_2_4} UpperCAmelCase : Optional[Any] = crop_size if crop_size is not None else {'''height''': 1_8, '''width''': 1_8} UpperCAmelCase : int = parent UpperCAmelCase : Union[str, Any] = batch_size UpperCAmelCase : List[Any] = num_channels UpperCAmelCase : Dict = image_size UpperCAmelCase : List[str] = min_resolution UpperCAmelCase : Optional[Any] = max_resolution UpperCAmelCase : Union[str, Any] = do_resize UpperCAmelCase : Dict = size UpperCAmelCase : Any = do_center_crop UpperCAmelCase : Union[str, Any] = crop_size UpperCAmelCase : List[str] = do_normalize UpperCAmelCase : Optional[Any] = image_mean UpperCAmelCase : Optional[Any] = image_std UpperCAmelCase : List[Any] = do_convert_rgb def __magic_name__ ( self : int ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_convert_rgb": self.do_convert_rgb, } def __magic_name__ ( self : Optional[Any], __A : Any=False, __A : str=False, __A : List[Any]=False ): assert not (numpify and torchify), "You cannot specify both numpy and PyTorch tensors at the same time" if equal_resolution: UpperCAmelCase : Dict = [] for i in range(self.batch_size ): image_inputs.append( np.random.randint( 2_5_5, size=(self.num_channels, self.max_resolution, self.max_resolution), dtype=np.uinta ) ) else: UpperCAmelCase : Tuple = [] for i in range(self.batch_size ): UpperCAmelCase , UpperCAmelCase : Tuple = np.random.choice(np.arange(self.min_resolution, self.max_resolution ), 2 ) image_inputs.append(np.random.randint(2_5_5, size=(self.num_channels, width, height), dtype=np.uinta ) ) if not numpify and not torchify: # PIL expects the channel dimension as last dimension UpperCAmelCase : Optional[int] = [Image.fromarray(np.moveaxis(__A, 0, -1 ) ) for x in image_inputs] if torchify: UpperCAmelCase : str = [torch.from_numpy(__A ) for x in image_inputs] return image_inputs @require_torch @require_vision class __UpperCAmelCase ( lowerCamelCase__ , unittest.TestCase ): UpperCamelCase = ChineseCLIPImageProcessor if is_vision_available() else None def __magic_name__ ( self : Dict ): UpperCAmelCase : Optional[Any] = ChineseCLIPImageProcessingTester(self, do_center_crop=__A ) @property def __magic_name__ ( self : Tuple ): return self.image_processor_tester.prepare_image_processor_dict() def __magic_name__ ( self : Any ): UpperCAmelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__A, '''do_resize''' ) ) self.assertTrue(hasattr(__A, '''size''' ) ) self.assertTrue(hasattr(__A, '''do_center_crop''' ) ) self.assertTrue(hasattr(__A, '''center_crop''' ) ) self.assertTrue(hasattr(__A, '''do_normalize''' ) ) self.assertTrue(hasattr(__A, '''image_mean''' ) ) self.assertTrue(hasattr(__A, '''image_std''' ) ) self.assertTrue(hasattr(__A, '''do_convert_rgb''' ) ) def __magic_name__ ( self : Optional[Any] ): UpperCAmelCase : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size, {'''height''': 2_2_4, '''width''': 2_2_4} ) self.assertEqual(image_processor.crop_size, {'''height''': 1_8, '''width''': 1_8} ) UpperCAmelCase : int = self.image_processing_class.from_dict(self.image_processor_dict, size=4_2, crop_size=8_4 ) self.assertEqual(image_processor.size, {'''shortest_edge''': 4_2} ) self.assertEqual(image_processor.crop_size, {'''height''': 8_4, '''width''': 8_4} ) def __magic_name__ ( self : Union[str, Any] ): pass def __magic_name__ ( self : Tuple ): # Initialize image_processing UpperCAmelCase : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase : Optional[int] = self.image_processor_tester.prepare_inputs(equal_resolution=__A ) for image in image_inputs: self.assertIsInstance(__A, Image.Image ) # Test not batched input UpperCAmelCase : str = image_processing(image_inputs[0], return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), ) # Test batched UpperCAmelCase : Any = image_processing(__A, return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), ) def __magic_name__ ( self : Optional[Any] ): # Initialize image_processing UpperCAmelCase : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase : List[Any] = self.image_processor_tester.prepare_inputs(equal_resolution=__A, numpify=__A ) for image in image_inputs: self.assertIsInstance(__A, np.ndarray ) # Test not batched input UpperCAmelCase : List[str] = image_processing(image_inputs[0], return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), ) # Test batched UpperCAmelCase : Dict = image_processing(__A, return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), ) def __magic_name__ ( self : Any ): # Initialize image_processing UpperCAmelCase : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase : Optional[Any] = self.image_processor_tester.prepare_inputs(equal_resolution=__A, torchify=__A ) for image in image_inputs: self.assertIsInstance(__A, torch.Tensor ) # Test not batched input UpperCAmelCase : str = image_processing(image_inputs[0], return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), ) # Test batched UpperCAmelCase : List[str] = image_processing(__A, return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), ) @require_torch @require_vision class __UpperCAmelCase ( lowerCamelCase__ , unittest.TestCase ): UpperCamelCase = ChineseCLIPImageProcessor if is_vision_available() else None def __magic_name__ ( self : Optional[Any] ): UpperCAmelCase : str = ChineseCLIPImageProcessingTester(self, num_channels=4, do_center_crop=__A ) UpperCAmelCase : Dict = 3 @property def __magic_name__ ( self : List[Any] ): return self.image_processor_tester.prepare_image_processor_dict() def __magic_name__ ( self : Dict ): UpperCAmelCase : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__A, '''do_resize''' ) ) self.assertTrue(hasattr(__A, '''size''' ) ) self.assertTrue(hasattr(__A, '''do_center_crop''' ) ) self.assertTrue(hasattr(__A, '''center_crop''' ) ) self.assertTrue(hasattr(__A, '''do_normalize''' ) ) self.assertTrue(hasattr(__A, '''image_mean''' ) ) self.assertTrue(hasattr(__A, '''image_std''' ) ) self.assertTrue(hasattr(__A, '''do_convert_rgb''' ) ) def __magic_name__ ( self : List[Any] ): pass def __magic_name__ ( self : Tuple ): # Initialize image_processing UpperCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase : Dict = self.image_processor_tester.prepare_inputs(equal_resolution=__A ) for image in image_inputs: self.assertIsInstance(__A, Image.Image ) # Test not batched input UpperCAmelCase : Tuple = image_processing(image_inputs[0], return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( 1, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), ) # Test batched UpperCAmelCase : Tuple = image_processing(__A, return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape, ( self.image_processor_tester.batch_size, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ), )
99
0
from ..utils import DummyObject, requires_backends class __UpperCAmelCase (metaclass=lowerCamelCase_ ): __snake_case : int = ['''flax''', '''transformers'''] def __init__( self: Tuple , *UpperCAmelCase_: Optional[Any] , **UpperCAmelCase_: Any ): '''simple docstring''' requires_backends(self , ["""flax""", """transformers"""] ) @classmethod def UpperCamelCase ( cls: Union[str, Any] , *UpperCAmelCase_: Tuple , **UpperCAmelCase_: Any ): '''simple docstring''' requires_backends(cls , ["""flax""", """transformers"""] ) @classmethod def UpperCamelCase ( cls: str , *UpperCAmelCase_: Dict , **UpperCAmelCase_: int ): '''simple docstring''' requires_backends(cls , ["""flax""", """transformers"""] ) class __UpperCAmelCase (metaclass=lowerCamelCase_ ): __snake_case : Any = ['''flax''', '''transformers'''] def __init__( self: Optional[int] , *UpperCAmelCase_: List[Any] , **UpperCAmelCase_: int ): '''simple docstring''' requires_backends(self , ["""flax""", """transformers"""] ) @classmethod def UpperCamelCase ( cls: Dict , *UpperCAmelCase_: Union[str, Any] , **UpperCAmelCase_: Tuple ): '''simple docstring''' requires_backends(cls , ["""flax""", """transformers"""] ) @classmethod def UpperCamelCase ( cls: Union[str, Any] , *UpperCAmelCase_: int , **UpperCAmelCase_: str ): '''simple docstring''' requires_backends(cls , ["""flax""", """transformers"""] ) class __UpperCAmelCase (metaclass=lowerCamelCase_ ): __snake_case : Union[str, Any] = ['''flax''', '''transformers'''] def __init__( self: Optional[int] , *UpperCAmelCase_: Any , **UpperCAmelCase_: Dict ): '''simple docstring''' requires_backends(self , ["""flax""", """transformers"""] ) @classmethod def UpperCamelCase ( cls: int , *UpperCAmelCase_: Union[str, Any] , **UpperCAmelCase_: Dict ): '''simple docstring''' requires_backends(cls , ["""flax""", """transformers"""] ) @classmethod def UpperCamelCase ( cls: List[Any] , *UpperCAmelCase_: List[str] , **UpperCAmelCase_: int ): '''simple docstring''' requires_backends(cls , ["""flax""", """transformers"""] ) class __UpperCAmelCase (metaclass=lowerCamelCase_ ): __snake_case : str = ['''flax''', '''transformers'''] def __init__( self: Optional[Any] , *UpperCAmelCase_: Optional[int] , **UpperCAmelCase_: Dict ): '''simple docstring''' requires_backends(self , ["""flax""", """transformers"""] ) @classmethod def UpperCamelCase ( cls: str , *UpperCAmelCase_: Optional[Any] , **UpperCAmelCase_: Optional[Any] ): '''simple docstring''' requires_backends(cls , ["""flax""", """transformers"""] ) @classmethod def UpperCamelCase ( cls: Dict , *UpperCAmelCase_: List[str] , **UpperCAmelCase_: List[str] ): '''simple docstring''' requires_backends(cls , ["""flax""", """transformers"""] )
306
import os import random import sys from . import cryptomath_module as cryptoMath # noqa: N812 from . import rabin_miller as rabinMiller # noqa: N812 def __lowercase ( ): print('Making key files...' ) make_key_files('rsa' , 1_0_2_4 ) print('Key files generation successful.' ) def __lowercase ( __lowerCAmelCase : int ): print('Generating prime p...' ) a__ = rabinMiller.generate_large_prime(__lowerCAmelCase ) print('Generating prime q...' ) a__ = rabinMiller.generate_large_prime(__lowerCAmelCase ) a__ = p * q print('Generating e that is relatively prime to (p - 1) * (q - 1)...' ) while True: a__ = random.randrange(2 ** (key_size - 1) , 2 ** (key_size) ) if cryptoMath.gcd(__lowerCAmelCase , (p - 1) * (q - 1) ) == 1: break print('Calculating d that is mod inverse of e...' ) a__ = cryptoMath.find_mod_inverse(__lowerCAmelCase , (p - 1) * (q - 1) ) a__ = (n, e) a__ = (n, d) return (public_key, private_key) def __lowercase ( __lowerCAmelCase : str , __lowerCAmelCase : int ): if os.path.exists(F'{name}_pubkey.txt' ) or os.path.exists(F'{name}_privkey.txt' ): print('\nWARNING:' ) print( F'"{name}_pubkey.txt" or "{name}_privkey.txt" already exists. \n' 'Use a different name or delete these files and re-run this program.' ) sys.exit() a__ , a__ = generate_key(__lowerCAmelCase ) print(F'\nWriting public key to file {name}_pubkey.txt...' ) with open(F'{name}_pubkey.txt' , 'w' ) as out_file: out_file.write(F'{key_size},{public_key[0]},{public_key[1]}' ) print(F'Writing private key to file {name}_privkey.txt...' ) with open(F'{name}_privkey.txt' , 'w' ) as out_file: out_file.write(F'{key_size},{private_key[0]},{private_key[1]}' ) if __name__ == "__main__": main()
240
0
'''simple docstring''' from typing import List import jiwer import jiwer.transforms as tr from packaging import version import datasets from datasets.config import PY_VERSION if PY_VERSION < version.parse("""3.8"""): import importlib_metadata else: import importlib.metadata as importlib_metadata snake_case__ = """""" if version.parse(importlib_metadata.version("""jiwer""")) < version.parse("""2.3.0"""): class UpperCamelCase_ (tr.AbstractTransform ): """simple docstring""" def __init__( self : str , _lowerCamelCase : str = " " ): """simple docstring""" A_ : Optional[Any] = sentence_delimiter def _a ( self : List[Any] , _lowerCamelCase : str ): """simple docstring""" return list(_lowerCamelCase ) def _a ( self : List[str] , _lowerCamelCase : List[str] ): """simple docstring""" A_ : Tuple = [] for sent_idx, sentence in enumerate(_lowerCamelCase ): chars.extend(self.process_string(_lowerCamelCase ) ) if self.sentence_delimiter is not None and self.sentence_delimiter != "" and sent_idx < len(_lowerCamelCase ) - 1: chars.append(self.sentence_delimiter ) return chars snake_case__ = tr.Compose( [tr.RemoveMultipleSpaces(), tr.Strip(), SentencesToListOfCharacters(SENTENCE_DELIMITER)] ) else: snake_case__ = tr.Compose( [ tr.RemoveMultipleSpaces(), tr.Strip(), tr.ReduceToSingleSentence(SENTENCE_DELIMITER), tr.ReduceToListOfListOfChars(), ] ) snake_case__ = """\\n@inproceedings{inproceedings,\n author = {Morris, Andrew and Maier, Viktoria and Green, Phil},\n year = {2004},\n month = {01},\n pages = {},\n title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.}\n}\n""" snake_case__ = """\\nCharacter error rate (CER) is a common metric of the performance of an automatic speech recognition system.\n\nCER is similar to Word Error Rate (WER), but operates on character instead of word. Please refer to docs of WER for further information.\n\nCharacter error rate can be computed as:\n\nCER = (S + D + I) / N = (S + D + I) / (S + D + C)\n\nwhere\n\nS is the number of substitutions,\nD is the number of deletions,\nI is the number of insertions,\nC is the number of correct characters,\nN is the number of characters in the reference (N=S+D+C).\n\nCER\'s output is not always a number between 0 and 1, in particular when there is a high number of insertions. This value is often associated to the percentage of characters that were incorrectly predicted. The lower the value, the better the\nperformance of the ASR system with a CER of 0 being a perfect score.\n""" snake_case__ = """\nComputes CER score of transcribed segments against references.\nArgs:\n references: list of references for each speech input.\n predictions: list of transcribtions to score.\n concatenate_texts: Whether or not to concatenate sentences before evaluation, set to True for more accurate result.\nReturns:\n (float): the character error rate\n\nExamples:\n\n >>> predictions = [\"this is the prediction\", \"there is an other sample\"]\n >>> references = [\"this is the reference\", \"there is another one\"]\n >>> cer = datasets.load_metric(\"cer\")\n >>> cer_score = cer.compute(predictions=predictions, references=references)\n >>> print(cer_score)\n 0.34146341463414637\n""" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class UpperCamelCase_ (datasets.Metric ): """simple docstring""" def _a ( self : Union[str, Any] ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/jitsi/jiwer/'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/Word_error_rate''', '''https://sites.google.com/site/textdigitisation/qualitymeasures/computingerrorrates''', ] , ) def _a ( self : Union[str, Any] , _lowerCamelCase : Tuple , _lowerCamelCase : Tuple , _lowerCamelCase : List[str]=False ): """simple docstring""" if concatenate_texts: return jiwer.compute_measures( _lowerCamelCase , _lowerCamelCase , truth_transform=_lowerCamelCase , hypothesis_transform=_lowerCamelCase , )["wer"] A_ : Dict = 0 A_ : Union[str, Any] = 0 for prediction, reference in zip(_lowerCamelCase , _lowerCamelCase ): A_ : str = jiwer.compute_measures( _lowerCamelCase , _lowerCamelCase , truth_transform=_lowerCamelCase , hypothesis_transform=_lowerCamelCase , ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
355
'''simple docstring''' def snake_case__ ( lowerCamelCase__ : list[int] , lowerCamelCase__ : list[int] , lowerCamelCase__ : int ) -> bool: return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(lowerCamelCase__ ) ) def snake_case__ ( lowerCamelCase__ : list[list[int]] , lowerCamelCase__ : int , lowerCamelCase__ : list[int] , lowerCamelCase__ : int ) -> bool: # Base Case if index == len(lowerCamelCase__ ): return True # Recursive Step for i in range(lowerCamelCase__ ): if valid_coloring(graph[index] , lowerCamelCase__ , lowerCamelCase__ ): # Color current vertex A_ : int = i # Validate coloring if util_color(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , index + 1 ): return True # Backtrack A_ : str = -1 return False def snake_case__ ( lowerCamelCase__ : list[list[int]] , lowerCamelCase__ : int ) -> list[int]: A_ : List[str] = [-1] * len(lowerCamelCase__ ) if util_color(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , 0 ): return colored_vertices return []
4
0
import os import sys a__ = os.path.join(os.path.dirname(__file__), """src""") sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) a__ = [ """torch""", """numpy""", """tokenizers""", """filelock""", """requests""", """tqdm""", """regex""", """sentencepiece""", """sacremoses""", """importlib_metadata""", """huggingface_hub""", ] @add_start_docstrings(AutoConfig.__doc__ ) def lowercase ( *SCREAMING_SNAKE_CASE__ : Any , **SCREAMING_SNAKE_CASE__ : int ) -> int: return AutoConfig.from_pretrained(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) @add_start_docstrings(AutoTokenizer.__doc__ ) def lowercase ( *SCREAMING_SNAKE_CASE__ : List[Any] , **SCREAMING_SNAKE_CASE__ : Any ) -> Optional[Any]: return AutoTokenizer.from_pretrained(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) @add_start_docstrings(AutoModel.__doc__ ) def lowercase ( *SCREAMING_SNAKE_CASE__ : Any , **SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> Optional[int]: return AutoModel.from_pretrained(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def lowercase ( *SCREAMING_SNAKE_CASE__ : int , **SCREAMING_SNAKE_CASE__ : Any ) -> Union[str, Any]: return AutoModelForCausalLM.from_pretrained(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def lowercase ( *SCREAMING_SNAKE_CASE__ : Union[str, Any] , **SCREAMING_SNAKE_CASE__ : str ) -> Optional[Any]: return AutoModelForMaskedLM.from_pretrained(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def lowercase ( *SCREAMING_SNAKE_CASE__ : Dict , **SCREAMING_SNAKE_CASE__ : List[str] ) -> str: return AutoModelForSequenceClassification.from_pretrained(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def lowercase ( *SCREAMING_SNAKE_CASE__ : int , **SCREAMING_SNAKE_CASE__ : Any ) -> Tuple: return AutoModelForQuestionAnswering.from_pretrained(*SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ )
317
def lowercase ( SCREAMING_SNAKE_CASE__ : str ) -> list: _snake_case : Optional[Any] = [0] * len(SCREAMING_SNAKE_CASE__ ) for i in range(1 , len(SCREAMING_SNAKE_CASE__ ) ): # use last results for better performance - dynamic programming _snake_case : Optional[Any] = prefix_result[i - 1] while j > 0 and input_string[i] != input_string[j]: _snake_case : List[Any] = prefix_result[j - 1] if input_string[i] == input_string[j]: j += 1 _snake_case : Optional[int] = j return prefix_result def lowercase ( SCREAMING_SNAKE_CASE__ : str ) -> int: return max(prefix_function(SCREAMING_SNAKE_CASE__ ) ) if __name__ == "__main__": import doctest doctest.testmod()
317
1
from __future__ import annotations from math import gcd def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 2 , SCREAMING_SNAKE_CASE_ = 1 , SCREAMING_SNAKE_CASE_ = 3 , ) -> int | None: # A value less than 2 can cause an infinite loop in the algorithm. if num < 2: raise ValueError('The input value cannot be less than 2' ) # Because of the relationship between ``f(f(x))`` and ``f(x)``, this # algorithm struggles to find factors that are divisible by two. # As a workaround, we specifically check for two and even inputs. # See: https://math.stackexchange.com/a/2856214/165820 if num > 2 and num % 2 == 0: return 2 # Pollard's Rho algorithm requires a function that returns pseudorandom # values between 0 <= X < ``num``. It doesn't need to be random in the # sense that the output value is cryptographically secure or difficult # to calculate, it only needs to be random in the sense that all output # values should be equally likely to appear. # For this reason, Pollard suggested using ``f(x) = (x**2 - 1) % num`` # However, the success of Pollard's algorithm isn't guaranteed and is # determined in part by the initial seed and the chosen random function. # To make retries easier, we will instead use ``f(x) = (x**2 + C) % num`` # where ``C`` is a value that we can modify between each attempt. def rand_fn(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: return (pow(SCREAMING_SNAKE_CASE_ , 2 ) + step) % modulus for _ in range(SCREAMING_SNAKE_CASE_ ): # These track the position within the cycle detection logic. lowerCAmelCase__ : int = seed lowerCAmelCase__ : Dict = seed while True: # At each iteration, the tortoise moves one step and the hare moves two. lowerCAmelCase__ : Dict = rand_fn(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Any = rand_fn(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Union[str, Any] = rand_fn(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # At some point both the tortoise and the hare will enter a cycle whose # length ``p`` is a divisor of ``num``. Once in that cycle, at some point # the tortoise and hare will end up on the same value modulo ``p``. # We can detect when this happens because the position difference between # the tortoise and the hare will share a common divisor with ``num``. lowerCAmelCase__ : List[str] = gcd(hare - tortoise , SCREAMING_SNAKE_CASE_ ) if divisor == 1: # No common divisor yet, just keep searching. continue else: # We found a common divisor! if divisor == num: # Unfortunately, the divisor is ``num`` itself and is useless. break else: # The divisor is a nontrivial factor of ``num``! return divisor # If we made it here, then this attempt failed. # We need to pick a new starting seed for the tortoise and hare # in addition to a new step value for the random function. # To keep this example implementation deterministic, the # new values will be generated based on currently available # values instead of using something like ``random.randint``. # We can use the hare's position as the new seed. # This is actually what Richard Brent's the "optimized" variant does. lowerCAmelCase__ : List[Any] = hare # The new step value for the random function can just be incremented. # At first the results will be similar to what the old function would # have produced, but the value will quickly diverge after a bit. step += 1 # We haven't found a divisor within the requested number of attempts. # We were unlucky or ``num`` itself is actually prime. return None if __name__ == "__main__": import argparse lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument( """num""", type=int, help="""The value to find a divisor of""", ) parser.add_argument( """--attempts""", type=int, default=3, help="""The number of attempts before giving up""", ) lowerCamelCase__ = parser.parse_args() lowerCamelCase__ = pollard_rho(args.num, attempts=args.attempts) if divisor is None: print(F"""{args.num} is probably prime""") else: lowerCamelCase__ = args.num // divisor print(F"""{args.num} = {divisor} * {quotient}""")
307
from __future__ import annotations def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> list[list[int]]: lowerCAmelCase__ : list[list[int]] = [] create_all_state(1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , [] , SCREAMING_SNAKE_CASE_ ) return result def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , ) -> None: if level == 0: total_list.append(current_list[:] ) return for i in range(SCREAMING_SNAKE_CASE_ , total_number - level + 2 ): current_list.append(SCREAMING_SNAKE_CASE_ ) create_all_state(i + 1 , SCREAMING_SNAKE_CASE_ , level - 1 , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) current_list.pop() def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> None: for i in total_list: print(*SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": lowerCamelCase__ = 4 lowerCamelCase__ = 2 lowerCamelCase__ = generate_all_combinations(n, k) print_all_state(total_list)
307
1
'''simple docstring''' from __future__ import annotations class a : def __init__( self , __magic_name__ ) -> Optional[int]: _a = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(__magic_name__ ) != 0: _a = len(rows[0] ) if cols == 0: raise error for row in rows: if len(__magic_name__ ) != cols: raise error for value in row: if not isinstance(__magic_name__ , (int, float) ): raise error _a = rows else: _a = [] def __UpperCAmelCase ( self ) -> list[list[int]]: return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def __UpperCAmelCase ( self ) -> int: return len(self.rows ) @property def __UpperCAmelCase ( self ) -> int: return len(self.rows[0] ) @property def __UpperCAmelCase ( self ) -> tuple[int, int]: return (self.num_rows, self.num_columns) @property def __UpperCAmelCase ( self ) -> bool: return self.order[0] == self.order[1] def __UpperCAmelCase ( self ) -> Matrix: _a = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(__magic_name__ ) def __UpperCAmelCase ( self ) -> int: if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def __UpperCAmelCase ( self ) -> bool: return bool(self.determinant() ) def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ ) -> int: _a = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(__magic_name__ ).determinant() def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ ) -> int: if (row + column) % 2 == 0: return self.get_minor(__magic_name__ , __magic_name__ ) return -1 * self.get_minor(__magic_name__ , __magic_name__ ) def __UpperCAmelCase ( self ) -> Matrix: return Matrix( [ [self.get_minor(__magic_name__ , __magic_name__ ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def __UpperCAmelCase ( self ) -> Matrix: return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def __UpperCAmelCase ( self ) -> Matrix: _a = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(__magic_name__ ) def __UpperCAmelCase ( self ) -> Matrix: _a = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self ) -> str: return str(self.rows ) def __str__( self ) -> str: if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(__magic_name__ ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ = None ) -> None: _a = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(__magic_name__ , __magic_name__ ): raise type_error for value in row: if not isinstance(__magic_name__ , (int, float) ): raise type_error if len(__magic_name__ ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(__magic_name__ ) else: _a = self.rows[0:position] + [row] + self.rows[position:] def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ = None ) -> None: _a = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(__magic_name__ , __magic_name__ ): raise type_error for value in column: if not isinstance(__magic_name__ , (int, float) ): raise type_error if len(__magic_name__ ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: _a = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: _a = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self , __magic_name__ ) -> bool: if not isinstance(__magic_name__ , __magic_name__ ): return NotImplemented return self.rows == other.rows def __ne__( self , __magic_name__ ) -> bool: return not self == other def __neg__( self ) -> Matrix: return self * -1 def __add__( self , __magic_name__ ) -> Matrix: if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self , __magic_name__ ) -> Matrix: if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self , __magic_name__ ) -> Matrix: if isinstance(__magic_name__ , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(__magic_name__ , __magic_name__ ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(__magic_name__ , __magic_name__ ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self , __magic_name__ ) -> Matrix: if not isinstance(__magic_name__ , __magic_name__ ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) _a = self for _ in range(other - 1 ): result *= self return result @classmethod def __UpperCAmelCase ( cls , __magic_name__ , __magic_name__ ) -> int: return sum(row[i] * column[i] for i in range(len(__magic_name__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
168
'''simple docstring''' import math class a : def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ ) -> int: _a = 0.0 _a = 0.0 for i in range(len(__magic_name__ ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def __UpperCAmelCase ( self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> list[list[int | float]]: for i in range(len(__magic_name__ ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def _A () -> None: '''simple docstring''' _a = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) _a = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training _a = SelfOrganizingMap() _a = 3 _a = 0.5 for _ in range(lowerCAmelCase__ ): for j in range(len(lowerCAmelCase__ ) ): # training sample _a = training_samples[j] # Compute the winning vector _a = self_organizing_map.get_winner(lowerCAmelCase__ , lowerCAmelCase__ ) # Update the winning vector _a = self_organizing_map.update(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # classify test sample _a = [0, 0, 0, 1] _a = self_organizing_map.get_winner(lowerCAmelCase__ , lowerCAmelCase__ ) # results print(f'Clusters that the test sample belongs to : {winner}' ) print(f'Weights that have been trained : {weights}' ) # running the main() function if __name__ == "__main__": main()
168
1
from __future__ import annotations from collections.abc import Iterator class __magic_name__ : def __init__( self , __snake_case ) -> None: '''simple docstring''' __a =value __a =None __a =None class __magic_name__ : def __init__( self , __snake_case ) -> None: '''simple docstring''' __a =tree def __magic_name__ ( self , __snake_case ) -> int: '''simple docstring''' if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self ) -> Iterator[int]: '''simple docstring''' yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
308
import json import os import unittest from transformers.models.blenderbot_small.tokenization_blenderbot_small import ( VOCAB_FILES_NAMES, BlenderbotSmallTokenizer, ) from ...test_tokenization_common import TokenizerTesterMixin class __magic_name__ ( lowerCAmelCase_ , unittest.TestCase ): SCREAMING_SNAKE_CASE = BlenderbotSmallTokenizer SCREAMING_SNAKE_CASE = False def __magic_name__ ( self ) -> Union[str, Any]: '''simple docstring''' super().setUp() __a =['__start__', 'adapt', 'act', 'ap@@', 'te', '__end__', '__unk__'] __a =dict(zip(__snake_case , range(len(__snake_case ) ) ) ) __a =['#version: 0.2', 'a p', 't e</w>', 'ap t</w>', 'a d', 'ad apt</w>', 'a c', 'ac t</w>', ''] __a ={'unk_token': '__unk__', 'bos_token': '__start__', 'eos_token': '__end__'} __a =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) __a =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(__snake_case ) + '\n' ) with open(self.merges_file , 'w' , encoding='utf-8' ) as fp: fp.write('\n'.join(__snake_case ) ) def __magic_name__ ( self , **__snake_case ) -> Any: '''simple docstring''' kwargs.update(self.special_tokens_map ) return BlenderbotSmallTokenizer.from_pretrained(self.tmpdirname , **__snake_case ) def __magic_name__ ( self , __snake_case ) -> List[Any]: '''simple docstring''' __a ='adapt act apte' __a ='adapt act apte' return input_text, output_text def __magic_name__ ( self ) -> str: '''simple docstring''' __a =BlenderbotSmallTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) __a ='adapt act apte' __a =['adapt', 'act', 'ap@@', 'te'] __a =tokenizer.tokenize(__snake_case ) self.assertListEqual(__snake_case , __snake_case ) __a =[tokenizer.bos_token] + tokens + [tokenizer.eos_token] __a =[0, 1, 2, 3, 4, 5] self.assertListEqual(tokenizer.convert_tokens_to_ids(__snake_case ) , __snake_case ) def __magic_name__ ( self ) -> str: '''simple docstring''' __a =BlenderbotSmallTokenizer.from_pretrained('facebook/blenderbot-90M' ) assert tok('sam' ).input_ids == [1384] __a ='I am a small frog.' __a =tok([src_text] , padding=__snake_case , truncation=__snake_case )['input_ids'] __a =tok.batch_decode(__snake_case , skip_special_tokens=__snake_case , clean_up_tokenization_spaces=__snake_case )[0] assert src_text != decoded # I wish it did! assert decoded == "i am a small frog ." def __magic_name__ ( self ) -> str: '''simple docstring''' __a =BlenderbotSmallTokenizer.from_pretrained('facebook/blenderbot-90M' ) __a ='I am a small frog .' __a ='.' __a =tok(__snake_case )['input_ids'] __a =tok(__snake_case )['input_ids'] assert encoded[-1] == encoded_dot[0]
308
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer UpperCAmelCase__ : Tuple = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast UpperCAmelCase__ : Any = TaTokenizerFast UpperCAmelCase__ : Dict = {"""configuration_mt5""": ["""MT5Config""", """MT5OnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ : Optional[Any] = [ """MT5EncoderModel""", """MT5ForConditionalGeneration""", """MT5ForQuestionAnswering""", """MT5Model""", """MT5PreTrainedModel""", """MT5Stack""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ : Optional[int] = ["""TFMT5EncoderModel""", """TFMT5ForConditionalGeneration""", """TFMT5Model"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase__ : str = ["""FlaxMT5EncoderModel""", """FlaxMT5ForConditionalGeneration""", """FlaxMT5Model"""] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys UpperCAmelCase__ : Optional[int] = _LazyModule( __name__, globals()["""__file__"""], _import_structure, extra_objects={"""MT5Tokenizer""": MTaTokenizer, """MT5TokenizerFast""": MTaTokenizerFast}, module_spec=__spec__, )
245
from __future__ import annotations import os import tempfile import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import is_tensorflow_text_available, is_tf_available from transformers.testing_utils import require_tensorflow_text, require_tf, slow from ..test_modeling_tf_common import floats_tensor from .test_framework_agnostic import GenerationIntegrationTestsMixin if is_tf_available(): import tensorflow as tf from transformers import ( AutoTokenizer, TFAutoModelForCausalLM, TFAutoModelForSeqaSeqLM, TFAutoModelForSpeechSeqaSeq, TFAutoModelForVisionaSeq, TFBartForConditionalGeneration, TFLogitsProcessorList, TFMinLengthLogitsProcessor, tf_top_k_top_p_filtering, ) if is_tensorflow_text_available(): import tensorflow_text as text @require_tf class a__ ( unittest.TestCase ): """simple docstring""" def _lowercase ( self : Union[str, Any] ) ->List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = tf.convert_to_tensor( [ [ 8.2_22_09_91, # 3rd highest value; idx. 0 -0.5_62_00_44, 5.23_22_97_52, 4.0_38_63_93, -6.8_79_83_78, -0.54_78_58_02, -3.2_01_21_53, 2.92_77_71_76, 1.88_17_19_53, 7.35_34_12_76, # 5th highest value; idx. 9 8.43_20_78_33, # 2nd highest value; idx. 10 -9.85_71_18_36, -5.96_20_92_36, -1.13_03_91_61, -7.1_11_52_94, -0.8_36_96_33, -5.3_18_64_08, 7.06_42_74_07, 0.81_36_93_44, -0.82_02_38_17, -5.9_17_97_96, 0.58_81_34_43, -6.99_77_84_38, 4.71_55_11_89, -0.18_77_16_37, 7.44_02_07_59, # 4th highest value; idx. 25 9.38_45_09_87, # 1st highest value; idx. 26 2.12_66_29_41, -9.32_56_20_38, 2.35_65_25_22, ], # cummulative prob of 5 highest values <= 0.6 [ 0.58_42_55_18, 4.53_13_92_38, -5.57_51_04_64, -6.28_03_06_99, -7.19_52_95_03, -4.02_12_25_51, 1.39_33_70_37, -6.06_70_70_57, 1.59_48_05_17, -9.64_31_19, 0.03_90_77_99, 0.67_23_17_62, -8.88_20_67_26, 6.27_11_59_22, # 4th highest value; idx. 13 2.28_52_07_23, 4.82_76_75_06, 4.30_42_13_68, 8.8_27_53_13, # 2nd highest value; idx. 17 5.44_02_99_58, # 5th highest value; idx. 18 -4.4_73_57_94, 7.38_57_95_36, # 3rd highest value; idx. 20 -2.91_05_16_63, 2.61_94_60_77, -2.5_67_47_62, -9.48_95_93_02, -4.02_92_26_45, -1.35_41_69_18, 9.67_70_23_23, # 1st highest value; idx. 27 -5.89_47_85_53, 1.85_37_04_67, ], # cummulative prob of 5 highest values <= 0.6 ] , dtype=tf.floataa , ) SCREAMING_SNAKE_CASE : List[Any] = tf.convert_to_tensor( [[0, 0], [0, 9], [0, 1_0], [0, 2_5], [0, 2_6], [1, 1_3], [1, 1_7], [1, 1_8], [1, 2_0], [1, 2_7]] , dtype=tf.intaa , ) # expected non filtered idx as noted above SCREAMING_SNAKE_CASE : Optional[int] = tf.convert_to_tensor( [8.22_20_99, 7.3_53_41_26, 8.43_20_78, 7.4_40_20_75, 9.3_84_51, 6.27_11_59, 8.82_75_31, 5.4_40_29_95, 7.3_85_79_56, 9.67_70_23] , dtype=tf.floataa , ) # expected non filtered values as noted above SCREAMING_SNAKE_CASE : Dict = tf_top_k_top_p_filtering(UpperCAmelCase__ , top_k=1_0 , top_p=0.6 , min_tokens_to_keep=4 ) SCREAMING_SNAKE_CASE : str = output[output != -float("""inf""" )] SCREAMING_SNAKE_CASE : Optional[int] = tf.cast( tf.where(tf.not_equal(UpperCAmelCase__ , tf.constant(-float("""inf""" ) , dtype=tf.floataa ) ) ) , dtype=tf.intaa , ) tf.debugging.assert_near(UpperCAmelCase__ , UpperCAmelCase__ , rtol=1e-12 ) tf.debugging.assert_equal(UpperCAmelCase__ , UpperCAmelCase__ ) @require_tf class a__ ( unittest.TestCase , UpperCAmelCase ): """simple docstring""" if is_tf_available(): UpperCAmelCase__ : Optional[Any] ={ """AutoModelForCausalLM""": TFAutoModelForCausalLM, """AutoModelForSpeechSeq2Seq""": TFAutoModelForSpeechSeqaSeq, """AutoModelForSeq2SeqLM""": TFAutoModelForSeqaSeqLM, """AutoModelForVision2Seq""": TFAutoModelForVisionaSeq, """LogitsProcessorList""": TFLogitsProcessorList, """MinLengthLogitsProcessor""": TFMinLengthLogitsProcessor, """create_tensor_fn""": tf.convert_to_tensor, """floats_tensor""": floats_tensor, """return_tensors""": """tf""", } @slow def _lowercase ( self : int ) ->List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = TFAutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) SCREAMING_SNAKE_CASE : str = 2 SCREAMING_SNAKE_CASE : Tuple = 2 class a__ ( tf.Module ): """simple docstring""" def __init__( self : Tuple , UpperCAmelCase__ : Optional[int] ) ->str: """simple docstring""" super(UpperCAmelCase__ , self ).__init__() SCREAMING_SNAKE_CASE : Optional[int] = model @tf.function( input_signature=( tf.TensorSpec((None, input_length) , tf.intaa , name="""input_ids""" ), tf.TensorSpec((None, input_length) , tf.intaa , name="""attention_mask""" ), ) , jit_compile=UpperCAmelCase__ , ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : List[str] ) ->List[str]: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = self.model.generate( input_ids=UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , max_new_tokens=UpperCAmelCase__ , return_dict_in_generate=UpperCAmelCase__ , ) return {"sequences": outputs["sequences"]} SCREAMING_SNAKE_CASE : Any = [[2, 0], [1_0_2, 1_0_3]] SCREAMING_SNAKE_CASE : Tuple = [[1, 0], [1, 1]] SCREAMING_SNAKE_CASE : Dict = DummyModel(model=UpperCAmelCase__ ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(UpperCAmelCase__ , UpperCAmelCase__ , signatures={"""serving_default""": dummy_model.serving} ) SCREAMING_SNAKE_CASE : Optional[int] = tf.saved_model.load(UpperCAmelCase__ ).signatures["""serving_default"""] for batch_size in range(1 , len(UpperCAmelCase__ ) + 1 ): SCREAMING_SNAKE_CASE : int = { """input_ids""": tf.constant(dummy_input_ids[:batch_size] ), """attention_mask""": tf.constant(dummy_attention_masks[:batch_size] ), } SCREAMING_SNAKE_CASE : Tuple = serving_func(**UpperCAmelCase__ )["""sequences"""] SCREAMING_SNAKE_CASE : List[str] = test_model.generate(**UpperCAmelCase__ , max_new_tokens=UpperCAmelCase__ ) tf.debugging.assert_equal(UpperCAmelCase__ , UpperCAmelCase__ ) @slow def _lowercase ( self : Dict ) ->int: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = TFAutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) SCREAMING_SNAKE_CASE : Any = 1 SCREAMING_SNAKE_CASE : int = 2 class a__ ( tf.Module ): """simple docstring""" def __init__( self : List[Any] , UpperCAmelCase__ : Optional[Any] ) ->Optional[int]: """simple docstring""" super(UpperCAmelCase__ , self ).__init__() SCREAMING_SNAKE_CASE : List[str] = model @tf.function( input_signature=( tf.TensorSpec((batch_size, None) , tf.intaa , name="""input_ids""" ), tf.TensorSpec((batch_size, None) , tf.intaa , name="""attention_mask""" ), ) , jit_compile=UpperCAmelCase__ , ) def _lowercase ( self : List[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : str ) ->Any: """simple docstring""" SCREAMING_SNAKE_CASE : int = self.model.generate( input_ids=UpperCAmelCase__ , attention_mask=UpperCAmelCase__ , max_new_tokens=UpperCAmelCase__ , return_dict_in_generate=UpperCAmelCase__ , ) return {"sequences": outputs["sequences"]} SCREAMING_SNAKE_CASE : List[Any] = [[2], [1_0_2, 1_0_3]] SCREAMING_SNAKE_CASE : List[Any] = [[1], [1, 1]] SCREAMING_SNAKE_CASE : Union[str, Any] = DummyModel(model=UpperCAmelCase__ ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(UpperCAmelCase__ , UpperCAmelCase__ , signatures={"""serving_default""": dummy_model.serving} ) SCREAMING_SNAKE_CASE : int = tf.saved_model.load(UpperCAmelCase__ ).signatures["""serving_default"""] for input_row in range(len(UpperCAmelCase__ ) ): SCREAMING_SNAKE_CASE : str = { """input_ids""": tf.constant([dummy_input_ids[input_row]] ), """attention_mask""": tf.constant([dummy_attention_masks[input_row]] ), } SCREAMING_SNAKE_CASE : List[str] = serving_func(**UpperCAmelCase__ )["""sequences"""] SCREAMING_SNAKE_CASE : List[Any] = test_model.generate(**UpperCAmelCase__ , max_new_tokens=UpperCAmelCase__ ) tf.debugging.assert_equal(UpperCAmelCase__ , UpperCAmelCase__ ) @slow @require_tensorflow_text def _lowercase ( self : Optional[Any] ) ->Tuple: """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: # file needed to load the TF tokenizer hf_hub_download(repo_id="""google/flan-t5-small""" , filename="""spiece.model""" , local_dir=UpperCAmelCase__ ) class a__ ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self : Optional[Any] ) ->List[str]: """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE : Any = text.SentencepieceTokenizer( model=tf.io.gfile.GFile(os.path.join(UpperCAmelCase__ , """spiece.model""" ) , """rb""" ).read() ) SCREAMING_SNAKE_CASE : Dict = TFAutoModelForSeqaSeqLM.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) def _lowercase ( self : int , UpperCAmelCase__ : Any , *UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : str ) ->int: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = self.tokenizer.tokenize(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : str = text.pad_model_inputs( UpperCAmelCase__ , max_seq_length=6_4 , pad_value=self.model.config.pad_token_id ) SCREAMING_SNAKE_CASE : Union[str, Any] = self.model.generate(input_ids=UpperCAmelCase__ , attention_mask=UpperCAmelCase__ ) return self.tokenizer.detokenize(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : str = CompleteSentenceTransformer() SCREAMING_SNAKE_CASE : Tuple = tf.keras.layers.Input(shape=(1,) , dtype=tf.string , name="""inputs""" ) SCREAMING_SNAKE_CASE : str = complete_model(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : Optional[int] = tf.keras.Model(UpperCAmelCase__ , UpperCAmelCase__ ) keras_model.save(UpperCAmelCase__ ) def _lowercase ( self : Optional[Any] ) ->List[str]: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = { """do_sample""": True, """num_beams""": 1, """top_p""": 0.7, """top_k""": 1_0, """temperature""": 0.7, } SCREAMING_SNAKE_CASE : Tuple = 1_4 SCREAMING_SNAKE_CASE : List[Any] = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) SCREAMING_SNAKE_CASE : List[Any] = """Hello, my dog is cute and""" SCREAMING_SNAKE_CASE : Tuple = tokenizer(UpperCAmelCase__ , return_tensors="""tf""" ) SCREAMING_SNAKE_CASE : Optional[int] = TFAutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-random-gpt2""" ) SCREAMING_SNAKE_CASE : Dict = 6_3_8 # forces the generation to happen on CPU, to avoid GPU-related quirks with tf.device(""":/CPU:0""" ): tf.random.set_seed(0 ) SCREAMING_SNAKE_CASE : int = model.generate(**UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , **UpperCAmelCase__ ) self.assertTrue(expectation == len(generated_tokens[0] ) ) SCREAMING_SNAKE_CASE : Dict = [6_3_8, 1_9_8] with tf.device(""":/CPU:0""" ): tf.random.set_seed(0 ) SCREAMING_SNAKE_CASE : Dict = model.generate(**UpperCAmelCase__ , eos_token_id=UpperCAmelCase__ , **UpperCAmelCase__ ) self.assertTrue(expectation == len(generated_tokens[0] ) ) def _lowercase ( self : str ) ->List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-bart""" ) SCREAMING_SNAKE_CASE : List[Any] = """Hugging Face is a technology company based in New York and Paris.""" SCREAMING_SNAKE_CASE : Optional[int] = bart_tokenizer(UpperCAmelCase__ , return_tensors="""tf""" ).input_ids SCREAMING_SNAKE_CASE : int = TFBartForConditionalGeneration.from_pretrained("""hf-internal-testing/tiny-random-bart""" ) SCREAMING_SNAKE_CASE : Optional[int] = bart_model.generate(UpperCAmelCase__ ).numpy() class a__ ( UpperCAmelCase ): """simple docstring""" def _lowercase ( self : Any , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Dict=None , **UpperCAmelCase__ : Dict ) ->List[str]: """simple docstring""" return super().call(UpperCAmelCase__ , **UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : Optional[int] = FakeBart.from_pretrained("""hf-internal-testing/tiny-random-bart""" ) SCREAMING_SNAKE_CASE : Optional[int] = bart_model.generate(UpperCAmelCase__ , foo="""bar""" ).numpy() self.assertTrue(np.array_equal(UpperCAmelCase__ , UpperCAmelCase__ ) ) class a__ ( bart_model.model.encoder.__class__ ): """simple docstring""" def _lowercase ( self : List[Any] , UpperCAmelCase__ : List[Any] , **UpperCAmelCase__ : Optional[int] ) ->Union[str, Any]: """simple docstring""" return super().call(UpperCAmelCase__ , **UpperCAmelCase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = FakeEncoder(bart_model.config , bart_model.model.shared ) SCREAMING_SNAKE_CASE : Tuple = fake_encoder # Normal generation still works (the output will be different because the encoder weights are different) SCREAMING_SNAKE_CASE : Tuple = bart_model.generate(UpperCAmelCase__ ).numpy() with self.assertRaises(UpperCAmelCase__ ): # FakeEncoder.call() accepts **kwargs -> no filtering -> value error due to unexpected input "foo" bart_model.generate(UpperCAmelCase__ , foo="""bar""" )
245
1
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() snake_case = logging.get_logger(__name__) def lowerCamelCase__ ( lowercase , lowercase=False ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''blocks.{i}.norm1.weight''', F'''deit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''blocks.{i}.norm1.bias''', F'''deit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append((F'''blocks.{i}.attn.proj.weight''', F'''deit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.attn.proj.bias''', F'''deit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''blocks.{i}.norm2.weight''', F'''deit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''blocks.{i}.norm2.bias''', F'''deit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.weight''', F'''deit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc1.bias''', F'''deit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.weight''', F'''deit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''blocks.{i}.mlp.fc2.bias''', F'''deit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "deit.embeddings.cls_token"), ("dist_token", "deit.embeddings.distillation_token"), ("patch_embed.proj.weight", "deit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "deit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "deit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" SCREAMING_SNAKE_CASE : Dict = [(pair[0], pair[1][4:]) if pair[1].startswith("deit" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("norm.weight", "deit.layernorm.weight"), ("norm.bias", "deit.layernorm.bias"), ("head.weight", "cls_classifier.weight"), ("head.bias", "cls_classifier.bias"), ("head_dist.weight", "distillation_classifier.weight"), ("head_dist.bias", "distillation_classifier.bias"), ] ) return rename_keys def lowerCamelCase__ ( lowercase , lowercase , lowercase=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: SCREAMING_SNAKE_CASE : Dict = "" else: SCREAMING_SNAKE_CASE : Tuple = "deit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) SCREAMING_SNAKE_CASE : int = state_dict.pop(F'''blocks.{i}.attn.qkv.weight''' ) SCREAMING_SNAKE_CASE : Union[str, Any] = state_dict.pop(F'''blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE : List[str] = in_proj_weight[ : config.hidden_size, : ] SCREAMING_SNAKE_CASE : Dict = in_proj_bias[: config.hidden_size] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] SCREAMING_SNAKE_CASE : Tuple = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] SCREAMING_SNAKE_CASE : Dict = in_proj_weight[ -config.hidden_size :, : ] SCREAMING_SNAKE_CASE : Optional[Any] = in_proj_bias[-config.hidden_size :] def lowerCamelCase__ ( lowercase , lowercase , lowercase ): """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = dct.pop(lowercase ) SCREAMING_SNAKE_CASE : List[Any] = val def lowerCamelCase__ ( ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = "http://images.cocodataset.org/val2017/000000039769.jpg" SCREAMING_SNAKE_CASE : Tuple = Image.open(requests.get(lowercase , stream=lowercase ).raw ) return im @torch.no_grad() def lowerCamelCase__ ( lowercase , lowercase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = DeiTConfig() # all deit models have fine-tuned heads SCREAMING_SNAKE_CASE : List[str] = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size SCREAMING_SNAKE_CASE : Tuple = 1000 SCREAMING_SNAKE_CASE : Dict = "huggingface/label-files" SCREAMING_SNAKE_CASE : Tuple = "imagenet-1k-id2label.json" SCREAMING_SNAKE_CASE : Union[str, Any] = json.load(open(hf_hub_download(lowercase , lowercase , repo_type="dataset" ) , "r" ) ) SCREAMING_SNAKE_CASE : Union[str, Any] = {int(lowercase ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE : Optional[Any] = idalabel SCREAMING_SNAKE_CASE : Dict = {v: k for k, v in idalabel.items()} SCREAMING_SNAKE_CASE : Union[str, Any] = int(deit_name[-6:-4] ) SCREAMING_SNAKE_CASE : int = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("tiny" ): SCREAMING_SNAKE_CASE : List[Any] = 192 SCREAMING_SNAKE_CASE : Union[str, Any] = 768 SCREAMING_SNAKE_CASE : Optional[int] = 12 SCREAMING_SNAKE_CASE : Tuple = 3 elif deit_name[9:].startswith("small" ): SCREAMING_SNAKE_CASE : str = 384 SCREAMING_SNAKE_CASE : Optional[Any] = 1536 SCREAMING_SNAKE_CASE : List[str] = 12 SCREAMING_SNAKE_CASE : int = 6 if deit_name[9:].startswith("base" ): pass elif deit_name[4:].startswith("large" ): SCREAMING_SNAKE_CASE : int = 1024 SCREAMING_SNAKE_CASE : Any = 4096 SCREAMING_SNAKE_CASE : int = 24 SCREAMING_SNAKE_CASE : Union[str, Any] = 16 # load original model from timm SCREAMING_SNAKE_CASE : Tuple = timm.create_model(lowercase , pretrained=lowercase ) timm_model.eval() # load state_dict of original model, remove and rename some keys SCREAMING_SNAKE_CASE : int = timm_model.state_dict() SCREAMING_SNAKE_CASE : Union[str, Any] = create_rename_keys(lowercase , lowercase ) for src, dest in rename_keys: rename_key(lowercase , lowercase , lowercase ) read_in_q_k_v(lowercase , lowercase , lowercase ) # load HuggingFace model SCREAMING_SNAKE_CASE : Union[str, Any] = DeiTForImageClassificationWithTeacher(lowercase ).eval() model.load_state_dict(lowercase ) # Check outputs on an image, prepared by DeiTImageProcessor SCREAMING_SNAKE_CASE : Union[str, Any] = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 SCREAMING_SNAKE_CASE : Dict = DeiTImageProcessor(size=lowercase , crop_size=config.image_size ) SCREAMING_SNAKE_CASE : Optional[int] = image_processor(images=prepare_img() , return_tensors="pt" ) SCREAMING_SNAKE_CASE : int = encoding["pixel_values"] SCREAMING_SNAKE_CASE : Optional[Any] = model(lowercase ) SCREAMING_SNAKE_CASE : Dict = timm_model(lowercase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(lowercase , outputs.logits , atol=1E-3 ) Path(lowercase ).mkdir(exist_ok=lowercase ) print(F'''Saving model {deit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(lowercase ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(lowercase ) if __name__ == "__main__": snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( """--deit_name""", default="""vit_deit_base_distilled_patch16_224""", type=str, help="""Name of the DeiT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) snake_case = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
363
import argparse import os import torch from transformers.utils import WEIGHTS_NAME snake_case = ["""small""", """medium""", """large"""] snake_case = """lm_head.decoder.weight""" snake_case = """lm_head.weight""" def lowerCamelCase__ ( lowercase , lowercase ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = torch.load(lowercase ) SCREAMING_SNAKE_CASE : Any = d.pop(lowercase ) os.makedirs(lowercase , exist_ok=lowercase ) torch.save(lowercase , os.path.join(lowercase , lowercase ) ) if __name__ == "__main__": snake_case = argparse.ArgumentParser() parser.add_argument("""--dialogpt_path""", default=""".""", type=str) snake_case = parser.parse_args() for MODEL in DIALOGPT_MODELS: snake_case = os.path.join(args.dialogpt_path, F"""{MODEL}_ft.pkl""") snake_case = F"""./DialoGPT-{MODEL}""" convert_dialogpt_checkpoint( checkpoint_path, pytorch_dump_folder_path, )
319
0
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCAmelCase__ : List[Any] = logging.get_logger(__name__) UpperCAmelCase__ : List[Any] = {'vocab_file': 'sentencepiece.bpe.model'} UpperCAmelCase__ : List[Any] = { 'vocab_file': { 'camembert-base': 'https://huggingface.co/camembert-base/resolve/main/sentencepiece.bpe.model', } } UpperCAmelCase__ : Optional[int] = { 'camembert-base': 5_1_2, } UpperCAmelCase__ : List[Any] = '▁' class lowerCAmelCase_ (a__ ): """simple docstring""" __UpperCamelCase : Optional[int] = VOCAB_FILES_NAMES __UpperCamelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase : Optional[Any] = ['''input_ids''', '''attention_mask'''] def __init__(self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__="<s>" , SCREAMING_SNAKE_CASE__="</s>" , SCREAMING_SNAKE_CASE__="</s>" , SCREAMING_SNAKE_CASE__="<s>" , SCREAMING_SNAKE_CASE__="<unk>" , SCREAMING_SNAKE_CASE__="<pad>" , SCREAMING_SNAKE_CASE__="<mask>" , SCREAMING_SNAKE_CASE__=["<s>NOTUSED", "</s>NOTUSED"] , SCREAMING_SNAKE_CASE__ = None , **SCREAMING_SNAKE_CASE__ , ) -> None: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = AddedToken(SCREAMING_SNAKE_CASE__ , lstrip=SCREAMING_SNAKE_CASE__ , rstrip=SCREAMING_SNAKE_CASE__ ) if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else mask_token SCREAMING_SNAKE_CASE__ : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=SCREAMING_SNAKE_CASE__ , eos_token=SCREAMING_SNAKE_CASE__ , unk_token=SCREAMING_SNAKE_CASE__ , sep_token=SCREAMING_SNAKE_CASE__ , cls_token=SCREAMING_SNAKE_CASE__ , pad_token=SCREAMING_SNAKE_CASE__ , mask_token=SCREAMING_SNAKE_CASE__ , additional_special_tokens=SCREAMING_SNAKE_CASE__ , sp_model_kwargs=self.sp_model_kwargs , **SCREAMING_SNAKE_CASE__ , ) SCREAMING_SNAKE_CASE__ : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(SCREAMING_SNAKE_CASE__ ) ) SCREAMING_SNAKE_CASE__ : Dict = vocab_file # HACK: These tokens were added by fairseq but don't seem to be actually used when duplicated in the actual # sentencepiece vocabulary (this is the case for <s> and </s> SCREAMING_SNAKE_CASE__ : Optional[int] = {"""<s>NOTUSED""": 0, """<pad>""": 1, """</s>NOTUSED""": 2, """<unk>""": 3} SCREAMING_SNAKE_CASE__ : Any = len(self.fairseq_tokens_to_ids ) SCREAMING_SNAKE_CASE__ : Any = len(self.sp_model ) + len(self.fairseq_tokens_to_ids ) SCREAMING_SNAKE_CASE__ : Any = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None ) -> List[int]: """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] SCREAMING_SNAKE_CASE__ : Optional[int] = [self.cls_token_id] SCREAMING_SNAKE_CASE__ : int = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=SCREAMING_SNAKE_CASE__ , token_ids_a=SCREAMING_SNAKE_CASE__ , already_has_special_tokens=SCREAMING_SNAKE_CASE__ ) if token_ids_a is None: return [1] + ([0] * len(SCREAMING_SNAKE_CASE__ )) + [1] return [1] + ([0] * len(SCREAMING_SNAKE_CASE__ )) + [1, 1] + ([0] * len(SCREAMING_SNAKE_CASE__ )) + [1] def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None ) -> List[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = [self.sep_token_id] SCREAMING_SNAKE_CASE__ : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __magic_name__ (self ) -> Dict: """simple docstring""" return len(self.fairseq_tokens_to_ids ) + len(self.sp_model ) def __magic_name__ (self ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[int] = {self.convert_ids_to_tokens(SCREAMING_SNAKE_CASE__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __magic_name__ (self , SCREAMING_SNAKE_CASE__ ) -> List[str]: """simple docstring""" return self.sp_model.encode(SCREAMING_SNAKE_CASE__ , out_type=SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ ) -> int: """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] elif self.sp_model.PieceToId(SCREAMING_SNAKE_CASE__ ) == 0: # Convert sentence piece unk token to fairseq unk token index return self.unk_token_id return self.fairseq_offset + self.sp_model.PieceToId(SCREAMING_SNAKE_CASE__ ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ ) -> Any: """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = [] SCREAMING_SNAKE_CASE__ : Tuple = """""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(SCREAMING_SNAKE_CASE__ ) + token SCREAMING_SNAKE_CASE__ : Dict = True SCREAMING_SNAKE_CASE__ : List[str] = [] else: current_sub_tokens.append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Dict = False out_string += self.sp_model.decode(SCREAMING_SNAKE_CASE__ ) return out_string.strip() def __getstate__(self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = self.__dict__.copy() SCREAMING_SNAKE_CASE__ : Any = None return state def __setstate__(self , SCREAMING_SNAKE_CASE__ ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): SCREAMING_SNAKE_CASE__ : Union[str, Any] = {} SCREAMING_SNAKE_CASE__ : Union[str, Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(SCREAMING_SNAKE_CASE__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return SCREAMING_SNAKE_CASE__ : int = os.path.join( SCREAMING_SNAKE_CASE__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(SCREAMING_SNAKE_CASE__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , SCREAMING_SNAKE_CASE__ ) elif not os.path.isfile(self.vocab_file ): with open(SCREAMING_SNAKE_CASE__ , """wb""" ) as fi: SCREAMING_SNAKE_CASE__ : Optional[int] = self.sp_model.serialized_model_proto() fi.write(SCREAMING_SNAKE_CASE__ ) return (out_vocab_file,)
25
'''simple docstring''' def UpperCAmelCase_ ( __lowerCamelCase : int ,__lowerCamelCase : int ): return int((input_a, input_a).count(0 ) == 0 ) def UpperCAmelCase_ ( ): assert and_gate(0 ,0 ) == 0 assert and_gate(0 ,1 ) == 0 assert and_gate(1 ,0 ) == 0 assert and_gate(1 ,1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
223
0
import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def UpperCAmelCase_ ( _A=None , _A=None ): '''simple docstring''' return field(default_factory=lambda: default , metadata=_A ) @dataclass class UpperCAmelCase__ : """simple docstring""" a = field( metadata={"help": "The csv file to plot."} , ) a = field( default=A__ , metadata={"help": "Whether to plot along batch size or sequence length. Defaults to sequence length."} , ) a = field( default=A__ , metadata={"help": "Whether the csv file has time results or memory results. Defaults to memory results."} , ) a = field( default=A__ , metadata={"help": "Disable logarithmic scale when plotting"} , ) a = field( default=A__ , metadata={ "help": "Whether the csv file has training results or inference results. Defaults to inference results." } , ) a = field( default=A__ , metadata={"help": "Filename under which the plot will be saved. If unused no plot is saved."} , ) a = list_field( default=A__ , metadata={"help": "List of model names that are used instead of the ones in the csv file."} ) def UpperCAmelCase_ ( _A ): '''simple docstring''' try: int(_A ) return True except ValueError: return False def UpperCAmelCase_ ( _A ): '''simple docstring''' try: float(_A ) return True except ValueError: return False class UpperCAmelCase__ : """simple docstring""" def __init__( self : Tuple , __lowerCamelCase : str ) -> str: SCREAMING_SNAKE_CASE__ = args SCREAMING_SNAKE_CASE__ = defaultdict(lambda: {"bsz": [], "seq_len": [], "result": {}} ) with open(self.args.csv_file , newline='''''' ) as csv_file: SCREAMING_SNAKE_CASE__ = csv.DictReader(__lowerCamelCase ) for row in reader: SCREAMING_SNAKE_CASE__ = row['''model'''] self.result_dict[model_name]["bsz"].append(int(row['''batch_size'''] ) ) self.result_dict[model_name]["seq_len"].append(int(row['''sequence_length'''] ) ) if can_convert_to_int(row['''result'''] ): # value is not None SCREAMING_SNAKE_CASE__ = int(row['''result'''] ) elif can_convert_to_float(row['''result'''] ): # value is not None SCREAMING_SNAKE_CASE__ = float(row['''result'''] ) def lowercase_ ( self : Union[str, Any] ) -> Dict: SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = plt.subplots() SCREAMING_SNAKE_CASE__ = '''Time usage''' if self.args.is_time else '''Memory usage''' SCREAMING_SNAKE_CASE__ = title_str + ''' for training''' if self.args.is_train else title_str + ''' for inference''' if not self.args.no_log_scale: # set logarithm scales ax.set_xscale('''log''' ) ax.set_yscale('''log''' ) for axis in [ax.xaxis, ax.yaxis]: axis.set_major_formatter(ScalarFormatter() ) for model_name_idx, model_name in enumerate(self.result_dict.keys() ): SCREAMING_SNAKE_CASE__ = sorted(set(self.result_dict[model_name]['''bsz'''] ) ) SCREAMING_SNAKE_CASE__ = sorted(set(self.result_dict[model_name]['''seq_len'''] ) ) SCREAMING_SNAKE_CASE__ = self.result_dict[model_name]['''result'''] ((SCREAMING_SNAKE_CASE__),(SCREAMING_SNAKE_CASE__)) = ( (batch_sizes, sequence_lengths) if self.args.plot_along_batch else (sequence_lengths, batch_sizes) ) SCREAMING_SNAKE_CASE__ = ( model_name if self.args.short_model_names is None else self.args.short_model_names[model_name_idx] ) for inner_loop_value in inner_loop_array: if self.args.plot_along_batch: SCREAMING_SNAKE_CASE__ = np.asarray( [results[(x, inner_loop_value)] for x in x_axis_array if (x, inner_loop_value) in results] , dtype=__lowerCamelCase , ) else: SCREAMING_SNAKE_CASE__ = np.asarray( [results[(inner_loop_value, x)] for x in x_axis_array if (inner_loop_value, x) in results] , dtype=np.floataa , ) ((SCREAMING_SNAKE_CASE__),(SCREAMING_SNAKE_CASE__)) = ( ('''batch_size''', '''len''') if self.args.plot_along_batch else ('''in #tokens''', '''bsz''') ) SCREAMING_SNAKE_CASE__ = np.asarray(__lowerCamelCase , __lowerCamelCase )[: len(__lowerCamelCase )] plt.scatter( __lowerCamelCase , __lowerCamelCase , label=f'''{label_model_name} - {inner_loop_label}: {inner_loop_value}''' ) plt.plot(__lowerCamelCase , __lowerCamelCase , '''--''' ) title_str += f''' {label_model_name} vs.''' SCREAMING_SNAKE_CASE__ = title_str[:-4] SCREAMING_SNAKE_CASE__ = '''Time in s''' if self.args.is_time else '''Memory in MB''' # plot plt.title(__lowerCamelCase ) plt.xlabel(__lowerCamelCase ) plt.ylabel(__lowerCamelCase ) plt.legend() if self.args.figure_png_file is not None: plt.savefig(self.args.figure_png_file ) else: plt.show() def UpperCAmelCase_ ( ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = HfArgumentParser(_A ) SCREAMING_SNAKE_CASE__ = parser.parse_args_into_dataclasses()[0] SCREAMING_SNAKE_CASE__ = Plot(args=_A ) plot.plot() if __name__ == "__main__": main()
218
import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 _SCREAMING_SNAKE_CASE : Tuple = data_utils.TransfoXLTokenizer _SCREAMING_SNAKE_CASE : Dict = data_utils.TransfoXLCorpus _SCREAMING_SNAKE_CASE : Union[str, Any] = data_utils _SCREAMING_SNAKE_CASE : Any = data_utils def UpperCAmelCase_ ( _A , _A , _A , _A ): '''simple docstring''' if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(_A , '''rb''' ) as fp: SCREAMING_SNAKE_CASE__ = pickle.load(_A , encoding='''latin1''' ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) SCREAMING_SNAKE_CASE__ = pytorch_dump_folder_path + '''/''' + VOCAB_FILES_NAMES['''pretrained_vocab_file'''] print(F'''Save vocabulary to {pytorch_vocab_dump_path}''' ) SCREAMING_SNAKE_CASE__ = corpus.vocab.__dict__ torch.save(_A , _A ) SCREAMING_SNAKE_CASE__ = corpus.__dict__ corpus_dict_no_vocab.pop('''vocab''' , _A ) SCREAMING_SNAKE_CASE__ = pytorch_dump_folder_path + '''/''' + CORPUS_NAME print(F'''Save dataset to {pytorch_dataset_dump_path}''' ) torch.save(_A , _A ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model SCREAMING_SNAKE_CASE__ = os.path.abspath(_A ) SCREAMING_SNAKE_CASE__ = os.path.abspath(_A ) print(F'''Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.''' ) # Initialise PyTorch model if transfo_xl_config_file == "": SCREAMING_SNAKE_CASE__ = TransfoXLConfig() else: SCREAMING_SNAKE_CASE__ = TransfoXLConfig.from_json_file(_A ) print(F'''Building PyTorch model from configuration: {config}''' ) SCREAMING_SNAKE_CASE__ = TransfoXLLMHeadModel(_A ) SCREAMING_SNAKE_CASE__ = load_tf_weights_in_transfo_xl(_A , _A , _A ) # Save pytorch-model SCREAMING_SNAKE_CASE__ = os.path.join(_A , _A ) SCREAMING_SNAKE_CASE__ = os.path.join(_A , _A ) print(F'''Save PyTorch model to {os.path.abspath(_A )}''' ) torch.save(model.state_dict() , _A ) print(F'''Save configuration file to {os.path.abspath(_A )}''' ) with open(_A , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE : Any = argparse.ArgumentParser() parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the folder to store the PyTorch model or dataset/vocab.''', ) parser.add_argument( '''--tf_checkpoint_path''', default='''''', type=str, help='''An optional path to a TensorFlow checkpoint path to be converted.''', ) parser.add_argument( '''--transfo_xl_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained BERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--transfo_xl_dataset_file''', default='''''', type=str, help='''An optional dataset file to be converted in a vocabulary.''', ) _SCREAMING_SNAKE_CASE : Union[str, Any] = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
218
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase: str = logging.get_logger(__name__) _lowercase: str = { # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class _lowercase ( lowerCamelCase__ ): """simple docstring""" __A = """megatron-bert""" def __init__(self , lowerCamelCase_=29056 , lowerCamelCase_=1024 , lowerCamelCase_=24 , lowerCamelCase_=16 , lowerCamelCase_=4096 , lowerCamelCase_="gelu" , lowerCamelCase_=0.1 , lowerCamelCase_=0.1 , lowerCamelCase_=512 , lowerCamelCase_=2 , lowerCamelCase_=0.02 , lowerCamelCase_=1E-1_2 , lowerCamelCase_=0 , lowerCamelCase_="absolute" , lowerCamelCase_=True , **lowerCamelCase_ , ): """simple docstring""" super().__init__(pad_token_id=snake_case__ , **snake_case__ ) a = vocab_size a = hidden_size a = num_hidden_layers a = num_attention_heads a = hidden_act a = intermediate_size a = hidden_dropout_prob a = attention_probs_dropout_prob a = max_position_embeddings a = type_vocab_size a = initializer_range a = layer_norm_eps a = position_embedding_type a = use_cache
227
__snake_case = '''Input must be a string of 8 numbers plus letter''' __snake_case = '''TRWAGMYFPDXBNJZSQVHLCKE''' def lowerCAmelCase_ ( __lowerCAmelCase )-> bool: '''simple docstring''' if not isinstance(__lowerCAmelCase , __lowerCAmelCase ): UpperCAmelCase : Optional[Any] =f'''Expected string as input, found {type(__lowerCAmelCase ).__name__}''' raise TypeError(__lowerCAmelCase ) UpperCAmelCase : List[Any] =spanish_id.replace('''-''' , '''''' ).upper() if len(__lowerCAmelCase ) != 9: raise ValueError(__lowerCAmelCase ) try: UpperCAmelCase : int =int(spanish_id_clean[0:8] ) UpperCAmelCase : Optional[int] =spanish_id_clean[8] except ValueError as ex: raise ValueError(__lowerCAmelCase ) from ex if letter.isdigit(): raise ValueError(__lowerCAmelCase ) return letter == LOOKUP_LETTERS[number % 23] if __name__ == "__main__": import doctest doctest.testmod()
348
0
'''simple docstring''' import inspect import unittest from transformers import RegNetConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import RegNetForImageClassification, RegNetModel from transformers.models.regnet.modeling_regnet import REGNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase__ : def __init__( self : Tuple ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Optional[int]=3 ,lowerCamelCase__ : int=32 ,lowerCamelCase__ : str=3 ,lowerCamelCase__ : int=10 ,lowerCamelCase__ : Tuple=[10, 20, 30, 40] ,lowerCamelCase__ : Optional[int]=[1, 1, 2, 1] ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : int=True ,lowerCamelCase__ : List[Any]="relu" ,lowerCamelCase__ : Tuple=3 ,lowerCamelCase__ : str=None ,): '''simple docstring''' _UpperCamelCase : Any = parent _UpperCamelCase : Optional[int] = batch_size _UpperCamelCase : List[str] = image_size _UpperCamelCase : List[Any] = num_channels _UpperCamelCase : Optional[int] = embeddings_size _UpperCamelCase : Union[str, Any] = hidden_sizes _UpperCamelCase : List[Any] = depths _UpperCamelCase : Tuple = is_training _UpperCamelCase : Optional[Any] = use_labels _UpperCamelCase : Union[str, Any] = hidden_act _UpperCamelCase : int = num_labels _UpperCamelCase : Optional[Any] = scope _UpperCamelCase : List[str] = len(lowerCamelCase__ ) def UpperCamelCase_ ( self : Any ): '''simple docstring''' _UpperCamelCase : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCamelCase : List[str] = None if self.use_labels: _UpperCamelCase : str = ids_tensor([self.batch_size] ,self.num_labels ) _UpperCamelCase : Optional[int] = self.get_config() return config, pixel_values, labels def UpperCamelCase_ ( self : str ): '''simple docstring''' return RegNetConfig( num_channels=self.num_channels ,embeddings_size=self.embeddings_size ,hidden_sizes=self.hidden_sizes ,depths=self.depths ,hidden_act=self.hidden_act ,num_labels=self.num_labels ,) def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : str ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : int = RegNetModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCamelCase : int = model(lowerCamelCase__ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) ,) def UpperCamelCase_ ( self : Union[str, Any] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Dict ): '''simple docstring''' _UpperCamelCase : Optional[int] = self.num_labels _UpperCamelCase : str = RegNetForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _UpperCamelCase : int = model(lowerCamelCase__ ,labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' _UpperCamelCase : Optional[int] = self.prepare_config_and_inputs() _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : int = config_and_inputs _UpperCamelCase : Union[str, Any] = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class lowercase__ ( lowercase , lowercase , unittest.TestCase ): lowercase__ = (RegNetModel, RegNetForImageClassification) if is_torch_available() else () lowercase__ = ( {"""feature-extraction""": RegNetModel, """image-classification""": RegNetForImageClassification} if is_torch_available() else {} ) lowercase__ = False lowercase__ = False lowercase__ = False lowercase__ = False def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _UpperCamelCase : Optional[Any] = RegNetModelTester(self ) _UpperCamelCase : Dict = ConfigTester(self ,config_class=lowerCamelCase__ ,has_text_modality=lowerCamelCase__ ) def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase_ ( self : Dict ): '''simple docstring''' return @unittest.skip(reason='RegNet does not use inputs_embeds' ) def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' pass @unittest.skip(reason='RegNet does not support input and output embeddings' ) def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' pass def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase , _UpperCamelCase : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase : List[str] = model_class(lowerCamelCase__ ) _UpperCamelCase : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCamelCase : Optional[int] = [*signature.parameters.keys()] _UpperCamelCase : Any = ['pixel_values'] self.assertListEqual(arg_names[:1] ,lowerCamelCase__ ) def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' _UpperCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' _UpperCamelCase , _UpperCamelCase : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCamelCase : Union[str, Any] = model_class(config=lowerCamelCase__ ) for name, module in model.named_modules(): if isinstance(lowerCamelCase__ ,(nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) ,msg=F'Parameter {name} of model {model_class} seems not properly initialized' ,) self.assertTrue( torch.all(module.bias == 0 ) ,msg=F'Parameter {name} of model {model_class} seems not properly initialized' ,) def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' def check_hidden_states_output(lowerCamelCase__ : Dict ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : List[Any] ): _UpperCamelCase : Union[str, Any] = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _UpperCamelCase : Dict = model(**self._prepare_for_class(lowerCamelCase__ ,lowerCamelCase__ ) ) _UpperCamelCase : List[Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _UpperCamelCase : int = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase__ ) ,expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[self.model_tester.image_size // 2, self.model_tester.image_size // 2] ,) _UpperCamelCase , _UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase : Any = ['basic', 'bottleneck'] for model_class in self.all_model_classes: for layer_type in layers_type: _UpperCamelCase : Dict = layer_type _UpperCamelCase : Dict = True check_hidden_states_output(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCamelCase : List[Any] = True check_hidden_states_output(lowerCamelCase__ ,lowerCamelCase__ ,lowerCamelCase__ ) def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' _UpperCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @slow def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' for model_name in REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase : Union[str, Any] = RegNetModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def A__ ( ): _UpperCamelCase : Union[str, Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class lowercase__ ( unittest.TestCase ): @cached_property def UpperCamelCase_ ( self : Dict ): '''simple docstring''' return ( AutoImageProcessor.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Optional[int] = RegNetForImageClassification.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowerCamelCase__ ) _UpperCamelCase : Any = self.default_image_processor _UpperCamelCase : Union[str, Any] = prepare_img() _UpperCamelCase : Optional[int] = image_processor(images=lowerCamelCase__ ,return_tensors='pt' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): _UpperCamelCase : Tuple = model(**lowerCamelCase__ ) # verify the logits _UpperCamelCase : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape ,lowerCamelCase__ ) _UpperCamelCase : Tuple = torch.tensor([-0.4_1_8_0, -1.5_0_5_1, -3.4_8_3_6] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowerCamelCase__ ,atol=1E-4 ) )
236
'''simple docstring''' import argparse import torch from transformers import LxmertConfig, LxmertForPreTraining, load_tf_weights_in_lxmert from transformers.utils import logging logging.set_verbosity_info() def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): # Initialise PyTorch model _UpperCamelCase : Any = LxmertConfig.from_json_file(UpperCAmelCase_ ) print(f'Building PyTorch model from configuration: {config}' ) _UpperCamelCase : int = LxmertForPreTraining(UpperCAmelCase_ ) # Load weights from tf checkpoint load_tf_weights_in_lxmert(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) # Save pytorch-model print(f'Save PyTorch model to {pytorch_dump_path}' ) torch.save(model.state_dict() , UpperCAmelCase_ ) if __name__ == "__main__": snake_case_ : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) snake_case_ : List[str] = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
236
1
"""simple docstring""" def lowercase (_lowerCAmelCase = 100_0000 ): __lowerCAmelCase = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , _lowerCAmelCase ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
301
"""simple docstring""" import importlib.util import json import os import warnings from dataclasses import dataclass, field import torch from ..training_args import TrainingArguments from ..utils import cached_property, is_sagemaker_dp_enabled, logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) def lowercase (): # Get the sagemaker specific mp parameters from smp_options variable. __lowerCAmelCase = os.getenv("""SM_HP_MP_PARAMETERS""" , """{}""" ) try: # Parse it and check the field "partitions" is included, it is required for model parallel. __lowerCAmelCase = json.loads(_lowerCAmelCase ) if "partitions" not in smp_options: return False except json.JSONDecodeError: return False # Get the sagemaker specific framework parameters from mpi_options variable. __lowerCAmelCase = os.getenv("""SM_FRAMEWORK_PARAMS""" , """{}""" ) try: # Parse it and check the field "sagemaker_distributed_dataparallel_enabled". __lowerCAmelCase = json.loads(_lowerCAmelCase ) if not mpi_options.get("""sagemaker_mpi_enabled""" , _lowerCAmelCase ): return False except json.JSONDecodeError: return False # Lastly, check if the `smdistributed` module is present. return importlib.util.find_spec("""smdistributed""" ) is not None if is_sagemaker_model_parallel_available(): import smdistributed.modelparallel.torch as smp smp.init() @dataclass class lowerCAmelCase_ ( A__ ): '''simple docstring''' _snake_case = field( default='''''' , metadata={'''help''': '''Used by the SageMaker launcher to send mp-specific args. Ignored in SageMakerTrainer'''} , ) def A__ ( self ) -> Tuple: super().__post_init__() warnings.warn( """`SageMakerTrainingArguments` is deprecated and will be removed in v5 of Transformers. You can use """ """`TrainingArguments` instead.""" , snake_case_ , ) @cached_property def A__ ( self ) -> "torch.device": logger.info("""PyTorch: setting up devices""" ) if torch.distributed.is_available() and torch.distributed.is_initialized() and self.local_rank == -1: logger.warning( """torch.distributed process group is initialized, but local_rank == -1. """ """In order to use Torch DDP, launch your script with `python -m torch.distributed.launch""" ) if self.no_cuda: __lowerCAmelCase = torch.device("""cpu""" ) __lowerCAmelCase = 0 elif is_sagemaker_model_parallel_available(): __lowerCAmelCase = smp.local_rank() __lowerCAmelCase = torch.device("""cuda""" , snake_case_ ) __lowerCAmelCase = 1 elif is_sagemaker_dp_enabled(): import smdistributed.dataparallel.torch.torch_smddp # noqa: F401 torch.distributed.init_process_group(backend="""smddp""" , timeout=self.ddp_timeout_delta ) __lowerCAmelCase = int(os.getenv("""SMDATAPARALLEL_LOCAL_RANK""" ) ) __lowerCAmelCase = torch.device("""cuda""" , self.local_rank ) __lowerCAmelCase = 1 elif self.local_rank == -1: # if n_gpu is > 1 we'll use nn.DataParallel. # If you only want to use a specific subset of GPUs use `CUDA_VISIBLE_DEVICES=0` # Explicitly set CUDA to the first (index 0) CUDA device, otherwise `set_device` will # trigger an error that a device index is missing. Index 0 takes into account the # GPUs available in the environment, so `CUDA_VISIBLE_DEVICES=1,2` with `cuda:0` # will use the first GPU in that env, i.e. GPU#1 __lowerCAmelCase = torch.device("""cuda:0""" if torch.cuda.is_available() else """cpu""" ) # Sometimes the line in the postinit has not been run before we end up here, so just checking we're not at # the default value. __lowerCAmelCase = torch.cuda.device_count() else: # Here, we'll use torch.distributed. # Initializes the distributed backend which will take care of synchronizing nodes/GPUs if not torch.distributed.is_initialized(): torch.distributed.init_process_group(backend="""nccl""" , timeout=self.ddp_timeout_delta ) __lowerCAmelCase = torch.device("""cuda""" , self.local_rank ) __lowerCAmelCase = 1 if device.type == "cuda": torch.cuda.set_device(snake_case_ ) return device @property def A__ ( self ) -> Dict: if is_sagemaker_model_parallel_available(): return smp.dp_size() return super().world_size @property def A__ ( self ) -> Optional[int]: return not is_sagemaker_model_parallel_available() @property def A__ ( self ) -> Tuple: return False
301
1
"""simple docstring""" from collections import defaultdict def _lowercase ( __lowerCAmelCase ) -> int: SCREAMING_SNAKE_CASE__ : List[Any] = 1 SCREAMING_SNAKE_CASE__ : Optional[Any] = True for v in tree[start]: if v not in visited: ret += dfs(__lowerCAmelCase ) if ret % 2 == 0: cuts.append(__lowerCAmelCase ) return ret def _lowercase ( ) -> Any: dfs(1 ) if __name__ == "__main__": a ,a :Union[str, Any] = 10, 9 a :Tuple = defaultdict(list) a :dict[int, bool] = {} a :list[int] = [] a :Tuple = 0 a :Optional[Any] = [(2, 1), (3, 1), (4, 3), (5, 2), (6, 1), (7, 2), (8, 6), (9, 8), (10, 8)] for u, v in edges: tree[u].append(v) tree[v].append(u) even_tree() print(len(cuts) - 1)
56
"""simple docstring""" import re from typing import Callable, List, Optional, Union import tensorflow as tf try: from tensorflow.keras.optimizers.legacy import Adam except ImportError: from tensorflow.keras.optimizers import Adam class __a (tf.keras.optimizers.schedules.LearningRateSchedule): '''simple docstring''' def __init__( self , _a , _a , _a , _a = 1.0 , _a = None , ) -> Union[str, Any]: """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE__ : List[Any] = initial_learning_rate SCREAMING_SNAKE_CASE__ : Tuple = warmup_steps SCREAMING_SNAKE_CASE__ : Optional[Any] = power SCREAMING_SNAKE_CASE__ : Optional[Any] = decay_schedule_fn SCREAMING_SNAKE_CASE__ : Any = name def __call__( self , _a ) -> List[Any]: """simple docstring""" with tf.name_scope(self.name or """WarmUp""" ) as name: # Implements polynomial warmup. i.e., if global_step < warmup_steps, the # learning rate will be `global_step/num_warmup_steps * init_lr`. SCREAMING_SNAKE_CASE__ : Optional[Any] = tf.cast(_a , tf.floataa ) SCREAMING_SNAKE_CASE__ : Optional[Any] = tf.cast(self.warmup_steps , tf.floataa ) SCREAMING_SNAKE_CASE__ : str = global_step_float / warmup_steps_float SCREAMING_SNAKE_CASE__ : Optional[int] = self.initial_learning_rate * tf.math.pow(_a , self.power ) return tf.cond( global_step_float < warmup_steps_float , lambda: warmup_learning_rate , lambda: self.decay_schedule_fn(step - self.warmup_steps ) , name=_a , ) def _a ( self ) -> List[Any]: """simple docstring""" return { "initial_learning_rate": self.initial_learning_rate, "decay_schedule_fn": self.decay_schedule_fn, "warmup_steps": self.warmup_steps, "power": self.power, "name": self.name, } def _lowercase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = 0.0 , __lowerCAmelCase = 0.9 , __lowerCAmelCase = 0.999 , __lowerCAmelCase = 1E-8 , __lowerCAmelCase = None , __lowerCAmelCase = None , __lowerCAmelCase = 0.0 , __lowerCAmelCase = 1.0 , __lowerCAmelCase = None , ) -> Dict: SCREAMING_SNAKE_CASE__ : Optional[int] = tf.keras.optimizers.schedules.PolynomialDecay( initial_learning_rate=__lowerCAmelCase , decay_steps=num_train_steps - num_warmup_steps , end_learning_rate=init_lr * min_lr_ratio , power=__lowerCAmelCase , ) if num_warmup_steps: SCREAMING_SNAKE_CASE__ : Dict = WarmUp( initial_learning_rate=__lowerCAmelCase , decay_schedule_fn=__lowerCAmelCase , warmup_steps=__lowerCAmelCase , ) if weight_decay_rate > 0.0: SCREAMING_SNAKE_CASE__ : int = AdamWeightDecay( learning_rate=__lowerCAmelCase , weight_decay_rate=__lowerCAmelCase , beta_a=__lowerCAmelCase , beta_a=__lowerCAmelCase , epsilon=__lowerCAmelCase , clipnorm=__lowerCAmelCase , global_clipnorm=__lowerCAmelCase , exclude_from_weight_decay=["""LayerNorm""", """layer_norm""", """bias"""] , include_in_weight_decay=__lowerCAmelCase , ) else: SCREAMING_SNAKE_CASE__ : int = tf.keras.optimizers.Adam( learning_rate=__lowerCAmelCase , beta_a=__lowerCAmelCase , beta_a=__lowerCAmelCase , epsilon=__lowerCAmelCase , clipnorm=__lowerCAmelCase , global_clipnorm=__lowerCAmelCase , ) # We return the optimizer and the LR scheduler in order to better track the # evolution of the LR independently of the optimizer. return optimizer, lr_schedule class __a (UpperCamelCase_): '''simple docstring''' def __init__( self , _a = 0.001 , _a = 0.9 , _a = 0.999 , _a = 1E-7 , _a = False , _a = 0.0 , _a = None , _a = None , _a = "AdamWeightDecay" , **_a , ) -> Union[str, Any]: """simple docstring""" super().__init__(_a , _a , _a , _a , _a , _a , **_a ) SCREAMING_SNAKE_CASE__ : Tuple = weight_decay_rate SCREAMING_SNAKE_CASE__ : Tuple = include_in_weight_decay SCREAMING_SNAKE_CASE__ : Dict = exclude_from_weight_decay @classmethod def _a ( cls , _a ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = {"""WarmUp""": WarmUp} return super(_a , cls ).from_config(_a , custom_objects=_a ) def _a ( self , _a , _a , _a ) -> str: """simple docstring""" super(_a , self )._prepare_local(_a , _a , _a ) SCREAMING_SNAKE_CASE__ : Tuple = tf.constant( self.weight_decay_rate , name="""adam_weight_decay_rate""" ) def _a ( self , _a , _a , _a ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = self._do_use_weight_decay(var.name ) if do_decay: return var.assign_sub( learning_rate * var * apply_state[(var.device, var.dtype.base_dtype)]["""weight_decay_rate"""] , use_locking=self._use_locking , ) return tf.no_op() def _a ( self , _a , _a=None , **_a ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : str = list(zip(*_a ) ) return super(_a , self ).apply_gradients(zip(_a , _a ) , name=_a , **_a ) def _a ( self , _a , _a , _a ) -> str: """simple docstring""" if apply_state is None: return self._decayed_lr_t[var_dtype], {} SCREAMING_SNAKE_CASE__ : Dict = apply_state or {} SCREAMING_SNAKE_CASE__ : List[str] = apply_state.get((var_device, var_dtype) ) if coefficients is None: SCREAMING_SNAKE_CASE__ : Union[str, Any] = self._fallback_apply_state(_a , _a ) SCREAMING_SNAKE_CASE__ : List[Any] = coefficients return coefficients["lr_t"], {"apply_state": apply_state} def _a ( self , _a , _a , _a=None ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Optional[int] = self._get_lr(var.device , var.dtype.base_dtype , _a ) SCREAMING_SNAKE_CASE__ : Any = self._decay_weights_op(_a , _a , _a ) with tf.control_dependencies([decay] ): return super(_a , self )._resource_apply_dense(_a , _a , **_a ) def _a ( self , _a , _a , _a , _a=None ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[str] = self._get_lr(var.device , var.dtype.base_dtype , _a ) SCREAMING_SNAKE_CASE__ : Dict = self._decay_weights_op(_a , _a , _a ) with tf.control_dependencies([decay] ): return super(_a , self )._resource_apply_sparse(_a , _a , _a , **_a ) def _a ( self ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = super().get_config() config.update({"""weight_decay_rate""": self.weight_decay_rate} ) return config def _a ( self , _a ) -> Tuple: """simple docstring""" if self.weight_decay_rate == 0: return False if self._include_in_weight_decay: for r in self._include_in_weight_decay: if re.search(_a , _a ) is not None: return True if self._exclude_from_weight_decay: for r in self._exclude_from_weight_decay: if re.search(_a , _a ) is not None: return False return True class __a (UpperCamelCase_): '''simple docstring''' def __init__( self ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = [] SCREAMING_SNAKE_CASE__ : List[str] = None @property def _a ( self ) -> str: """simple docstring""" if self._accum_steps is None: SCREAMING_SNAKE_CASE__ : Dict = tf.Variable( tf.constant(0 , dtype=tf.intaa ) , trainable=_a , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) return self._accum_steps.value() @property def _a ( self ) -> List[str]: """simple docstring""" if not self._gradients: raise ValueError("""The accumulator should be called first to initialize the gradients""" ) return [gradient.value() if gradient is not None else gradient for gradient in self._gradients] def __call__( self , _a ) -> str: """simple docstring""" if not self._gradients: SCREAMING_SNAKE_CASE__ : Optional[Any] = self.step # Create the step variable. self._gradients.extend( [ tf.Variable( tf.zeros_like(_a ) , trainable=_a , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) if gradient is not None else gradient for gradient in gradients ] ) if len(_a ) != len(self._gradients ): raise ValueError(f'''Expected {len(self._gradients )} gradients, but got {len(_a )}''' ) for accum_gradient, gradient in zip(self._gradients , _a ): if accum_gradient is not None and gradient is not None: accum_gradient.assign_add(_a ) self._accum_steps.assign_add(1 ) def _a ( self ) -> Any: """simple docstring""" if not self._gradients: return self._accum_steps.assign(0 ) for gradient in self._gradients: if gradient is not None: gradient.assign(tf.zeros_like(_a ) )
56
1
import logging import os from logging import ( CRITICAL, # NOQA DEBUG, # NOQA ERROR, # NOQA FATAL, # NOQA INFO, # NOQA NOTSET, # NOQA WARN, # NOQA WARNING, # NOQA ) from typing import Optional from tqdm import auto as tqdm_lib A__: Optional[int] = { '''debug''': logging.DEBUG, '''info''': logging.INFO, '''warning''': logging.WARNING, '''error''': logging.ERROR, '''critical''': logging.CRITICAL, } A__: int = logging.WARNING def lowerCAmelCase_ ( ): UpperCamelCase__: Optional[int] = os.getenv("DATASETS_VERBOSITY" ,A_) if env_level_str: if env_level_str in log_levels: return log_levels[env_level_str] else: logging.getLogger().warning( F"Unknown option DATASETS_VERBOSITY={env_level_str}, " F"has to be one of: { ', '.join(log_levels.keys()) }") return _default_log_level def lowerCAmelCase_ ( ): return __name__.split(".")[0] def lowerCAmelCase_ ( ): return logging.getLogger(_get_library_name()) def lowerCAmelCase_ ( ): # Apply our default configuration to the library root logger. UpperCamelCase__: Tuple = _get_library_root_logger() library_root_logger.setLevel(_get_default_logging_level()) def lowerCAmelCase_ ( ): UpperCamelCase__: Tuple = _get_library_root_logger() library_root_logger.setLevel(logging.NOTSET) def lowerCAmelCase_ ( A_ = None): if name is None: UpperCamelCase__: Optional[Any] = _get_library_name() return logging.getLogger(A_) def lowerCAmelCase_ ( ): return _get_library_root_logger().getEffectiveLevel() def lowerCAmelCase_ ( A_): _get_library_root_logger().setLevel(A_) def lowerCAmelCase_ ( ): return set_verbosity(A_) def lowerCAmelCase_ ( ): return set_verbosity(A_) def lowerCAmelCase_ ( ): return set_verbosity(A_) def lowerCAmelCase_ ( ): return set_verbosity(A_) def lowerCAmelCase_ ( ): UpperCamelCase__: List[Any] = False def lowerCAmelCase_ ( ): UpperCamelCase__: List[str] = True # Configure the library root logger at the module level (singleton-like) _configure_library_root_logger() class _a : """simple docstring""" def __init__( self: int , *__lowerCamelCase: Tuple , **__lowerCamelCase: str ): # pylint: disable=unused-argument '''simple docstring''' UpperCamelCase__: int = args[0] if args else None def __iter__( self: Optional[Any] ): '''simple docstring''' return iter(self._iterator ) def __getattr__( self: Dict , __lowerCamelCase: Any ): '''simple docstring''' def empty_fn(*__lowerCamelCase: Any , **__lowerCamelCase: Optional[Any] ): # pylint: disable=unused-argument return return empty_fn def __enter__( self: str ): '''simple docstring''' return self def __exit__( self: Tuple , __lowerCamelCase: List[str] , __lowerCamelCase: Any , __lowerCamelCase: List[Any] ): '''simple docstring''' return A__: Tuple = True class _a : """simple docstring""" def __call__( self: Any , *__lowerCamelCase: List[str] , __lowerCamelCase: List[Any]=False , **__lowerCamelCase: Union[str, Any] ): '''simple docstring''' if _tqdm_active and not disable: return tqdm_lib.tqdm(*__lowerCamelCase , **__lowerCamelCase ) else: return EmptyTqdm(*__lowerCamelCase , **__lowerCamelCase ) def UpperCAmelCase_ ( self: List[str] , *__lowerCamelCase: List[str] , **__lowerCamelCase: Tuple ): '''simple docstring''' UpperCamelCase__: Optional[int] = None if _tqdm_active: return tqdm_lib.tqdm.set_lock(*__lowerCamelCase , **__lowerCamelCase ) def UpperCAmelCase_ ( self: int ): '''simple docstring''' if _tqdm_active: return tqdm_lib.tqdm.get_lock() A__: Optional[Any] = _tqdm_cls() def lowerCAmelCase_ ( ): global _tqdm_active return bool(_tqdm_active) def lowerCAmelCase_ ( ): global _tqdm_active UpperCamelCase__: int = True def lowerCAmelCase_ ( ): global _tqdm_active UpperCamelCase__: str = False
149
from typing import Dict, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import flip_channel_order, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract A__: str = logging.get_logger(__name__) def lowerCAmelCase_ ( A_ ,A_ ,A_): return [ int(10_00 * (box[0] / width)), int(10_00 * (box[1] / height)), int(10_00 * (box[2] / width)), int(10_00 * (box[3] / height)), ] def lowerCAmelCase_ ( A_ ,A_ ,A_ = None): UpperCamelCase__: List[str] = tesseract_config if tesseract_config is not None else "" # apply OCR UpperCamelCase__: Optional[int] = to_pil_image(A_) UpperCamelCase__ , UpperCamelCase__: Tuple = pil_image.size UpperCamelCase__: List[Any] = pytesseract.image_to_data(A_ ,lang=A_ ,output_type="dict" ,config=A_) UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__: Dict = data["text"], data["left"], data["top"], data["width"], data["height"] # filter empty words and corresponding coordinates UpperCamelCase__: List[Any] = [idx for idx, word in enumerate(A_) if not word.strip()] UpperCamelCase__: Union[str, Any] = [word for idx, word in enumerate(A_) if idx not in irrelevant_indices] UpperCamelCase__: Dict = [coord for idx, coord in enumerate(A_) if idx not in irrelevant_indices] UpperCamelCase__: List[Any] = [coord for idx, coord in enumerate(A_) if idx not in irrelevant_indices] UpperCamelCase__: Optional[int] = [coord for idx, coord in enumerate(A_) if idx not in irrelevant_indices] UpperCamelCase__: Optional[Any] = [coord for idx, coord in enumerate(A_) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format UpperCamelCase__: List[str] = [] for x, y, w, h in zip(A_ ,A_ ,A_ ,A_): UpperCamelCase__: str = [x, y, x + w, y + h] actual_boxes.append(A_) # finally, normalize the bounding boxes UpperCamelCase__: Union[str, Any] = [] for box in actual_boxes: normalized_boxes.append(normalize_box(A_ ,A_ ,A_)) assert len(A_) == len(A_), "Not as many words as there are bounding boxes" return words, normalized_boxes class _a ( UpperCamelCase__): """simple docstring""" UpperCamelCase__ = ["""pixel_values"""] def __init__( self: int , __lowerCamelCase: bool = True , __lowerCamelCase: Dict[str, int] = None , __lowerCamelCase: PILImageResampling = PILImageResampling.BILINEAR , __lowerCamelCase: bool = True , __lowerCamelCase: Optional[str] = None , __lowerCamelCase: Optional[str] = "" , **__lowerCamelCase: str , ): '''simple docstring''' super().__init__(**__lowerCamelCase ) UpperCamelCase__: Optional[Any] = size if size is not None else {"height": 224, "width": 224} UpperCamelCase__: Dict = get_size_dict(__lowerCamelCase ) UpperCamelCase__: Optional[Any] = do_resize UpperCamelCase__: Optional[int] = size UpperCamelCase__: int = resample UpperCamelCase__: str = apply_ocr UpperCamelCase__: List[Any] = ocr_lang UpperCamelCase__: List[Any] = tesseract_config def UpperCAmelCase_ ( self: int , __lowerCamelCase: np.ndarray , __lowerCamelCase: Dict[str, int] , __lowerCamelCase: PILImageResampling = PILImageResampling.BILINEAR , __lowerCamelCase: Optional[Union[str, ChannelDimension]] = None , **__lowerCamelCase: str , ): '''simple docstring''' UpperCamelCase__: int = get_size_dict(__lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(F"The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}" ) UpperCamelCase__: int = (size["height"], size["width"]) return resize(__lowerCamelCase , size=__lowerCamelCase , resample=__lowerCamelCase , data_format=__lowerCamelCase , **__lowerCamelCase ) def UpperCAmelCase_ ( self: Dict , __lowerCamelCase: ImageInput , __lowerCamelCase: bool = None , __lowerCamelCase: Dict[str, int] = None , __lowerCamelCase: PILImageResampling = None , __lowerCamelCase: bool = None , __lowerCamelCase: Optional[str] = None , __lowerCamelCase: Optional[str] = None , __lowerCamelCase: Optional[Union[str, TensorType]] = None , __lowerCamelCase: ChannelDimension = ChannelDimension.FIRST , **__lowerCamelCase: str , ): '''simple docstring''' UpperCamelCase__: str = do_resize if do_resize is not None else self.do_resize UpperCamelCase__: Any = size if size is not None else self.size UpperCamelCase__: Union[str, Any] = get_size_dict(__lowerCamelCase ) UpperCamelCase__: Tuple = resample if resample is not None else self.resample UpperCamelCase__: int = apply_ocr if apply_ocr is not None else self.apply_ocr UpperCamelCase__: Any = ocr_lang if ocr_lang is not None else self.ocr_lang UpperCamelCase__: Optional[int] = tesseract_config if tesseract_config is not None else self.tesseract_config UpperCamelCase__: Any = make_list_of_images(__lowerCamelCase ) if not valid_images(__lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None: raise ValueError("Size must be specified if do_resize is True." ) # All transformations expect numpy arrays. UpperCamelCase__: Union[str, Any] = [to_numpy_array(__lowerCamelCase ) for image in images] if apply_ocr: requires_backends(self , "pytesseract" ) UpperCamelCase__: str = [] UpperCamelCase__: Optional[Any] = [] for image in images: UpperCamelCase__ , UpperCamelCase__: Any = apply_tesseract(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) words_batch.append(__lowerCamelCase ) boxes_batch.append(__lowerCamelCase ) if do_resize: UpperCamelCase__: List[Any] = [self.resize(image=__lowerCamelCase , size=__lowerCamelCase , resample=__lowerCamelCase ) for image in images] # flip color channels from RGB to BGR (as Detectron2 requires this) UpperCamelCase__: List[str] = [flip_channel_order(__lowerCamelCase ) for image in images] UpperCamelCase__: Optional[Any] = [to_channel_dimension_format(__lowerCamelCase , __lowerCamelCase ) for image in images] UpperCamelCase__: int = BatchFeature(data={"pixel_values": images} , tensor_type=__lowerCamelCase ) if apply_ocr: UpperCamelCase__: Dict = words_batch UpperCamelCase__: Tuple = boxes_batch return data
149
1
import copy import tempfile import unittest from huggingface_hub import HfFolder, delete_repo from parameterized import parameterized from requests.exceptions import HTTPError from transformers import AutoConfig, GenerationConfig from transformers.testing_utils import TOKEN, USER, is_staging_test class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @parameterized.expand([(None,), ('foo.json',)] ) def lowercase ( self : Dict , lowerCAmelCase_ : Optional[Any] ) -> Union[str, Any]: __lowerCAmelCase = GenerationConfig( do_sample=lowerCAmelCase_ , temperature=0.7 , length_penalty=1.0 , bad_words_ids=[[1, 2, 3], [4, 5]] , ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCAmelCase_ , config_name=lowerCAmelCase_ ) __lowerCAmelCase = GenerationConfig.from_pretrained(lowerCAmelCase_ , config_name=lowerCAmelCase_ ) # Checks parameters that were specified self.assertEqual(loaded_config.do_sample , lowerCAmelCase_ ) self.assertEqual(loaded_config.temperature , 0.7 ) self.assertEqual(loaded_config.length_penalty , 1.0 ) self.assertEqual(loaded_config.bad_words_ids , [[1, 2, 3], [4, 5]] ) # Checks parameters that were not specified (defaults) self.assertEqual(loaded_config.top_k , 5_0 ) self.assertEqual(loaded_config.max_length , 2_0 ) self.assertEqual(loaded_config.max_time , lowerCAmelCase_ ) def lowercase ( self : Dict ) -> Optional[int]: __lowerCAmelCase = AutoConfig.from_pretrained('gpt2' ) __lowerCAmelCase = GenerationConfig.from_model_config(lowerCAmelCase_ ) __lowerCAmelCase = GenerationConfig() # The generation config has loaded a few non-default parameters from the model config self.assertNotEqual(lowerCAmelCase_ , lowerCAmelCase_ ) # One of those parameters is eos_token_id -- check if it matches self.assertNotEqual(generation_config_from_model.eos_token_id , default_generation_config.eos_token_id ) self.assertEqual(generation_config_from_model.eos_token_id , model_config.eos_token_id ) def lowercase ( self : str ) -> Optional[Any]: __lowerCAmelCase = GenerationConfig() __lowerCAmelCase = { 'max_new_tokens': 1_0_2_4, 'foo': 'bar', } __lowerCAmelCase = copy.deepcopy(lowerCAmelCase_ ) __lowerCAmelCase = generation_config.update(**lowerCAmelCase_ ) # update_kwargs was not modified (no side effects) self.assertEqual(lowerCAmelCase_ , lowerCAmelCase_ ) # update_kwargs was used to update the config on valid attributes self.assertEqual(generation_config.max_new_tokens , 1_0_2_4 ) # `.update()` returns a dictionary of unused kwargs self.assertEqual(lowerCAmelCase_ , {'foo': 'bar'} ) def lowercase ( self : Optional[Any] ) -> Optional[Any]: __lowerCAmelCase = GenerationConfig() __lowerCAmelCase = 'bar' with tempfile.TemporaryDirectory('test-generation-config' ) as tmp_dir: generation_config.save_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = GenerationConfig.from_pretrained(lowerCAmelCase_ ) # update_kwargs was used to update the config on valid attributes self.assertEqual(new_config.foo , 'bar' ) __lowerCAmelCase = GenerationConfig.from_model_config(lowerCAmelCase_ ) assert not hasattr(lowerCAmelCase_ , 'foo' ) # no new kwargs should be initialized if from config def lowercase ( self : Any ) -> Optional[Any]: __lowerCAmelCase = GenerationConfig() self.assertEqual(default_config.temperature , 1.0 ) self.assertEqual(default_config.do_sample , lowerCAmelCase_ ) self.assertEqual(default_config.num_beams , 1 ) __lowerCAmelCase = GenerationConfig( do_sample=lowerCAmelCase_ , temperature=0.7 , length_penalty=1.0 , bad_words_ids=[[1, 2, 3], [4, 5]] , ) self.assertEqual(config.temperature , 0.7 ) self.assertEqual(config.do_sample , lowerCAmelCase_ ) self.assertEqual(config.num_beams , 1 ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCAmelCase_ ) __lowerCAmelCase = GenerationConfig.from_pretrained(lowerCAmelCase_ , temperature=1.0 ) self.assertEqual(loaded_config.temperature , 1.0 ) self.assertEqual(loaded_config.do_sample , lowerCAmelCase_ ) self.assertEqual(loaded_config.num_beams , 1 ) # default value @is_staging_test class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @classmethod def lowercase ( cls : Optional[int] ) -> Tuple: __lowerCAmelCase = TOKEN HfFolder.save_token(lowerCAmelCase_ ) @classmethod def lowercase ( cls : List[Any] ) -> Dict: try: delete_repo(token=cls._token , repo_id='test-generation-config' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-generation-config-org' ) except HTTPError: pass def lowercase ( self : Union[str, Any] ) -> List[Any]: __lowerCAmelCase = GenerationConfig( do_sample=lowerCAmelCase_ , temperature=0.7 , length_penalty=1.0 , ) config.push_to_hub('test-generation-config' , use_auth_token=self._token ) __lowerCAmelCase = GenerationConfig.from_pretrained(f"""{USER}/test-generation-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCAmelCase_ , getattr(lowerCAmelCase_ , lowerCAmelCase_ ) ) # Reset repo delete_repo(token=self._token , repo_id='test-generation-config' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCAmelCase_ , repo_id='test-generation-config' , push_to_hub=lowerCAmelCase_ , use_auth_token=self._token ) __lowerCAmelCase = GenerationConfig.from_pretrained(f"""{USER}/test-generation-config""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCAmelCase_ , getattr(lowerCAmelCase_ , lowerCAmelCase_ ) ) def lowercase ( self : str ) -> Any: __lowerCAmelCase = GenerationConfig( do_sample=lowerCAmelCase_ , temperature=0.7 , length_penalty=1.0 , ) config.push_to_hub('valid_org/test-generation-config-org' , use_auth_token=self._token ) __lowerCAmelCase = GenerationConfig.from_pretrained('valid_org/test-generation-config-org' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCAmelCase_ , getattr(lowerCAmelCase_ , lowerCAmelCase_ ) ) # Reset repo delete_repo(token=self._token , repo_id='valid_org/test-generation-config-org' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCAmelCase_ , repo_id='valid_org/test-generation-config-org' , push_to_hub=lowerCAmelCase_ , use_auth_token=self._token ) __lowerCAmelCase = GenerationConfig.from_pretrained('valid_org/test-generation-config-org' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCAmelCase_ , getattr(lowerCAmelCase_ , lowerCAmelCase_ ) )
360
import builtins import sys from ...utils.imports import _is_package_available from . import cursor, input from .helpers import Direction, clear_line, forceWrite, linebreak, move_cursor, reset_cursor, writeColor from .keymap import KEYMAP _snake_case : Union[str, Any] = False try: _snake_case : Tuple = _is_package_available('google.colab') except ModuleNotFoundError: pass @input.register class _UpperCAmelCase : """simple docstring""" def __init__( self : Union[str, Any] , lowerCAmelCase_ : str = None , lowerCAmelCase_ : list = [] ) -> Optional[int]: __lowerCAmelCase = 0 __lowerCAmelCase = choices __lowerCAmelCase = prompt if sys.platform == "win32": __lowerCAmelCase = '*' else: __lowerCAmelCase = '➔ ' def lowercase ( self : Dict , lowerCAmelCase_ : Dict , lowerCAmelCase_ : str = "" ) -> Any: if sys.platform != "win32": writeColor(self.choices[index] , 3_2 , lowerCAmelCase_ ) else: forceWrite(self.choices[index] , lowerCAmelCase_ ) def lowercase ( self : List[Any] , lowerCAmelCase_ : int ) -> List[Any]: if index == self.position: forceWrite(f""" {self.arrow_char} """ ) self.write_choice(lowerCAmelCase_ ) else: forceWrite(f""" {self.choices[index]}""" ) reset_cursor() def lowercase ( self : Dict , lowerCAmelCase_ : Direction , lowerCAmelCase_ : int = 1 ) -> Union[str, Any]: __lowerCAmelCase = self.position if direction == Direction.DOWN: if self.position + 1 >= len(self.choices ): return self.position += num_spaces else: if self.position - 1 < 0: return self.position -= num_spaces clear_line() self.print_choice(lowerCAmelCase_ ) move_cursor(lowerCAmelCase_ , direction.name ) self.print_choice(self.position ) @input.mark(KEYMAP['up'] ) def lowercase ( self : Optional[int] ) -> Tuple: self.move_direction(Direction.UP ) @input.mark(KEYMAP['down'] ) def lowercase ( self : str ) -> Optional[Any]: self.move_direction(Direction.DOWN ) @input.mark(KEYMAP['newline'] ) def lowercase ( self : str ) -> Any: move_cursor(len(self.choices ) - self.position , 'DOWN' ) return self.position @input.mark(KEYMAP['interrupt'] ) def lowercase ( self : Dict ) -> Optional[Any]: move_cursor(len(self.choices ) - self.position , 'DOWN' ) raise KeyboardInterrupt @input.mark_multiple(*[KEYMAP[str(lowerCAmelCase_ )] for number in range(1_0 )] ) def lowercase ( self : List[Any] ) -> Optional[int]: __lowerCAmelCase = int(chr(self.current_selection ) ) __lowerCAmelCase = index - self.position if index == self.position: return if index < len(self.choices ): if self.position > index: self.move_direction(Direction.UP , -movement ) elif self.position < index: self.move_direction(Direction.DOWN , lowerCAmelCase_ ) else: return else: return def lowercase ( self : Any , lowerCAmelCase_ : int = 0 ) -> int: if self.prompt: linebreak() forceWrite(self.prompt , '\n' ) if in_colab: forceWrite('Please input a choice index (starting from 0), and press enter' , '\n' ) else: forceWrite('Please select a choice using the arrow or number keys, and selecting with enter' , '\n' ) __lowerCAmelCase = default_choice for i in range(len(self.choices ) ): self.print_choice(lowerCAmelCase_ ) forceWrite('\n' ) move_cursor(len(self.choices ) - self.position , 'UP' ) with cursor.hide(): while True: if in_colab: try: __lowerCAmelCase = int(builtins.input() ) except ValueError: __lowerCAmelCase = default_choice else: __lowerCAmelCase = self.handle_input() if choice is not None: reset_cursor() for _ in range(len(self.choices ) + 1 ): move_cursor(1 , 'UP' ) clear_line() self.write_choice(lowerCAmelCase_ , '\n' ) return choice
207
0
'''simple docstring''' from PIL import Image def _A ( A__ , A__ ): """simple docstring""" __lowercase = (259 * (level + 255)) / (255 * (259 - level)) def contrast(A__ ) -> int: return int(128 + factor * (c - 128) ) return img.point(A__ ) if __name__ == "__main__": # Load image with Image.open('''image_data/lena.jpg''') as img: # Change contrast to 170 lowerCAmelCase__ = change_contrast(img, 170) cont_img.save('''image_data/lena_high_contrast.png''', format='''png''')
104
"""simple docstring""" import argparse import torch from transformers import GPTaConfig, GPTaModel, load_tf_weights_in_gpta from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def _A (__a , __a , __a ) -> Dict: """simple docstring""" if gpta_config_file == "": SCREAMING_SNAKE_CASE_ : Optional[Any] = GPTaConfig() else: SCREAMING_SNAKE_CASE_ : Tuple = GPTaConfig.from_json_file(__a ) SCREAMING_SNAKE_CASE_ : Optional[int] = GPTaModel(__a ) # Load weights from numpy load_tf_weights_in_gpta(__a , __a , __a ) # Save pytorch-model SCREAMING_SNAKE_CASE_ : List[str] = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME SCREAMING_SNAKE_CASE_ : List[Any] = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(f'Save PyTorch model to {pytorch_weights_dump_path}' ) torch.save(model.state_dict() , __a ) print(f'Save configuration file to {pytorch_config_dump_path}' ) with open(__a , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": UpperCAmelCase_ : str = argparse.ArgumentParser() # Required parameters parser.add_argument( """--gpt2_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--gpt2_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained OpenAI model. \n""" """This specifies the model architecture.""" ), ) UpperCAmelCase_ : Union[str, Any] = parser.parse_args() convert_gpta_checkpoint_to_pytorch(args.gpta_checkpoint_path, args.gpta_config_file, args.pytorch_dump_folder_path)
91
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ ={ 'configuration_xlm_roberta_xl': [ 'XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMRobertaXLConfig', 'XLMRobertaXLOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ =[ 'XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMRobertaXLForCausalLM', 'XLMRobertaXLForMaskedLM', 'XLMRobertaXLForMultipleChoice', 'XLMRobertaXLForQuestionAnswering', 'XLMRobertaXLForSequenceClassification', 'XLMRobertaXLForTokenClassification', 'XLMRobertaXLModel', 'XLMRobertaXLPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaXLConfig, XLMRobertaXLOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaXLForCausalLM, XLMRobertaXLForMaskedLM, XLMRobertaXLForMultipleChoice, XLMRobertaXLForQuestionAnswering, XLMRobertaXLForSequenceClassification, XLMRobertaXLForTokenClassification, XLMRobertaXLModel, XLMRobertaXLPreTrainedModel, ) else: import sys lowercase__ =_LazyModule(__name__, globals()['__file__'], _import_structure)
90
from math import pi, sqrt def __UpperCamelCase ( lowerCAmelCase__ : float ): if num <= 0: raise ValueError('''math domain error''' ) if num > 1_71.5: raise OverflowError('''math range error''' ) elif num - int(lowerCAmelCase__ ) not in (0, 0.5): raise NotImplementedError('''num must be an integer or a half-integer''' ) elif num == 0.5: return sqrt(lowerCAmelCase__ ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def __UpperCamelCase ( ): assert gamma(0.5 ) == sqrt(lowerCAmelCase__ ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() lowercase__ =1.0 while num: lowercase__ =float(input('Gamma of: ')) print(F"""gamma({num}) = {gamma(num)}""") print('\nEnter 0 to exit...')
90
1
import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class _a ( UpperCamelCase__): """simple docstring""" UpperCamelCase__ = (UnCLIPScheduler,) def UpperCAmelCase_ ( self: Tuple , **__lowerCamelCase: List[Any] ): '''simple docstring''' UpperCamelCase__: Optional[Any] = { "num_train_timesteps": 1000, "variance_type": "fixed_small_log", "clip_sample": True, "clip_sample_range": 1.0, "prediction_type": "epsilon", } config.update(**__lowerCamelCase ) return config def UpperCAmelCase_ ( self: Optional[Any] ): '''simple docstring''' for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=__lowerCamelCase ) def UpperCAmelCase_ ( self: Tuple ): '''simple docstring''' for variance in ["fixed_small_log", "learned_range"]: self.check_over_configs(variance_type=__lowerCamelCase ) def UpperCAmelCase_ ( self: List[Any] ): '''simple docstring''' for clip_sample in [True, False]: self.check_over_configs(clip_sample=__lowerCamelCase ) def UpperCAmelCase_ ( self: Tuple ): '''simple docstring''' for clip_sample_range in [1, 5, 10, 20]: self.check_over_configs(clip_sample_range=__lowerCamelCase ) def UpperCAmelCase_ ( self: Optional[Any] ): '''simple docstring''' for prediction_type in ["epsilon", "sample"]: self.check_over_configs(prediction_type=__lowerCamelCase ) def UpperCAmelCase_ ( self: Optional[Any] ): '''simple docstring''' for time_step in [0, 500, 999]: for prev_timestep in [None, 5, 100, 250, 500, 750]: if prev_timestep is not None and prev_timestep >= time_step: continue self.check_over_forward(time_step=__lowerCamelCase , prev_timestep=__lowerCamelCase ) def UpperCAmelCase_ ( self: Optional[int] ): '''simple docstring''' UpperCamelCase__: Optional[int] = self.scheduler_classes[0] UpperCamelCase__: Tuple = self.get_scheduler_config(variance_type="fixed_small_log" ) UpperCamelCase__: Union[str, Any] = scheduler_class(**__lowerCamelCase ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 1.00_00e-10 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487 ) - 0.0_549_625 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999 ) - 0.9_994_987 ) ) < 1e-5 def UpperCAmelCase_ ( self: int ): '''simple docstring''' UpperCamelCase__: Optional[Any] = self.scheduler_classes[0] UpperCamelCase__: str = self.get_scheduler_config(variance_type="learned_range" ) UpperCamelCase__: List[str] = scheduler_class(**__lowerCamelCase ) UpperCamelCase__: str = 0.5 assert scheduler._get_variance(1 , predicted_variance=__lowerCamelCase ) - -10.1_712_790 < 1e-5 assert scheduler._get_variance(487 , predicted_variance=__lowerCamelCase ) - -5.7_998_052 < 1e-5 assert scheduler._get_variance(999 , predicted_variance=__lowerCamelCase ) - -0.0_010_011 < 1e-5 def UpperCAmelCase_ ( self: Dict ): '''simple docstring''' UpperCamelCase__: Tuple = self.scheduler_classes[0] UpperCamelCase__: str = self.get_scheduler_config() UpperCamelCase__: Union[str, Any] = scheduler_class(**__lowerCamelCase ) UpperCamelCase__: List[Any] = scheduler.timesteps UpperCamelCase__: List[str] = self.dummy_model() UpperCamelCase__: Any = self.dummy_sample_deter UpperCamelCase__: Any = torch.manual_seed(0 ) for i, t in enumerate(__lowerCamelCase ): # 1. predict noise residual UpperCamelCase__: str = model(__lowerCamelCase , __lowerCamelCase ) # 2. predict previous mean of sample x_t-1 UpperCamelCase__: List[str] = scheduler.step(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , generator=__lowerCamelCase ).prev_sample UpperCamelCase__: str = pred_prev_sample UpperCamelCase__: Optional[Any] = torch.sum(torch.abs(__lowerCamelCase ) ) UpperCamelCase__: int = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 252.2_682_495 ) < 1e-2 assert abs(result_mean.item() - 0.3_284_743 ) < 1e-3 def UpperCAmelCase_ ( self: Tuple ): '''simple docstring''' UpperCamelCase__: Dict = self.scheduler_classes[0] UpperCamelCase__: Union[str, Any] = self.get_scheduler_config() UpperCamelCase__: Dict = scheduler_class(**__lowerCamelCase ) scheduler.set_timesteps(25 ) UpperCamelCase__: Dict = scheduler.timesteps UpperCamelCase__: Union[str, Any] = self.dummy_model() UpperCamelCase__: Optional[int] = self.dummy_sample_deter UpperCamelCase__: Dict = torch.manual_seed(0 ) for i, t in enumerate(__lowerCamelCase ): # 1. predict noise residual UpperCamelCase__: List[Any] = model(__lowerCamelCase , __lowerCamelCase ) if i + 1 == timesteps.shape[0]: UpperCamelCase__: List[str] = None else: UpperCamelCase__: Optional[int] = timesteps[i + 1] # 2. predict previous mean of sample x_t-1 UpperCamelCase__: Optional[Any] = scheduler.step( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , prev_timestep=__lowerCamelCase , generator=__lowerCamelCase ).prev_sample UpperCamelCase__: Union[str, Any] = pred_prev_sample UpperCamelCase__: Optional[int] = torch.sum(torch.abs(__lowerCamelCase ) ) UpperCamelCase__: List[str] = torch.mean(torch.abs(__lowerCamelCase ) ) assert abs(result_sum.item() - 258.2_044_983 ) < 1e-2 assert abs(result_mean.item() - 0.3_362_038 ) < 1e-3 def UpperCAmelCase_ ( self: Optional[int] ): '''simple docstring''' pass def UpperCAmelCase_ ( self: Union[str, Any] ): '''simple docstring''' pass
149
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available A__: str = {'''configuration_yolos''': ['''YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''YolosConfig''', '''YolosOnnxConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__: Union[str, Any] = ['''YolosFeatureExtractor'''] A__: Optional[int] = ['''YolosImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__: str = [ '''YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''YolosForObjectDetection''', '''YolosModel''', '''YolosPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys A__: Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
149
1
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class UpperCamelCase__( unittest.TestCase ): def __init__( self ,__UpperCAmelCase ,__UpperCAmelCase=7 ,__UpperCAmelCase=3 ,__UpperCAmelCase=18 ,__UpperCAmelCase=30 ,__UpperCAmelCase=4_00 ,__UpperCAmelCase=True ,__UpperCAmelCase=None ,__UpperCAmelCase=True ,__UpperCAmelCase=False ,__UpperCAmelCase=True ,__UpperCAmelCase=True ,__UpperCAmelCase=[0.5, 0.5, 0.5] ,__UpperCAmelCase=[0.5, 0.5, 0.5] ,) -> Any: A__ = parent A__ = batch_size A__ = num_channels A__ = image_size A__ = min_resolution A__ = max_resolution A__ = do_resize A__ = size if size is not None else {'height': 18, 'width': 20} A__ = do_thumbnail A__ = do_align_axis A__ = do_pad A__ = do_normalize A__ = image_mean A__ = image_std def snake_case__ ( self ) -> List[Any]: return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class UpperCamelCase__( __A , unittest.TestCase ): lowerCAmelCase__ : int = DonutImageProcessor if is_vision_available() else None def snake_case__ ( self ) -> int: A__ = DonutImageProcessingTester(self ) @property def snake_case__ ( self ) -> List[str]: return self.image_processor_tester.prepare_image_processor_dict() def snake_case__ ( self ) -> List[str]: A__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__UpperCAmelCase ,'do_resize' ) ) self.assertTrue(hasattr(__UpperCAmelCase ,'size' ) ) self.assertTrue(hasattr(__UpperCAmelCase ,'do_thumbnail' ) ) self.assertTrue(hasattr(__UpperCAmelCase ,'do_align_long_axis' ) ) self.assertTrue(hasattr(__UpperCAmelCase ,'do_pad' ) ) self.assertTrue(hasattr(__UpperCAmelCase ,'do_normalize' ) ) self.assertTrue(hasattr(__UpperCAmelCase ,'image_mean' ) ) self.assertTrue(hasattr(__UpperCAmelCase ,'image_std' ) ) def snake_case__ ( self ) -> Dict: A__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{'height': 18, 'width': 20} ) A__ = self.image_processing_class.from_dict(self.image_processor_dict ,size=42 ) self.assertEqual(image_processor.size ,{'height': 42, 'width': 42} ) # Previous config had dimensions in (width, height) order A__ = self.image_processing_class.from_dict(self.image_processor_dict ,size=(42, 84) ) self.assertEqual(image_processor.size ,{'height': 84, 'width': 42} ) def snake_case__ ( self ) -> Dict: pass @is_flaky() def snake_case__ ( self ) -> Dict: # Initialize image_processing A__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A__ = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase ,Image.Image ) # Test not batched input A__ = image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) ,) # Test batched A__ = image_processing(__UpperCAmelCase ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) ,) @is_flaky() def snake_case__ ( self ) -> Optional[Any]: # Initialize image_processing A__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A__ = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__UpperCAmelCase ,numpify=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase ,np.ndarray ) # Test not batched input A__ = image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) ,) # Test batched A__ = image_processing(__UpperCAmelCase ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) ,) @is_flaky() def snake_case__ ( self ) -> int: # Initialize image_processing A__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A__ = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__UpperCAmelCase ,torchify=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase ,torch.Tensor ) # Test not batched input A__ = image_processing(image_inputs[0] ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) ,) # Test batched A__ = image_processing(__UpperCAmelCase ,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) ,)
154
"""simple docstring""" import os import jsonlines import numpy as np from tqdm import tqdm __lowerCamelCase = 20_48 __lowerCamelCase = 40_96 __lowerCamelCase = 42 __lowerCamelCase = os.environ.pop("PROCESS_TRAIN", "false") __lowerCamelCase = {"null": 0, "short": 1, "long": 2, "yes": 3, "no": 4} def UpperCAmelCase ( UpperCamelCase__ ): """simple docstring""" def choose_first(UpperCamelCase__ , UpperCamelCase__=False ): assert isinstance(UpperCamelCase__ , UpperCamelCase__ ) if len(UpperCamelCase__ ) == 1: A__ = answer[0] return {k: [answer[k]] for k in answer} if is_long_answer else answer for a in answer: if is_long_answer: A__ = {k: [a[k]] for k in a} if len(a['start_token'] ) > 0: break return a A__ = {'id': example['id']} A__ = example['annotations'] A__ = annotation['yes_no_answer'] if 0 in yes_no_answer or 1 in yes_no_answer: A__ = ['yes'] if 1 in yes_no_answer else ['no'] A__ = A__ = [] A__ = A__ = [] A__ = ['<cls>'] else: A__ = ['short'] A__ = choose_first(annotation['short_answers'] ) if len(out['start_token'] ) == 0: # answer will be long if short is not available A__ = ['long'] A__ = choose_first(annotation['long_answer'] , is_long_answer=UpperCamelCase__ ) A__ = [] answer.update(UpperCamelCase__ ) # disregard some samples if len(answer['start_token'] ) > 1 or answer["start_token"] == answer["end_token"]: A__ = True else: A__ = False A__ = ['start_token', 'end_token', 'start_byte', 'end_byte', 'text'] if not all(isinstance(answer[k] , UpperCamelCase__ ) for k in cols ): raise ValueError('Issue in ID' , example['id'] ) return answer def UpperCAmelCase ( UpperCamelCase__ , UpperCamelCase__=False ): """simple docstring""" A__ = _get_single_answer(UpperCamelCase__ ) # bytes are of no use del answer["start_byte"] del answer["end_byte"] # handle yes_no answers explicitly if answer["category"][0] in ["yes", "no"]: # category is list with one element A__ = example['document']['tokens'] A__ = [] for i in range(len(doc['token'] ) ): if not doc["is_html"][i]: context.append(doc['token'][i] ) return { "context": " ".join(UpperCamelCase__ ), "answer": { "start_token": -100, # ignore index in cross-entropy "end_token": -100, # ignore index in cross-entropy "category": answer["category"], "span": answer["category"], # extra }, } # later, help in removing all no answers if answer["start_token"] == [-1]: return { "context": "None", "answer": { "start_token": -1, "end_token": -1, "category": "null", "span": "None", # extra }, } # handling normal samples A__ = ['start_token', 'end_token'] answer.update({k: answer[k][0] if len(answer[k] ) > 0 else answer[k] for k in cols} ) # e.g. [10] == 10 A__ = example['document']['tokens'] A__ = answer['start_token'] A__ = answer['end_token'] A__ = [] for i in range(len(doc['token'] ) ): if not doc["is_html"][i]: context.append(doc['token'][i] ) else: if answer["start_token"] > i: start_token -= 1 if answer["end_token"] > i: end_token -= 1 A__ = ' '.join(context[start_token:end_token] ) # checking above code if assertion: A__ = doc['is_html'][answer['start_token'] : answer['end_token']] A__ = doc['token'][answer['start_token'] : answer['end_token']] A__ = ' '.join([old[i] for i in range(len(UpperCamelCase__ ) ) if not is_html[i]] ) if new != old: print('ID:' , example['id'] ) print('New:' , UpperCamelCase__ , end='\n' ) print('Old:' , UpperCamelCase__ , end='\n\n' ) return { "context": " ".join(UpperCamelCase__ ), "answer": { "start_token": start_token, "end_token": end_token - 1, # this makes it inclusive "category": answer["category"], # either long or short "span": new, # extra }, } def UpperCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=2_048 , UpperCamelCase__=4_096 , UpperCamelCase__=True ): """simple docstring""" A__ = get_context_and_ans(UpperCamelCase__ , assertion=UpperCamelCase__ ) A__ = out['answer'] # later, removing these samples if answer["start_token"] == -1: return { "example_id": example["id"], "input_ids": [[-1]], "labels": { "start_token": [-1], "end_token": [-1], "category": ["null"], }, } A__ = tokenizer(example['question']['text'] , out['context'] ).input_ids A__ = input_ids.index(tokenizer.sep_token_id ) + 1 # return yes/no if answer["category"][0] in ["yes", "no"]: # category is list with one element A__ = [] A__ = [] A__ = input_ids[:q_len] A__ = range(UpperCamelCase__ , len(UpperCamelCase__ ) , max_length - doc_stride ) for i in doc_start_indices: A__ = i + max_length - q_len A__ = input_ids[i:end_index] inputs.append(q_indices + slice ) category.append(answer['category'][0] ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": [-100] * len(UpperCamelCase__ ), "end_token": [-100] * len(UpperCamelCase__ ), "category": category, }, } A__ = out['context'].split() A__ = splitted_context[answer['end_token']] A__ = len( tokenizer( ' '.join(splitted_context[: answer['start_token']] ) , add_special_tokens=UpperCamelCase__ , ).input_ids ) A__ = len( tokenizer(' '.join(splitted_context[: answer['end_token']] ) , add_special_tokens=UpperCamelCase__ ).input_ids ) answer["start_token"] += q_len answer["end_token"] += q_len # fixing end token A__ = len(tokenizer(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ).input_ids ) if num_sub_tokens > 1: answer["end_token"] += num_sub_tokens - 1 A__ = input_ids[answer['start_token'] : answer['end_token'] + 1] # right & left are inclusive A__ = answer['start_token'] A__ = answer['end_token'] if assertion: A__ = tokenizer.decode(UpperCamelCase__ ) if answer["span"] != new: print('ISSUE IN TOKENIZATION' ) print('OLD:' , answer['span'] ) print('NEW:' , UpperCamelCase__ , end='\n\n' ) if len(UpperCamelCase__ ) <= max_length: return { "example_id": example["id"], "input_ids": [input_ids], "labels": { "start_token": [answer["start_token"]], "end_token": [answer["end_token"]], "category": answer["category"], }, } A__ = input_ids[:q_len] A__ = range(UpperCamelCase__ , len(UpperCamelCase__ ) , max_length - doc_stride ) A__ = [] A__ = [] A__ = [] A__ = [] # null, yes, no, long, short for i in doc_start_indices: A__ = i + max_length - q_len A__ = input_ids[i:end_index] inputs.append(q_indices + slice ) assert len(inputs[-1] ) <= max_length, "Issue in truncating length" if start_token >= i and end_token <= end_index - 1: A__ = start_token - i + q_len A__ = end_token - i + q_len answers_category.append(answer['category'][0] ) # ["short"] -> "short" else: A__ = -100 A__ = -100 answers_category.append('null' ) A__ = inputs[-1][start_token : end_token + 1] answers_start_token.append(UpperCamelCase__ ) answers_end_token.append(UpperCamelCase__ ) if assertion: if new != old and new != [tokenizer.cls_token_id]: print('ISSUE in strided for ID:' , example['id'] ) print('New:' , tokenizer.decode(UpperCamelCase__ ) ) print('Old:' , tokenizer.decode(UpperCamelCase__ ) , end='\n\n' ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": answers_start_token, "end_token": answers_end_token, "category": answers_category, }, } def UpperCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=2_048 , UpperCamelCase__=4_096 , UpperCamelCase__=False ): """simple docstring""" A__ = get_strided_contexts_and_ans( UpperCamelCase__ , UpperCamelCase__ , doc_stride=UpperCamelCase__ , max_length=UpperCamelCase__ , assertion=UpperCamelCase__ , ) return example def UpperCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ): """simple docstring""" with jsonlines.open(UpperCamelCase__ , 'a' ) as writer: for example in tqdm(UpperCamelCase__ , total=len(UpperCamelCase__ ) , desc='Saving samples ... ' ): A__ = example['labels'] for ids, start, end, cat in zip( example['input_ids'] , labels['start_token'] , labels['end_token'] , labels['category'] , ): if start == -1 and end == -1: continue # leave waste samples with no answer if cat == "null" and np.random.rand() < 0.6: continue # removing 50 % samples writer.write( { 'input_ids': ids, 'start_token': start, 'end_token': end, 'category': CATEGORY_MAPPING[cat], } ) if __name__ == "__main__": from datasets import load_dataset from transformers import BigBirdTokenizer __lowerCamelCase = load_dataset("natural_questions") __lowerCamelCase = BigBirdTokenizer.from_pretrained("google/bigbird-roberta-base") __lowerCamelCase = data["train" if PROCESS_TRAIN == "true" else "validation"] __lowerCamelCase = { "tokenizer": tokenizer, "doc_stride": DOC_STRIDE, "max_length": MAX_LENGTH, "assertion": False, } __lowerCamelCase = data.map(prepare_inputs, fn_kwargs=fn_kwargs) __lowerCamelCase = data.remove_columns(["annotations", "document", "id", "question"]) print(data) np.random.seed(SEED) __lowerCamelCase = "nq-training.jsonl" if PROCESS_TRAIN == "true" else "nq-validation.jsonl" save_to_disk(data, file_name=cache_file_name)
154
1
'''simple docstring''' import argparse import collections import os import re import tempfile import pandas as pd from datasets import Dataset from huggingface_hub import hf_hub_download, upload_folder from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/update_metadata.py lowerCamelCase = """src/transformers""" # This is to make sure the transformers module imported is the one in the repo. lowerCamelCase = direct_transformers_import(TRANSFORMERS_PATH) # Regexes that match TF/Flax/PT model names. lowerCamelCase = re.compile(R"""TF(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") lowerCamelCase = re.compile(R"""Flax(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # Will match any TF or Flax model too so need to be in an else branch afterthe two previous regexes. lowerCamelCase = re.compile(R"""(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)""") # Fill this with tuples (pipeline_tag, model_mapping, auto_model) lowerCamelCase = [ ("""pretraining""", """MODEL_FOR_PRETRAINING_MAPPING_NAMES""", """AutoModelForPreTraining"""), ("""feature-extraction""", """MODEL_MAPPING_NAMES""", """AutoModel"""), ("""audio-classification""", """MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForAudioClassification"""), ("""text-generation""", """MODEL_FOR_CAUSAL_LM_MAPPING_NAMES""", """AutoModelForCausalLM"""), ("""automatic-speech-recognition""", """MODEL_FOR_CTC_MAPPING_NAMES""", """AutoModelForCTC"""), ("""image-classification""", """MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForImageClassification"""), ("""image-segmentation""", """MODEL_FOR_IMAGE_SEGMENTATION_MAPPING_NAMES""", """AutoModelForImageSegmentation"""), ("""fill-mask""", """MODEL_FOR_MASKED_LM_MAPPING_NAMES""", """AutoModelForMaskedLM"""), ("""object-detection""", """MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES""", """AutoModelForObjectDetection"""), ( """zero-shot-object-detection""", """MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING_NAMES""", """AutoModelForZeroShotObjectDetection""", ), ("""question-answering""", """MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES""", """AutoModelForQuestionAnswering"""), ("""text2text-generation""", """MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES""", """AutoModelForSeq2SeqLM"""), ("""text-classification""", """MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForSequenceClassification"""), ("""automatic-speech-recognition""", """MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES""", """AutoModelForSpeechSeq2Seq"""), ( """table-question-answering""", """MODEL_FOR_TABLE_QUESTION_ANSWERING_MAPPING_NAMES""", """AutoModelForTableQuestionAnswering""", ), ("""token-classification""", """MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForTokenClassification"""), ("""multiple-choice""", """MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES""", """AutoModelForMultipleChoice"""), ( """next-sentence-prediction""", """MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES""", """AutoModelForNextSentencePrediction""", ), ( """audio-frame-classification""", """MODEL_FOR_AUDIO_FRAME_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForAudioFrameClassification""", ), ("""audio-xvector""", """MODEL_FOR_AUDIO_XVECTOR_MAPPING_NAMES""", """AutoModelForAudioXVector"""), ( """document-question-answering""", """MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES""", """AutoModelForDocumentQuestionAnswering""", ), ( """visual-question-answering""", """MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING_NAMES""", """AutoModelForVisualQuestionAnswering""", ), ("""image-to-text""", """MODEL_FOR_FOR_VISION_2_SEQ_MAPPING_NAMES""", """AutoModelForVision2Seq"""), ( """zero-shot-image-classification""", """MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForZeroShotImageClassification""", ), ("""depth-estimation""", """MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES""", """AutoModelForDepthEstimation"""), ("""video-classification""", """MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING_NAMES""", """AutoModelForVideoClassification"""), ("""mask-generation""", """MODEL_FOR_MASK_GENERATION_MAPPING_NAMES""", """AutoModelForMaskGeneration"""), ] def _A ( _lowerCAmelCase ): """simple docstring""" __lowercase =re.finditer('.+?(?:(?<=[a-z])(?=[A-Z])|(?<=[A-Z])(?=[A-Z][a-z])|$)' , _lowerCAmelCase ) return [m.group(0 ) for m in matches] def _A ( ): """simple docstring""" __lowercase =transformers_module.models.auto.configuration_auto.CONFIG_MAPPING_NAMES __lowercase ={ config.replace('Config' , '' ): model_type for model_type, config in config_maping_names.items() } # Dictionaries flagging if each model prefix has a backend in PT/TF/Flax. __lowercase =collections.defaultdict(_lowerCAmelCase ) __lowercase =collections.defaultdict(_lowerCAmelCase ) __lowercase =collections.defaultdict(_lowerCAmelCase ) # Let's lookup through all transformers object (once) and find if models are supported by a given backend. for attr_name in dir(_lowerCAmelCase ): __lowercase =None if _re_tf_models.match(_lowerCAmelCase ) is not None: __lowercase =tf_models __lowercase =_re_tf_models.match(_lowerCAmelCase ).groups()[0] elif _re_flax_models.match(_lowerCAmelCase ) is not None: __lowercase =flax_models __lowercase =_re_flax_models.match(_lowerCAmelCase ).groups()[0] elif _re_pt_models.match(_lowerCAmelCase ) is not None: __lowercase =pt_models __lowercase =_re_pt_models.match(_lowerCAmelCase ).groups()[0] if lookup_dict is not None: while len(_lowerCAmelCase ) > 0: if attr_name in model_prefix_to_model_type: __lowercase =True break # Try again after removing the last word in the name __lowercase =''.join(camel_case_split(_lowerCAmelCase )[:-1] ) __lowercase =set(list(pt_models.keys() ) + list(tf_models.keys() ) + list(flax_models.keys() ) ) __lowercase =list(_lowerCAmelCase ) all_models.sort() __lowercase ={'model_type': all_models} __lowercase =[pt_models[t] for t in all_models] __lowercase =[tf_models[t] for t in all_models] __lowercase =[flax_models[t] for t in all_models] # Now let's use the auto-mapping names to make sure __lowercase ={} for t in all_models: if t in transformers_module.models.auto.processing_auto.PROCESSOR_MAPPING_NAMES: __lowercase ='AutoProcessor' elif t in transformers_module.models.auto.tokenization_auto.TOKENIZER_MAPPING_NAMES: __lowercase ='AutoTokenizer' elif t in transformers_module.models.auto.feature_extraction_auto.FEATURE_EXTRACTOR_MAPPING_NAMES: __lowercase ='AutoFeatureExtractor' else: # Default to AutoTokenizer if a model has nothing, for backward compatibility. __lowercase ='AutoTokenizer' __lowercase =[processors[t] for t in all_models] return pd.DataFrame(_lowerCAmelCase ) def _A ( _lowerCAmelCase ): """simple docstring""" __lowercase =[ transformers_module.models.auto.modeling_auto, transformers_module.models.auto.modeling_tf_auto, transformers_module.models.auto.modeling_flax_auto, ] for pipeline_tag, model_mapping, auto_class in PIPELINE_TAGS_AND_AUTO_MODELS: __lowercase =[model_mapping, f"""TF_{model_mapping}""", f"""FLAX_{model_mapping}"""] __lowercase =[auto_class, f"""TF_{auto_class}""", f"""Flax_{auto_class}"""] # Loop through all three frameworks for module, cls, mapping in zip(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): # The type of pipeline may not exist in this framework if not hasattr(_lowerCAmelCase , _lowerCAmelCase ): continue # First extract all model_names __lowercase =[] for name in getattr(_lowerCAmelCase , _lowerCAmelCase ).values(): if isinstance(_lowerCAmelCase , _lowerCAmelCase ): model_names.append(_lowerCAmelCase ) else: model_names.extend(list(_lowerCAmelCase ) ) # Add pipeline tag and auto model class for those models table.update({model_name: (pipeline_tag, cls) for model_name in model_names} ) return table def _A ( _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =get_frameworks_table() __lowercase =Dataset.from_pandas(_lowerCAmelCase ) __lowercase =hf_hub_download( 'huggingface/transformers-metadata' , 'pipeline_tags.json' , repo_type='dataset' , token=_lowerCAmelCase ) __lowercase =Dataset.from_json(_lowerCAmelCase ) __lowercase ={ tags_dataset[i]['model_class']: (tags_dataset[i]['pipeline_tag'], tags_dataset[i]['auto_class']) for i in range(len(_lowerCAmelCase ) ) } __lowercase =update_pipeline_and_auto_class_table(_lowerCAmelCase ) # Sort the model classes to avoid some nondeterministic updates to create false update commits. __lowercase =sorted(table.keys() ) __lowercase =pd.DataFrame( { 'model_class': model_classes, 'pipeline_tag': [table[m][0] for m in model_classes], 'auto_class': [table[m][1] for m in model_classes], } ) __lowercase =Dataset.from_pandas(_lowerCAmelCase ) with tempfile.TemporaryDirectory() as tmp_dir: frameworks_dataset.to_json(os.path.join(_lowerCAmelCase , 'frameworks.json' ) ) tags_dataset.to_json(os.path.join(_lowerCAmelCase , 'pipeline_tags.json' ) ) if commit_sha is not None: __lowercase =( f"""Update with commit {commit_sha}\n\nSee: """ f"""https://github.com/huggingface/transformers/commit/{commit_sha}""" ) else: __lowercase ='Update' upload_folder( repo_id='huggingface/transformers-metadata' , folder_path=_lowerCAmelCase , repo_type='dataset' , token=_lowerCAmelCase , commit_message=_lowerCAmelCase , ) def _A ( ): """simple docstring""" __lowercase ={tag: cls for tag, _, cls in PIPELINE_TAGS_AND_AUTO_MODELS} __lowercase =transformers_module.pipelines.SUPPORTED_TASKS __lowercase =[] for key in pipeline_tasks: if key not in in_table: __lowercase =pipeline_tasks[key]['pt'] if isinstance(_lowerCAmelCase , (list, tuple) ): __lowercase =model[0] __lowercase =model.__name__ if model not in in_table.values(): missing.append(_lowerCAmelCase ) if len(_lowerCAmelCase ) > 0: __lowercase =', '.join(_lowerCAmelCase ) raise ValueError( 'The following pipeline tags are not present in the `PIPELINE_TAGS_AND_AUTO_MODELS` constant inside ' f"""`utils/update_metadata.py`: {msg}. Please add them!""" ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() parser.add_argument("""--token""", type=str, help="""The token to use to push to the transformers-metadata dataset.""") parser.add_argument("""--commit_sha""", type=str, help="""The sha of the commit going with this update.""") parser.add_argument("""--check-only""", action="""store_true""", help="""Activate to just check all pipelines are present.""") lowerCamelCase = parser.parse_args() if args.check_only: check_pipeline_tags() else: update_metadata(args.token, args.commit_sha)
166
'''simple docstring''' from __future__ import annotations def _A ( _lowerCAmelCase ): """simple docstring""" __lowercase =[True] * limit __lowercase =False __lowercase =False __lowercase =True for i in range(3 , int(limit**0.5 + 1 ) , 2 ): __lowercase =i * 2 while index < limit: __lowercase =False __lowercase =index + i __lowercase =[2] for i in range(3 , _lowerCAmelCase , 2 ): if is_prime[i]: primes.append(_lowerCAmelCase ) return primes def _A ( _lowerCAmelCase = 1_000_000 ): """simple docstring""" __lowercase =prime_sieve(_lowerCAmelCase ) __lowercase =0 __lowercase =0 for i in range(len(_lowerCAmelCase ) ): for j in range(i + length , len(_lowerCAmelCase ) ): __lowercase =sum(primes[i:j] ) if sol >= ceiling: break if sol in primes: __lowercase =j - i __lowercase =sol return largest if __name__ == "__main__": print(f"{solution() = }")
166
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _snake_case = {'configuration_glpn': ['GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GLPNConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['GLPNFeatureExtractor'] _snake_case = ['GLPNImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ 'GLPN_PRETRAINED_MODEL_ARCHIVE_LIST', 'GLPNForDepthEstimation', 'GLPNLayer', 'GLPNModel', 'GLPNPreTrainedModel', ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
324
"""simple docstring""" _snake_case = 8.31_44_62 # Unit - J mol-1 K-1 def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' if moles < 0 or kelvin < 0 or volume < 0: raise ValueError("""Invalid inputs. Enter positive value.""" ) return moles * kelvin * UNIVERSAL_GAS_CONSTANT / volume def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' if moles < 0 or kelvin < 0 or pressure < 0: raise ValueError("""Invalid inputs. Enter positive value.""" ) return moles * kelvin * UNIVERSAL_GAS_CONSTANT / pressure if __name__ == "__main__": from doctest import testmod testmod()
324
1
from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class _a ( UpperCamelCase__ ): _lowercase : List[str] = '''new-model''' if is_tf_available(): class _a ( UpperCamelCase__ ): _lowercase : str = NewModelConfig @require_tf class _a ( unittest.TestCase ): @slow def lowerCamelCase_ ( self: Union[str, Any] ) -> Optional[Any]: """simple docstring""" lowercase__ = '''bert-base-cased''' lowercase__ = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) lowercase__ = TFAutoModel.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow def lowerCamelCase_ ( self: str ) -> Dict: """simple docstring""" lowercase__ = '''bert-base-cased''' lowercase__ = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) lowercase__ = TFAutoModelForPreTraining.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow def lowerCamelCase_ ( self: Optional[Any] ) -> Dict: """simple docstring""" for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) lowercase__ = TFAutoModelForCausalLM.from_pretrained(UpperCamelCase_ ) lowercase__ , lowercase__ = TFAutoModelForCausalLM.from_pretrained(UpperCamelCase_ , output_loading_info=UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow def lowerCamelCase_ ( self: Any ) -> List[Any]: """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) lowercase__ = TFAutoModelWithLMHead.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow def lowerCamelCase_ ( self: Optional[Any] ) -> Tuple: """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) lowercase__ = TFAutoModelForMaskedLM.from_pretrained(UpperCamelCase_ ) lowercase__ , lowercase__ = TFAutoModelForMaskedLM.from_pretrained(UpperCamelCase_ , output_loading_info=UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow def lowerCamelCase_ ( self: List[str] ) -> Optional[Any]: """simple docstring""" for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained(UpperCamelCase_ ) lowercase__ , lowercase__ = TFAutoModelForSeqaSeqLM.from_pretrained(UpperCamelCase_ , output_loading_info=UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow def lowerCamelCase_ ( self: List[Any] ) -> Tuple: """simple docstring""" for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) lowercase__ = TFAutoModelForSequenceClassification.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow def lowerCamelCase_ ( self: Optional[int] ) -> int: """simple docstring""" for model_name in ["bert-base-uncased"]: lowercase__ = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) lowercase__ = TFAutoModelForQuestionAnswering.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) @slow @require_tensorflow_probability def lowerCamelCase_ ( self: Tuple ) -> Any: """simple docstring""" for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: lowercase__ = AutoConfig.from_pretrained(UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) lowercase__ = TFAutoModelForTableQuestionAnswering.from_pretrained(UpperCamelCase_ ) lowercase__ , lowercase__ = TFAutoModelForTableQuestionAnswering.from_pretrained( UpperCamelCase_ , output_loading_info=UpperCamelCase_ ) self.assertIsNotNone(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) def lowerCamelCase_ ( self: Union[str, Any] ) -> str: """simple docstring""" lowercase__ = TFAutoModelWithLMHead.from_pretrained(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(model.num_parameters() , 14_410 ) self.assertEqual(model.num_parameters(only_trainable=UpperCamelCase_ ) , 14_410 ) def lowerCamelCase_ ( self: List[Any] ) -> Optional[Any]: """simple docstring""" lowercase__ = TFAutoModelWithLMHead.from_pretrained(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) self.assertEqual(model.num_parameters() , 14_410 ) self.assertEqual(model.num_parameters(only_trainable=UpperCamelCase_ ) , 14_410 ) def lowerCamelCase_ ( self: Optional[Any] ) -> Any: """simple docstring""" lowercase__ = TFAutoModel.from_pretrained('''sgugger/funnel-random-tiny''' ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) lowercase__ = copy.deepcopy(model.config ) lowercase__ = ['''FunnelBaseModel'''] lowercase__ = TFAutoModel.from_config(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(UpperCamelCase_ ) lowercase__ = TFAutoModel.from_pretrained(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) def lowerCamelCase_ ( self: Tuple ) -> Any: """simple docstring""" try: AutoConfig.register('''new-model''' , UpperCamelCase_ ) lowercase__ = [ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__ ): # Wrong config class will raise an error with self.assertRaises(UpperCamelCase_ ): auto_class.register(UpperCamelCase_ , UpperCamelCase_ ) auto_class.register(UpperCamelCase_ , UpperCamelCase_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(UpperCamelCase_ ): auto_class.register(UpperCamelCase_ , UpperCamelCase_ ) # Now that the config is registered, it can be used as any other config with the auto-API lowercase__ = BertModelTester(self ).get_config() lowercase__ = NewModelConfig(**tiny_config.to_dict() ) lowercase__ = auto_class.from_config(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(UpperCamelCase_ ) lowercase__ = auto_class.from_pretrained(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def lowerCamelCase_ ( self: Dict ) -> Optional[Any]: """simple docstring""" with self.assertRaisesRegex( UpperCamelCase_ , '''bert-base is not a local folder and is not a valid model identifier''' ): lowercase__ = TFAutoModel.from_pretrained('''bert-base''' ) def lowerCamelCase_ ( self: Optional[int] ) -> Any: """simple docstring""" with self.assertRaisesRegex( UpperCamelCase_ , r'''aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)''' ): lowercase__ = TFAutoModel.from_pretrained(UpperCamelCase_ , revision='''aaaaaa''' ) def lowerCamelCase_ ( self: Optional[Any] ) -> Dict: """simple docstring""" with self.assertRaisesRegex( UpperCamelCase_ , '''hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin''' , ): lowercase__ = TFAutoModel.from_pretrained('''hf-internal-testing/config-no-model''' ) def lowerCamelCase_ ( self: Any ) -> Optional[int]: """simple docstring""" with self.assertRaisesRegex(UpperCamelCase_ , '''Use `from_pt=True` to load this model''' ): lowercase__ = TFAutoModel.from_pretrained('''hf-internal-testing/tiny-bert-pt-only''' ) def lowerCamelCase_ ( self: str ) -> Union[str, Any]: """simple docstring""" lowercase__ = TFAutoModel.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) with RequestCounter() as counter: lowercase__ = TFAutoModel.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 ) # With a sharded checkpoint lowercase__ = TFAutoModel.from_pretrained('''ArthurZ/tiny-random-bert-sharded''' ) with RequestCounter() as counter: lowercase__ = TFAutoModel.from_pretrained('''ArthurZ/tiny-random-bert-sharded''' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
110
def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" return credit_card_number.startswith(('''34''', '''35''', '''37''', '''4''', '''5''', '''6''') ) def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = credit_card_number lowercase__ = 0 lowercase__ = len(SCREAMING_SNAKE_CASE ) - 2 for i in range(SCREAMING_SNAKE_CASE , -1 , -2 ): # double the value of every second digit lowercase__ = int(cc_number[i] ) digit *= 2 # If doubling of a number results in a two digit number # i.e greater than 9(e.g., 6 × 2 = 12), # then add the digits of the product (e.g., 12: 1 + 2 = 3, 15: 1 + 5 = 6), # to get a single digit number. if digit > 9: digit %= 10 digit += 1 lowercase__ = cc_number[:i] + str(SCREAMING_SNAKE_CASE ) + cc_number[i + 1 :] total += digit # Sum up the remaining digits for i in range(len(SCREAMING_SNAKE_CASE ) - 1 , -1 , -2 ): total += int(cc_number[i] ) return total % 10 == 0 def _a ( SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = f'{credit_card_number} is an invalid credit card number because' if not credit_card_number.isdigit(): print(f'{error_message} it has nonnumerical characters.' ) return False if not 13 <= len(SCREAMING_SNAKE_CASE ) <= 16: print(f'{error_message} of its length.' ) return False if not validate_initial_digits(SCREAMING_SNAKE_CASE ): print(f'{error_message} of its first two digits.' ) return False if not luhn_validation(SCREAMING_SNAKE_CASE ): print(f'{error_message} it fails the Luhn check.' ) return False print(f'{credit_card_number} is a valid credit card number.' ) return True if __name__ == "__main__": import doctest doctest.testmod() validate_credit_card_number('4111111111111111') validate_credit_card_number('32323')
110
1
'''simple docstring''' import sys from pathlib import Path UpperCamelCase__ : Optional[Any] = Path(__file__).resolve().parents[3] / 'src' sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa import io # noqa import itertools # noqa import json # noqa import os # noqa import unittest # noqa from copy import deepcopy # noqa from parameterized import parameterized # noqa from transformers import TrainingArguments, is_torch_available # noqa from transformers.deepspeed import is_deepspeed_available # noqa from transformers.file_utils import WEIGHTS_NAME # noqa from transformers.testing_utils import ( # noqa CaptureLogger, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, mockenv_context, require_deepspeed, require_torch_gpu, require_torch_multi_gpu, slow, ) from transformers.trainer_utils import set_seed # noqa set_seed(42) UpperCamelCase__ : Tuple = {'base': 'patrickvonplaten/wav2vec2_tiny_random', 'robust': 'patrickvonplaten/wav2vec2_tiny_random_robust'} UpperCamelCase__ : Optional[Any] = 'zero2' UpperCamelCase__ : Optional[int] = 'zero3' UpperCamelCase__ : Dict = [ZEROa, ZEROa] def UpperCAmelCase ( a_ , a_ , a_ ) -> int: """simple docstring""" A_ : int = parameterized.to_safe_name("""_""".join(str(a_ ) for x in param.args ) ) return F"{func.__name__}_{param_based_name}" # Cartesian-product of zero stages with models to test UpperCamelCase__ : Tuple = list(itertools.product(stages, models.keys())) @slow @require_deepspeed @require_torch_gpu class _lowerCAmelCase ( __A ): """simple docstring""" @parameterized.expand(_lowerCamelCase , name_func=_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> Tuple: self.run_and_check( stage=_lowerCamelCase , model=_lowerCamelCase , distributed=_lowerCamelCase , fpaa=_lowerCamelCase , ) @require_torch_multi_gpu @parameterized.expand(_lowerCamelCase , name_func=_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> Optional[int]: self.run_and_check( stage=_lowerCamelCase , model=_lowerCamelCase , distributed=_lowerCamelCase , fpaa=_lowerCamelCase , ) @parameterized.expand(_lowerCamelCase , name_func=_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> Dict: self.run_and_check( stage=_lowerCamelCase , model=_lowerCamelCase , distributed=_lowerCamelCase , fpaa=_lowerCamelCase , ) @require_torch_multi_gpu @parameterized.expand(_lowerCamelCase , name_func=_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> int: self.run_and_check( stage=_lowerCamelCase , model=_lowerCamelCase , distributed=_lowerCamelCase , fpaa=_lowerCamelCase , ) def UpperCAmelCase_ ( self , _lowerCamelCase ) -> Optional[Any]: # XXX: run_asr is premature and doesn't save any results # so all we check for now is that the process didn't fail pass def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = 10 , _lowerCamelCase = True , _lowerCamelCase = True , _lowerCamelCase = True , ) -> List[str]: A_ : Union[str, Any] = models[model] A_ : Tuple = self.run_trainer( stage=_lowerCamelCase , model_name=_lowerCamelCase , eval_steps=_lowerCamelCase , num_train_epochs=1 , distributed=_lowerCamelCase , fpaa=_lowerCamelCase , ) self.do_checks(_lowerCamelCase ) return output_dir def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = 10 , _lowerCamelCase = 1 , _lowerCamelCase = True , _lowerCamelCase = True , ) -> Any: A_ : Dict = self.get_auto_remove_tmp_dir("""./xxx""" , after=_lowerCamelCase ) A_ : str = F"\n --model_name_or_path {model_name}\n --dataset_name hf-internal-testing/librispeech_asr_dummy\n --dataset_config_name clean\n --train_split_name validation\n --validation_split_name validation\n --output_dir {output_dir}\n --num_train_epochs {str(_lowerCamelCase )}\n --per_device_train_batch_size 2\n --per_device_eval_batch_size 2\n --evaluation_strategy steps\n --learning_rate 5e-4\n --warmup_steps 8\n --orthography timit\n --preprocessing_num_workers 1\n --group_by_length\n --freeze_feature_extractor\n --report_to none\n --save_steps 0\n --eval_steps {eval_steps}\n --report_to none\n ".split() if fpaa: args.extend(["""--fp16"""] ) # currently ds_config_wav2vec2_zero.json requires "zero_optimization.find_unused_parameters": true, # hence the separate config files A_ : List[str] = F"--deepspeed {self.test_file_dir_str}/ds_config_wav2vec2_{stage}.json".split() A_ : Union[str, Any] = [F"{self.examples_dir_str}/research_projects/wav2vec2/run_asr.py"] A_ : Tuple = self.get_launcher(_lowerCamelCase ) A_ : Optional[int] = launcher + script + args + ds_args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(_lowerCamelCase , env=self.get_env() ) return output_dir def UpperCAmelCase_ ( self , _lowerCamelCase=False ) -> Any: # 1. explicitly set --num_nodes=1 just in case these tests end up run on a multi-node setup # - it won't be able to handle that # 2. for now testing with just 2 gpus max (since some quality tests may give different # results with mode gpus because we use very little data) A_ : int = min(2 , get_gpu_count() ) if distributed else 1 return F"deepspeed --num_nodes 1 --num_gpus {num_gpus}".split()
357
'''simple docstring''' import numpy as np import pandas as pd from sklearn.preprocessing import Normalizer from sklearn.svm import SVR from statsmodels.tsa.statespace.sarimax import SARIMAX def UpperCAmelCase ( a_ , a_ , a_ , a_ , a_ ) -> float: """simple docstring""" A_ : Optional[Any] = np.array([[1, item, train_mtch[i]] for i, item in enumerate(a_ )] ) A_ : Optional[Any] = np.array(a_ ) A_ : Optional[int] = np.dot(np.dot(np.linalg.inv(np.dot(x.transpose() , a_ ) ) , x.transpose() ) , a_ ) return abs(beta[0] + test_dt[0] * beta[1] + test_mtch[0] + beta[2] ) def UpperCAmelCase ( a_ , a_ , a_ ) -> float: """simple docstring""" A_ : List[str] = (1, 2, 1) A_ : Tuple = (1, 1, 0, 7) A_ : List[Any] = SARIMAX( a_ , exog=a_ , order=a_ , seasonal_order=a_ ) A_ : Tuple = model.fit(disp=a_ , maxiter=6_0_0 , method="""nm""" ) A_ : List[Any] = model_fit.predict(1 , len(a_ ) , exog=[test_match] ) return result[0] def UpperCAmelCase ( a_ , a_ , a_ ) -> float: """simple docstring""" A_ : int = SVR(kernel="""rbf""" , C=1 , gamma=0.1 , epsilon=0.1 ) regressor.fit(a_ , a_ ) A_ : Tuple = regressor.predict(a_ ) return y_pred[0] def UpperCAmelCase ( a_ ) -> float: """simple docstring""" train_user.sort() A_ : Any = np.percentile(a_ , 2_5 ) A_ : Union[str, Any] = np.percentile(a_ , 7_5 ) A_ : str = qa - qa A_ : List[Any] = qa - (iqr * 0.1) return low_lim def UpperCAmelCase ( a_ , a_ ) -> bool: """simple docstring""" A_ : Dict = 0 A_ : Optional[Any] = 0 for i in list_vote: if i > actual_result: A_ : Optional[Any] = not_safe + 1 else: if abs(abs(a_ ) - abs(a_ ) ) <= 0.1: safe += 1 else: not_safe += 1 return safe > not_safe if __name__ == "__main__": # data_input_df = pd.read_csv("ex_data.csv", header=None) UpperCamelCase__ : List[str] = [[18_231, 0.0, 1], [22_621, 1.0, 2], [15_675, 0.0, 3], [23_583, 1.0, 4]] UpperCamelCase__ : Optional[Any] = pd.DataFrame( data_input, columns=['total_user', 'total_even', 'days'] ) UpperCamelCase__ : Union[str, Any] = Normalizer().fit_transform(data_input_df.values) # split data UpperCamelCase__ : List[Any] = normalize_df[:, 2].tolist() UpperCamelCase__ : Tuple = normalize_df[:, 0].tolist() UpperCamelCase__ : Union[str, Any] = normalize_df[:, 1].tolist() # for svr (input variable = total date and total match) UpperCamelCase__ : Any = normalize_df[:, [1, 2]].tolist() UpperCamelCase__ : Optional[int] = x[: len(x) - 1] UpperCamelCase__ : Optional[Any] = x[len(x) - 1 :] # for linear regression & sarimax UpperCamelCase__ : Optional[int] = total_date[: len(total_date) - 1] UpperCamelCase__ : str = total_user[: len(total_user) - 1] UpperCamelCase__ : Tuple = total_match[: len(total_match) - 1] UpperCamelCase__ : List[str] = total_date[len(total_date) - 1 :] UpperCamelCase__ : List[Any] = total_user[len(total_user) - 1 :] UpperCamelCase__ : Dict = total_match[len(total_match) - 1 :] # voting system with forecasting UpperCamelCase__ : List[Any] = [ linear_regression_prediction( trn_date, trn_user, trn_match, tst_date, tst_match ), sarimax_predictor(trn_user, trn_match, tst_match), support_vector_regressor(x_train, x_test, trn_user), ] # check the safety of today's data UpperCamelCase__ : Tuple = '' if data_safety_checker(res_vote, tst_user) else 'not ' print('Today\'s data is {not_str}safe.')
164
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : Optional[Any] = {"""vocab_file""": """vocab.txt"""} _UpperCAmelCase : Union[str, Any] = { """vocab_file""": { """YituTech/conv-bert-base""": """https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt""", """YituTech/conv-bert-medium-small""": ( """https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt""" ), """YituTech/conv-bert-small""": """https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt""", } } _UpperCAmelCase : Dict = { """YituTech/conv-bert-base""": 5_12, """YituTech/conv-bert-medium-small""": 5_12, """YituTech/conv-bert-small""": 5_12, } _UpperCAmelCase : List[str] = { """YituTech/conv-bert-base""": {"""do_lower_case""": True}, """YituTech/conv-bert-medium-small""": {"""do_lower_case""": True}, """YituTech/conv-bert-small""": {"""do_lower_case""": True}, } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = ConvBertTokenizer def __init__( self : Optional[Any] , UpperCAmelCase : int=None , UpperCAmelCase : Dict=None , UpperCAmelCase : Optional[Any]=True , UpperCAmelCase : Tuple="[UNK]" , UpperCAmelCase : Optional[int]="[SEP]" , UpperCAmelCase : List[Any]="[PAD]" , UpperCAmelCase : List[Any]="[CLS]" , UpperCAmelCase : Union[str, Any]="[MASK]" , UpperCAmelCase : Any=True , UpperCAmelCase : int=None , **UpperCAmelCase : Union[str, Any] , ) -> List[str]: super().__init__( UpperCAmelCase , tokenizer_file=UpperCAmelCase , do_lower_case=UpperCAmelCase , unk_token=UpperCAmelCase , sep_token=UpperCAmelCase , pad_token=UpperCAmelCase , cls_token=UpperCAmelCase , mask_token=UpperCAmelCase , tokenize_chinese_chars=UpperCAmelCase , strip_accents=UpperCAmelCase , **UpperCAmelCase , ) lowerCamelCase__ : int = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , UpperCAmelCase ) != do_lower_case or normalizer_state.get('strip_accents' , UpperCAmelCase ) != strip_accents or normalizer_state.get('handle_chinese_chars' , UpperCAmelCase ) != tokenize_chinese_chars ): lowerCamelCase__ : Union[str, Any] = getattr(UpperCAmelCase , normalizer_state.pop('type' ) ) lowerCamelCase__ : Dict = do_lower_case lowerCamelCase__ : Dict = strip_accents lowerCamelCase__ : Union[str, Any] = tokenize_chinese_chars lowerCamelCase__ : Tuple = normalizer_class(**UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = do_lower_case def A_ ( self : Union[str, Any] , UpperCAmelCase : str , UpperCAmelCase : Any=None ) -> Dict: lowerCamelCase__ : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def A_ ( self : Union[str, Any] , UpperCAmelCase : List[int] , UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: lowerCamelCase__ : List[Any] = [self.sep_token_id] lowerCamelCase__ : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A_ ( self : List[str] , UpperCAmelCase : str , UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: lowerCamelCase__ : List[Any] = self._tokenizer.model.save(UpperCAmelCase , name=UpperCAmelCase ) return tuple(UpperCAmelCase )
50
'''simple docstring''' import os __snake_case ={"""I""": 1, """V""": 5, """X""": 10, """L""": 50, """C""": 100, """D""": 500, """M""": 1_000} def a_ ( lowerCamelCase : str ): lowerCAmelCase = 0 lowerCAmelCase = 0 while index < len(lowerCamelCase ) - 1: lowerCAmelCase = SYMBOLS[numerals[index]] lowerCAmelCase = SYMBOLS[numerals[index + 1]] if current_value < next_value: total_value -= current_value else: total_value += current_value index += 1 total_value += SYMBOLS[numerals[index]] return total_value def a_ ( lowerCamelCase : int ): lowerCAmelCase = '' lowerCAmelCase = num // 1000 numerals += m_count * "M" num %= 1000 lowerCAmelCase = num // 100 if c_count == 9: numerals += "CM" c_count -= 9 elif c_count == 4: numerals += "CD" c_count -= 4 if c_count >= 5: numerals += "D" c_count -= 5 numerals += c_count * "C" num %= 100 lowerCAmelCase = num // 10 if x_count == 9: numerals += "XC" x_count -= 9 elif x_count == 4: numerals += "XL" x_count -= 4 if x_count >= 5: numerals += "L" x_count -= 5 numerals += x_count * "X" num %= 10 if num == 9: numerals += "IX" num -= 9 elif num == 4: numerals += "IV" num -= 4 if num >= 5: numerals += "V" num -= 5 numerals += num * "I" return numerals def a_ ( lowerCamelCase : str = "/p089_roman.txt" ): lowerCAmelCase = 0 with open(os.path.dirname(lowerCamelCase ) + roman_numerals_filename ) as filea: lowerCAmelCase = filea.readlines() for line in lines: lowerCAmelCase = line.strip() lowerCAmelCase = parse_roman_numerals(lowerCamelCase ) lowerCAmelCase = generate_roman_numerals(lowerCamelCase ) savings += len(lowerCamelCase ) - len(lowerCamelCase ) return savings if __name__ == "__main__": print(F'''{solution() = }''')
4
0
"""simple docstring""" from __future__ import annotations from collections import deque class a_ : '''simple docstring''' def __init__(self, lowerCamelCase_ ): '''simple docstring''' lowerCamelCase__ : list[dict] = [] self.adlist.append( {'value': '', 'next_states': [], 'fail_state': 0, 'output': []} ) for keyword in keywords: self.add_keyword(lowerCamelCase_ ) self.set_fail_transitions() def a__ (self, lowerCamelCase_, lowerCamelCase_ ): '''simple docstring''' for state in self.adlist[current_state]["next_states"]: if char == self.adlist[state]["value"]: return state return None def a__ (self, lowerCamelCase_ ): '''simple docstring''' lowerCamelCase__ : Union[str, Any] = 0 for character in keyword: lowerCamelCase__ : int = self.find_next_state(lowerCamelCase_, lowerCamelCase_ ) if next_state is None: self.adlist.append( { 'value': character, 'next_states': [], 'fail_state': 0, 'output': [], } ) self.adlist[current_state]["next_states"].append(len(self.adlist ) - 1 ) lowerCamelCase__ : Tuple = len(self.adlist ) - 1 else: lowerCamelCase__ : Optional[int] = next_state self.adlist[current_state]["output"].append(lowerCamelCase_ ) def a__ (self ): '''simple docstring''' lowerCamelCase__ : deque = deque() for node in self.adlist[0]["next_states"]: q.append(lowerCamelCase_ ) lowerCamelCase__ : str = 0 while q: lowerCamelCase__ : Optional[Any] = q.popleft() for child in self.adlist[r]["next_states"]: q.append(lowerCamelCase_ ) lowerCamelCase__ : Dict = self.adlist[r]['fail_state'] while ( self.find_next_state(lowerCamelCase_, self.adlist[child]['value'] ) is None and state != 0 ): lowerCamelCase__ : Optional[int] = self.adlist[state]['fail_state'] lowerCamelCase__ : Optional[int] = self.find_next_state( lowerCamelCase_, self.adlist[child]['value'] ) if self.adlist[child]["fail_state"] is None: lowerCamelCase__ : Optional[int] = 0 lowerCamelCase__ : int = ( self.adlist[child]['output'] + self.adlist[self.adlist[child]['fail_state']]['output'] ) def a__ (self, lowerCamelCase_ ): '''simple docstring''' lowerCamelCase__ : dict = {} # returns a dict with keywords and list of its occurrences lowerCamelCase__ : Tuple = 0 for i in range(len(lowerCamelCase_ ) ): while ( self.find_next_state(lowerCamelCase_, string[i] ) is None and current_state != 0 ): lowerCamelCase__ : Dict = self.adlist[current_state]['fail_state'] lowerCamelCase__ : Union[str, Any] = self.find_next_state(lowerCamelCase_, string[i] ) if next_state is None: lowerCamelCase__ : Union[str, Any] = 0 else: lowerCamelCase__ : Dict = next_state for key in self.adlist[current_state]["output"]: if key not in result: lowerCamelCase__ : List[Any] = [] result[key].append(i - len(lowerCamelCase_ ) + 1 ) return result if __name__ == "__main__": import doctest doctest.testmod()
316
"""simple docstring""" import numpy as np def lowerCamelCase_ ( _lowerCamelCase ): return (2 / (1 + np.exp(-2 * vector ))) - 1 if __name__ == "__main__": import doctest doctest.testmod()
316
1
from __future__ import annotations def a_ ( _A ) -> bool: """simple docstring""" snake_case__ = len(_A ) # We need to create solution object to save path. snake_case__ = [[0 for _ in range(_A )] for _ in range(_A )] snake_case__ = run_maze(_A , 0 , 0 , _A ) if solved: print('\n'.join(str(_A ) for row in solutions ) ) else: print('No solution exists!' ) return solved def a_ ( _A , _A , _A , _A ) -> bool: """simple docstring""" snake_case__ = len(_A ) # Final check point. if i == j == (size - 1): snake_case__ = 1 return True snake_case__ = (not i < 0) and (not j < 0) # Check lower bounds snake_case__ = (i < size) and (j < size) # Check upper bounds if lower_flag and upper_flag: # check for already visited and block points. snake_case__ = (not solutions[i][j]) and (not maze[i][j]) if block_flag: # check visited snake_case__ = 1 # check for directions if ( run_maze(_A , i + 1 , _A , _A ) or run_maze(_A , _A , j + 1 , _A ) or run_maze(_A , i - 1 , _A , _A ) or run_maze(_A , _A , j - 1 , _A ) ): return True snake_case__ = 0 return False return False if __name__ == "__main__": import doctest doctest.testmod()
307
from typing import TYPE_CHECKING from ...utils import _LazyModule __UpperCamelCase : Any = {"""tokenization_byt5""": ["""ByT5Tokenizer"""]} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys __UpperCamelCase : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
307
1
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase : Union[str, Any] = logging.get_logger(__name__) lowerCAmelCase : Any = { """microsoft/trocr-base-handwritten""": ( """https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json""" ), # See all TrOCR models at https://huggingface.co/models?filter=trocr } class __lowercase ( UpperCAmelCase_ ): """simple docstring""" _UpperCAmelCase : Any = '''trocr''' _UpperCAmelCase : Optional[Any] = ['''past_key_values'''] _UpperCAmelCase : List[Any] = { '''num_attention_heads''': '''decoder_attention_heads''', '''hidden_size''': '''d_model''', '''num_hidden_layers''': '''decoder_layers''', } def __init__( self : Any , lowerCAmelCase__ : Dict=5_0265 , lowerCAmelCase__ : Union[str, Any]=1024 , lowerCAmelCase__ : Union[str, Any]=12 , lowerCAmelCase__ : List[str]=16 , lowerCAmelCase__ : Optional[Any]=4096 , lowerCAmelCase__ : Any="gelu" , lowerCAmelCase__ : Optional[int]=512 , lowerCAmelCase__ : int=0.1 , lowerCAmelCase__ : str=0.0 , lowerCAmelCase__ : str=0.0 , lowerCAmelCase__ : List[Any]=2 , lowerCAmelCase__ : List[Any]=0.02 , lowerCAmelCase__ : List[Any]=0.0 , lowerCAmelCase__ : Optional[int]=True , lowerCAmelCase__ : Tuple=False , lowerCAmelCase__ : str=True , lowerCAmelCase__ : Optional[int]=True , lowerCAmelCase__ : Any=1 , lowerCAmelCase__ : int=0 , lowerCAmelCase__ : Union[str, Any]=2 , **lowerCAmelCase__ : List[str] , ): SCREAMING_SNAKE_CASE_: Union[str, Any] = vocab_size SCREAMING_SNAKE_CASE_: int = d_model SCREAMING_SNAKE_CASE_: str = decoder_layers SCREAMING_SNAKE_CASE_: List[str] = decoder_attention_heads SCREAMING_SNAKE_CASE_: List[Any] = decoder_ffn_dim SCREAMING_SNAKE_CASE_: Tuple = activation_function SCREAMING_SNAKE_CASE_: Dict = max_position_embeddings SCREAMING_SNAKE_CASE_: Optional[Any] = dropout SCREAMING_SNAKE_CASE_: List[str] = attention_dropout SCREAMING_SNAKE_CASE_: Union[str, Any] = activation_dropout SCREAMING_SNAKE_CASE_: int = init_std SCREAMING_SNAKE_CASE_: str = decoder_layerdrop SCREAMING_SNAKE_CASE_: int = use_cache SCREAMING_SNAKE_CASE_: Dict = scale_embedding SCREAMING_SNAKE_CASE_: List[Any] = use_learned_position_embeddings SCREAMING_SNAKE_CASE_: Optional[int] = layernorm_embedding super().__init__( pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , decoder_start_token_id=lowerCAmelCase__ , **lowerCAmelCase__ , )
127
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file lowerCAmelCase : List[str] = """Run commands across TPU VMs for initial setup before running `accelerate launch`.""" def A_ ( _UpperCAmelCase=None ): if subparsers is not None: SCREAMING_SNAKE_CASE_: Optional[Any] = subparsers.add_parser("tpu-config" , description=_description ) else: SCREAMING_SNAKE_CASE_: Optional[int] = argparse.ArgumentParser("Accelerate tpu-config command" , description=_description ) # Core arguments SCREAMING_SNAKE_CASE_: Any = parser.add_argument_group( "Config Arguments" , "Arguments that can be configured through `accelerate config`." ) config_args.add_argument( "--config_file" , type=_UpperCAmelCase , default=_UpperCAmelCase , help="Path to the config file to use for accelerate." , ) config_args.add_argument( "--tpu_name" , default=_UpperCAmelCase , help="The name of the TPU to use. If not specified, will use the TPU specified in the config file." , ) config_args.add_argument( "--tpu_zone" , default=_UpperCAmelCase , help="The zone of the TPU to use. If not specified, will use the zone specified in the config file." , ) SCREAMING_SNAKE_CASE_: Tuple = parser.add_argument_group("TPU Arguments" , "Arguments for options ran inside the TPU." ) pod_args.add_argument( "--use_alpha" , action="store_true" , help="Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`." , ) pod_args.add_argument( "--command_file" , default=_UpperCAmelCase , help="The path to the file containing the commands to run on the pod on startup." , ) pod_args.add_argument( "--command" , action="append" , nargs="+" , help="A command to run on the pod. Can be passed multiple times." , ) pod_args.add_argument( "--install_accelerate" , action="store_true" , help="Whether to install accelerate on the pod. Defaults to False." , ) pod_args.add_argument( "--accelerate_version" , default="latest" , help="The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify 'dev' to install from GitHub." , ) pod_args.add_argument( "--debug" , action="store_true" , help="If set, will print the command that would be run instead of running it." ) if subparsers is not None: parser.set_defaults(func=_UpperCAmelCase ) return parser def A_ ( _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Tuple = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(_UpperCAmelCase ): SCREAMING_SNAKE_CASE_: List[Any] = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: SCREAMING_SNAKE_CASE_: Any = defaults.command_file if not args.command and defaults.commands is not None: SCREAMING_SNAKE_CASE_: int = defaults.commands if not args.tpu_name: SCREAMING_SNAKE_CASE_: List[str] = defaults.tpu_name if not args.tpu_zone: SCREAMING_SNAKE_CASE_: Dict = defaults.tpu_zone if args.accelerate_version == "dev": SCREAMING_SNAKE_CASE_: Tuple = "git+https://github.com/huggingface/accelerate.git" elif args.accelerate_version == "latest": SCREAMING_SNAKE_CASE_: str = "accelerate -U" elif isinstance(parse(args.accelerate_version ) , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: List[str] = f"accelerate=={args.accelerate_version}" if not args.command_file and not args.command: raise ValueError("You must specify either a command file or a command to run on the pod." ) if args.command_file: with open(args.command_file , "r" ) as f: SCREAMING_SNAKE_CASE_: Dict = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: List[str] = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate SCREAMING_SNAKE_CASE_: Dict = ["cd /usr/share"] if args.install_accelerate: new_cmd += [f"pip install {args.accelerate_version}"] new_cmd += args.command SCREAMING_SNAKE_CASE_: List[str] = "; ".join(_UpperCAmelCase ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess SCREAMING_SNAKE_CASE_: int = ["gcloud"] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(f"Running {' '.join(_UpperCAmelCase )}" ) return subprocess.run(_UpperCAmelCase ) print("Successfully setup pod." ) def A_ ( ): SCREAMING_SNAKE_CASE_: List[str] = tpu_command_parser() SCREAMING_SNAKE_CASE_: List[Any] = parser.parse_args() tpu_command_launcher(_UpperCAmelCase )
127
1
'''simple docstring''' import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def snake_case_ ( __SCREAMING_SNAKE_CASE : bytes , __SCREAMING_SNAKE_CASE : int ): """simple docstring""" lowercase_ : Any = F'''{sampling_rate}''' lowercase_ : Union[str, Any] = '''1''' lowercase_ : Optional[Any] = '''f32le''' lowercase_ : Optional[Any] = [ '''ffmpeg''', '''-i''', '''pipe:0''', '''-ac''', ac, '''-ar''', ar, '''-f''', format_for_conversion, '''-hide_banner''', '''-loglevel''', '''quiet''', '''pipe:1''', ] try: with subprocess.Popen(__SCREAMING_SNAKE_CASE , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: lowercase_ : List[str] = ffmpeg_process.communicate(__SCREAMING_SNAKE_CASE ) except FileNotFoundError as error: raise ValueError('''ffmpeg was not found but is required to load audio files from filename''' ) from error lowercase_ : int = output_stream[0] lowercase_ : Dict = np.frombuffer(__SCREAMING_SNAKE_CASE , np.floataa ) if audio.shape[0] == 0: raise ValueError('''Malformed soundfile''' ) return audio def snake_case_ ( __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : float , __SCREAMING_SNAKE_CASE : str = "f32le" , ): """simple docstring""" lowercase_ : str = F'''{sampling_rate}''' lowercase_ : Optional[Any] = '''1''' if format_for_conversion == "s16le": lowercase_ : Any = 2 elif format_for_conversion == "f32le": lowercase_ : Optional[int] = 4 else: raise ValueError(F'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) lowercase_ : List[Any] = platform.system() if system == "Linux": lowercase_ : Optional[int] = '''alsa''' lowercase_ : Dict = '''default''' elif system == "Darwin": lowercase_ : int = '''avfoundation''' lowercase_ : str = ''':0''' elif system == "Windows": lowercase_ : int = '''dshow''' lowercase_ : Optional[Any] = '''default''' lowercase_ : Dict = [ '''ffmpeg''', '''-f''', format_, '''-i''', input_, '''-ac''', ac, '''-ar''', ar, '''-f''', format_for_conversion, '''-fflags''', '''nobuffer''', '''-hide_banner''', '''-loglevel''', '''quiet''', '''pipe:1''', ] lowercase_ : Any = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample lowercase_ : int = _ffmpeg_stream(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) for item in iterator: yield item def snake_case_ ( __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : float , __SCREAMING_SNAKE_CASE : Optional[int] = None , __SCREAMING_SNAKE_CASE : Optional[Union[Tuple[float, float], float]] = None , __SCREAMING_SNAKE_CASE : str = "f32le" , ): """simple docstring""" if stream_chunk_s is not None: lowercase_ : List[str] = stream_chunk_s else: lowercase_ : Optional[int] = chunk_length_s lowercase_ : List[Any] = ffmpeg_microphone(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , format_for_conversion=__SCREAMING_SNAKE_CASE ) if format_for_conversion == "s16le": lowercase_ : Tuple = np.intaa lowercase_ : Dict = 2 elif format_for_conversion == "f32le": lowercase_ : List[Any] = np.floataa lowercase_ : List[Any] = 4 else: raise ValueError(F'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: lowercase_ : List[str] = chunk_length_s / 6 lowercase_ : Any = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(__SCREAMING_SNAKE_CASE , (int, float) ): lowercase_ : Tuple = [stride_length_s, stride_length_s] lowercase_ : str = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample lowercase_ : int = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample lowercase_ : Tuple = datetime.datetime.now() lowercase_ : Dict = datetime.timedelta(seconds=__SCREAMING_SNAKE_CASE ) for item in chunk_bytes_iter(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , stride=(stride_left, stride_right) , stream=__SCREAMING_SNAKE_CASE ): # Put everything back in numpy scale lowercase_ : int = np.frombuffer(item['''raw'''] , dtype=__SCREAMING_SNAKE_CASE ) lowercase_ : Union[str, Any] = ( item['''stride'''][0] // size_of_sample, item['''stride'''][1] // size_of_sample, ) lowercase_ : Optional[Any] = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def snake_case_ ( __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : Tuple[int, int] , __SCREAMING_SNAKE_CASE : bool = False ): """simple docstring""" lowercase_ : Optional[int] = B'''''' lowercase_ , lowercase_ : Dict = stride if stride_left + stride_right >= chunk_len: raise ValueError( F'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) lowercase_ : Any = 0 for raw in iterator: acc += raw if stream and len(__SCREAMING_SNAKE_CASE ) < chunk_len: lowercase_ : List[str] = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(__SCREAMING_SNAKE_CASE ) >= chunk_len: # We are flushing the accumulator lowercase_ : Tuple = (_stride_left, stride_right) lowercase_ : List[str] = {'''raw''': acc[:chunk_len], '''stride''': stride} if stream: lowercase_ : int = False yield item lowercase_ : Optional[Any] = stride_left lowercase_ : Union[str, Any] = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(__SCREAMING_SNAKE_CASE ) > stride_left: lowercase_ : Union[str, Any] = {'''raw''': acc, '''stride''': (_stride_left, 0)} if stream: lowercase_ : Union[str, Any] = False yield item def snake_case_ ( __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : int ): """simple docstring""" lowercase_ : str = 2**24 # 16Mo try: with subprocess.Popen(__SCREAMING_SNAKE_CASE , stdout=subprocess.PIPE , bufsize=__SCREAMING_SNAKE_CASE ) as ffmpeg_process: while True: lowercase_ : Any = ffmpeg_process.stdout.read(__SCREAMING_SNAKE_CASE ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('''ffmpeg was not found but is required to stream audio files from filename''' ) from error
93
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters _lowercase : Any = (7_2_0, 1_2_8_0) # Height, Width _lowercase : List[Any] = (0.4, 0.6) # if height or width lower than this scale, drop it. _lowercase : str = 1 / 1_0_0 _lowercase : Any = "" _lowercase : Union[str, Any] = "" _lowercase : Optional[int] = "" _lowercase : List[Any] = 2_5_0 def snake_case_ ( ): """simple docstring""" lowercase_ , lowercase_ : Any = get_dataset(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) for index in range(__SCREAMING_SNAKE_CASE ): lowercase_ : str = random.sample(range(len(__SCREAMING_SNAKE_CASE ) ) , 4 ) lowercase_ , lowercase_ , lowercase_ : Any = update_image_and_anno( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , filter_scale=__SCREAMING_SNAKE_CASE , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' lowercase_ : int = random_chars(32 ) lowercase_ : str = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] lowercase_ : int = F'''{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}''' cva.imwrite(F'''{file_root}.jpg''' , __SCREAMING_SNAKE_CASE , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F'''Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}''' ) lowercase_ : List[Any] = [] for anno in new_annos: lowercase_ : List[Any] = anno[3] - anno[1] lowercase_ : List[str] = anno[4] - anno[2] lowercase_ : Dict = anno[1] + width / 2 lowercase_ : Dict = anno[2] + height / 2 lowercase_ : int = F'''{anno[0]} {x_center} {y_center} {width} {height}''' annos_list.append(__SCREAMING_SNAKE_CASE ) with open(F'''{file_root}.txt''' , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def snake_case_ ( __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : str ): """simple docstring""" lowercase_ : Optional[Any] = [] lowercase_ : Optional[Any] = [] for label_file in glob.glob(os.path.join(__SCREAMING_SNAKE_CASE , '''*.txt''' ) ): lowercase_ : int = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(__SCREAMING_SNAKE_CASE ) as in_file: lowercase_ : List[str] = in_file.readlines() lowercase_ : Optional[Any] = os.path.join(__SCREAMING_SNAKE_CASE , F'''{label_name}.jpg''' ) lowercase_ : Optional[int] = [] for obj_list in obj_lists: lowercase_ : List[str] = obj_list.rstrip('''\n''' ).split(''' ''' ) lowercase_ : Optional[int] = float(obj[1] ) - float(obj[3] ) / 2 lowercase_ : Any = float(obj[2] ) - float(obj[4] ) / 2 lowercase_ : str = float(obj[1] ) + float(obj[3] ) / 2 lowercase_ : List[str] = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(__SCREAMING_SNAKE_CASE ) labels.append(__SCREAMING_SNAKE_CASE ) return img_paths, labels def snake_case_ ( __SCREAMING_SNAKE_CASE : list , __SCREAMING_SNAKE_CASE : list , __SCREAMING_SNAKE_CASE : list[int] , __SCREAMING_SNAKE_CASE : tuple[int, int] , __SCREAMING_SNAKE_CASE : tuple[float, float] , __SCREAMING_SNAKE_CASE : float = 0.0 , ): """simple docstring""" lowercase_ : List[Any] = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) lowercase_ : Tuple = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) lowercase_ : List[Any] = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) lowercase_ : Optional[int] = int(scale_x * output_size[1] ) lowercase_ : Dict = int(scale_y * output_size[0] ) lowercase_ : Union[str, Any] = [] lowercase_ : List[Any] = [] for i, index in enumerate(__SCREAMING_SNAKE_CASE ): lowercase_ : Union[str, Any] = all_img_list[index] path_list.append(__SCREAMING_SNAKE_CASE ) lowercase_ : int = all_annos[index] lowercase_ : Dict = cva.imread(__SCREAMING_SNAKE_CASE ) if i == 0: # top-left lowercase_ : Optional[Any] = cva.resize(__SCREAMING_SNAKE_CASE , (divid_point_x, divid_point_y) ) lowercase_ : Tuple = img for bbox in img_annos: lowercase_ : Optional[int] = bbox[1] * scale_x lowercase_ : Optional[Any] = bbox[2] * scale_y lowercase_ : str = bbox[3] * scale_x lowercase_ : Tuple = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right lowercase_ : Dict = cva.resize(__SCREAMING_SNAKE_CASE , (output_size[1] - divid_point_x, divid_point_y) ) lowercase_ : Dict = img for bbox in img_annos: lowercase_ : int = scale_x + bbox[1] * (1 - scale_x) lowercase_ : Dict = bbox[2] * scale_y lowercase_ : Optional[int] = scale_x + bbox[3] * (1 - scale_x) lowercase_ : int = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left lowercase_ : List[Any] = cva.resize(__SCREAMING_SNAKE_CASE , (divid_point_x, output_size[0] - divid_point_y) ) lowercase_ : List[str] = img for bbox in img_annos: lowercase_ : Any = bbox[1] * scale_x lowercase_ : Optional[int] = scale_y + bbox[2] * (1 - scale_y) lowercase_ : str = bbox[3] * scale_x lowercase_ : Optional[int] = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right lowercase_ : int = cva.resize( __SCREAMING_SNAKE_CASE , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) lowercase_ : List[str] = img for bbox in img_annos: lowercase_ : int = scale_x + bbox[1] * (1 - scale_x) lowercase_ : Any = scale_y + bbox[2] * (1 - scale_y) lowercase_ : Optional[Any] = scale_x + bbox[3] * (1 - scale_x) lowercase_ : int = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: lowercase_ : Optional[Any] = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def snake_case_ ( __SCREAMING_SNAKE_CASE : int ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" lowercase_ : Any = ascii_lowercase + digits return "".join(random.choice(__SCREAMING_SNAKE_CASE ) for _ in range(__SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": main() print("DONE ✅")
93
1
import argparse import logging import os import sys import numpy as np import onnxruntime import torch from bart_onnx.generation_onnx import BARTBeamSearchGenerator from bart_onnx.reduce_onnx_size import remove_dup_initializers import transformers from transformers import BartForConditionalGeneration, BartTokenizer logging.basicConfig( format="%(asctime)s | %(levelname)s | %(name)s | [%(filename)s:%(lineno)d] %(message)s", datefmt="%Y-%m-%d %H:%M:%S", level=os.environ.get("LOGLEVEL", "INFO").upper(), stream=sys.stdout, ) _snake_case : Dict = logging.getLogger(__name__) _snake_case : List[Any] = {"facebook/bart-base": BartForConditionalGeneration} _snake_case : List[str] = {"facebook/bart-base": BartTokenizer} def lowerCAmelCase_ ( ): __snake_case : Tuple = argparse.ArgumentParser(description="Export Bart model + Beam Search to ONNX graph." ) parser.add_argument( "--validation_file" , type=__lowerCamelCase , default=__lowerCamelCase , help="A csv or a json file containing the validation data." ) parser.add_argument( "--max_length" , type=__lowerCamelCase , default=5 , help="The maximum total input sequence length after tokenization." , ) parser.add_argument( "--num_beams" , type=__lowerCamelCase , default=__lowerCamelCase , help=( "Number of beams to use for evaluation. This argument will be " "passed to ``model.generate``, which is used during ``evaluate`` and ``predict``." ) , ) parser.add_argument( "--model_name_or_path" , type=__lowerCamelCase , help="Path to pretrained model or model identifier from huggingface.co/models." , required=__lowerCamelCase , ) parser.add_argument( "--config_name" , type=__lowerCamelCase , default=__lowerCamelCase , help="Pretrained config name or path if not the same as model_name" , ) parser.add_argument( "--device" , type=__lowerCamelCase , default="cpu" , help="Device where the model will be run" , ) parser.add_argument("--output_file_path" , type=__lowerCamelCase , default=__lowerCamelCase , help="Where to store the final ONNX file." ) __snake_case : Optional[int] = parser.parse_args() return args def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase="cpu" ): __snake_case : str = model_dict[model_name].from_pretrained(__lowerCamelCase ).to(__lowerCamelCase ) __snake_case : int = tokenizer_dict[model_name].from_pretrained(__lowerCamelCase ) if model_name in ["facebook/bart-base"]: __snake_case : Any = 0 __snake_case : Any = None __snake_case : List[Any] = 0 return huggingface_model, tokenizer def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): model.eval() __snake_case : Dict = None __snake_case : Dict = torch.jit.script(BARTBeamSearchGenerator(__lowerCamelCase ) ) with torch.no_grad(): __snake_case : Any = "My friends are cool but they eat too many carbs." __snake_case : Optional[Any] = tokenizer([ARTICLE_TO_SUMMARIZE] , max_length=1_0_2_4 , return_tensors="pt" ).to(model.device ) __snake_case : Optional[int] = model.generate( inputs["input_ids"] , attention_mask=inputs["attention_mask"] , num_beams=__lowerCamelCase , max_length=__lowerCamelCase , early_stopping=__lowerCamelCase , decoder_start_token_id=model.config.decoder_start_token_id , ) torch.onnx.export( __lowerCamelCase , ( inputs["input_ids"], inputs["attention_mask"], num_beams, max_length, model.config.decoder_start_token_id, ) , __lowerCamelCase , opset_version=1_4 , input_names=["input_ids", "attention_mask", "num_beams", "max_length", "decoder_start_token_id"] , output_names=["output_ids"] , dynamic_axes={ "input_ids": {0: "batch", 1: "seq"}, "output_ids": {0: "batch", 1: "seq_out"}, } , example_outputs=__lowerCamelCase , ) logger.info("Model exported to {}".format(__lowerCamelCase ) ) __snake_case : Union[str, Any] = remove_dup_initializers(os.path.abspath(__lowerCamelCase ) ) logger.info("Deduplicated and optimized model written to {}".format(__lowerCamelCase ) ) __snake_case : Dict = onnxruntime.InferenceSession(__lowerCamelCase ) __snake_case : Any = ort_sess.run( __lowerCamelCase , { "input_ids": inputs["input_ids"].cpu().numpy(), "attention_mask": inputs["attention_mask"].cpu().numpy(), "num_beams": np.array(__lowerCamelCase ), "max_length": np.array(__lowerCamelCase ), "decoder_start_token_id": np.array(model.config.decoder_start_token_id ), } , ) np.testing.assert_allclose(summary_ids.cpu().numpy() , ort_out[0] , rtol=1e-3 , atol=1e-3 ) logger.info("Model outputs from torch and ONNX Runtime are similar." ) logger.info("Success." ) def lowerCAmelCase_ ( ): __snake_case : Union[str, Any] = parse_args() __snake_case : List[str] = 5 __snake_case : Optional[int] = 4 # Make one log on every process with the configuration for debugging. logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , level=logging.INFO , ) logger.setLevel(logging.INFO ) transformers.utils.logging.set_verbosity_error() __snake_case : Any = torch.device(args.device ) __snake_case , __snake_case : str = load_model_tokenizer(args.model_name_or_path , __lowerCamelCase ) if model.config.decoder_start_token_id is None: raise ValueError("Make sure that `config.decoder_start_token_id` is correctly defined" ) model.to(__lowerCamelCase ) if args.max_length: __snake_case : int = args.max_length if args.num_beams: __snake_case : Tuple = args.num_beams if args.output_file_path: __snake_case : List[Any] = args.output_file_path else: __snake_case : Union[str, Any] = "BART.onnx" logger.info("Exporting model to ONNX" ) export_and_validate_model(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": main()
134
import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case : Union[str, Any] = logging.get_logger(__name__) _snake_case : List[str] = { "facebook/encodec_24khz": "https://huggingface.co/facebook/encodec_24khz/resolve/main/config.json", "facebook/encodec_48khz": "https://huggingface.co/facebook/encodec_48khz/resolve/main/config.json", } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Tuple = "encodec" def __init__( self : Any , lowerCamelCase : Optional[int]=[1.5, 3.0, 6.0, 12.0, 24.0] , lowerCamelCase : List[str]=24000 , lowerCamelCase : int=1 , lowerCamelCase : Optional[int]=False , lowerCamelCase : Dict=None , lowerCamelCase : Tuple=None , lowerCamelCase : Optional[int]=128 , lowerCamelCase : Optional[int]=32 , lowerCamelCase : List[str]=1 , lowerCamelCase : str=[8, 5, 4, 2] , lowerCamelCase : List[str]="weight_norm" , lowerCamelCase : Any=7 , lowerCamelCase : Tuple=7 , lowerCamelCase : int=3 , lowerCamelCase : int=2 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : List[Any]="reflect" , lowerCamelCase : Union[str, Any]=2 , lowerCamelCase : Optional[int]=2 , lowerCamelCase : int=1.0 , lowerCamelCase : Optional[Any]=1024 , lowerCamelCase : Optional[Any]=None , lowerCamelCase : str=True , **lowerCamelCase : Dict , ) -> Any: __snake_case : Tuple = target_bandwidths __snake_case : Union[str, Any] = sampling_rate __snake_case : Union[str, Any] = audio_channels __snake_case : Dict = normalize __snake_case : List[Any] = chunk_length_s __snake_case : Tuple = overlap __snake_case : Optional[int] = hidden_size __snake_case : List[Any] = num_filters __snake_case : Union[str, Any] = num_residual_layers __snake_case : Optional[int] = upsampling_ratios __snake_case : List[str] = norm_type __snake_case : Optional[int] = kernel_size __snake_case : Dict = last_kernel_size __snake_case : Tuple = residual_kernel_size __snake_case : List[Any] = dilation_growth_rate __snake_case : Optional[int] = use_causal_conv __snake_case : Tuple = pad_mode __snake_case : Union[str, Any] = compress __snake_case : Union[str, Any] = num_lstm_layers __snake_case : int = trim_right_ratio __snake_case : Tuple = codebook_size __snake_case : Optional[Any] = codebook_dim if codebook_dim is not None else hidden_size __snake_case : int = use_conv_shortcut if self.norm_type not in ["weight_norm", "time_group_norm"]: raise ValueError( F'self.norm_type must be one of `"weight_norm"`, `"time_group_norm"`), got {self.norm_type}' ) super().__init__(**lowerCamelCase ) @property def __snake_case ( self : int ) -> Optional[int]: if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def __snake_case ( self : Union[str, Any] ) -> Optional[int]: if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) @property def __snake_case ( self : Optional[Any] ) -> int: __snake_case : Union[str, Any] = np.prod(self.upsampling_ratios ) return math.ceil(self.sampling_rate / hop_length ) @property def __snake_case ( self : Optional[Any] ) -> int: return int(1000 * self.target_bandwidths[-1] // (self.frame_rate * 10) )
134
1
'''simple docstring''' import os import pytest from transformers.dynamic_module_utils import get_imports SCREAMING_SNAKE_CASE_: Tuple ='\nimport os\n' SCREAMING_SNAKE_CASE_: Any ='\ndef foo():\n import os\n return False\n' SCREAMING_SNAKE_CASE_: List[str] ='\ndef foo():\n def bar():\n if True:\n import os\n return False\n return bar()\n' SCREAMING_SNAKE_CASE_: str ='\nimport os\n\ntry:\n import bar\nexcept ImportError:\n raise ValueError()\n' SCREAMING_SNAKE_CASE_: Tuple ='\nimport os\n\ndef foo():\n try:\n import bar\n except ImportError:\n raise ValueError()\n' SCREAMING_SNAKE_CASE_: List[Any] ='\nimport os\n\ntry:\n import bar\nexcept (ImportError, AttributeError):\n raise ValueError()\n' SCREAMING_SNAKE_CASE_: str ='\nimport os\n\ntry:\n import bar\nexcept ImportError as e:\n raise ValueError()\n' SCREAMING_SNAKE_CASE_: Any ='\nimport os\n\ntry:\n import bar\nexcept:\n raise ValueError()\n' SCREAMING_SNAKE_CASE_: Any ='\nimport os\n\ntry:\n import bar\n import baz\nexcept ImportError:\n raise ValueError()\n' SCREAMING_SNAKE_CASE_: List[Any] ='\nimport os\n\ntry:\n import bar\n import baz\nexcept ImportError:\n x = 1\n raise ValueError()\n' SCREAMING_SNAKE_CASE_: Optional[int] =[ TOP_LEVEL_IMPORT, IMPORT_IN_FUNCTION, DEEPLY_NESTED_IMPORT, TOP_LEVEL_TRY_IMPORT, GENERIC_EXCEPT_IMPORT, MULTILINE_TRY_IMPORT, MULTILINE_BOTH_IMPORT, MULTIPLE_EXCEPTS_IMPORT, EXCEPT_AS_IMPORT, TRY_IMPORT_IN_FUNCTION, ] @pytest.mark.parametrize("case" , __lowercase ) def lowerCAmelCase_ ( snake_case_ : Any , snake_case_ : List[Any] ) -> Dict: '''simple docstring''' UpperCAmelCase_ = os.path.join(__lowercase , "test_file.py" ) with open(__lowercase , "w" ) as _tmp_file: _tmp_file.write(__lowercase ) UpperCAmelCase_ = get_imports(__lowercase ) assert parsed_imports == ["os"]
1
'''simple docstring''' import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt'''} UpperCamelCase = { '''vocab_file''': { '''allenai/longformer-base-4096''': '''https://huggingface.co/allenai/longformer-base-4096/resolve/main/vocab.json''', '''allenai/longformer-large-4096''': ( '''https://huggingface.co/allenai/longformer-large-4096/resolve/main/vocab.json''' ), '''allenai/longformer-large-4096-finetuned-triviaqa''': ( '''https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/vocab.json''' ), '''allenai/longformer-base-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/vocab.json''' ), '''allenai/longformer-large-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/vocab.json''' ), }, '''merges_file''': { '''allenai/longformer-base-4096''': '''https://huggingface.co/allenai/longformer-base-4096/resolve/main/merges.txt''', '''allenai/longformer-large-4096''': ( '''https://huggingface.co/allenai/longformer-large-4096/resolve/main/merges.txt''' ), '''allenai/longformer-large-4096-finetuned-triviaqa''': ( '''https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/merges.txt''' ), '''allenai/longformer-base-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/merges.txt''' ), '''allenai/longformer-large-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/merges.txt''' ), }, } UpperCamelCase = { '''allenai/longformer-base-4096''': 4096, '''allenai/longformer-large-4096''': 4096, '''allenai/longformer-large-4096-finetuned-triviaqa''': 4096, '''allenai/longformer-base-4096-extra.pos.embd.only''': 4096, '''allenai/longformer-large-4096-extra.pos.embd.only''': 4096, } @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def SCREAMING_SNAKE_CASE( ) -> Dict: A: Dict = ( list(range(ord('''!''' ) , ord('''~''' ) + 1 ) ) + list(range(ord('''¡''' ) , ord('''¬''' ) + 1 ) ) + list(range(ord('''®''' ) , ord('''ÿ''' ) + 1 ) ) ) A: Union[str, Any] = bs[:] A: List[str] = 0 for b in range(2**8 ): if b not in bs: bs.append(__lowercase ) cs.append(2**8 + n ) n += 1 A: List[Any] = [chr(__lowercase ) for n in cs] return dict(zip(__lowercase , __lowercase ) ) def SCREAMING_SNAKE_CASE( __lowercase ) -> Optional[int]: A: Optional[Any] = set() A: Tuple = word[0] for char in word[1:]: pairs.add((prev_char, char) ) A: List[Any] = char return pairs class lowerCAmelCase_ ( UpperCAmelCase_ ): '''simple docstring''' UpperCamelCase_ : int = VOCAB_FILES_NAMES UpperCamelCase_ : int = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ : int = ["""input_ids""", """attention_mask"""] def __init__( self : int , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : str="replace" , SCREAMING_SNAKE_CASE_ : str="<s>" , SCREAMING_SNAKE_CASE_ : Any="</s>" , SCREAMING_SNAKE_CASE_ : int="</s>" , SCREAMING_SNAKE_CASE_ : List[Any]="<s>" , SCREAMING_SNAKE_CASE_ : str="<unk>" , SCREAMING_SNAKE_CASE_ : Dict="<pad>" , SCREAMING_SNAKE_CASE_ : Dict="<mask>" , SCREAMING_SNAKE_CASE_ : Union[str, Any]=False , **SCREAMING_SNAKE_CASE_ : Tuple , ) -> List[str]: '''simple docstring''' A: int = AddedToken(SCREAMING_SNAKE_CASE_ , lstrip=SCREAMING_SNAKE_CASE_ , rstrip=SCREAMING_SNAKE_CASE_ ) if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) else bos_token A: Dict = AddedToken(SCREAMING_SNAKE_CASE_ , lstrip=SCREAMING_SNAKE_CASE_ , rstrip=SCREAMING_SNAKE_CASE_ ) if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) else eos_token A: int = AddedToken(SCREAMING_SNAKE_CASE_ , lstrip=SCREAMING_SNAKE_CASE_ , rstrip=SCREAMING_SNAKE_CASE_ ) if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) else sep_token A: Dict = AddedToken(SCREAMING_SNAKE_CASE_ , lstrip=SCREAMING_SNAKE_CASE_ , rstrip=SCREAMING_SNAKE_CASE_ ) if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) else cls_token A: Any = AddedToken(SCREAMING_SNAKE_CASE_ , lstrip=SCREAMING_SNAKE_CASE_ , rstrip=SCREAMING_SNAKE_CASE_ ) if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) else unk_token A: str = AddedToken(SCREAMING_SNAKE_CASE_ , lstrip=SCREAMING_SNAKE_CASE_ , rstrip=SCREAMING_SNAKE_CASE_ ) if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it A: Dict = AddedToken(SCREAMING_SNAKE_CASE_ , lstrip=SCREAMING_SNAKE_CASE_ , rstrip=SCREAMING_SNAKE_CASE_ ) if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) else mask_token super().__init__( errors=SCREAMING_SNAKE_CASE_ , bos_token=SCREAMING_SNAKE_CASE_ , eos_token=SCREAMING_SNAKE_CASE_ , unk_token=SCREAMING_SNAKE_CASE_ , sep_token=SCREAMING_SNAKE_CASE_ , cls_token=SCREAMING_SNAKE_CASE_ , pad_token=SCREAMING_SNAKE_CASE_ , mask_token=SCREAMING_SNAKE_CASE_ , add_prefix_space=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ , ) with open(SCREAMING_SNAKE_CASE_ , encoding='''utf-8''' ) as vocab_handle: A: str = json.load(SCREAMING_SNAKE_CASE_ ) A: str = {v: k for k, v in self.encoder.items()} A: Union[str, Any] = errors # how to handle errors in decoding A: Optional[int] = bytes_to_unicode() A: Union[str, Any] = {v: k for k, v in self.byte_encoder.items()} with open(SCREAMING_SNAKE_CASE_ , encoding='''utf-8''' ) as merges_handle: A: int = merges_handle.read().split('''\n''' )[1:-1] A: str = [tuple(merge.split() ) for merge in bpe_merges] A: Any = dict(zip(SCREAMING_SNAKE_CASE_ , range(len(SCREAMING_SNAKE_CASE_ ) ) ) ) A: Union[str, Any] = {} A: Tuple = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions A: Dict = re.compile(R'''\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+''' ) @property def _snake_case ( self : int ) -> List[Any]: '''simple docstring''' return len(self.encoder ) def _snake_case ( self : Optional[Any] ) -> int: '''simple docstring''' return dict(self.encoder , **self.added_tokens_encoder ) def _snake_case ( self : str , SCREAMING_SNAKE_CASE_ : Optional[int] ) -> Optional[Any]: '''simple docstring''' if token in self.cache: return self.cache[token] A: str = tuple(SCREAMING_SNAKE_CASE_ ) A: str = get_pairs(SCREAMING_SNAKE_CASE_ ) if not pairs: return token while True: A: Dict = min(SCREAMING_SNAKE_CASE_ , key=lambda SCREAMING_SNAKE_CASE_ : self.bpe_ranks.get(SCREAMING_SNAKE_CASE_ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break A , A: Optional[Any] = bigram A: Tuple = [] A: List[Any] = 0 while i < len(SCREAMING_SNAKE_CASE_ ): try: A: Union[str, Any] = word.index(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) A: int = j if word[i] == first and i < len(SCREAMING_SNAKE_CASE_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 A: Optional[Any] = tuple(SCREAMING_SNAKE_CASE_ ) A: Any = new_word if len(SCREAMING_SNAKE_CASE_ ) == 1: break else: A: Union[str, Any] = get_pairs(SCREAMING_SNAKE_CASE_ ) A: str = ''' '''.join(SCREAMING_SNAKE_CASE_ ) A: str = word return word def _snake_case ( self : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> Optional[int]: '''simple docstring''' A: Dict = [] for token in re.findall(self.pat , SCREAMING_SNAKE_CASE_ ): A: Tuple = ''''''.join( self.byte_encoder[b] for b in token.encode('''utf-8''' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(SCREAMING_SNAKE_CASE_ ).split(''' ''' ) ) return bpe_tokens def _snake_case ( self : List[Any] , SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> Optional[Any]: '''simple docstring''' return self.encoder.get(SCREAMING_SNAKE_CASE_ , self.encoder.get(self.unk_token ) ) def _snake_case ( self : Tuple , SCREAMING_SNAKE_CASE_ : Optional[Any] ) -> str: '''simple docstring''' return self.decoder.get(SCREAMING_SNAKE_CASE_ ) def _snake_case ( self : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Optional[int] ) -> Tuple: '''simple docstring''' A: Optional[int] = ''''''.join(SCREAMING_SNAKE_CASE_ ) A: Tuple = bytearray([self.byte_decoder[c] for c in text] ).decode('''utf-8''' , errors=self.errors ) return text def _snake_case ( self : int , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(SCREAMING_SNAKE_CASE_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return A: Union[str, Any] = os.path.join( SCREAMING_SNAKE_CASE_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) A: int = os.path.join( SCREAMING_SNAKE_CASE_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(SCREAMING_SNAKE_CASE_ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=SCREAMING_SNAKE_CASE_ , ensure_ascii=SCREAMING_SNAKE_CASE_ ) + '''\n''' ) A: Any = 0 with open(SCREAMING_SNAKE_CASE_ , '''w''' , encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda SCREAMING_SNAKE_CASE_ : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ''' Please check that the tokenizer is not corrupted!''' ) A: Union[str, Any] = token_index writer.write(''' '''.join(SCREAMING_SNAKE_CASE_ ) + '''\n''' ) index += 1 return vocab_file, merge_file def _snake_case ( self : List[Any] , SCREAMING_SNAKE_CASE_ : List[int] , SCREAMING_SNAKE_CASE_ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] A: int = [self.cls_token_id] A: str = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _snake_case ( self : Tuple , SCREAMING_SNAKE_CASE_ : List[int] , SCREAMING_SNAKE_CASE_ : Optional[List[int]] = None , SCREAMING_SNAKE_CASE_ : bool = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=SCREAMING_SNAKE_CASE_ , token_ids_a=SCREAMING_SNAKE_CASE_ , already_has_special_tokens=SCREAMING_SNAKE_CASE_ ) if token_ids_a is None: return [1] + ([0] * len(SCREAMING_SNAKE_CASE_ )) + [1] return [1] + ([0] * len(SCREAMING_SNAKE_CASE_ )) + [1, 1] + ([0] * len(SCREAMING_SNAKE_CASE_ )) + [1] def _snake_case ( self : Tuple , SCREAMING_SNAKE_CASE_ : List[int] , SCREAMING_SNAKE_CASE_ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' A: Dict = [self.sep_token_id] A: Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _snake_case ( self : int , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Dict=False , **SCREAMING_SNAKE_CASE_ : Optional[int] ) -> int: '''simple docstring''' A: Tuple = kwargs.pop('''add_prefix_space''' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(SCREAMING_SNAKE_CASE_ ) > 0 and not text[0].isspace()): A: List[Any] = ''' ''' + text return (text, kwargs)
319
0
import itertools import string from collections.abc import Generator, Iterable def __lowerCAmelCase (SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE )-> Generator[tuple[str, ...], None, None]: """simple docstring""" snake_case_ = iter(SCREAMING_SNAKE_CASE ) while True: snake_case_ = tuple(itertools.islice(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) if not chunk: return yield chunk def __lowerCAmelCase (SCREAMING_SNAKE_CASE )-> str: """simple docstring""" snake_case_ = ''''''.join([c.upper() for c in dirty if c in string.ascii_letters] ) snake_case_ = '''''' if len(SCREAMING_SNAKE_CASE ) < 2: return dirty for i in range(len(SCREAMING_SNAKE_CASE ) - 1 ): clean += dirty[i] if dirty[i] == dirty[i + 1]: clean += "X" clean += dirty[-1] if len(SCREAMING_SNAKE_CASE ) & 1: clean += "X" return clean def __lowerCAmelCase (SCREAMING_SNAKE_CASE )-> list[str]: """simple docstring""" snake_case_ = '''ABCDEFGHIKLMNOPQRSTUVWXYZ''' # we're using a list instead of a '2d' array because it makes the math # for setting up the table and doing the actual encoding/decoding simpler snake_case_ = [] # copy key chars into the table if they are in `alphabet` ignoring duplicates for char in key.upper(): if char not in table and char in alphabet: table.append(SCREAMING_SNAKE_CASE ) # fill the rest of the table in with the remaining alphabet chars for char in alphabet: if char not in table: table.append(SCREAMING_SNAKE_CASE ) return table def __lowerCAmelCase (SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE )-> str: """simple docstring""" snake_case_ = generate_table(SCREAMING_SNAKE_CASE ) snake_case_ = prepare_input(SCREAMING_SNAKE_CASE ) snake_case_ = '''''' # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(SCREAMING_SNAKE_CASE , 2 ): snake_case_ , snake_case_ = divmod(table.index(SCREAMING_SNAKE_CASE ) , 5 ) snake_case_ , snake_case_ = divmod(table.index(SCREAMING_SNAKE_CASE ) , 5 ) if rowa == rowa: ciphertext += table[rowa * 5 + (cola + 1) % 5] ciphertext += table[rowa * 5 + (cola + 1) % 5] elif cola == cola: ciphertext += table[((rowa + 1) % 5) * 5 + cola] ciphertext += table[((rowa + 1) % 5) * 5 + cola] else: # rectangle ciphertext += table[rowa * 5 + cola] ciphertext += table[rowa * 5 + cola] return ciphertext def __lowerCAmelCase (SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE )-> str: """simple docstring""" snake_case_ = generate_table(SCREAMING_SNAKE_CASE ) snake_case_ = '''''' # https://en.wikipedia.org/wiki/Playfair_cipher#Description for chara, chara in chunker(SCREAMING_SNAKE_CASE , 2 ): snake_case_ , snake_case_ = divmod(table.index(SCREAMING_SNAKE_CASE ) , 5 ) snake_case_ , snake_case_ = divmod(table.index(SCREAMING_SNAKE_CASE ) , 5 ) if rowa == rowa: plaintext += table[rowa * 5 + (cola - 1) % 5] plaintext += table[rowa * 5 + (cola - 1) % 5] elif cola == cola: plaintext += table[((rowa - 1) % 5) * 5 + cola] plaintext += table[((rowa - 1) % 5) * 5 + cola] else: # rectangle plaintext += table[rowa * 5 + cola] plaintext += table[rowa * 5 + cola] return plaintext
267
# Function to print upper half of diamond (pyramid) def __lowerCAmelCase (SCREAMING_SNAKE_CASE )-> Dict: """simple docstring""" for i in range(0 , SCREAMING_SNAKE_CASE ): for _ in range(0 , n - i - 1 ): # printing spaces print(''' ''' , end='''''' ) for _ in range(0 , i + 1 ): # printing stars print('''* ''' , end='''''' ) print() def __lowerCAmelCase (SCREAMING_SNAKE_CASE )-> Dict: """simple docstring""" for i in range(SCREAMING_SNAKE_CASE , 0 , -1 ): for _ in range(SCREAMING_SNAKE_CASE , 0 , -1 ): # printing stars print('''* ''' , end='''''' ) print() for _ in range(n - i + 1 , 0 , -1 ): # printing spaces print(''' ''' , end='''''' ) def __lowerCAmelCase (SCREAMING_SNAKE_CASE )-> Any: """simple docstring""" if n <= 0: print(''' ... .... nothing printing :(''' ) return floyd(SCREAMING_SNAKE_CASE ) # upper half reverse_floyd(SCREAMING_SNAKE_CASE ) # lower half if __name__ == "__main__": print(r"""| /\ | |- | |- |--| |\ /| |-""") print(r"""|/ \| |- |_ |_ |__| | \/ | |_""") UpperCAmelCase = 1 while K: UpperCAmelCase = int(input("""enter the number and , and see the magic : """)) print() pretty_print(user_number) UpperCAmelCase = int(input("""press 0 to exit... and 1 to continue...""")) print("""Good Bye...""")
267
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __UpperCamelCase : Optional[Any] = { '''configuration_rag''': ['''RagConfig'''], '''retrieval_rag''': ['''RagRetriever'''], '''tokenization_rag''': ['''RagTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Tuple = [ '''RagModel''', '''RagPreTrainedModel''', '''RagSequenceForGeneration''', '''RagTokenForGeneration''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCamelCase : Union[str, Any] = [ '''TFRagModel''', '''TFRagPreTrainedModel''', '''TFRagSequenceForGeneration''', '''TFRagTokenForGeneration''', ] if TYPE_CHECKING: from .configuration_rag import RagConfig from .retrieval_rag import RagRetriever from .tokenization_rag import RagTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rag import RagModel, RagPreTrainedModel, RagSequenceForGeneration, RagTokenForGeneration try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rag import ( TFRagModel, TFRagPreTrainedModel, TFRagSequenceForGeneration, TFRagTokenForGeneration, ) else: import sys __UpperCamelCase : Dict = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
106
"""simple docstring""" import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput __UpperCamelCase : Optional[Any] = '''scheduler_config.json''' class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" lowercase__ = 1 lowercase__ = 2 lowercase__ = 3 lowercase__ = 4 lowercase__ = 5 @dataclass class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" lowercase__ = 42 class SCREAMING_SNAKE_CASE : """simple docstring""" lowercase__ = SCHEDULER_CONFIG_NAME lowercase__ = ["dtype"] lowercase__ = [] lowercase__ = True @classmethod def __lowerCAmelCase ( cls : List[Any] ,lowercase_ : Dict[str, Any] = None ,lowercase_ : Optional[str] = None ,lowercase_ : Optional[int]=False ,**lowercase_ : Any ,): lowerCAmelCase__ ,lowerCAmelCase__ : Dict = cls.load_config( pretrained_model_name_or_path=lowercase_ ,subfolder=lowercase_ ,return_unused_kwargs=lowercase_ ,**lowercase_ ,) lowerCAmelCase__ ,lowerCAmelCase__ : Union[str, Any] = cls.from_config(lowercase_ ,return_unused_kwargs=lowercase_ ,**lowercase_ ) if hasattr(lowercase_ ,'''create_state''' ) and getattr(lowercase_ ,'''has_state''' ,lowercase_ ): lowerCAmelCase__ : List[Any] = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def __lowerCAmelCase ( self : Tuple ,lowercase_ : Union[str, os.PathLike] ,lowercase_ : bool = False ,**lowercase_ : str ): self.save_config(save_directory=lowercase_ ,push_to_hub=lowercase_ ,**lowercase_ ) @property def __lowerCAmelCase ( self : List[str] ): return self._get_compatibles() @classmethod def __lowerCAmelCase ( cls : List[Any] ): lowerCAmelCase__ : Tuple = list(set([cls.__name__] + cls._compatibles ) ) lowerCAmelCase__ : Tuple = importlib.import_module(__name__.split('''.''' )[0] ) lowerCAmelCase__ : Union[str, Any] = [ getattr(lowercase_ ,lowercase_ ) for c in compatible_classes_str if hasattr(lowercase_ ,lowercase_ ) ] return compatible_classes def __SCREAMING_SNAKE_CASE ( A_ , A_ ): assert len(A_ ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(A_ ) - x.ndim) ) , A_ ) def __SCREAMING_SNAKE_CASE ( A_ , A_=0.999 , A_=jnp.floataa ): def alpha_bar(A_ ): return math.cos((time_step + 0.008) / 1.008 * math.pi / 2 ) ** 2 lowerCAmelCase__ : Optional[Any] = [] for i in range(A_ ): lowerCAmelCase__ : str = i / num_diffusion_timesteps lowerCAmelCase__ : Union[str, Any] = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(A_ ) / alpha_bar(A_ ) , A_ ) ) return jnp.array(A_ , dtype=A_ ) @flax.struct.dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" lowercase__ = 42 lowercase__ = 42 lowercase__ = 42 @classmethod def __lowerCAmelCase ( cls : Union[str, Any] ,lowercase_ : List[Any] ): lowerCAmelCase__ : Optional[int] = scheduler.config if config.trained_betas is not None: lowerCAmelCase__ : Any = jnp.asarray(config.trained_betas ,dtype=scheduler.dtype ) elif config.beta_schedule == "linear": lowerCAmelCase__ : Union[str, Any] = jnp.linspace(config.beta_start ,config.beta_end ,config.num_train_timesteps ,dtype=scheduler.dtype ) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. lowerCAmelCase__ : int = ( jnp.linspace( config.beta_start**0.5 ,config.beta_end**0.5 ,config.num_train_timesteps ,dtype=scheduler.dtype ) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule lowerCAmelCase__ : List[Any] = betas_for_alpha_bar(config.num_train_timesteps ,dtype=scheduler.dtype ) else: raise NotImplementedError( F'beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}' ) lowerCAmelCase__ : str = 1.0 - betas lowerCAmelCase__ : Union[str, Any] = jnp.cumprod(lowercase_ ,axis=0 ) return cls( alphas=lowercase_ ,betas=lowercase_ ,alphas_cumprod=lowercase_ ,) def __SCREAMING_SNAKE_CASE ( A_ , A_ , A_ , A_ ): lowerCAmelCase__ : Any = state.alphas_cumprod lowerCAmelCase__ : Optional[Any] = alphas_cumprod[timesteps] ** 0.5 lowerCAmelCase__ : Tuple = sqrt_alpha_prod.flatten() lowerCAmelCase__ : str = broadcast_to_shape_from_left(A_ , original_samples.shape ) lowerCAmelCase__ : Optional[Any] = (1 - alphas_cumprod[timesteps]) ** 0.5 lowerCAmelCase__ : Optional[Any] = sqrt_one_minus_alpha_prod.flatten() lowerCAmelCase__ : Optional[int] = broadcast_to_shape_from_left(A_ , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def __SCREAMING_SNAKE_CASE ( A_ , A_ , A_ , A_ ): lowerCAmelCase__ ,lowerCAmelCase__ : List[Any] = get_sqrt_alpha_prod(A_ , A_ , A_ , A_ ) lowerCAmelCase__ : str = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def __SCREAMING_SNAKE_CASE ( A_ , A_ , A_ , A_ ): lowerCAmelCase__ ,lowerCAmelCase__ : List[Any] = get_sqrt_alpha_prod(A_ , A_ , A_ , A_ ) lowerCAmelCase__ : Union[str, Any] = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
106
1
from typing import TYPE_CHECKING from ...utils import _LazyModule a : Dict = {'tokenization_wav2vec2_phoneme': ['Wav2Vec2PhonemeCTCTokenizer']} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys a : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
366
'''simple docstring''' import unittest from diffusers import FlaxAutoencoderKL from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax from .test_modeling_common_flax import FlaxModelTesterMixin if is_flax_available(): import jax @require_flax class a ( _lowerCamelCase , unittest.TestCase ): snake_case_ = FlaxAutoencoderKL @property def A_ ( self : List[Any] ): snake_case_ = 4 snake_case_ = 3 snake_case_ = (32, 32) snake_case_ = jax.random.PRNGKey(0 ) snake_case_ = jax.random.uniform(lowercase_ , ((batch_size, num_channels) + sizes) ) return {"sample": image, "prng_key": prng_key} def A_ ( self : Tuple ): snake_case_ = { '''block_out_channels''': [32, 64], '''in_channels''': 3, '''out_channels''': 3, '''down_block_types''': ['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''], '''up_block_types''': ['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''], '''latent_channels''': 4, } snake_case_ = self.dummy_input return init_dict, inputs_dict
72
0
"""simple docstring""" def __UpperCAmelCase ( UpperCAmelCase_ : list[int] , UpperCAmelCase_ : list[int] , UpperCAmelCase_ : int ) -> bool: '''simple docstring''' return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(UpperCAmelCase_ ) ) def __UpperCAmelCase ( UpperCAmelCase_ : list[list[int]] , UpperCAmelCase_ : int , UpperCAmelCase_ : list[int] , UpperCAmelCase_ : int ) -> bool: '''simple docstring''' if index == len(UpperCAmelCase_ ): return True # Recursive Step for i in range(UpperCAmelCase_ ): if valid_coloring(graph[index] , UpperCAmelCase_ , UpperCAmelCase_ ): # Color current vertex __snake_case : int = i # Validate coloring if util_color(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , index + 1 ): return True # Backtrack __snake_case : Optional[Any] = -1 return False def __UpperCAmelCase ( UpperCAmelCase_ : list[list[int]] , UpperCAmelCase_ : int ) -> list[int]: '''simple docstring''' __snake_case : List[str] = [-1] * len(UpperCAmelCase_ ) if util_color(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , 0 ): return colored_vertices return []
172
"""simple docstring""" # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers _a : Any= "3" print("Python version:", sys.version) print("transformers version:", transformers.__version__) try: import torch print("Torch version:", torch.__version__) print("Cuda available:", torch.cuda.is_available()) print("Cuda version:", torch.version.cuda) print("CuDNN version:", torch.backends.cudnn.version()) print("Number of GPUs available:", torch.cuda.device_count()) print("NCCL version:", torch.cuda.nccl.version()) except ImportError: print("Torch version:", None) try: import deepspeed print("DeepSpeed version:", deepspeed.__version__) except ImportError: print("DeepSpeed version:", None) try: import tensorflow as tf print("TensorFlow version:", tf.__version__) print("TF GPUs available:", bool(tf.config.list_physical_devices("GPU"))) print("Number of TF GPUs available:", len(tf.config.list_physical_devices("GPU"))) except ImportError: print("TensorFlow version:", None)
172
1
import torch from diffusers import EulerDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE__ ( __SCREAMING_SNAKE_CASE ): __SCREAMING_SNAKE_CASE = (EulerDiscreteScheduler,) __SCREAMING_SNAKE_CASE = 10 def UpperCamelCase ( self,**__lowerCamelCase ): A__ = { "num_train_timesteps": 1100, "beta_start": 0.0001, "beta_end": 0.02, "beta_schedule": "linear", } config.update(**_SCREAMING_SNAKE_CASE ) return config def UpperCamelCase ( self ): for timesteps in [10, 50, 100, 1000]: self.check_over_configs(num_train_timesteps=_SCREAMING_SNAKE_CASE ) def UpperCamelCase ( self ): for beta_start, beta_end in zip([0.00001, 0.0001, 0.001],[0.0002, 0.002, 0.02] ): self.check_over_configs(beta_start=_SCREAMING_SNAKE_CASE,beta_end=_SCREAMING_SNAKE_CASE ) def UpperCamelCase ( self ): for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=_SCREAMING_SNAKE_CASE ) def UpperCamelCase ( self ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_SCREAMING_SNAKE_CASE ) def UpperCamelCase ( self ): A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config() A__ = scheduler_class(**_SCREAMING_SNAKE_CASE ) scheduler.set_timesteps(self.num_inference_steps ) A__ = torch.manual_seed(0 ) A__ = self.dummy_model() A__ = self.dummy_sample_deter * scheduler.init_noise_sigma A__ = sample.to(_SCREAMING_SNAKE_CASE ) for i, t in enumerate(scheduler.timesteps ): A__ = scheduler.scale_model_input(_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE ) A__ = model(_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE ) A__ = scheduler.step(_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE,generator=_SCREAMING_SNAKE_CASE ) A__ = output.prev_sample A__ = torch.sum(torch.abs(_SCREAMING_SNAKE_CASE ) ) A__ = torch.mean(torch.abs(_SCREAMING_SNAKE_CASE ) ) assert abs(result_sum.item() - 10.0807 ) < 1E-2 assert abs(result_mean.item() - 0.0131 ) < 1E-3 def UpperCamelCase ( self ): A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config(prediction_type='''v_prediction''' ) A__ = scheduler_class(**_SCREAMING_SNAKE_CASE ) scheduler.set_timesteps(self.num_inference_steps ) A__ = torch.manual_seed(0 ) A__ = self.dummy_model() A__ = self.dummy_sample_deter * scheduler.init_noise_sigma A__ = sample.to(_SCREAMING_SNAKE_CASE ) for i, t in enumerate(scheduler.timesteps ): A__ = scheduler.scale_model_input(_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE ) A__ = model(_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE ) A__ = scheduler.step(_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE,generator=_SCREAMING_SNAKE_CASE ) A__ = output.prev_sample A__ = torch.sum(torch.abs(_SCREAMING_SNAKE_CASE ) ) A__ = torch.mean(torch.abs(_SCREAMING_SNAKE_CASE ) ) assert abs(result_sum.item() - 0.0002 ) < 1E-2 assert abs(result_mean.item() - 2.26_76E-06 ) < 1E-3 def UpperCamelCase ( self ): A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config() A__ = scheduler_class(**_SCREAMING_SNAKE_CASE ) scheduler.set_timesteps(self.num_inference_steps,device=_SCREAMING_SNAKE_CASE ) A__ = torch.manual_seed(0 ) A__ = self.dummy_model() A__ = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() A__ = sample.to(_SCREAMING_SNAKE_CASE ) for t in scheduler.timesteps: A__ = scheduler.scale_model_input(_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE ) A__ = model(_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE ) A__ = scheduler.step(_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE,generator=_SCREAMING_SNAKE_CASE ) A__ = output.prev_sample A__ = torch.sum(torch.abs(_SCREAMING_SNAKE_CASE ) ) A__ = torch.mean(torch.abs(_SCREAMING_SNAKE_CASE ) ) assert abs(result_sum.item() - 10.0807 ) < 1E-2 assert abs(result_mean.item() - 0.0131 ) < 1E-3 def UpperCamelCase ( self ): A__ = self.scheduler_classes[0] A__ = self.get_scheduler_config() A__ = scheduler_class(**_SCREAMING_SNAKE_CASE,use_karras_sigmas=_SCREAMING_SNAKE_CASE ) scheduler.set_timesteps(self.num_inference_steps,device=_SCREAMING_SNAKE_CASE ) A__ = torch.manual_seed(0 ) A__ = self.dummy_model() A__ = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() A__ = sample.to(_SCREAMING_SNAKE_CASE ) for t in scheduler.timesteps: A__ = scheduler.scale_model_input(_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE ) A__ = model(_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE ) A__ = scheduler.step(_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE,_SCREAMING_SNAKE_CASE,generator=_SCREAMING_SNAKE_CASE ) A__ = output.prev_sample A__ = torch.sum(torch.abs(_SCREAMING_SNAKE_CASE ) ) A__ = torch.mean(torch.abs(_SCREAMING_SNAKE_CASE ) ) assert abs(result_sum.item() - 124.52299499511719 ) < 1E-2 assert abs(result_mean.item() - 0.16213932633399963 ) < 1E-3
363
from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class SCREAMING_SNAKE_CASE__ ( UpperCamelCase__ ): __SCREAMING_SNAKE_CASE = ['''image_processor''', '''tokenizer'''] __SCREAMING_SNAKE_CASE = '''Pix2StructImageProcessor''' __SCREAMING_SNAKE_CASE = ('''T5Tokenizer''', '''T5TokenizerFast''') def __init__( self,__lowerCamelCase,__lowerCamelCase ): A__ = False super().__init__(__lowerCamelCase,__lowerCamelCase ) def __call__( self,__lowerCamelCase=None,__lowerCamelCase = None,__lowerCamelCase = True,__lowerCamelCase = False,__lowerCamelCase = None,__lowerCamelCase = None,__lowerCamelCase = 2048,__lowerCamelCase = 0,__lowerCamelCase = None,__lowerCamelCase = None,__lowerCamelCase = False,__lowerCamelCase = False,__lowerCamelCase = False,__lowerCamelCase = False,__lowerCamelCase = False,__lowerCamelCase = True,__lowerCamelCase = None,**__lowerCamelCase,): if images is None and text is None: raise ValueError('''You have to specify either images or text.''' ) # Get only text if images is None and not self.image_processor.is_vqa: A__ = self.tokenizer A__ = self.tokenizer( text=__lowerCamelCase,add_special_tokens=__lowerCamelCase,padding=__lowerCamelCase,truncation=__lowerCamelCase,max_length=__lowerCamelCase,stride=__lowerCamelCase,pad_to_multiple_of=__lowerCamelCase,return_attention_mask=__lowerCamelCase,return_overflowing_tokens=__lowerCamelCase,return_special_tokens_mask=__lowerCamelCase,return_offsets_mapping=__lowerCamelCase,return_token_type_ids=__lowerCamelCase,return_length=__lowerCamelCase,verbose=__lowerCamelCase,return_tensors=__lowerCamelCase,**__lowerCamelCase,) return text_encoding if not self.image_processor.is_vqa: # add pixel_values A__ = self.image_processor( __lowerCamelCase,return_tensors=__lowerCamelCase,max_patches=__lowerCamelCase,**__lowerCamelCase ) else: # add pixel_values and bbox A__ = self.image_processor( __lowerCamelCase,return_tensors=__lowerCamelCase,max_patches=__lowerCamelCase,header_text=__lowerCamelCase,**__lowerCamelCase ) if text is not None and not self.image_processor.is_vqa: A__ = self.tokenizer( text=__lowerCamelCase,add_special_tokens=__lowerCamelCase,padding=__lowerCamelCase,truncation=__lowerCamelCase,max_length=__lowerCamelCase,stride=__lowerCamelCase,pad_to_multiple_of=__lowerCamelCase,return_attention_mask=__lowerCamelCase,return_overflowing_tokens=__lowerCamelCase,return_special_tokens_mask=__lowerCamelCase,return_offsets_mapping=__lowerCamelCase,return_token_type_ids=__lowerCamelCase,return_length=__lowerCamelCase,verbose=__lowerCamelCase,return_tensors=__lowerCamelCase,**__lowerCamelCase,) if "attention_mask" in text_encoding: A__ = text_encoding.pop('''attention_mask''' ) if "input_ids" in text_encoding: A__ = text_encoding.pop('''input_ids''' ) else: A__ = None if text_encoding is not None: encoding_image_processor.update(__lowerCamelCase ) return encoding_image_processor def UpperCamelCase ( self,*__lowerCamelCase,**__lowerCamelCase ): return self.tokenizer.batch_decode(*__lowerCamelCase,**__lowerCamelCase ) def UpperCamelCase ( self,*__lowerCamelCase,**__lowerCamelCase ): return self.tokenizer.decode(*__lowerCamelCase,**__lowerCamelCase ) @property def UpperCamelCase ( self ): A__ = self.tokenizer.model_input_names A__ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
39
0
'''simple docstring''' from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class a ( _lowerCamelCase ): snake_case_ = ["image_processor"] snake_case_ = "SamImageProcessor" def __init__( self : List[Any] , lowercase_ : List[Any] ): super().__init__(lowercase_ ) snake_case_ = self.image_processor snake_case_ = -10 snake_case_ = self.image_processor.size['''longest_edge'''] def __call__( self : List[str] , lowercase_ : Tuple=None , lowercase_ : Tuple=None , lowercase_ : Optional[int]=None , lowercase_ : Dict=None , lowercase_ : Optional[Union[str, TensorType]] = None , **lowercase_ : Any , ): snake_case_ = self.image_processor( lowercase_ , return_tensors=lowercase_ , **lowercase_ , ) # pop arguments that are not used in the foward but used nevertheless snake_case_ = encoding_image_processor['''original_sizes'''] if hasattr(lowercase_ , '''numpy''' ): # Checks if Torch or TF tensor snake_case_ = original_sizes.numpy() snake_case_ ,snake_case_ ,snake_case_ = self._check_and_preprocess_points( input_points=lowercase_ , input_labels=lowercase_ , input_boxes=lowercase_ , ) snake_case_ = self._normalize_and_convert( lowercase_ , lowercase_ , input_points=lowercase_ , input_labels=lowercase_ , input_boxes=lowercase_ , return_tensors=lowercase_ , ) return encoding_image_processor def A_ ( self : Dict , lowercase_ : Union[str, Any] , lowercase_ : str , lowercase_ : Optional[Any]=None , lowercase_ : Tuple=None , lowercase_ : Optional[Any]=None , lowercase_ : Optional[Any]="pt" , ): if input_points is not None: if len(lowercase_ ) != len(lowercase_ ): snake_case_ = [ self._normalize_coordinates(self.target_size , lowercase_ , original_sizes[0] ) for point in input_points ] else: snake_case_ = [ self._normalize_coordinates(self.target_size , lowercase_ , lowercase_ ) for point, original_size in zip(lowercase_ , lowercase_ ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: snake_case_ ,snake_case_ = self._pad_points_and_labels(lowercase_ , lowercase_ ) snake_case_ = np.array(lowercase_ ) if input_labels is not None: snake_case_ = np.array(lowercase_ ) if input_boxes is not None: if len(lowercase_ ) != len(lowercase_ ): snake_case_ = [ self._normalize_coordinates(self.target_size , lowercase_ , original_sizes[0] , is_bounding_box=lowercase_ ) for box in input_boxes ] else: snake_case_ = [ self._normalize_coordinates(self.target_size , lowercase_ , lowercase_ , is_bounding_box=lowercase_ ) for box, original_size in zip(lowercase_ , lowercase_ ) ] snake_case_ = np.array(lowercase_ ) if input_boxes is not None: if return_tensors == "pt": snake_case_ = torch.from_numpy(lowercase_ ) # boxes batch size of 1 by default snake_case_ = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": snake_case_ = tf.convert_to_tensor(lowercase_ ) # boxes batch size of 1 by default snake_case_ = tf.expand_dims(lowercase_ , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({'''input_boxes''': input_boxes} ) if input_points is not None: if return_tensors == "pt": snake_case_ = torch.from_numpy(lowercase_ ) # point batch size of 1 by default snake_case_ = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": snake_case_ = tf.convert_to_tensor(lowercase_ ) # point batch size of 1 by default snake_case_ = tf.expand_dims(lowercase_ , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({'''input_points''': input_points} ) if input_labels is not None: if return_tensors == "pt": snake_case_ = torch.from_numpy(lowercase_ ) # point batch size of 1 by default snake_case_ = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": snake_case_ = tf.convert_to_tensor(lowercase_ ) # point batch size of 1 by default snake_case_ = tf.expand_dims(lowercase_ , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({'''input_labels''': input_labels} ) return encoding_image_processor def A_ ( self : Union[str, Any] , lowercase_ : Any , lowercase_ : Dict ): snake_case_ = max([point.shape[0] for point in input_points] ) snake_case_ = [] for i, point in enumerate(lowercase_ ): if point.shape[0] != expected_nb_points: snake_case_ = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) snake_case_ = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(lowercase_ ) snake_case_ = processed_input_points return input_points, input_labels def A_ ( self : int , lowercase_ : int , lowercase_ : np.ndarray , lowercase_ : List[str] , lowercase_ : Tuple=False ): snake_case_ ,snake_case_ = original_size snake_case_ ,snake_case_ = self.image_processor._get_preprocess_shape(lowercase_ , longest_edge=lowercase_ ) snake_case_ = deepcopy(lowercase_ ).astype(lowercase_ ) if is_bounding_box: snake_case_ = coords.reshape(-1 , 2 , 2 ) snake_case_ = coords[..., 0] * (new_w / old_w) snake_case_ = coords[..., 1] * (new_h / old_h) if is_bounding_box: snake_case_ = coords.reshape(-1 , 4 ) return coords def A_ ( self : Any , lowercase_ : int=None , lowercase_ : int=None , lowercase_ : Optional[Any]=None , ): if input_points is not None: if hasattr(lowercase_ , '''numpy''' ): # Checks for TF or Torch tensor snake_case_ = input_points.numpy().tolist() if not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_points[0] , lowercase_ ): raise ValueError('''Input points must be a list of list of floating points.''' ) snake_case_ = [np.array(lowercase_ ) for input_point in input_points] else: snake_case_ = None if input_labels is not None: if hasattr(lowercase_ , '''numpy''' ): snake_case_ = input_labels.numpy().tolist() if not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_labels[0] , lowercase_ ): raise ValueError('''Input labels must be a list of list integers.''' ) snake_case_ = [np.array(lowercase_ ) for label in input_labels] else: snake_case_ = None if input_boxes is not None: if hasattr(lowercase_ , '''numpy''' ): snake_case_ = input_boxes.numpy().tolist() if ( not isinstance(lowercase_ , lowercase_ ) or not isinstance(input_boxes[0] , lowercase_ ) or not isinstance(input_boxes[0][0] , lowercase_ ) ): raise ValueError('''Input boxes must be a list of list of list of floating points.''' ) snake_case_ = [np.array(lowercase_ ).astype(np.floataa ) for box in input_boxes] else: snake_case_ = None return input_points, input_labels, input_boxes @property def A_ ( self : Optional[int] ): snake_case_ = self.image_processor.model_input_names return list(dict.fromkeys(lowercase_ ) ) def A_ ( self : Optional[Any] , *lowercase_ : Optional[Any] , **lowercase_ : List[Any] ): return self.image_processor.post_process_masks(*lowercase_ , **lowercase_ )
56
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging a : Dict = logging.get_logger(__name__) a : List[str] = { 'Helsinki-NLP/opus-mt-en-de': 'https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json', # See all Marian models at https://huggingface.co/models?filter=marian } class a ( _lowerCamelCase ): snake_case_ = "marian" snake_case_ = ["past_key_values"] snake_case_ = {"num_attention_heads": "encoder_attention_heads", "hidden_size": "d_model"} def __init__( self : List[Any] , lowercase_ : Optional[Any]=5_8101 , lowercase_ : Dict=None , lowercase_ : List[str]=1024 , lowercase_ : Optional[Any]=12 , lowercase_ : int=4096 , lowercase_ : Any=16 , lowercase_ : Optional[int]=12 , lowercase_ : str=4096 , lowercase_ : Union[str, Any]=16 , lowercase_ : Dict=0.0 , lowercase_ : Union[str, Any]=0.0 , lowercase_ : Optional[Any]=True , lowercase_ : Union[str, Any]=True , lowercase_ : int="gelu" , lowercase_ : Dict=1024 , lowercase_ : int=0.1 , lowercase_ : Tuple=0.0 , lowercase_ : Tuple=0.0 , lowercase_ : Tuple=0.02 , lowercase_ : int=5_8100 , lowercase_ : Optional[Any]=False , lowercase_ : Any=5_8100 , lowercase_ : Optional[int]=0 , lowercase_ : Dict=0 , lowercase_ : List[str]=True , **lowercase_ : Any , ): snake_case_ = vocab_size snake_case_ = decoder_vocab_size or vocab_size snake_case_ = max_position_embeddings snake_case_ = d_model snake_case_ = encoder_ffn_dim snake_case_ = encoder_layers snake_case_ = encoder_attention_heads snake_case_ = decoder_ffn_dim snake_case_ = decoder_layers snake_case_ = decoder_attention_heads snake_case_ = dropout snake_case_ = attention_dropout snake_case_ = activation_dropout snake_case_ = activation_function snake_case_ = init_std snake_case_ = encoder_layerdrop snake_case_ = decoder_layerdrop snake_case_ = use_cache snake_case_ = encoder_layers snake_case_ = scale_embedding # scale factor will be sqrt(d_model) if True snake_case_ = share_encoder_decoder_embeddings super().__init__( pad_token_id=lowercase_ , eos_token_id=lowercase_ , is_encoder_decoder=lowercase_ , decoder_start_token_id=lowercase_ , forced_eos_token_id=lowercase_ , **lowercase_ , ) class a ( _lowerCamelCase ): @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def A_ ( self : Union[str, Any] ): if self.task in ["default", "seq2seq-lm"]: snake_case_ = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: snake_case_ = {0: '''batch'''} snake_case_ = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: snake_case_ = {0: '''batch''', 1: '''decoder_sequence'''} snake_case_ = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(lowercase_ , direction='''inputs''' ) elif self.task == "causal-lm": # TODO: figure this case out. snake_case_ = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ] ) if self.use_past: snake_case_ ,snake_case_ = self.num_layers for i in range(lowercase_ ): snake_case_ = {0: '''batch''', 2: '''past_sequence + sequence'''} snake_case_ = {0: '''batch''', 2: '''past_sequence + sequence'''} else: snake_case_ = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ] ) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def A_ ( self : Dict ): if self.task in ["default", "seq2seq-lm"]: snake_case_ = super().outputs else: snake_case_ = super(lowercase_ , self ).outputs if self.use_past: snake_case_ ,snake_case_ = self.num_layers for i in range(lowercase_ ): snake_case_ = {0: '''batch''', 2: '''past_sequence + sequence'''} snake_case_ = {0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def A_ ( self : Dict , lowercase_ : PreTrainedTokenizer , lowercase_ : int = -1 , lowercase_ : int = -1 , lowercase_ : bool = False , lowercase_ : Optional[TensorType] = None , ): snake_case_ = self._generate_dummy_inputs_for_encoder_and_decoder( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) # Generate decoder inputs snake_case_ = seq_length if not self.use_past else 1 snake_case_ = self._generate_dummy_inputs_for_encoder_and_decoder( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) snake_case_ = {F"decoder_{name}": tensor for name, tensor in decoder_inputs.items()} snake_case_ = dict(**lowercase_ , **lowercase_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch snake_case_ ,snake_case_ = common_inputs['''input_ids'''].shape snake_case_ = common_inputs['''decoder_input_ids'''].shape[1] snake_case_ ,snake_case_ = self.num_attention_heads snake_case_ = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) snake_case_ = decoder_seq_length + 3 snake_case_ = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) snake_case_ = torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(lowercase_ , lowercase_ )] , dim=1 ) snake_case_ = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered snake_case_ ,snake_case_ = self.num_layers snake_case_ = min(lowercase_ , lowercase_ ) snake_case_ = max(lowercase_ , lowercase_ ) - min_num_layers snake_case_ = '''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(lowercase_ ): common_inputs["past_key_values"].append( ( torch.zeros(lowercase_ ), torch.zeros(lowercase_ ), torch.zeros(lowercase_ ), torch.zeros(lowercase_ ), ) ) # TODO: test this. snake_case_ = encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(lowercase_ , lowercase_ ): common_inputs["past_key_values"].append((torch.zeros(lowercase_ ), torch.zeros(lowercase_ )) ) return common_inputs def A_ ( self : Union[str, Any] , lowercase_ : PreTrainedTokenizer , lowercase_ : int = -1 , lowercase_ : int = -1 , lowercase_ : bool = False , lowercase_ : Optional[TensorType] = None , ): snake_case_ = self._generate_dummy_inputs_for_encoder_and_decoder( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch snake_case_ ,snake_case_ = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values snake_case_ = seqlen + 2 snake_case_ ,snake_case_ = self.num_layers snake_case_ ,snake_case_ = self.num_attention_heads snake_case_ = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) snake_case_ = common_inputs['''attention_mask'''].dtype snake_case_ = torch.cat( [common_inputs['''attention_mask'''], torch.ones(lowercase_ , lowercase_ , dtype=lowercase_ )] , dim=1 ) snake_case_ = [ (torch.zeros(lowercase_ ), torch.zeros(lowercase_ )) for _ in range(lowercase_ ) ] return common_inputs def A_ ( self : List[str] , lowercase_ : PreTrainedTokenizer , lowercase_ : int = -1 , lowercase_ : int = -1 , lowercase_ : bool = False , lowercase_ : Optional[TensorType] = None , ): # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX snake_case_ = compute_effective_axis_dimension( lowercase_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX snake_case_ = tokenizer.num_special_tokens_to_add(lowercase_ ) snake_case_ = compute_effective_axis_dimension( lowercase_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=lowercase_ ) # Generate dummy inputs according to compute batch and sequence snake_case_ = [''' '''.join([tokenizer.unk_token] ) * seq_length] * batch_size snake_case_ = dict(tokenizer(lowercase_ , return_tensors=lowercase_ ) ) return common_inputs def A_ ( self : Any , lowercase_ : PreTrainedTokenizer , lowercase_ : int = -1 , lowercase_ : int = -1 , lowercase_ : bool = False , lowercase_ : Optional[TensorType] = None , ): if self.task in ["default", "seq2seq-lm"]: snake_case_ = self._generate_dummy_inputs_for_default_and_seqaseq_lm( lowercase_ , batch_size=lowercase_ , seq_length=lowercase_ , is_pair=lowercase_ , framework=lowercase_ ) else: snake_case_ = self._generate_dummy_inputs_for_causal_lm( lowercase_ , batch_size=lowercase_ , seq_length=lowercase_ , is_pair=lowercase_ , framework=lowercase_ ) return common_inputs def A_ ( self : Dict , lowercase_ : List[str] , lowercase_ : List[str] , lowercase_ : int , lowercase_ : List[str] ): if self.task in ["default", "seq2seq-lm"]: snake_case_ = super()._flatten_past_key_values_(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) else: snake_case_ = super(lowercase_ , self )._flatten_past_key_values_( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) @property def A_ ( self : List[str] ): return 1e-4
56
1
import argparse import datetime def snake_case_ (__A : Optional[int] ) -> Optional[int]: __lowerCAmelCase = { """0""": """Sunday""", """1""": """Monday""", """2""": """Tuesday""", """3""": """Wednesday""", """4""": """Thursday""", """5""": """Friday""", """6""": """Saturday""", } __lowerCAmelCase = {0: 1, 1: 2, 2: 3, 3: 4, 4: 5, 5: 6, 6: 0} # Validate if not 0 < len(__A ) < 1_1: raise ValueError("""Must be 10 characters long""" ) # Get month __lowerCAmelCase = int(date_input[0] + date_input[1] ) # Validate if not 0 < m < 1_3: raise ValueError("""Month must be between 1 - 12""" ) __lowerCAmelCase = date_input[2] # Validate if sep_a not in ["-", "/"]: raise ValueError("""Date separator must be \'-\' or \'/\'""" ) # Get day __lowerCAmelCase = int(date_input[3] + date_input[4] ) # Validate if not 0 < d < 3_2: raise ValueError("""Date must be between 1 - 31""" ) # Get second separator __lowerCAmelCase = date_input[5] # Validate if sep_a not in ["-", "/"]: raise ValueError("""Date separator must be \'-\' or \'/\'""" ) # Get year __lowerCAmelCase = int(date_input[6] + date_input[7] + date_input[8] + date_input[9] ) # Arbitrary year range if not 4_5 < y < 8_5_0_0: raise ValueError( """Year out of range. There has to be some sort of limit...right?""" ) # Get datetime obj for validation __lowerCAmelCase = datetime.date(int(__A ) , int(__A ) , int(__A ) ) # Start math if m <= 2: __lowerCAmelCase = y - 1 __lowerCAmelCase = m + 1_2 # maths var __lowerCAmelCase = int(str(__A )[:2] ) __lowerCAmelCase = int(str(__A )[2:] ) __lowerCAmelCase = int(2.6 * m - 5.39 ) __lowerCAmelCase = int(c / 4 ) __lowerCAmelCase = int(k / 4 ) __lowerCAmelCase = int(d + k ) __lowerCAmelCase = int(t + u + v + x ) __lowerCAmelCase = int(z - (2 * c) ) __lowerCAmelCase = round(w % 7 ) # End math # Validate math if f != convert_datetime_days[dt_ck.weekday()]: raise AssertionError("""The date was evaluated incorrectly. Contact developer.""" ) # Response __lowerCAmelCase = f'''Your date {date_input}, is a {days[str(__A )]}!''' return response if __name__ == "__main__": import doctest doctest.testmod() __UpperCAmelCase = argparse.ArgumentParser( description=( """Find out what day of the week nearly any date is or was. Enter """ """date as a string in the mm-dd-yyyy or mm/dd/yyyy format""" ) ) parser.add_argument( """date_input""", type=str, help="""Date as a string (mm-dd-yyyy or mm/dd/yyyy)""" ) __UpperCAmelCase = parser.parse_args() zeller(args.date_input)
359
from __future__ import annotations def snake_case_ (__A : list[float] , __A : list[float] ) -> float: __lowerCAmelCase : Union[str, Any] = sorted(numsa + numsa ) __lowerCAmelCase ,__lowerCAmelCase : Optional[int] = divmod(len(__A ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() __UpperCAmelCase = [float(x) for x in input("""Enter the elements of first array: """).split()] __UpperCAmelCase = [float(x) for x in input("""Enter the elements of second array: """).split()] print(F'The median of two arrays is: {median_of_two_arrays(array_a, array_a)}')
139
0
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_squeezebert import SqueezeBertTokenizer lowercase_ = logging.get_logger(__name__) lowercase_ = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} lowercase_ = { 'vocab_file': { 'squeezebert/squeezebert-uncased': ( 'https://huggingface.co/squeezebert/squeezebert-uncased/resolve/main/vocab.txt' ), 'squeezebert/squeezebert-mnli': 'https://huggingface.co/squeezebert/squeezebert-mnli/resolve/main/vocab.txt', 'squeezebert/squeezebert-mnli-headless': ( 'https://huggingface.co/squeezebert/squeezebert-mnli-headless/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'squeezebert/squeezebert-uncased': ( 'https://huggingface.co/squeezebert/squeezebert-uncased/resolve/main/tokenizer.json' ), 'squeezebert/squeezebert-mnli': ( 'https://huggingface.co/squeezebert/squeezebert-mnli/resolve/main/tokenizer.json' ), 'squeezebert/squeezebert-mnli-headless': ( 'https://huggingface.co/squeezebert/squeezebert-mnli-headless/resolve/main/tokenizer.json' ), }, } lowercase_ = { 'squeezebert/squeezebert-uncased': 512, 'squeezebert/squeezebert-mnli': 512, 'squeezebert/squeezebert-mnli-headless': 512, } lowercase_ = { 'squeezebert/squeezebert-uncased': {'do_lower_case': True}, 'squeezebert/squeezebert-mnli': {'do_lower_case': True}, 'squeezebert/squeezebert-mnli-headless': {'do_lower_case': True}, } class snake_case ( _lowercase ): '''simple docstring''' A_ : Dict = VOCAB_FILES_NAMES A_ : Dict = PRETRAINED_VOCAB_FILES_MAP A_ : List[Any] = PRETRAINED_INIT_CONFIGURATION A_ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ : List[str] = SqueezeBertTokenizer def __init__( self : List[Any], _lowerCamelCase : Optional[Any]=None, _lowerCamelCase : Any=None, _lowerCamelCase : Optional[Any]=True, _lowerCamelCase : str="[UNK]", _lowerCamelCase : Optional[int]="[SEP]", _lowerCamelCase : Dict="[PAD]", _lowerCamelCase : Tuple="[CLS]", _lowerCamelCase : Dict="[MASK]", _lowerCamelCase : Tuple=True, _lowerCamelCase : Tuple=None, **_lowerCamelCase : int, ): '''simple docstring''' super().__init__( A_, tokenizer_file=A_, do_lower_case=A_, unk_token=A_, sep_token=A_, pad_token=A_, cls_token=A_, mask_token=A_, tokenize_chinese_chars=A_, strip_accents=A_, **A_, ) __A = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''', A_ ) != do_lower_case or normalizer_state.get('''strip_accents''', A_ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''', A_ ) != tokenize_chinese_chars ): __A = getattr(A_, normalizer_state.pop('''type''' ) ) __A = do_lower_case __A = strip_accents __A = tokenize_chinese_chars __A = normalizer_class(**A_ ) __A = do_lower_case def _SCREAMING_SNAKE_CASE ( self : List[Any], _lowerCamelCase : List[str], _lowerCamelCase : Optional[int]=None ): '''simple docstring''' __A = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _SCREAMING_SNAKE_CASE ( self : str, _lowerCamelCase : List[int], _lowerCamelCase : Optional[List[int]] = None ): '''simple docstring''' __A = [self.sep_token_id] __A = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _SCREAMING_SNAKE_CASE ( self : int, _lowerCamelCase : str, _lowerCamelCase : Optional[str] = None ): '''simple docstring''' __A = self._tokenizer.model.save(A_, name=A_ ) return tuple(A_ )
266
"""simple docstring""" import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed _lowercase = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F"""{bindir}/../../examples/pytorch/translation"""): from run_translation import main # noqa set_seed(42) _lowercase = '''sshleifer/student_marian_en_ro_6_1''' _lowercase = '''sshleifer/tiny-mbart''' @require_torch class lowerCAmelCase_ ( _lowercase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self : Any ,A_ : Union[str, Any]=False ,A_ : Optional[int]=None ,A_ : List[str]=True ,A_ : Tuple=True ,A_ : Union[str, Any]=True ,A_ : List[str]=True ,) -> Tuple: A = self.run_trainer( eval_steps=1 ,max_len=12 ,model_name=A_ ,num_train_epochs=1 ,distributed=A_ ,extra_args_str=A_ ,predict_with_generate=A_ ,do_train=A_ ,do_eval=A_ ,do_predict=A_ ,) A = TrainerState.load_from_json(os.path.join(A_ ,'trainer_state.json' ) ).log_history if not do_eval: return A = [log for log in logs if 'eval_loss' in log.keys()] A = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats A = eval_metrics[-1] assert isinstance(last_step_stats['eval_bleu'] ,A_ ) assert not math.isnan(float(last_step_stats['eval_loss'] ) ), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def _SCREAMING_SNAKE_CASE ( self : str ) -> Dict: self.run_seqaseq_quick() @require_torch_multi_gpu def _SCREAMING_SNAKE_CASE ( self : int ) -> int: self.run_seqaseq_quick(distributed=A_ ) @require_torch_multi_gpu def _SCREAMING_SNAKE_CASE ( self : str ) -> List[Any]: self.run_seqaseq_quick(distributed=A_ ) @unittest.skip('Requires an update of the env running those tests' ) @require_torch_multi_gpu @require_fairscale def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Dict: self.run_seqaseq_quick(distributed=A_ ,extra_args_str='--sharded_ddp simple' ) @unittest.skip('Requires an update of the env running those tests' ) @require_torch_multi_gpu @require_fairscale def _SCREAMING_SNAKE_CASE ( self : Any ) -> int: self.run_seqaseq_quick(distributed=A_ ,extra_args_str='--sharded_ddp simple --fp16' ) @unittest.skip('Requires an update of the env running those tests' ) @require_torch_multi_gpu @require_fairscale def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> List[Any]: self.run_seqaseq_quick(distributed=A_ ,extra_args_str='--sharded_ddp zero_dp_2' ,predict_with_generate=A_ ) @unittest.skip('Requires an update of the env running those tests' ) @require_torch_multi_gpu @require_fairscale def _SCREAMING_SNAKE_CASE ( self : Dict ) -> Dict: self.run_seqaseq_quick( distributed=A_ ,extra_args_str='--sharded_ddp zero_dp_2 --fp16' ,predict_with_generate=A_ ) @require_apex @require_torch_gpu def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> Optional[Any]: # XXX: apex breaks the trainer if it's run twice e.g. run_seq2seq.main() from the same # program and it breaks other tests that run from the same pytest worker, therefore until this is # sorted out it must be run only in an external program, that is distributed=True in this # test and only under one or more gpus - if we want cpu will need to make a special test # # specifically to the problem traced it to self.optimizer.step() - if it's run 2nd time via # 2nd main() call it botches the future eval. # self.run_seqaseq_quick(distributed=A_ ,extra_args_str='--fp16 --fp16_backend=apex' ) # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=A_ ,extra_args_str='--fp16 --fp16_backend=apex' ) @parameterized.expand(['base', 'low', 'high', 'mixed'] ) @require_torch_multi_gpu def _SCREAMING_SNAKE_CASE ( self : str ,A_ : Dict ) -> List[str]: # as each sub-test is slow-ish split into multiple sub-tests to avoid CI timeout A = { # test with the default log_level - should be info and thus log info once 'base': {'extra_args_str': '', 'n_matches': 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes 'low': {'extra_args_str': '--log_level debug --log_level_replica debug', 'n_matches': 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica 'high': {'extra_args_str': '--log_level error --log_level_replica debug', 'n_matches': 1}, # test with high log_level and log_level_replica - should be quiet on all processes 'mixed': {'extra_args_str': '--log_level error --log_level_replica error', 'n_matches': 0}, } A = experiments[experiment_id] A = {'distributed': True, 'predict_with_generate': False, 'do_eval': False, 'do_predict': False} A = 'Running training' with CaptureStderr() as cl: self.run_seqaseq_quick(**A_ ,extra_args_str=data['extra_args_str'] ) A = len(re.findall(A_ ,cl.err ) ) self.assertEqual(A_ ,data['n_matches'] ) @slow def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> str: A = self.run_trainer( eval_steps=2 ,max_len=128 ,model_name=A_ ,learning_rate=3e-4 ,num_train_epochs=10 ,distributed=A_ ,) # Check metrics A = TrainerState.load_from_json(os.path.join(A_ ,'trainer_state.json' ) ).log_history A = [log for log in logs if 'eval_loss' in log.keys()] A = eval_metrics[0] A = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats['eval_bleu'] ,A_ ) # test if do_predict saves generations and metrics A = os.listdir(A_ ) A = {os.path.basename(A_ ) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Optional[int]: from transformers.training_args import OptimizerNames def train_and_return_metrics(A_ : str ) -> Tuple[int, float]: A = '--skip_memory_metrics 0' A = self.run_trainer( max_len=128 ,model_name=A_ ,learning_rate=3e-4 ,num_train_epochs=1 ,optim=A_ ,distributed=A_ ,extra_args_str=A_ ,do_eval=A_ ,do_predict=A_ ,n_gpus_to_use=1 ,) # Check metrics A = TrainerState.load_from_json(Path(A_ ,'trainer_state.json' ) ).log_history A = int(logs[0]['train_mem_gpu_peaked_delta'] / 2**20 ) A = int(logs[0]['train_mem_gpu_alloc_delta'] / 2**20 ) A = logs[0]['train_loss'] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss A , A , A = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value ) A , A , A = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value ) A = gpu_alloc_mem_orig - gpu_alloc_mem_bnb A = gpu_peak_mem_orig + gpu_alloc_mem_orig A = gpu_peak_mem_bnb + gpu_alloc_mem_bnb A = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings A = 120 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( A_ ,A_ ,'should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got' F' a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and' F' gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB' ,) self.assertGreater( A_ ,A_ ,'should use ~150MB less total gpu memory with BNB, compared to without it for this model but got' F' a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and' F' gpu_total_mem_bnb={gpu_total_mem_bnb}MB' ,) self.assertEqual( A_ ,A_ ,F'loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}' ) def _SCREAMING_SNAKE_CASE ( self : List[str] ,A_ : int ,A_ : str ,A_ : int ,A_ : float = 3e-3 ,A_ : str = "adafactor" ,A_ : bool = False ,A_ : str = None ,A_ : int = 0 ,A_ : bool = True ,A_ : bool = True ,A_ : bool = True ,A_ : bool = True ,A_ : int = None ,) -> Dict: A = self.test_file_dir / '../fixtures/tests_samples/wmt_en_ro' A = self.get_auto_remove_tmp_dir() A = F'\n --model_name_or_path {model_name}\n --train_file {data_dir}/train.json\n --validation_file {data_dir}/val.json\n --test_file {data_dir}/test.json\n --output_dir {output_dir}\n --overwrite_output_dir\n --max_train_samples 8\n --max_source_length {max_len}\n --max_target_length {max_len}\n --do_train\n --num_train_epochs {str(A_ )}\n --per_device_train_batch_size 4\n --learning_rate {learning_rate}\n --warmup_steps 8\n --logging_steps 0\n --logging_strategy no\n --save_steps {str(A_ )}\n --group_by_length\n --label_smoothing_factor 0.1\n --target_lang ro_RO\n --source_lang en_XX\n '.split() A = F'\n --do_eval\n --per_device_eval_batch_size 4\n --max_eval_samples 8\n --val_max_target_length {max_len}\n --evaluation_strategy steps\n --eval_steps {str(A_ )}\n '.split() A = '\n --do_predict\n '.split() A = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += F'--optim {optim}'.split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: A = get_gpu_count() A = get_torch_dist_unique_port() A = F'\n -m torch.distributed.run\n --nproc_per_node={n_gpus_to_use}\n --master_port={master_port}\n {self.examples_dir_str}/pytorch/translation/run_translation.py\n '.split() A = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(A_ ,env=self.get_env() ) else: A = ['run_translation.py'] + args with patch.object(A_ ,'argv' ,A_ ): main() return output_dir
74
0
import copy import inspect import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import TimesformerConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, TimesformerForVideoClassification, TimesformerModel, ) from transformers.models.timesformer.modeling_timesformer import TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from transformers import VideoMAEImageProcessor class __SCREAMING_SNAKE_CASE : def __init__( self , __lowerCAmelCase , __lowerCAmelCase=13 , __lowerCAmelCase=10 , __lowerCAmelCase=3 , __lowerCAmelCase=2 , __lowerCAmelCase=2 , __lowerCAmelCase=True , __lowerCAmelCase=True , __lowerCAmelCase=32 , __lowerCAmelCase=5 , __lowerCAmelCase=4 , __lowerCAmelCase=37 , __lowerCAmelCase="gelu" , __lowerCAmelCase=0.1 , __lowerCAmelCase=0.1 , __lowerCAmelCase=10 , __lowerCAmelCase=0.02 , __lowerCAmelCase="divided_space_time" , __lowerCAmelCase=None , ): UpperCamelCase__ = parent UpperCamelCase__ = batch_size UpperCamelCase__ = image_size UpperCamelCase__ = num_channels UpperCamelCase__ = patch_size UpperCamelCase__ = num_frames UpperCamelCase__ = is_training UpperCamelCase__ = use_labels UpperCamelCase__ = hidden_size UpperCamelCase__ = num_hidden_layers UpperCamelCase__ = num_attention_heads UpperCamelCase__ = intermediate_size UpperCamelCase__ = hidden_act UpperCamelCase__ = hidden_dropout_prob UpperCamelCase__ = attention_probs_dropout_prob UpperCamelCase__ = attention_type UpperCamelCase__ = initializer_range UpperCamelCase__ = scope UpperCamelCase__ = num_labels # in TimeSformer, the number of spatial tokens equals num_frames * num_patches per frame + 1 CLS token UpperCamelCase__ = (image_size // patch_size) ** 2 UpperCamelCase__ = (num_frames) * self.num_patches_per_frame + 1 def _lowerCamelCase ( self ): UpperCamelCase__ = floats_tensor( [self.batch_size, self.num_frames, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase__ = None if self.use_labels: UpperCamelCase__ = ids_tensor([self.batch_size] , self.num_labels ) UpperCamelCase__ = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self ): UpperCamelCase__ = TimesformerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_frames=self.num_frames , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , attention_type=self.attention_type , ) UpperCamelCase__ = self.num_labels return config def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): UpperCamelCase__ = TimesformerModel(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() UpperCamelCase__ = model(__lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): UpperCamelCase__ = TimesformerForVideoClassification(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() UpperCamelCase__ = model(__lowerCAmelCase ) # verify the logits shape UpperCamelCase__ = torch.Size((self.batch_size, self.num_labels) ) self.parent.assertEqual(result.logits.shape , __lowerCAmelCase ) def _lowerCamelCase ( self ): UpperCamelCase__ = self.prepare_config_and_inputs() UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = config_and_inputs UpperCamelCase__ = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( _a , _a , unittest.TestCase ): snake_case : Any = (TimesformerModel, TimesformerForVideoClassification) if is_torch_available() else () snake_case : Optional[int] = ( {"""feature-extraction""": TimesformerModel, """video-classification""": TimesformerForVideoClassification} if is_torch_available() else {} ) snake_case : Union[str, Any] = False snake_case : Dict = False snake_case : List[str] = False snake_case : List[str] = False def _lowerCamelCase ( self ): UpperCamelCase__ = TimesformerModelTester(self ) UpperCamelCase__ = ConfigTester( self , config_class=__lowerCAmelCase , has_text_modality=__lowerCAmelCase , hidden_size=37 ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False ): UpperCamelCase__ = copy.deepcopy(__lowerCAmelCase ) if return_labels: if model_class in get_values(__lowerCAmelCase ): UpperCamelCase__ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__lowerCAmelCase ) return inputs_dict def _lowerCamelCase ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="""TimeSformer does not use inputs_embeds""" ) def _lowerCamelCase ( self ): pass def _lowerCamelCase ( self ): UpperCamelCase__ , UpperCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase__ = model_class(__lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCamelCase__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCAmelCase , nn.Linear ) ) def _lowerCamelCase ( self ): UpperCamelCase__ , UpperCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase__ = model_class(__lowerCAmelCase ) UpperCamelCase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase__ = [*signature.parameters.keys()] UpperCamelCase__ = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCAmelCase ) def _lowerCamelCase ( self ): UpperCamelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCAmelCase ) def _lowerCamelCase ( self ): UpperCamelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_video_classification(*__lowerCAmelCase ) @slow def _lowerCamelCase ( self ): for model_name in TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase__ = TimesformerModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) def _lowerCamelCase ( self ): if not self.has_attentions: pass else: UpperCamelCase__ , UpperCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCamelCase__ = True for model_class in self.all_model_classes: UpperCamelCase__ = self.model_tester.seq_length UpperCamelCase__ = self.model_tester.num_frames UpperCamelCase__ = True UpperCamelCase__ = False UpperCamelCase__ = True UpperCamelCase__ = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() with torch.no_grad(): UpperCamelCase__ = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) UpperCamelCase__ = outputs.attentions self.assertEqual(len(__lowerCAmelCase ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] UpperCamelCase__ = True UpperCamelCase__ = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() with torch.no_grad(): UpperCamelCase__ = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) UpperCamelCase__ = outputs.attentions self.assertEqual(len(__lowerCAmelCase ) , self.model_tester.num_hidden_layers ) # attentions has shape (batch_size x num_frames) x num_heads x (num_patches per frame + 1) x (num_patches per frame + 1) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len // num_frames + 1, seq_len // num_frames + 1] , ) UpperCamelCase__ = len(__lowerCAmelCase ) # Check attention is always last and order is fine UpperCamelCase__ = True UpperCamelCase__ = True UpperCamelCase__ = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() with torch.no_grad(): UpperCamelCase__ = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) self.assertEqual(out_len + 1 , len(__lowerCAmelCase ) ) UpperCamelCase__ = outputs.attentions self.assertEqual(len(__lowerCAmelCase ) , self.model_tester.num_hidden_layers ) # attentions has shape (batch_size x num_frames) x num_heads x (num_patches per frame + 1) x (num_patches per frame + 1) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len // num_frames + 1, seq_len // num_frames + 1] , ) def _lowerCamelCase ( self ): def check_hidden_states_output(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): UpperCamelCase__ = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() with torch.no_grad(): UpperCamelCase__ = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) UpperCamelCase__ = outputs.hidden_states UpperCamelCase__ = self.model_tester.num_hidden_layers + 1 self.assertEqual(len(__lowerCAmelCase ) , __lowerCAmelCase ) UpperCamelCase__ = self.model_tester.seq_length self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) UpperCamelCase__ , UpperCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase__ = True check_hidden_states_output(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase__ = True check_hidden_states_output(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def _UpperCamelCase (): """simple docstring""" UpperCamelCase__ = hf_hub_download( repo_id="""hf-internal-testing/spaghetti-video""" , filename="""eating_spaghetti.npy""" , repo_type="""dataset""" ) UpperCamelCase__ = np.load(a__ ) return list(a__ ) @require_torch @require_vision class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self ): # logits were tested with a different mean and std, so we use the same here return ( VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) if is_vision_available() else None ) @slow def _lowerCamelCase ( self ): UpperCamelCase__ = TimesformerForVideoClassification.from_pretrained("""facebook/timesformer-base-finetuned-k400""" ).to( __lowerCAmelCase ) UpperCamelCase__ = self.default_image_processor UpperCamelCase__ = prepare_video() UpperCamelCase__ = image_processor(video[:8] , return_tensors="""pt""" ).to(__lowerCAmelCase ) # forward pass with torch.no_grad(): UpperCamelCase__ = model(**__lowerCAmelCase ) # verify the logits UpperCamelCase__ = torch.Size((1, 400) ) self.assertEqual(outputs.logits.shape , __lowerCAmelCase ) UpperCamelCase__ = torch.tensor([-0.3016, -0.7713, -0.4205] ).to(__lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCAmelCase , atol=1E-4 ) )
87
import json import os import torch from diffusers import UNetaDModel os.makedirs("hub/hopper-medium-v2/unet/hor32", exist_ok=True) os.makedirs("hub/hopper-medium-v2/unet/hor128", exist_ok=True) os.makedirs("hub/hopper-medium-v2/value_function", exist_ok=True) def _UpperCamelCase (a__ :int ): """simple docstring""" if hor == 128: UpperCamelCase__ = ("""DownResnetBlock1D""", """DownResnetBlock1D""", """DownResnetBlock1D""") UpperCamelCase__ = (32, 128, 256) UpperCamelCase__ = ("""UpResnetBlock1D""", """UpResnetBlock1D""") elif hor == 32: UpperCamelCase__ = ("""DownResnetBlock1D""", """DownResnetBlock1D""", """DownResnetBlock1D""", """DownResnetBlock1D""") UpperCamelCase__ = (32, 64, 128, 256) UpperCamelCase__ = ("""UpResnetBlock1D""", """UpResnetBlock1D""", """UpResnetBlock1D""") UpperCamelCase__ = torch.load(f"""/Users/bglickenhaus/Documents/diffuser/temporal_unet-hopper-mediumv2-hor{hor}.torch""" ) UpperCamelCase__ = model.state_dict() UpperCamelCase__ = { """down_block_types""": down_block_types, """block_out_channels""": block_out_channels, """up_block_types""": up_block_types, """layers_per_block""": 1, """use_timestep_embedding""": True, """out_block_type""": """OutConv1DBlock""", """norm_num_groups""": 8, """downsample_each_block""": False, """in_channels""": 14, """out_channels""": 14, """extra_in_channels""": 0, """time_embedding_type""": """positional""", """flip_sin_to_cos""": False, """freq_shift""": 1, """sample_size""": 6_5536, """mid_block_type""": """MidResTemporalBlock1D""", """act_fn""": """mish""", } UpperCamelCase__ = UNetaDModel(**a__ ) print(f"""length of state dict: {len(state_dict.keys() )}""" ) print(f"""length of value function dict: {len(hf_value_function.state_dict().keys() )}""" ) UpperCamelCase__ = dict(zip(model.state_dict().keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): UpperCamelCase__ = state_dict.pop(a__ ) hf_value_function.load_state_dict(a__ ) torch.save(hf_value_function.state_dict() , f"""hub/hopper-medium-v2/unet/hor{hor}/diffusion_pytorch_model.bin""" ) with open(f"""hub/hopper-medium-v2/unet/hor{hor}/config.json""" , """w""" ) as f: json.dump(a__ , a__ ) def _UpperCamelCase (): """simple docstring""" UpperCamelCase__ = { """in_channels""": 14, """down_block_types""": ("""DownResnetBlock1D""", """DownResnetBlock1D""", """DownResnetBlock1D""", """DownResnetBlock1D"""), """up_block_types""": (), """out_block_type""": """ValueFunction""", """mid_block_type""": """ValueFunctionMidBlock1D""", """block_out_channels""": (32, 64, 128, 256), """layers_per_block""": 1, """downsample_each_block""": True, """sample_size""": 6_5536, """out_channels""": 14, """extra_in_channels""": 0, """time_embedding_type""": """positional""", """use_timestep_embedding""": True, """flip_sin_to_cos""": False, """freq_shift""": 1, """norm_num_groups""": 8, """act_fn""": """mish""", } UpperCamelCase__ = torch.load("""/Users/bglickenhaus/Documents/diffuser/value_function-hopper-mediumv2-hor32.torch""" ) UpperCamelCase__ = model UpperCamelCase__ = UNetaDModel(**a__ ) print(f"""length of state dict: {len(state_dict.keys() )}""" ) print(f"""length of value function dict: {len(hf_value_function.state_dict().keys() )}""" ) UpperCamelCase__ = dict(zip(state_dict.keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): UpperCamelCase__ = state_dict.pop(a__ ) hf_value_function.load_state_dict(a__ ) torch.save(hf_value_function.state_dict() , """hub/hopper-medium-v2/value_function/diffusion_pytorch_model.bin""" ) with open("""hub/hopper-medium-v2/value_function/config.json""" , """w""" ) as f: json.dump(a__ , a__ ) if __name__ == "__main__": unet(32) # unet(128) value_function()
87
1
import os import numpy import onnx def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: lowercase : int = a.name lowercase : Any = b.name lowercase : Optional[Any] = """""" lowercase : Dict = """""" lowercase : int = a == b lowercase : int = name_a lowercase : List[str] = name_b return res def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> str: for i, input_name in enumerate(node_proto.input ): if input_name == name: node_proto.input.insert(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) node_proto.input.pop(i + 1 ) if node_proto.op_type == "If": _graph_replace_input_with(node_proto.attribute[0].g , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) _graph_replace_input_with(node_proto.attribute[1].g , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) if node_proto.op_type == "Loop": _graph_replace_input_with(node_proto.attribute[0].g , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: for n in graph_proto.node: _node_replace_input_with(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> List[Any]: lowercase : Any = list(model.graph.initializer ) lowercase : Dict = list(model_without_ext.graph.initializer ) for i, ref_i in ind_to_replace: assert inits_with_data[i].name == inits[i].name assert inits_with_data[ref_i].name == inits[ref_i].name assert i > ref_i lowercase : Union[str, Any] = inits[i].name lowercase : Dict = inits[ref_i].name model_without_ext.graph.initializer.remove(inits[i] ) # for n in model.graph.node: _graph_replace_input_with(model_without_ext.graph , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def _snake_case( SCREAMING_SNAKE_CASE__ ) -> List[str]: lowercase : Union[str, Any] = os.path.dirname(SCREAMING_SNAKE_CASE__ ) lowercase : Dict = os.path.basename(SCREAMING_SNAKE_CASE__ ) lowercase : str = onnx.load(os.path.join(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) lowercase : List[str] = list(model.graph.initializer ) lowercase : Tuple = set() lowercase : int = {} lowercase : Optional[Any] = [] lowercase : Dict = 0 for i in range(len(SCREAMING_SNAKE_CASE__ ) ): if i in dup_set: continue for j in range(i + 1 , len(SCREAMING_SNAKE_CASE__ ) ): if j in dup_set: continue if _is_equal_tensor_proto(inits[i] , inits[j] ): dup_set.add(SCREAMING_SNAKE_CASE__ ) dup_set.add(SCREAMING_SNAKE_CASE__ ) lowercase : int = inits[j].data_type lowercase : Optional[int] = numpy.prod(inits[j].dims ) if dtype == 1: mem_size *= 4 elif dtype == 6: mem_size *= 4 elif dtype == 7 or dtype == 11: mem_size *= 8 else: print("""unexpected data type: """ , SCREAMING_SNAKE_CASE__ ) total_reduced_size += mem_size lowercase : Tuple = inits[i].name lowercase : int = inits[j].name if name_i in dup_map: dup_map[name_i].append(SCREAMING_SNAKE_CASE__ ) else: lowercase : List[str] = [name_j] ind_to_replace.append((j, i) ) print("""total reduced size: """ , total_reduced_size / 1_024 / 1_024 / 1_024 , """GB""" ) lowercase : str = sorted(SCREAMING_SNAKE_CASE__ ) _remove_dup_initializers_from_model(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) lowercase : Optional[Any] = """optimized_""" + model_file_name lowercase : Dict = os.path.join(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) onnx.save(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return new_model
20
import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": lowercase__ : Union[str, Any] = argparse.ArgumentParser() parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the output model.") parser.add_argument( "--txt2img_unclip", default="kakaobrain/karlo-v1-alpha", type=str, required=False, help="The pretrained txt2img unclip.", ) lowercase__ : Any = parser.parse_args() lowercase__ : Union[str, Any] = UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) lowercase__ : List[str] = CLIPImageProcessor() lowercase__ : Optional[Any] = CLIPVisionModelWithProjection.from_pretrained("openai/clip-vit-large-patch14") lowercase__ : Optional[Any] = UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
328
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import BeitConfig, BeitForImageClassification, BeitForMaskedImageModeling, BeitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase_ =logging.get_logger(__name__) def a_ ( _lowercase , _lowercase=False , _lowercase=False ): _UpperCamelCase : Tuple = '''backbone.''' if is_semantic else '''''' _UpperCamelCase : List[str] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""{prefix}blocks.{i}.norm1.weight""", F"""beit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""{prefix}blocks.{i}.norm1.bias""", F"""beit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append( (F"""{prefix}blocks.{i}.attn.proj.weight""", F"""beit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append( (F"""{prefix}blocks.{i}.attn.proj.bias""", F"""beit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""{prefix}blocks.{i}.norm2.weight""", F"""beit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""{prefix}blocks.{i}.norm2.bias""", F"""beit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""{prefix}blocks.{i}.mlp.fc1.weight""", F"""beit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""{prefix}blocks.{i}.mlp.fc1.bias""", F"""beit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""{prefix}blocks.{i}.mlp.fc2.weight""", F"""beit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""{prefix}blocks.{i}.mlp.fc2.bias""", F"""beit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ (F"""{prefix}cls_token""", '''beit.embeddings.cls_token'''), (F"""{prefix}patch_embed.proj.weight""", '''beit.embeddings.patch_embeddings.projection.weight'''), (F"""{prefix}patch_embed.proj.bias""", '''beit.embeddings.patch_embeddings.projection.bias'''), (F"""{prefix}pos_embed""", '''beit.embeddings.position_embeddings'''), ] ) if has_lm_head: # mask token + layernorm rename_keys.extend( [ ('''mask_token''', '''beit.embeddings.mask_token'''), ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ] ) else: # layernorm + classification head rename_keys.extend( [ ('''fc_norm.weight''', '''beit.pooler.layernorm.weight'''), ('''fc_norm.bias''', '''beit.pooler.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def a_ ( _lowercase , _lowercase , _lowercase=False , _lowercase=False ): for i in range(config.num_hidden_layers ): _UpperCamelCase : List[str] = '''backbone.''' if is_semantic else '''''' # queries, keys and values _UpperCamelCase : List[str] = state_dict.pop(F"""{prefix}blocks.{i}.attn.qkv.weight""" ) _UpperCamelCase : Dict = state_dict.pop(F"""{prefix}blocks.{i}.attn.q_bias""" ) _UpperCamelCase : Union[str, Any] = state_dict.pop(F"""{prefix}blocks.{i}.attn.v_bias""" ) _UpperCamelCase : int = in_proj_weight[ : config.hidden_size, : ] _UpperCamelCase : Union[str, Any] = q_bias _UpperCamelCase : Optional[int] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _UpperCamelCase : int = in_proj_weight[ -config.hidden_size :, : ] _UpperCamelCase : Tuple = v_bias # gamma_1 and gamma_2 # we call them lambda because otherwise they are renamed when using .from_pretrained _UpperCamelCase : Dict = state_dict.pop(F"""{prefix}blocks.{i}.gamma_1""" ) _UpperCamelCase : List[str] = state_dict.pop(F"""{prefix}blocks.{i}.gamma_2""" ) _UpperCamelCase : Optional[int] = gamma_a _UpperCamelCase : int = gamma_a def a_ ( _lowercase , _lowercase , _lowercase ): _UpperCamelCase : str = dct.pop(_UpperCamelCase ) _UpperCamelCase : Union[str, Any] = val def a_ ( ): _UpperCamelCase : Any = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _UpperCamelCase : Optional[Any] = Image.open(requests.get(_UpperCamelCase , stream=_UpperCamelCase ).raw ) return im @torch.no_grad() def a_ ( _lowercase , _lowercase , _lowercase=False ): _UpperCamelCase : Union[str, Any] = False if '''rvlcdip''' in checkpoint_url else True _UpperCamelCase : str = BeitConfig(use_absolute_position_embeddings=_UpperCamelCase , use_mask_token=_UpperCamelCase ) # size of the architecture if "large" in checkpoint_url or "dit-l" in checkpoint_url: _UpperCamelCase : List[str] = 1024 _UpperCamelCase : Optional[Any] = 4096 _UpperCamelCase : Optional[Any] = 24 _UpperCamelCase : Tuple = 16 # labels if "rvlcdip" in checkpoint_url: _UpperCamelCase : Any = 16 _UpperCamelCase : int = '''huggingface/label-files''' _UpperCamelCase : Optional[Any] = '''rvlcdip-id2label.json''' _UpperCamelCase : Optional[int] = json.load(open(hf_hub_download(_UpperCamelCase , _UpperCamelCase , repo_type='''dataset''' ) , '''r''' ) ) _UpperCamelCase : Any = {int(_UpperCamelCase ): v for k, v in idalabel.items()} _UpperCamelCase : Union[str, Any] = idalabel _UpperCamelCase : List[Any] = {v: k for k, v in idalabel.items()} # load state_dict of original model, remove and rename some keys _UpperCamelCase : int = torch.hub.load_state_dict_from_url(_UpperCamelCase , map_location='''cpu''' )['''model'''] _UpperCamelCase : Optional[int] = create_rename_keys(_UpperCamelCase , has_lm_head=_UpperCamelCase ) for src, dest in rename_keys: rename_key(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) read_in_q_k_v(_UpperCamelCase , _UpperCamelCase , has_lm_head=_UpperCamelCase ) # load HuggingFace model _UpperCamelCase : Dict = BeitForMaskedImageModeling(_UpperCamelCase ) if has_lm_head else BeitForImageClassification(_UpperCamelCase ) model.eval() model.load_state_dict(_UpperCamelCase ) # Check outputs on an image _UpperCamelCase : List[str] = BeitImageProcessor( size=config.image_size , resample=PILImageResampling.BILINEAR , do_center_crop=_UpperCamelCase ) _UpperCamelCase : Optional[int] = prepare_img() _UpperCamelCase : Tuple = image_processor(images=_UpperCamelCase , return_tensors='''pt''' ) _UpperCamelCase : int = encoding['''pixel_values'''] _UpperCamelCase : Tuple = model(_UpperCamelCase ) _UpperCamelCase : Tuple = outputs.logits # verify logits _UpperCamelCase : Dict = [1, 16] if '''rvlcdip''' in checkpoint_url else [1, 196, 8192] assert logits.shape == torch.Size(_UpperCamelCase ), "Shape of logits not as expected" Path(_UpperCamelCase ).mkdir(exist_ok=_UpperCamelCase ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCamelCase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_UpperCamelCase ) if push_to_hub: if has_lm_head: _UpperCamelCase : List[Any] = '''dit-base''' if '''base''' in checkpoint_url else '''dit-large''' else: _UpperCamelCase : Optional[int] = '''dit-base-finetuned-rvlcdip''' if '''dit-b''' in checkpoint_url else '''dit-large-finetuned-rvlcdip''' image_processor.push_to_hub( repo_path_or_name=Path(_UpperCamelCase , _UpperCamelCase ) , organization='''nielsr''' , commit_message='''Add image processor''' , use_temp_dir=_UpperCamelCase , ) model.push_to_hub( repo_path_or_name=Path(_UpperCamelCase , _UpperCamelCase ) , organization='''nielsr''' , commit_message='''Add model''' , use_temp_dir=_UpperCamelCase , ) if __name__ == "__main__": UpperCamelCase_ =argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://layoutlm.blob.core.windows.net/dit/dit-pts/dit-base-224-p16-500k-62d53a.pth""", type=str, help="""URL to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", ) UpperCamelCase_ =parser.parse_args() convert_dit_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
359
"""simple docstring""" from ..utils import is_flax_available, is_torch_available if is_torch_available(): from .autoencoder_kl import AutoencoderKL from .controlnet import ControlNetModel from .dual_transformer_ad import DualTransformeraDModel from .modeling_utils import ModelMixin from .prior_transformer import PriorTransformer from .ta_film_transformer import TaFilmDecoder from .transformer_ad import TransformeraDModel from .unet_ad import UNetaDModel from .unet_ad import UNetaDModel from .unet_ad_condition import UNetaDConditionModel from .unet_ad_condition import UNetaDConditionModel from .vq_model import VQModel if is_flax_available(): from .controlnet_flax import FlaxControlNetModel from .unet_ad_condition_flax import FlaxUNetaDConditionModel from .vae_flax import FlaxAutoencoderKL
128
0
'''simple docstring''' def a ( lowerCamelCase__ = 1_00 ): '''simple docstring''' A_ : int = n * (n + 1) * (2 * n + 1) / 6 A_ : Dict = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(F"{solution() = }")
206
import os # Precomputes a list of the 100 first triangular numbers lowerCAmelCase__ = [int(0.5 * n * (n + 1)) for n in range(1, 1_0_1)] def __lowerCamelCase ( ): """simple docstring""" lowercase__ : str = os.path.dirname(os.path.realpath(lowerCamelCase__ ) ) lowercase__ : Optional[Any] = os.path.join(lowerCamelCase__ , "words.txt" ) lowercase__ : int = "" with open(lowerCamelCase__ ) as f: lowercase__ : Any = f.readline() lowercase__ : Any = [word.strip("\"" ) for word in words.strip("\r\n" ).split("," )] lowercase__ : str = [ word for word in [sum(ord(lowerCamelCase__ ) - 64 for x in word ) for word in words] if word in TRIANGULAR_NUMBERS ] return len(lowerCamelCase__ ) if __name__ == "__main__": print(solution())
130
0
"""simple docstring""" import contextlib from multiprocessing import Pool, RLock from tqdm.auto import tqdm from ..utils import experimental, logging __A = logging.get_logger(__name__) class _snake_case : snake_case__ = None @experimental def lowercase_ ( _lowerCamelCase: Optional[Any] , _lowerCamelCase: List[Any] , _lowerCamelCase: Tuple , _lowerCamelCase: int , _lowerCamelCase: str , _lowerCamelCase: Union[str, Any] , _lowerCamelCase: Optional[Any] ) -> List[str]: '''simple docstring''' if ParallelBackendConfig.backend_name is None: return _map_with_multiprocessing_pool( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) return _map_with_joblib(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def lowercase_ ( _lowerCamelCase: int , _lowerCamelCase: Optional[int] , _lowerCamelCase: Optional[Any] , _lowerCamelCase: List[Any] , _lowerCamelCase: List[Any] , _lowerCamelCase: int , _lowerCamelCase: List[Any] ) -> List[str]: '''simple docstring''' __lowerCamelCase : Any = num_proc if num_proc <= len(_lowerCamelCase ) else len(_lowerCamelCase ) __lowerCamelCase : str = [] # We organize the splits ourselve (contiguous splits) for index in range(_lowerCamelCase ): __lowerCamelCase : Dict = len(_lowerCamelCase ) // num_proc __lowerCamelCase : Union[str, Any] = len(_lowerCamelCase ) % num_proc __lowerCamelCase : Tuple = div * index + min(_lowerCamelCase , _lowerCamelCase ) __lowerCamelCase : Optional[Any] = start + div + (1 if index < mod else 0) split_kwds.append((function, iterable[start:end], types, index, disable_tqdm, desc) ) if len(_lowerCamelCase ) != sum(len(i[1] ) for i in split_kwds ): raise ValueError( F"""Error dividing inputs iterable among processes. """ F"""Total number of objects {len(_lowerCamelCase )}, """ F"""length: {sum(len(i[1] ) for i in split_kwds )}""" ) logger.info( F"""Spawning {num_proc} processes for {len(_lowerCamelCase )} objects in slices of {[len(i[1] ) for i in split_kwds]}""" ) __lowerCamelCase , __lowerCamelCase : Optional[Any] = None, None if not disable_tqdm: __lowerCamelCase , __lowerCamelCase : Optional[int] = (RLock(),), tqdm.set_lock with Pool(_lowerCamelCase , initargs=_lowerCamelCase , initializer=_lowerCamelCase ) as pool: __lowerCamelCase : Dict = pool.map(_lowerCamelCase , _lowerCamelCase ) logger.info(F"""Finished {num_proc} processes""" ) __lowerCamelCase : Union[str, Any] = [obj for proc_res in mapped for obj in proc_res] logger.info(F"""Unpacked {len(_lowerCamelCase )} objects""" ) return mapped def lowercase_ ( _lowerCamelCase: Any , _lowerCamelCase: str , _lowerCamelCase: Union[str, Any] , _lowerCamelCase: Any , _lowerCamelCase: Any , _lowerCamelCase: List[str] , _lowerCamelCase: Union[str, Any] ) -> Optional[Any]: '''simple docstring''' import joblib with joblib.parallel_backend(ParallelBackendConfig.backend_name , n_jobs=_lowerCamelCase ): return joblib.Parallel()( joblib.delayed(_lowerCamelCase )((function, obj, types, None, True, None) ) for obj in iterable ) @experimental @contextlib.contextmanager def lowercase_ ( _lowerCamelCase: str ) -> int: '''simple docstring''' __lowerCamelCase : Optional[int] = backend_name if backend_name == "spark": from joblibspark import register_spark register_spark() # TODO: call create_cache_and_write_probe if "download" in steps # TODO: raise NotImplementedError when Dataset.map etc is called try: yield finally: __lowerCamelCase : int = None
64
"""simple docstring""" def lowercase_ ( _lowerCamelCase: int = 100 ) -> int: '''simple docstring''' __lowerCamelCase : Optional[Any] = set() __lowerCamelCase : Union[str, Any] = 0 __lowerCamelCase : Optional[Any] = n + 1 # maximum limit for a in range(2 , _lowerCamelCase ): for b in range(2 , _lowerCamelCase ): __lowerCamelCase : Union[str, Any] = a**b # calculates the current power collect_powers.add(_lowerCamelCase ) # adds the result to the set return len(_lowerCamelCase ) if __name__ == "__main__": print('''Number of terms ''', solution(int(str(input()).strip())))
64
1
"""simple docstring""" import os from collections.abc import Iterator def _snake_case ( snake_case__ : str = "." ): for dir_path, dir_names, filenames in os.walk(snake_case__ ): A = [d for d in dir_names if d != 'scripts' and d[0] not in '._'] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(snake_case__ )[1] in (".py", ".ipynb"): yield os.path.join(snake_case__ , snake_case__ ).lstrip('./' ) def _snake_case ( snake_case__ : Union[str, Any] ): return F'{i * " "}*' if i else "\n##" def _snake_case ( snake_case__ : str , snake_case__ : str ): A = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(snake_case__ ) or old_parts[i] != new_part) and new_part: print(F'{md_prefix(snake_case__ )} {new_part.replace("_" , " " ).title()}' ) return new_path def _snake_case ( snake_case__ : str = "." ): A = '' for filepath in sorted(good_file_paths(snake_case__ ) ): A , A = os.path.split(snake_case__ ) if filepath != old_path: A = print_path(snake_case__ , snake_case__ ) A = (filepath.count(os.sep ) + 1) if filepath else 0 A = F'{filepath}/{filename}'.replace(' ' , '%20' ) A = os.path.splitext(filename.replace('_' , ' ' ).title() )[0] print(F'{md_prefix(snake_case__ )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md('''.''')
74
"""simple docstring""" import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self : int ,A_ : List[Any] ) -> Optional[Any]: for model_result in results.values(): for batch_size, sequence_length in zip(model_result['bs'] ,model_result['ss'] ): A = model_result['result'][batch_size][sequence_length] self.assertIsNotNone(A_ ) def _SCREAMING_SNAKE_CASE ( self : int ) -> Optional[Any]: A = 'sshleifer/tiny-gpt2' A = PyTorchBenchmarkArguments( models=[MODEL_ID] ,training=A_ ,inference=A_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=A_ ,) A = PyTorchBenchmark(A_ ) A = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Optional[int]: A = 'sgugger/tiny-distilbert-classification' A = PyTorchBenchmarkArguments( models=[MODEL_ID] ,training=A_ ,inference=A_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=A_ ,only_pretrain_model=A_ ,) A = PyTorchBenchmark(A_ ) A = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> List[str]: A = 'sshleifer/tiny-gpt2' A = PyTorchBenchmarkArguments( models=[MODEL_ID] ,training=A_ ,inference=A_ ,torchscript=A_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=A_ ,) A = PyTorchBenchmark(A_ ) A = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == 'cpu' ,'Cant do half precision' ) def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> List[Any]: A = 'sshleifer/tiny-gpt2' A = PyTorchBenchmarkArguments( models=[MODEL_ID] ,training=A_ ,inference=A_ ,fpaa=A_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=A_ ,) A = PyTorchBenchmark(A_ ) A = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Optional[Any]: A = 'sshleifer/tiny-gpt2' A = AutoConfig.from_pretrained(A_ ) # set architectures equal to `None` A = None A = PyTorchBenchmarkArguments( models=[MODEL_ID] ,training=A_ ,inference=A_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=A_ ,) A = PyTorchBenchmark(A_ ,configs=[config] ) A = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> Optional[int]: A = 'sshleifer/tiny-gpt2' A = PyTorchBenchmarkArguments( models=[MODEL_ID] ,training=A_ ,inference=A_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=A_ ,) A = PyTorchBenchmark(A_ ) A = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == 'cpu' ,'Can\'t do half precision' ) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> List[Any]: A = 'sshleifer/tiny-gpt2' A = PyTorchBenchmarkArguments( models=[MODEL_ID] ,training=A_ ,inference=A_ ,sequence_lengths=[8] ,batch_sizes=[1] ,fpaa=A_ ,multi_process=A_ ,) A = PyTorchBenchmark(A_ ) A = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> Optional[int]: A = 'sshleifer/tiny-gpt2' A = AutoConfig.from_pretrained(A_ ) A = PyTorchBenchmarkArguments( models=[MODEL_ID] ,training=A_ ,inference=A_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=A_ ,) A = PyTorchBenchmark(A_ ,configs=[config] ) A = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _SCREAMING_SNAKE_CASE ( self : Any ) -> List[Any]: A = 'sshleifer/tinier_bart' A = AutoConfig.from_pretrained(A_ ) A = PyTorchBenchmarkArguments( models=[MODEL_ID] ,training=A_ ,inference=A_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=A_ ,) A = PyTorchBenchmark(A_ ,configs=[config] ) A = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _SCREAMING_SNAKE_CASE ( self : Any ) -> List[str]: A = 'sshleifer/tiny-gpt2' A = AutoConfig.from_pretrained(A_ ) A = PyTorchBenchmarkArguments( models=[MODEL_ID] ,training=A_ ,inference=A_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=A_ ,) A = PyTorchBenchmark(A_ ,configs=[config] ) A = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _SCREAMING_SNAKE_CASE ( self : Any ) -> List[str]: A = 'sshleifer/tinier_bart' A = AutoConfig.from_pretrained(A_ ) A = PyTorchBenchmarkArguments( models=[MODEL_ID] ,training=A_ ,inference=A_ ,sequence_lengths=[8] ,batch_sizes=[1] ,multi_process=A_ ,) A = PyTorchBenchmark(A_ ,configs=[config] ) A = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _SCREAMING_SNAKE_CASE ( self : Dict ) -> Dict: A = 'sshleifer/tiny-gpt2' with tempfile.TemporaryDirectory() as tmp_dir: A = PyTorchBenchmarkArguments( models=[MODEL_ID] ,training=A_ ,inference=A_ ,save_to_csv=A_ ,sequence_lengths=[8] ,batch_sizes=[1] ,inference_time_csv_file=os.path.join(A_ ,'inf_time.csv' ) ,train_memory_csv_file=os.path.join(A_ ,'train_mem.csv' ) ,inference_memory_csv_file=os.path.join(A_ ,'inf_mem.csv' ) ,train_time_csv_file=os.path.join(A_ ,'train_time.csv' ) ,env_info_csv_file=os.path.join(A_ ,'env.csv' ) ,multi_process=A_ ,) A = PyTorchBenchmark(A_ ) benchmark.run() self.assertTrue(Path(os.path.join(A_ ,'inf_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(A_ ,'train_time.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(A_ ,'inf_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(A_ ,'train_mem.csv' ) ).exists() ) self.assertTrue(Path(os.path.join(A_ ,'env.csv' ) ).exists() ) def _SCREAMING_SNAKE_CASE ( self : Dict ) -> List[str]: A = 'sshleifer/tiny-gpt2' def _check_summary_is_not_empty(A_ : Optional[int] ): self.assertTrue(hasattr(A_ ,'sequential' ) ) self.assertTrue(hasattr(A_ ,'cumulative' ) ) self.assertTrue(hasattr(A_ ,'current' ) ) self.assertTrue(hasattr(A_ ,'total' ) ) with tempfile.TemporaryDirectory() as tmp_dir: A = PyTorchBenchmarkArguments( models=[MODEL_ID] ,training=A_ ,inference=A_ ,sequence_lengths=[8] ,batch_sizes=[1] ,log_filename=os.path.join(A_ ,'log.txt' ) ,log_print=A_ ,trace_memory_line_by_line=A_ ,multi_process=A_ ,) A = PyTorchBenchmark(A_ ) A = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(A_ ,'log.txt' ) ).exists() )
74
1
"""simple docstring""" import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCamelCase__: def __init__( self: Union[str, Any] , UpperCamelCase_: str , UpperCamelCase_: Union[str, Any]=2 , UpperCamelCase_: Dict=8 , UpperCamelCase_: List[Any]=True , UpperCamelCase_: Dict=True , UpperCamelCase_: Tuple=True , UpperCamelCase_: Any=True , UpperCamelCase_: Dict=99 , UpperCamelCase_: List[Any]=16 , UpperCamelCase_: Union[str, Any]=5 , UpperCamelCase_: List[Any]=2 , UpperCamelCase_: Union[str, Any]=36 , UpperCamelCase_: Union[str, Any]="gelu" , UpperCamelCase_: Any=0.0 , UpperCamelCase_: Union[str, Any]=0.0 , UpperCamelCase_: Dict=5_12 , UpperCamelCase_: Union[str, Any]=16 , UpperCamelCase_: Optional[int]=2 , UpperCamelCase_: Tuple=0.02 , UpperCamelCase_: Optional[int]=3 , UpperCamelCase_: str=4 , UpperCamelCase_: List[Any]=None , ): __lowerCamelCase = parent __lowerCamelCase = batch_size __lowerCamelCase = seq_length __lowerCamelCase = is_training __lowerCamelCase = use_input_mask __lowerCamelCase = use_token_type_ids __lowerCamelCase = use_labels __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_act __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = type_sequence_label_size __lowerCamelCase = initializer_range __lowerCamelCase = num_labels __lowerCamelCase = num_choices __lowerCamelCase = scope def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCamelCase = None if self.use_input_mask: __lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCamelCase = None if self.use_token_type_ids: __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None if self.use_labels: __lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCamelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCAmelCase__ ( self: Any ): return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , ) def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = self.get_config() __lowerCamelCase = 3_00 return config def lowerCAmelCase__ ( self: List[str] ): ( ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ) = self.prepare_config_and_inputs() __lowerCamelCase = True __lowerCamelCase = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: List[str] , UpperCamelCase_: Any , UpperCamelCase_: Any , UpperCamelCase_: List[str] , UpperCamelCase_: Dict , UpperCamelCase_: List[str] , UpperCamelCase_: Tuple ): __lowerCamelCase = MraModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowerCamelCase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) __lowerCamelCase = model(_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) __lowerCamelCase = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__ ( self: int , UpperCamelCase_: Tuple , UpperCamelCase_: int , UpperCamelCase_: Any , UpperCamelCase_: str , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: int , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Optional[int] , ): __lowerCamelCase = True __lowerCamelCase = MraModel(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowerCamelCase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , encoder_attention_mask=_lowerCAmelCase , ) __lowerCamelCase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , ) __lowerCamelCase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: Tuple , UpperCamelCase_: Dict , UpperCamelCase_: List[Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: List[str] , UpperCamelCase_: List[Any] , UpperCamelCase_: List[Any] ): __lowerCamelCase = MraForMaskedLM(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowerCamelCase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: str , UpperCamelCase_: Optional[int] , UpperCamelCase_: str , UpperCamelCase_: str , UpperCamelCase_: List[str] , UpperCamelCase_: List[str] , UpperCamelCase_: List[Any] ): __lowerCamelCase = MraForQuestionAnswering(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowerCamelCase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , start_positions=_lowerCAmelCase , end_positions=_lowerCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCAmelCase__ ( self: List[Any] , UpperCamelCase_: Tuple , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Dict , UpperCamelCase_: List[str] , UpperCamelCase_: Tuple , UpperCamelCase_: Tuple , UpperCamelCase_: Optional[Any] ): __lowerCamelCase = self.num_labels __lowerCamelCase = MraForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowerCamelCase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Any , UpperCamelCase_: Dict , UpperCamelCase_: Optional[Any] , UpperCamelCase_: str , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: List[Any] ): __lowerCamelCase = self.num_labels __lowerCamelCase = MraForTokenClassification(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowerCamelCase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCAmelCase__ ( self: int , UpperCamelCase_: str , UpperCamelCase_: Any , UpperCamelCase_: Dict , UpperCamelCase_: Tuple , UpperCamelCase_: Tuple , UpperCamelCase_: Dict , UpperCamelCase_: Optional[int] ): __lowerCamelCase = self.num_choices __lowerCamelCase = MraForMultipleChoice(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() __lowerCamelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCamelCase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCamelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCamelCase = model( _lowerCAmelCase , attention_mask=_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = self.prepare_config_and_inputs() ( ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ( __lowerCamelCase ), ) = config_and_inputs __lowerCamelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class lowerCamelCase__( __lowerCamelCase , unittest.TestCase): UpperCAmelCase__ : Dict = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) UpperCAmelCase__ : int = False UpperCAmelCase__ : Optional[Any] = False UpperCAmelCase__ : Optional[int] = False UpperCAmelCase__ : Any = False UpperCAmelCase__ : Optional[Any] = () def lowerCAmelCase__ ( self: Tuple ): __lowerCamelCase = MraModelTester(self ) __lowerCamelCase = ConfigTester(self , config_class=_lowerCAmelCase , hidden_size=37 ) def lowerCAmelCase__ ( self: Tuple ): self.config_tester.run_common_tests() def lowerCAmelCase__ ( self: List[Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def lowerCAmelCase__ ( self: Any ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __lowerCamelCase = type self.model_tester.create_and_check_model(*_lowerCAmelCase ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_lowerCAmelCase ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*_lowerCAmelCase ) def lowerCAmelCase__ ( self: str ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_lowerCAmelCase ) def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_lowerCAmelCase ) def lowerCAmelCase__ ( self: Optional[Any] ): __lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_lowerCAmelCase ) @slow def lowerCAmelCase__ ( self: Union[str, Any] ): for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCamelCase = MraModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @unittest.skip(reason="""MRA does not output attentions""" ) def lowerCAmelCase__ ( self: Dict ): return @require_torch class lowerCamelCase__( unittest.TestCase): @slow def lowerCAmelCase__ ( self: int ): __lowerCamelCase = MraModel.from_pretrained("""uw-madison/mra-base-512-4""" ) __lowerCamelCase = torch.arange(2_56 ).unsqueeze(0 ) with torch.no_grad(): __lowerCamelCase = model(_lowerCAmelCase )[0] __lowerCamelCase = torch.Size((1, 2_56, 7_68) ) self.assertEqual(output.shape , _lowerCAmelCase ) __lowerCamelCase = torch.tensor( [[[-0.0140, 0.0830, -0.0381], [0.1546, 0.1402, 0.0220], [0.1162, 0.0851, 0.0165]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , _lowerCAmelCase , atol=1E-4 ) ) @slow def lowerCAmelCase__ ( self: str ): __lowerCamelCase = MraForMaskedLM.from_pretrained("""uw-madison/mra-base-512-4""" ) __lowerCamelCase = torch.arange(2_56 ).unsqueeze(0 ) with torch.no_grad(): __lowerCamelCase = model(_lowerCAmelCase )[0] __lowerCamelCase = 5_02_65 __lowerCamelCase = torch.Size((1, 2_56, vocab_size) ) self.assertEqual(output.shape , _lowerCAmelCase ) __lowerCamelCase = torch.tensor( [[[9.2595, -3.6038, 11.8819], [9.3869, -3.2693, 11.0956], [11.8524, -3.4938, 13.1210]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , _lowerCAmelCase , atol=1E-4 ) ) @slow def lowerCAmelCase__ ( self: int ): __lowerCamelCase = MraForMaskedLM.from_pretrained("""uw-madison/mra-base-4096-8-d3""" ) __lowerCamelCase = torch.arange(40_96 ).unsqueeze(0 ) with torch.no_grad(): __lowerCamelCase = model(_lowerCAmelCase )[0] __lowerCamelCase = 5_02_65 __lowerCamelCase = torch.Size((1, 40_96, vocab_size) ) self.assertEqual(output.shape , _lowerCAmelCase ) __lowerCamelCase = torch.tensor( [[[5.4789, -2.3564, 7.5064], [7.9067, -1.3369, 9.9668], [9.0712, -1.8106, 7.0380]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , _lowerCAmelCase , atol=1E-4 ) )
353
UpperCAmelCase_ = {'a': ['c', 'b'], 'b': ['d', 'e'], 'c': [], 'd': [], 'e': []} UpperCAmelCase_ = ['a', 'b', 'c', 'd', 'e'] def lowerCamelCase__ ( A__ : Union[str, Any] , A__ : Optional[int] , A__ : str ): '''simple docstring''' __lowerCamelCase = start # add current to visited visited.append(A__ ) __lowerCamelCase = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __lowerCamelCase = topological_sort(A__ , A__ , A__ ) # if all neighbors visited add current to sort sort.append(A__ ) # if all vertices haven't been visited select a new one to visit if len(A__ ) != len(A__ ): for vertice in vertices: if vertice not in visited: __lowerCamelCase = topological_sort(A__ , A__ , A__ ) # return sort return sort if __name__ == "__main__": UpperCAmelCase_ = topological_sort('a', [], []) print(sort)
29
0
class __lowerCAmelCase : def __init__( self : Dict , snake_case__ : Optional[int] , snake_case__ : Any ): """simple docstring""" _UpperCAmelCase = name _UpperCAmelCase = val def __str__( self : Tuple ): """simple docstring""" return F"""{self.__class__.__name__}({self.name}, {self.val})""" def __lt__( self : List[str] , snake_case__ : int ): """simple docstring""" return self.val < other.val class __lowerCAmelCase : def __init__( self : Union[str, Any] , snake_case__ : str ): """simple docstring""" _UpperCAmelCase = {} _UpperCAmelCase = {} _UpperCAmelCase = self.build_heap(lowerCamelCase__ ) def __getitem__( self : int , snake_case__ : List[Any] ): """simple docstring""" return self.get_value(lowerCamelCase__ ) def UpperCamelCase ( self : Dict , snake_case__ : Optional[int] ): """simple docstring""" return (idx - 1) // 2 def UpperCamelCase ( self : Optional[int] , snake_case__ : Any ): """simple docstring""" return idx * 2 + 1 def UpperCamelCase ( self : Any , snake_case__ : Dict ): """simple docstring""" return idx * 2 + 2 def UpperCamelCase ( self : Dict , snake_case__ : str ): """simple docstring""" return self.heap_dict[key] def UpperCamelCase ( self : Any , snake_case__ : int ): """simple docstring""" _UpperCAmelCase = len(lowerCamelCase__ ) - 1 _UpperCAmelCase = self.get_parent_idx(lowerCamelCase__ ) for idx, i in enumerate(lowerCamelCase__ ): _UpperCAmelCase = idx _UpperCAmelCase = i.val for i in range(lowerCamelCase__ , -1 , -1 ): self.sift_down(lowerCamelCase__ , lowerCamelCase__ ) return array def UpperCamelCase ( self : str , snake_case__ : Any , snake_case__ : Any ): """simple docstring""" while True: _UpperCAmelCase = self.get_left_child_idx(lowerCamelCase__ ) # noqa: E741 _UpperCAmelCase = self.get_right_child_idx(lowerCamelCase__ ) _UpperCAmelCase = idx if l < len(lowerCamelCase__ ) and array[l] < array[idx]: _UpperCAmelCase = l if r < len(lowerCamelCase__ ) and array[r] < array[smallest]: _UpperCAmelCase = r if smallest != idx: _UpperCAmelCase , _UpperCAmelCase = array[smallest], array[idx] ( ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ) = ( self.idx_of_element[array[smallest]], self.idx_of_element[array[idx]], ) _UpperCAmelCase = smallest else: break def UpperCamelCase ( self : Tuple , snake_case__ : Dict ): """simple docstring""" _UpperCAmelCase = self.get_parent_idx(lowerCamelCase__ ) while p >= 0 and self.heap[p] > self.heap[idx]: _UpperCAmelCase , _UpperCAmelCase = self.heap[idx], self.heap[p] _UpperCAmelCase , _UpperCAmelCase = ( self.idx_of_element[self.heap[idx]], self.idx_of_element[self.heap[p]], ) _UpperCAmelCase = p _UpperCAmelCase = self.get_parent_idx(lowerCamelCase__ ) def UpperCamelCase ( self : Tuple ): """simple docstring""" return self.heap[0] def UpperCamelCase ( self : List[str] ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = self.heap[-1], self.heap[0] _UpperCAmelCase , _UpperCAmelCase = ( self.idx_of_element[self.heap[-1]], self.idx_of_element[self.heap[0]], ) _UpperCAmelCase = self.heap.pop() del self.idx_of_element[x] self.sift_down(0 , self.heap ) return x def UpperCamelCase ( self : List[str] , snake_case__ : Union[str, Any] ): """simple docstring""" self.heap.append(lowerCamelCase__ ) _UpperCAmelCase = len(self.heap ) - 1 _UpperCAmelCase = node.val self.sift_up(len(self.heap ) - 1 ) def UpperCamelCase ( self : Optional[int] ): """simple docstring""" return len(self.heap ) == 0 def UpperCamelCase ( self : Dict , snake_case__ : Tuple , snake_case__ : Union[str, Any] ): """simple docstring""" assert ( self.heap[self.idx_of_element[node]].val > new_value ), "newValue must be less that current value" _UpperCAmelCase = new_value _UpperCAmelCase = new_value self.sift_up(self.idx_of_element[node] ) lowercase_ : int = Node('R', -1) lowercase_ : str = Node('B', 6) lowercase_ : Dict = Node('A', 3) lowercase_ : Any = Node('X', 1) lowercase_ : int = Node('E', 4) # Use one of these two ways to generate Min-Heap # Generating Min-Heap from array lowercase_ : Optional[Any] = MinHeap([r, b, a, x, e]) # Generating Min-Heap by Insert method # myMinHeap.insert(a) # myMinHeap.insert(b) # myMinHeap.insert(x) # myMinHeap.insert(r) # myMinHeap.insert(e) # Before print('Min Heap - before decrease key') for i in my_min_heap.heap: print(i) print('Min Heap - After decrease key of node [B -> -17]') my_min_heap.decrease_key(b, -17) # After for i in my_min_heap.heap: print(i) if __name__ == "__main__": import doctest doctest.testmod()
133
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available __A = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = ["GPTSw3Tokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys __A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
164
0
import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class lowerCamelCase (unittest.TestCase ): """simple docstring""" def A_ ( self : str ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = mock.Mock() SCREAMING_SNAKE_CASE__ : int = 5_0_0 SCREAMING_SNAKE_CASE__ : List[str] = {} SCREAMING_SNAKE_CASE__ : Tuple = HTTPError SCREAMING_SNAKE_CASE__ : Any = {} # Download this model to make sure it's in the cache. SCREAMING_SNAKE_CASE__ : Optional[int] = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request", return_value=_UpperCAmelCase ) as mock_head: SCREAMING_SNAKE_CASE__ : Optional[int] = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def A_ ( self : Union[str, Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[str] = mock.Mock() SCREAMING_SNAKE_CASE__ : Any = 5_0_0 SCREAMING_SNAKE_CASE__ : int = {} SCREAMING_SNAKE_CASE__ : Optional[Any] = HTTPError SCREAMING_SNAKE_CASE__ : Dict = {} # Download this model to make sure it's in the cache. SCREAMING_SNAKE_CASE__ : List[str] = GPTaTokenizerFast.from_pretrained("gpt2" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request", return_value=_UpperCAmelCase ) as mock_head: SCREAMING_SNAKE_CASE__ : Tuple = GPTaTokenizerFast.from_pretrained("gpt2" ) # This check we did call the fake head request mock_head.assert_called() def A_ ( self : Tuple ) -> List[str]: """simple docstring""" # This test is for deprecated behavior and can be removed in v5 try: SCREAMING_SNAKE_CASE__ : Dict = tempfile.mktemp() with open(_UpperCAmelCase, "wb" ) as f: http_get("https://huggingface.co/albert-base-v1/resolve/main/spiece.model", _UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : List[Any] = AlbertTokenizer.from_pretrained(_UpperCAmelCase ) finally: os.remove(_UpperCAmelCase ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile("tokenizer.json" ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open("tokenizer.json", "wb" ) as f: http_get("https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json", _UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : Tuple = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size, 1_0_0_0 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove("tokenizer.json" ) def A_ ( self : str ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[str] = AlbertTokenizer.from_pretrained("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" ) @is_staging_test class lowerCamelCase (unittest.TestCase ): """simple docstring""" UpperCAmelCase_ = ["[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "bla", "blou"] @classmethod def A_ ( cls : str ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = TOKEN HfFolder.save_token(_UpperCAmelCase ) @classmethod def A_ ( cls : List[str] ) -> int: """simple docstring""" try: delete_repo(token=cls._token, repo_id="test-tokenizer" ) except HTTPError: pass try: delete_repo(token=cls._token, repo_id="valid_org/test-tokenizer-org" ) except HTTPError: pass try: delete_repo(token=cls._token, repo_id="test-dynamic-tokenizer" ) except HTTPError: pass def A_ ( self : Any ) -> str: """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE__ : Optional[Any] = os.path.join(_UpperCAmelCase, "vocab.txt" ) with open(_UpperCAmelCase, "w", encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) SCREAMING_SNAKE_CASE__ : Tuple = BertTokenizer(_UpperCAmelCase ) tokenizer.push_to_hub("test-tokenizer", use_auth_token=self._token ) SCREAMING_SNAKE_CASE__ : Optional[int] = BertTokenizer.from_pretrained(F'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab, tokenizer.vocab ) # Reset repo delete_repo(token=self._token, repo_id="test-tokenizer" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(_UpperCAmelCase, repo_id="test-tokenizer", push_to_hub=_UpperCAmelCase, use_auth_token=self._token ) SCREAMING_SNAKE_CASE__ : str = BertTokenizer.from_pretrained(F'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab, tokenizer.vocab ) def A_ ( self : List[Any] ) -> str: """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE__ : List[str] = os.path.join(_UpperCAmelCase, "vocab.txt" ) with open(_UpperCAmelCase, "w", encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) SCREAMING_SNAKE_CASE__ : str = BertTokenizer(_UpperCAmelCase ) tokenizer.push_to_hub("valid_org/test-tokenizer-org", use_auth_token=self._token ) SCREAMING_SNAKE_CASE__ : int = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab, tokenizer.vocab ) # Reset repo delete_repo(token=self._token, repo_id="valid_org/test-tokenizer-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( _UpperCAmelCase, repo_id="valid_org/test-tokenizer-org", push_to_hub=_UpperCAmelCase, use_auth_token=self._token ) SCREAMING_SNAKE_CASE__ : Dict = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab, tokenizer.vocab ) @require_tokenizers def A_ ( self : int ) -> Tuple: """simple docstring""" CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE__ : int = os.path.join(_UpperCAmelCase, "vocab.txt" ) with open(_UpperCAmelCase, "w", encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) SCREAMING_SNAKE_CASE__ : Dict = CustomTokenizer(_UpperCAmelCase ) # No fast custom tokenizer tokenizer.push_to_hub("test-dynamic-tokenizer", use_auth_token=self._token ) SCREAMING_SNAKE_CASE__ : Optional[Any] = AutoTokenizer.from_pretrained(F'''{USER}/test-dynamic-tokenizer''', trust_remote_code=_UpperCAmelCase ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__, "CustomTokenizer" ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE__ : int = os.path.join(_UpperCAmelCase, "vocab.txt" ) with open(_UpperCAmelCase, "w", encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) SCREAMING_SNAKE_CASE__ : Optional[Any] = BertTokenizerFast.from_pretrained(_UpperCAmelCase ) bert_tokenizer.save_pretrained(_UpperCAmelCase ) SCREAMING_SNAKE_CASE__ : List[Any] = CustomTokenizerFast.from_pretrained(_UpperCAmelCase ) tokenizer.push_to_hub("test-dynamic-tokenizer", use_auth_token=self._token ) SCREAMING_SNAKE_CASE__ : Optional[int] = AutoTokenizer.from_pretrained(F'''{USER}/test-dynamic-tokenizer''', trust_remote_code=_UpperCAmelCase ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__, "CustomTokenizerFast" ) SCREAMING_SNAKE_CASE__ : List[str] = AutoTokenizer.from_pretrained( F'''{USER}/test-dynamic-tokenizer''', use_fast=_UpperCAmelCase, trust_remote_code=_UpperCAmelCase ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__, "CustomTokenizer" ) class lowerCamelCase (unittest.TestCase ): """simple docstring""" def A_ ( self : List[str] ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = Trie() trie.add("Hello 友達" ) self.assertEqual(trie.data, {"H": {"e": {"l": {"l": {"o": {" ": {"友": {"達": {"": 1}}}}}}}}} ) trie.add("Hello" ) trie.data self.assertEqual(trie.data, {"H": {"e": {"l": {"l": {"o": {"": 1, " ": {"友": {"達": {"": 1}}}}}}}}} ) def A_ ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[str] = Trie() self.assertEqual(trie.split("[CLS] This is a extra_id_100" ), ["[CLS] This is a extra_id_100"] ) trie.add("[CLS]" ) trie.add("extra_id_1" ) trie.add("extra_id_100" ) self.assertEqual(trie.split("[CLS] This is a extra_id_100" ), ["[CLS]", " This is a ", "extra_id_100"] ) def A_ ( self : int ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = Trie() trie.add("A" ) self.assertEqual(trie.split("ABC" ), ["A", "BC"] ) self.assertEqual(trie.split("BCA" ), ["BC", "A"] ) def A_ ( self : Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : Optional[Any] = Trie() trie.add("TOKEN]" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ), ["This is something ", "[SPECIAL_TOKEN]"] ) def A_ ( self : List[str] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = Trie() trie.add("A" ) trie.add("P" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ), ["This is something ", "[SPECIAL_TOKEN]"] ) def A_ ( self : Tuple ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = Trie() trie.add("AB" ) trie.add("B" ) trie.add("C" ) self.assertEqual(trie.split("ABC" ), ["AB", "C"] ) def A_ ( self : Union[str, Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : Any = Trie() trie.add("ABC" ) trie.add("B" ) trie.add("CD" ) self.assertEqual(trie.split("ABCD" ), ["ABC", "D"] ) def A_ ( self : Optional[int] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = Trie() SCREAMING_SNAKE_CASE__ : Optional[Any] = trie.cut_text("ABC", [0, 0, 2, 1, 2, 3] ) self.assertEqual(_UpperCAmelCase, ["AB", "C"] )
353
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _lowerCamelCase : int = logging.get_logger(__name__) _lowerCamelCase : Optional[Any] = '''▁''' _lowerCamelCase : Dict = {'''vocab_file''': '''sentencepiece.bpe.model'''} _lowerCamelCase : int = { '''vocab_file''': { '''xlm-roberta-base''': '''https://huggingface.co/xlm-roberta-base/resolve/main/sentencepiece.bpe.model''', '''xlm-roberta-large''': '''https://huggingface.co/xlm-roberta-large/resolve/main/sentencepiece.bpe.model''', '''xlm-roberta-large-finetuned-conll02-dutch''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/sentencepiece.bpe.model''' ), '''xlm-roberta-large-finetuned-conll02-spanish''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/sentencepiece.bpe.model''' ), '''xlm-roberta-large-finetuned-conll03-english''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/sentencepiece.bpe.model''' ), '''xlm-roberta-large-finetuned-conll03-german''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/sentencepiece.bpe.model''' ), } } _lowerCamelCase : Optional[Any] = { '''xlm-roberta-base''': 5_1_2, '''xlm-roberta-large''': 5_1_2, '''xlm-roberta-large-finetuned-conll02-dutch''': 5_1_2, '''xlm-roberta-large-finetuned-conll02-spanish''': 5_1_2, '''xlm-roberta-large-finetuned-conll03-english''': 5_1_2, '''xlm-roberta-large-finetuned-conll03-german''': 5_1_2, } class lowerCamelCase (__lowerCamelCase ): """simple docstring""" UpperCAmelCase_ = VOCAB_FILES_NAMES UpperCAmelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase_ = ["input_ids", "attention_mask"] def __init__( self : Dict, _UpperCAmelCase : str, _UpperCAmelCase : Optional[int]="<s>", _UpperCAmelCase : Optional[int]="</s>", _UpperCAmelCase : Dict="</s>", _UpperCAmelCase : List[Any]="<s>", _UpperCAmelCase : Union[str, Any]="<unk>", _UpperCAmelCase : List[Any]="<pad>", _UpperCAmelCase : str="<mask>", _UpperCAmelCase : Optional[Dict[str, Any]] = None, **_UpperCAmelCase : List[Any], ) -> None: """simple docstring""" # Mask token behave like a normal word, i.e. include the space before it SCREAMING_SNAKE_CASE__ : int = AddedToken(_UpperCAmelCase, lstrip=_UpperCAmelCase, rstrip=_UpperCAmelCase ) if isinstance(_UpperCAmelCase, _UpperCAmelCase ) else mask_token SCREAMING_SNAKE_CASE__ : List[str] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_UpperCAmelCase, eos_token=_UpperCAmelCase, unk_token=_UpperCAmelCase, sep_token=_UpperCAmelCase, cls_token=_UpperCAmelCase, pad_token=_UpperCAmelCase, mask_token=_UpperCAmelCase, sp_model_kwargs=self.sp_model_kwargs, **_UpperCAmelCase, ) SCREAMING_SNAKE_CASE__ : str = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_UpperCAmelCase ) ) SCREAMING_SNAKE_CASE__ : Tuple = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token SCREAMING_SNAKE_CASE__ : List[str] = {"<s>": 0, "<pad>": 1, "</s>": 2, "<unk>": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab SCREAMING_SNAKE_CASE__ : Dict = 1 SCREAMING_SNAKE_CASE__ : int = len(self.sp_model ) + self.fairseq_offset SCREAMING_SNAKE_CASE__ : Optional[int] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : str ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : List[Any] = self.__dict__.copy() SCREAMING_SNAKE_CASE__ : List[Any] = None SCREAMING_SNAKE_CASE__ : Dict = self.sp_model.serialized_model_proto() return state def __setstate__( self : int, _UpperCAmelCase : List[Any] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE__ : Tuple = d # for backward compatibility if not hasattr(self, "sp_model_kwargs" ): SCREAMING_SNAKE_CASE__ : Dict = {} SCREAMING_SNAKE_CASE__ : Union[str, Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def A_ ( self : Any, _UpperCAmelCase : List[int], _UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] SCREAMING_SNAKE_CASE__ : List[str] = [self.cls_token_id] SCREAMING_SNAKE_CASE__ : Union[str, Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A_ ( self : List[Any], _UpperCAmelCase : List[int], _UpperCAmelCase : Optional[List[int]] = None, _UpperCAmelCase : bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_UpperCAmelCase, token_ids_a=_UpperCAmelCase, already_has_special_tokens=_UpperCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(_UpperCAmelCase )) + [1] return [1] + ([0] * len(_UpperCAmelCase )) + [1, 1] + ([0] * len(_UpperCAmelCase )) + [1] def A_ ( self : Union[str, Any], _UpperCAmelCase : List[int], _UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = [self.sep_token_id] SCREAMING_SNAKE_CASE__ : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def A_ ( self : List[str] ) -> List[str]: """simple docstring""" return len(self.sp_model ) + self.fairseq_offset + 1 # Add the <mask> token def A_ ( self : List[Any] ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = {self.convert_ids_to_tokens(_UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def A_ ( self : List[str], _UpperCAmelCase : str ) -> List[str]: """simple docstring""" return self.sp_model.encode(_UpperCAmelCase, out_type=_UpperCAmelCase ) def A_ ( self : Optional[Any], _UpperCAmelCase : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.sp_model.PieceToId(_UpperCAmelCase ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def A_ ( self : Tuple, _UpperCAmelCase : List[str] ) -> List[str]: """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def A_ ( self : Any, _UpperCAmelCase : int ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = "".join(_UpperCAmelCase ).replace(_UpperCAmelCase, " " ).strip() return out_string def A_ ( self : Union[str, Any], _UpperCAmelCase : str, _UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(_UpperCAmelCase ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return SCREAMING_SNAKE_CASE__ : Optional[Any] = os.path.join( _UpperCAmelCase, (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file, _UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(_UpperCAmelCase, "wb" ) as fi: SCREAMING_SNAKE_CASE__ : Any = self.sp_model.serialized_model_proto() fi.write(_UpperCAmelCase ) return (out_vocab_file,)
191
0
"""simple docstring""" from __future__ import annotations class lowerCAmelCase_ : """simple docstring""" def __init__( self , lowerCAmelCase ): """simple docstring""" snake_case = data snake_case = None snake_case = None def lowerCAmelCase__ ( _UpperCamelCase : Node | None ) -> None: # In Order traversal of the tree """simple docstring""" if tree: display(tree.left ) print(tree.data ) display(tree.right ) def lowerCAmelCase__ ( _UpperCamelCase : Node | None ) -> int: """simple docstring""" return 1 + max(depth_of_tree(tree.left ) , depth_of_tree(tree.right ) ) if tree else 0 def lowerCAmelCase__ ( _UpperCamelCase : Node ) -> bool: """simple docstring""" if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def lowerCAmelCase__ ( ) -> None: # Main function for testing. """simple docstring""" snake_case = Node(1 ) snake_case = Node(2 ) snake_case = Node(3 ) snake_case = Node(4 ) snake_case = Node(5 ) snake_case = Node(6 ) snake_case = Node(7 ) snake_case = Node(8 ) snake_case = Node(9 ) print(is_full_binary_tree(_UpperCamelCase ) ) print(depth_of_tree(_UpperCamelCase ) ) print('Tree is: ' ) display(_UpperCamelCase ) if __name__ == "__main__": main()
150
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) def lowerCAmelCase__ ( _UpperCamelCase : Optional[Any] , _UpperCamelCase : Optional[int]=False ) -> Optional[Any]: """simple docstring""" snake_case = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""deit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""deit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""deit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""deit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""deit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""deit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""deit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""deit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""deit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""deit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('cls_token', 'deit.embeddings.cls_token'), ('dist_token', 'deit.embeddings.distillation_token'), ('patch_embed.proj.weight', 'deit.embeddings.patch_embeddings.projection.weight'), ('patch_embed.proj.bias', 'deit.embeddings.patch_embeddings.projection.bias'), ('pos_embed', 'deit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ('pre_logits.fc.weight', 'pooler.dense.weight'), ('pre_logits.fc.bias', 'pooler.dense.bias'), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" snake_case = [(pair[0], pair[1][4:]) if pair[1].startswith('deit' ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ('norm.weight', 'deit.layernorm.weight'), ('norm.bias', 'deit.layernorm.bias'), ('head.weight', 'cls_classifier.weight'), ('head.bias', 'cls_classifier.bias'), ('head_dist.weight', 'distillation_classifier.weight'), ('head_dist.bias', 'distillation_classifier.bias'), ] ) return rename_keys def lowerCAmelCase__ ( _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Tuple , _UpperCamelCase : Tuple=False ) -> Union[str, Any]: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: snake_case = '' else: snake_case = 'deit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) snake_case = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) snake_case = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict snake_case = in_proj_weight[ : config.hidden_size, : ] snake_case = in_proj_bias[: config.hidden_size] snake_case = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] snake_case = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] snake_case = in_proj_weight[ -config.hidden_size :, : ] snake_case = in_proj_bias[-config.hidden_size :] def lowerCAmelCase__ ( _UpperCamelCase : List[Any] , _UpperCamelCase : Dict , _UpperCamelCase : int ) -> Any: """simple docstring""" snake_case = dct.pop(_UpperCamelCase ) snake_case = val def lowerCAmelCase__ ( ) -> Dict: """simple docstring""" snake_case = 'http://images.cocodataset.org/val2017/000000039769.jpg' snake_case = Image.open(requests.get(_UpperCamelCase , stream=_UpperCamelCase ).raw ) return im @torch.no_grad() def lowerCAmelCase__ ( _UpperCamelCase : Dict , _UpperCamelCase : int ) -> Optional[Any]: """simple docstring""" snake_case = DeiTConfig() # all deit models have fine-tuned heads snake_case = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size snake_case = 1_0_0_0 snake_case = 'huggingface/label-files' snake_case = 'imagenet-1k-id2label.json' snake_case = json.load(open(hf_hub_download(_UpperCamelCase , _UpperCamelCase , repo_type='dataset' ) , 'r' ) ) snake_case = {int(_UpperCamelCase ): v for k, v in idalabel.items()} snake_case = idalabel snake_case = {v: k for k, v in idalabel.items()} snake_case = int(deit_name[-6:-4] ) snake_case = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith('tiny' ): snake_case = 1_9_2 snake_case = 7_6_8 snake_case = 1_2 snake_case = 3 elif deit_name[9:].startswith('small' ): snake_case = 3_8_4 snake_case = 1_5_3_6 snake_case = 1_2 snake_case = 6 if deit_name[9:].startswith('base' ): pass elif deit_name[4:].startswith('large' ): snake_case = 1_0_2_4 snake_case = 4_0_9_6 snake_case = 2_4 snake_case = 1_6 # load original model from timm snake_case = timm.create_model(_UpperCamelCase , pretrained=_UpperCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys snake_case = timm_model.state_dict() snake_case = create_rename_keys(_UpperCamelCase , _UpperCamelCase ) for src, dest in rename_keys: rename_key(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) read_in_q_k_v(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) # load HuggingFace model snake_case = DeiTForImageClassificationWithTeacher(_UpperCamelCase ).eval() model.load_state_dict(_UpperCamelCase ) # Check outputs on an image, prepared by DeiTImageProcessor snake_case = int( (2_5_6 / 2_2_4) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 snake_case = DeiTImageProcessor(size=_UpperCamelCase , crop_size=config.image_size ) snake_case = image_processor(images=prepare_img() , return_tensors='pt' ) snake_case = encoding['pixel_values'] snake_case = model(_UpperCamelCase ) snake_case = timm_model(_UpperCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_UpperCamelCase , outputs.logits , atol=1e-3 ) Path(_UpperCamelCase ).mkdir(exist_ok=_UpperCamelCase ) print(f"""Saving model {deit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCamelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_UpperCamelCase ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--deit_name", default="vit_deit_base_distilled_patch16_224", type=str, help="Name of the DeiT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) SCREAMING_SNAKE_CASE__ = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
150
1
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation lowerCamelCase_ : List[Any] = logging.get_logger(__name__) lowerCamelCase_ : int = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} lowerCamelCase_ : int = { """tokenizer_file""": { """EleutherAI/gpt-neox-20b""": """https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/tokenizer.json""", }, } lowerCamelCase_ : List[Any] = { """gpt-neox-20b""": 2_048, } class a__ ( __snake_case ): A__ : List[str] = VOCAB_FILES_NAMES A__ : Any = PRETRAINED_VOCAB_FILES_MAP A__ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A__ : Union[str, Any] = ['input_ids', 'attention_mask'] def __init__( self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase="<|endoftext|>" , UpperCAmelCase="<|endoftext|>" , UpperCAmelCase="<|endoftext|>" , UpperCAmelCase=False , **UpperCAmelCase , ) -> Union[str, Any]: super().__init__( UpperCAmelCase , UpperCAmelCase , tokenizer_file=UpperCAmelCase , unk_token=UpperCAmelCase , bos_token=UpperCAmelCase , eos_token=UpperCAmelCase , add_prefix_space=UpperCAmelCase , **UpperCAmelCase , ) __a = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , UpperCAmelCase ) != add_prefix_space: __a = getattr(UpperCAmelCase , pre_tok_state.pop('type' ) ) __a = add_prefix_space __a = pre_tok_class(**UpperCAmelCase ) __a = add_prefix_space def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase , UpperCAmelCase = None ) -> Tuple[str]: __a = self._tokenizer.model.save(UpperCAmelCase , name=UpperCAmelCase ) return tuple(UpperCAmelCase ) def __SCREAMING_SNAKE_CASE ( self , UpperCAmelCase ) -> List[int]: __a = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(UpperCAmelCase , add_special_tokens=UpperCAmelCase ) + [self.eos_token_id] ) if len(UpperCAmelCase ) > self.model_max_length: __a = input_ids[-self.model_max_length :] return input_ids
369
from __future__ import annotations import numpy as np def lowerCAmelCase( __lowerCamelCase ): return np.maximum(0 , __lowerCamelCase ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
197
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCAmelCase : List[str] = { 'configuration_mobilenet_v2': [ 'MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MobileNetV2Config', 'MobileNetV2OnnxConfig', ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Dict = ['MobileNetV2FeatureExtractor'] UpperCAmelCase : Dict = ['MobileNetV2ImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Any = [ 'MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST', 'MobileNetV2ForImageClassification', 'MobileNetV2ForSemanticSegmentation', 'MobileNetV2Model', 'MobileNetV2PreTrainedModel', 'load_tf_weights_in_mobilenet_v2', ] if TYPE_CHECKING: from .configuration_mobilenet_va import ( MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileNetVaConfig, MobileNetVaOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilenet_va import MobileNetVaFeatureExtractor from .image_processing_mobilenet_va import MobileNetVaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilenet_va import ( MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel, MobileNetVaPreTrainedModel, load_tf_weights_in_mobilenet_va, ) else: import sys UpperCAmelCase : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
267
'''simple docstring''' import logging import os import threading import time try: import warnings except ImportError: UpperCAmelCase : Optional[Any] = None try: import msvcrt except ImportError: UpperCAmelCase : List[Any] = None try: import fcntl except ImportError: UpperCAmelCase : int = None # Backward compatibility # ------------------------------------------------ try: TimeoutError except NameError: UpperCAmelCase : Union[str, Any] = OSError # Data # ------------------------------------------------ UpperCAmelCase : List[Any] = [ 'Timeout', 'BaseFileLock', 'WindowsFileLock', 'UnixFileLock', 'SoftFileLock', 'FileLock', ] UpperCAmelCase : Tuple = '3.0.12' UpperCAmelCase : str = None def a__ ( ): """simple docstring""" global _logger __SCREAMING_SNAKE_CASE = _logger or logging.getLogger(__name__ ) return _logger class lowerCAmelCase__ ( a ): """simple docstring""" def __init__( self : Any , __SCREAMING_SNAKE_CASE : Optional[Any] ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = lock_file return None def __str__( self : str ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = f'The file lock \'{self.lock_file}\' could not be acquired.' return temp class lowerCAmelCase__ : """simple docstring""" def __init__( self : Tuple , __SCREAMING_SNAKE_CASE : List[Any] ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = lock return None def __enter__( self : List[str] ) -> List[Any]: """simple docstring""" return self.lock def __exit__( self : Tuple , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : Dict , __SCREAMING_SNAKE_CASE : Optional[Any] ) -> List[str]: """simple docstring""" self.lock.release() return None class lowerCAmelCase__ : """simple docstring""" def __init__( self : List[Any] , __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : str=-1 , __SCREAMING_SNAKE_CASE : Optional[Any]=None ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE = max_filename_length if max_filename_length is not None else 255 # Hash the filename if it's too long __SCREAMING_SNAKE_CASE = self.hash_filename_if_too_long(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # The path to the lock file. __SCREAMING_SNAKE_CASE = lock_file # The file descriptor for the *_lock_file* as it is returned by the # os.open() function. # This file lock is only NOT None, if the object currently holds the # lock. __SCREAMING_SNAKE_CASE = None # The default timeout value. __SCREAMING_SNAKE_CASE = timeout # We use this lock primarily for the lock counter. __SCREAMING_SNAKE_CASE = threading.Lock() # The lock counter is used for implementing the nested locking # mechanism. Whenever the lock is acquired, the counter is increased and # the lock is only released, when this value is 0 again. __SCREAMING_SNAKE_CASE = 0 return None @property def UpperCAmelCase__ ( self : List[Any] ) -> List[str]: """simple docstring""" return self._lock_file @property def UpperCAmelCase__ ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" return self._timeout @timeout.setter def UpperCAmelCase__ ( self : List[str] , __SCREAMING_SNAKE_CASE : Dict ) -> Optional[int]: """simple docstring""" __SCREAMING_SNAKE_CASE = float(__SCREAMING_SNAKE_CASE ) return None def UpperCAmelCase__ ( self : Optional[Any] ) -> List[Any]: """simple docstring""" raise NotImplementedError() def UpperCAmelCase__ ( self : Union[str, Any] ) -> Any: """simple docstring""" raise NotImplementedError() @property def UpperCAmelCase__ ( self : Union[str, Any] ) -> int: """simple docstring""" return self._lock_file_fd is not None def UpperCAmelCase__ ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Optional[Any]=None , __SCREAMING_SNAKE_CASE : Optional[int]=0.05 ) -> Optional[Any]: """simple docstring""" if timeout is None: __SCREAMING_SNAKE_CASE = self.timeout # Increment the number right at the beginning. # We can still undo it, if something fails. with self._thread_lock: self._lock_counter += 1 __SCREAMING_SNAKE_CASE = id(self ) __SCREAMING_SNAKE_CASE = self._lock_file __SCREAMING_SNAKE_CASE = time.time() try: while True: with self._thread_lock: if not self.is_locked: logger().debug(f'Attempting to acquire lock {lock_id} on {lock_filename}' ) self._acquire() if self.is_locked: logger().debug(f'Lock {lock_id} acquired on {lock_filename}' ) break elif timeout >= 0 and time.time() - start_time > timeout: logger().debug(f'Timeout on acquiring lock {lock_id} on {lock_filename}' ) raise Timeout(self._lock_file ) else: logger().debug( f'Lock {lock_id} not acquired on {lock_filename}, waiting {poll_intervall} seconds ...' ) time.sleep(__SCREAMING_SNAKE_CASE ) except: # noqa # Something did go wrong, so decrement the counter. with self._thread_lock: __SCREAMING_SNAKE_CASE = max(0 , self._lock_counter - 1 ) raise return _Acquire_ReturnProxy(lock=self ) def UpperCAmelCase__ ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Optional[Any]=False ) -> Dict: """simple docstring""" with self._thread_lock: if self.is_locked: self._lock_counter -= 1 if self._lock_counter == 0 or force: __SCREAMING_SNAKE_CASE = id(self ) __SCREAMING_SNAKE_CASE = self._lock_file logger().debug(f'Attempting to release lock {lock_id} on {lock_filename}' ) self._release() __SCREAMING_SNAKE_CASE = 0 logger().debug(f'Lock {lock_id} released on {lock_filename}' ) return None def __enter__( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" self.acquire() return self def __exit__( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Any ) -> Tuple: """simple docstring""" self.release() return None def __del__( self : str ) -> Union[str, Any]: """simple docstring""" self.release(force=__SCREAMING_SNAKE_CASE ) return None def UpperCAmelCase__ ( self : Optional[int] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : int ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE = os.path.basename(__SCREAMING_SNAKE_CASE ) if len(__SCREAMING_SNAKE_CASE ) > max_length and max_length > 0: __SCREAMING_SNAKE_CASE = os.path.dirname(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = str(hash(__SCREAMING_SNAKE_CASE ) ) __SCREAMING_SNAKE_CASE = filename[: max_length - len(__SCREAMING_SNAKE_CASE ) - 8] + """...""" + hashed_filename + """.lock""" return os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) else: return path class lowerCAmelCase__ ( a ): """simple docstring""" def __init__( self : Tuple , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : Dict=-1 , __SCREAMING_SNAKE_CASE : Dict=None ) -> List[Any]: """simple docstring""" from .file_utils import relative_to_absolute_path super().__init__(__SCREAMING_SNAKE_CASE , timeout=__SCREAMING_SNAKE_CASE , max_filename_length=__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = """\\\\?\\""" + relative_to_absolute_path(self.lock_file ) def UpperCAmelCase__ ( self : Dict ) -> Optional[int]: """simple docstring""" __SCREAMING_SNAKE_CASE = os.O_RDWR | os.O_CREAT | os.O_TRUNC try: __SCREAMING_SNAKE_CASE = os.open(self._lock_file , __SCREAMING_SNAKE_CASE ) except OSError: pass else: try: msvcrt.locking(__SCREAMING_SNAKE_CASE , msvcrt.LK_NBLCK , 1 ) except OSError: os.close(__SCREAMING_SNAKE_CASE ) else: __SCREAMING_SNAKE_CASE = fd return None def UpperCAmelCase__ ( self : Tuple ) -> List[str]: """simple docstring""" __SCREAMING_SNAKE_CASE = self._lock_file_fd __SCREAMING_SNAKE_CASE = None msvcrt.locking(__SCREAMING_SNAKE_CASE , msvcrt.LK_UNLCK , 1 ) os.close(__SCREAMING_SNAKE_CASE ) try: os.remove(self._lock_file ) # Probably another instance of the application # that acquired the file lock. except OSError: pass return None class lowerCAmelCase__ ( a ): """simple docstring""" def __init__( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : Union[str, Any]=-1 , __SCREAMING_SNAKE_CASE : Optional[Any]=None ) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE = os.statvfs(os.path.dirname(__SCREAMING_SNAKE_CASE ) ).f_namemax super().__init__(__SCREAMING_SNAKE_CASE , timeout=__SCREAMING_SNAKE_CASE , max_filename_length=__SCREAMING_SNAKE_CASE ) def UpperCAmelCase__ ( self : Optional[int] ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = os.O_RDWR | os.O_CREAT | os.O_TRUNC __SCREAMING_SNAKE_CASE = os.open(self._lock_file , __SCREAMING_SNAKE_CASE ) try: fcntl.flock(__SCREAMING_SNAKE_CASE , fcntl.LOCK_EX | fcntl.LOCK_NB ) except OSError: os.close(__SCREAMING_SNAKE_CASE ) else: __SCREAMING_SNAKE_CASE = fd return None def UpperCAmelCase__ ( self : List[Any] ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE = self._lock_file_fd __SCREAMING_SNAKE_CASE = None fcntl.flock(__SCREAMING_SNAKE_CASE , fcntl.LOCK_UN ) os.close(__SCREAMING_SNAKE_CASE ) return None class lowerCAmelCase__ ( a ): """simple docstring""" def UpperCAmelCase__ ( self : List[str] ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = os.O_WRONLY | os.O_CREAT | os.O_EXCL | os.O_TRUNC try: __SCREAMING_SNAKE_CASE = os.open(self._lock_file , __SCREAMING_SNAKE_CASE ) except OSError: pass else: __SCREAMING_SNAKE_CASE = fd return None def UpperCAmelCase__ ( self : int ) -> Optional[int]: """simple docstring""" os.close(self._lock_file_fd ) __SCREAMING_SNAKE_CASE = None try: os.remove(self._lock_file ) # The file is already deleted and that's what we want. except OSError: pass return None UpperCAmelCase : Dict = None if msvcrt: UpperCAmelCase : Optional[int] = WindowsFileLock elif fcntl: UpperCAmelCase : Optional[Any] = UnixFileLock else: UpperCAmelCase : int = SoftFileLock if warnings is not None: warnings.warn('only soft file lock is available')
267
1
'''simple docstring''' from collections import defaultdict from math import gcd def snake_case_ (UpperCamelCase : int = 150_0000 ): '''simple docstring''' _a = defaultdict(UpperCamelCase ) _a = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1 , UpperCamelCase , 2 ): if gcd(UpperCamelCase , UpperCamelCase ) > 1: continue _a = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(UpperCamelCase , limit + 1 , UpperCamelCase ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F'''{solution() = }''')
179
'''simple docstring''' from __future__ import annotations def snake_case_ (UpperCamelCase : list[int] ): '''simple docstring''' if not nums: return 0 _a = nums[0] _a = 0 for num in nums[1:]: _a , _a = ( max_excluding + num, max(UpperCamelCase , UpperCamelCase ), ) return max(UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
179
1
from binascii import hexlify from hashlib import shaaaa from os import urandom # RFC 3526 - More Modular Exponential (MODP) Diffie-Hellman groups for # Internet Key Exchange (IKE) https://tools.ietf.org/html/rfc3526 snake_case__ : Tuple = { # 1536-bit 5: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA237327FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 2048-bit 14: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AACAA68FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 3072-bit 15: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A93AD2CAFFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 4096-bit 16: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7' + '88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA' + '2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6' + '287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED' + '1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9' + '93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934063199' + 'FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 6144-bit 17: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD129024E08' + '8A67CC74020BBEA63B139B22514A08798E3404DDEF9519B3CD3A431B' + '302B0A6DF25F14374FE1356D6D51C245E485B576625E7EC6F44C42E9' + 'A637ED6B0BFF5CB6F406B7EDEE386BFB5A899FA5AE9F24117C4B1FE6' + '49286651ECE45B3DC2007CB8A163BF0598DA48361C55D39A69163FA8' + 'FD24CF5F83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3BE39E772C' + '180E86039B2783A2EC07A28FB5C55DF06F4C52C9DE2BCBF695581718' + '3995497CEA956AE515D2261898FA051015728E5A8AAAC42DAD33170D' + '04507A33A85521ABDF1CBA64ECFB850458DBEF0A8AEA71575D060C7D' + 'B3970F85A6E1E4C7ABF5AE8CDB0933D71E8C94E04A25619DCEE3D226' + '1AD2EE6BF12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB3143DB5BFC' + 'E0FD108E4B82D120A92108011A723C12A787E6D788719A10BDBA5B26' + '99C327186AF4E23C1A946834B6150BDA2583E9CA2AD44CE8DBBBC2DB' + '04DE8EF92E8EFC141FBECAA6287C59474E6BC05D99B2964FA090C3A2' + '233BA186515BE7ED1F612970CEE2D7AFB81BDD762170481CD0069127' + 'D5B05AA993B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492' + '36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BDF8FF9406' + 'AD9E530EE5DB382F413001AEB06A53ED9027D831179727B0865A8918' + 'DA3EDBEBCF9B14ED44CE6CBACED4BB1BDB7F1447E6CC254B33205151' + '2BD7AF426FB8F401378CD2BF5983CA01C64B92ECF032EA15D1721D03' + 'F482D7CE6E74FEF6D55E702F46980C82B5A84031900B1C9E59E7C97F' + 'BEC7E8F323A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA' + 'CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE32806A1D58B' + 'B7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55CDA56C9EC2EF29632' + '387FE8D76E3C0468043E8F663F4860EE12BF2D5B0B7474D6E694F91E' + '6DCC4024FFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, # 8192-bit 18: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7' + '88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA' + '2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6' + '287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED' + '1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9' + '93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492' + '36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BD' + 'F8FF9406AD9E530EE5DB382F413001AEB06A53ED9027D831' + '179727B0865A8918DA3EDBEBCF9B14ED44CE6CBACED4BB1B' + 'DB7F1447E6CC254B332051512BD7AF426FB8F401378CD2BF' + '5983CA01C64B92ECF032EA15D1721D03F482D7CE6E74FEF6' + 'D55E702F46980C82B5A84031900B1C9E59E7C97FBEC7E8F3' + '23A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA' + 'CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE328' + '06A1D58BB7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55C' + 'DA56C9EC2EF29632387FE8D76E3C0468043E8F663F4860EE' + '12BF2D5B0B7474D6E694F91E6DBE115974A3926F12FEE5E4' + '38777CB6A932DF8CD8BEC4D073B931BA3BC832B68D9DD300' + '741FA7BF8AFC47ED2576F6936BA424663AAB639C5AE4F568' + '3423B4742BF1C978238F16CBE39D652DE3FDB8BEFC848AD9' + '22222E04A4037C0713EB57A81A23F0C73473FC646CEA306B' + '4BCBC8862F8385DDFA9D4B7FA2C087E879683303ED5BDD3A' + '062B3CF5B3A278A66D2A13F83F44F82DDF310EE074AB6A36' + '4597E899A0255DC164F31CC50846851DF9AB48195DED7EA1' + 'B1D510BD7EE74D73FAF36BC31ECFA268359046F4EB879F92' + '4009438B481C6CD7889A002ED5EE382BC9190DA6FC026E47' + '9558E4475677E9AA9E3050E2765694DFC81F56E880B96E71' + '60C980DD98EDD3DFFFFFFFFFFFFFFFFF', base=16, ), 'generator': 2, }, } class A_ : def __init__(self :Dict , _UpperCamelCase :int = 14 )-> None: if group not in primes: raise ValueError('''Unsupported Group''' ) __A = primes[group]['''prime'''] __A = primes[group]['''generator'''] __A = int(hexlify(urandom(32 ) ) , base=16 ) def _lowerCAmelCase (self :Optional[int] )-> str: return hex(self.__private_key )[2:] def _lowerCAmelCase (self :Tuple )-> str: __A = pow(self.generator , self.__private_key , self.prime ) return hex(_UpperCamelCase )[2:] def _lowerCAmelCase (self :str , _UpperCamelCase :int )-> bool: # check if the other public key is valid based on NIST SP800-56 return ( 2 <= key <= self.prime - 2 and pow(_UpperCamelCase , (self.prime - 1) // 2 , self.prime ) == 1 ) def _lowerCAmelCase (self :str , _UpperCamelCase :str )-> str: __A = int(_UpperCamelCase , base=16 ) if not self.is_valid_public_key(_UpperCamelCase ): raise ValueError('''Invalid public key''' ) __A = pow(_UpperCamelCase , self.__private_key , self.prime ) return shaaaa(str(_UpperCamelCase ).encode() ).hexdigest() @staticmethod def _lowerCAmelCase (_UpperCamelCase :int , _UpperCamelCase :int )-> bool: # check if the other public key is valid based on NIST SP800-56 return ( 2 <= remote_public_key_str <= prime - 2 and pow(_UpperCamelCase , (prime - 1) // 2 , _UpperCamelCase ) == 1 ) @staticmethod def _lowerCAmelCase (_UpperCamelCase :str , _UpperCamelCase :str , _UpperCamelCase :int = 14 )-> str: __A = int(_UpperCamelCase , base=16 ) __A = int(_UpperCamelCase , base=16 ) __A = primes[group]['''prime'''] if not DiffieHellman.is_valid_public_key_static(_UpperCamelCase , _UpperCamelCase ): raise ValueError('''Invalid public key''' ) __A = pow(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) return shaaaa(str(_UpperCamelCase ).encode() ).hexdigest() if __name__ == "__main__": import doctest doctest.testmod()
117
def _a ( lowerCamelCase: dict ) -> bool: '''simple docstring''' __A = set() # To detect a back edge, keep track of vertices currently in the recursion stack __A = set() return any( node not in visited and depth_first_search(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) for node in graph ) def _a ( lowerCamelCase: dict , lowerCamelCase: int , lowerCamelCase: set , lowerCamelCase: set ) -> bool: '''simple docstring''' visited.add(lowerCamelCase ) rec_stk.add(lowerCamelCase ) for node in graph[vertex]: if node not in visited: if depth_first_search(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(lowerCamelCase ) return False if __name__ == "__main__": from doctest import testmod testmod()
117
1
import torch from transformers import CamembertForMaskedLM, CamembertTokenizer def __magic_name__ ( __lowerCAmelCase : List[Any] , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : str , __lowerCAmelCase : int=5 ) -> Tuple: # Adapted from https://github.com/pytorch/fairseq/blob/master/fairseq/models/roberta/hub_interface.py assert masked_input.count('''<mask>''' ) == 1 __lowerCamelCase = torch.tensor(tokenizer.encode(__lowerCAmelCase , add_special_tokens=__lowerCAmelCase ) ).unsqueeze(0 ) # Batch size 1 __lowerCamelCase = model(__lowerCAmelCase )[0] # The last hidden-state is the first element of the output tuple __lowerCamelCase = (input_ids.squeeze() == tokenizer.mask_token_id).nonzero().item() __lowerCamelCase = logits[0, masked_index, :] __lowerCamelCase = logits.softmax(dim=0 ) __lowerCamelCase , __lowerCamelCase = prob.topk(k=__lowerCAmelCase , dim=0 ) __lowerCamelCase = ''' '''.join( [tokenizer.convert_ids_to_tokens(indices[i].item() ) for i in range(len(__lowerCAmelCase ) )] ) __lowerCamelCase = tokenizer.mask_token __lowerCamelCase = [] for index, predicted_token_bpe in enumerate(topk_predicted_token_bpe.split(''' ''' ) ): __lowerCamelCase = predicted_token_bpe.replace('''\u2581''' , ''' ''' ) if " {0}".format(__lowerCAmelCase ) in masked_input: topk_filled_outputs.append( ( masked_input.replace(''' {0}'''.format(__lowerCAmelCase ) , __lowerCAmelCase ), values[index].item(), predicted_token, ) ) else: topk_filled_outputs.append( ( masked_input.replace(__lowerCAmelCase , __lowerCAmelCase ), values[index].item(), predicted_token, ) ) return topk_filled_outputs SCREAMING_SNAKE_CASE__ : Tuple = CamembertTokenizer.from_pretrained("camembert-base") SCREAMING_SNAKE_CASE__ : Union[str, Any] = CamembertForMaskedLM.from_pretrained("camembert-base") model.eval() SCREAMING_SNAKE_CASE__ : Any = "Le camembert est <mask> :)" print(fill_mask(masked_input, model, tokenizer, topk=3))
339
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE__ : List[Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : str = { "facebook/xmod-base": "https://huggingface.co/facebook/xmod-base/resolve/main/config.json", "facebook/xmod-large-prenorm": "https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json", "facebook/xmod-base-13-125k": "https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json", "facebook/xmod-base-30-125k": "https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json", "facebook/xmod-base-30-195k": "https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json", "facebook/xmod-base-60-125k": "https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json", "facebook/xmod-base-60-265k": "https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json", "facebook/xmod-base-75-125k": "https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json", "facebook/xmod-base-75-269k": "https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json", } class lowerCAmelCase__ ( __lowercase ): a__ : Dict = """xmod""" def __init__( self : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any]=3_05_22 , SCREAMING_SNAKE_CASE__ : str=7_68 , SCREAMING_SNAKE_CASE__ : int=12 , SCREAMING_SNAKE_CASE__ : Dict=12 , SCREAMING_SNAKE_CASE__ : List[str]=30_72 , SCREAMING_SNAKE_CASE__ : List[Any]="gelu" , SCREAMING_SNAKE_CASE__ : Dict=0.1 , SCREAMING_SNAKE_CASE__ : int=0.1 , SCREAMING_SNAKE_CASE__ : List[str]=5_12 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=2 , SCREAMING_SNAKE_CASE__ : List[Any]=0.02 , SCREAMING_SNAKE_CASE__ : Optional[Any]=1e-12 , SCREAMING_SNAKE_CASE__ : List[str]=1 , SCREAMING_SNAKE_CASE__ : Union[str, Any]=0 , SCREAMING_SNAKE_CASE__ : int=2 , SCREAMING_SNAKE_CASE__ : Any="absolute" , SCREAMING_SNAKE_CASE__ : Optional[Any]=True , SCREAMING_SNAKE_CASE__ : Union[str, Any]=None , SCREAMING_SNAKE_CASE__ : Optional[Any]=False , SCREAMING_SNAKE_CASE__ : Tuple=2 , SCREAMING_SNAKE_CASE__ : Optional[Any]=False , SCREAMING_SNAKE_CASE__ : Optional[int]=True , SCREAMING_SNAKE_CASE__ : Optional[Any]=True , SCREAMING_SNAKE_CASE__ : Dict=("en_XX",) , SCREAMING_SNAKE_CASE__ : Optional[Any]=None , **SCREAMING_SNAKE_CASE__ : int , ) -> str: super().__init__(pad_token_id=SCREAMING_SNAKE_CASE__ , bos_token_id=SCREAMING_SNAKE_CASE__ , eos_token_id=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = vocab_size __lowerCamelCase = hidden_size __lowerCamelCase = num_hidden_layers __lowerCamelCase = num_attention_heads __lowerCamelCase = hidden_act __lowerCamelCase = intermediate_size __lowerCamelCase = hidden_dropout_prob __lowerCamelCase = attention_probs_dropout_prob __lowerCamelCase = max_position_embeddings __lowerCamelCase = type_vocab_size __lowerCamelCase = initializer_range __lowerCamelCase = layer_norm_eps __lowerCamelCase = position_embedding_type __lowerCamelCase = use_cache __lowerCamelCase = classifier_dropout __lowerCamelCase = pre_norm __lowerCamelCase = adapter_reduction_factor __lowerCamelCase = adapter_layer_norm __lowerCamelCase = adapter_reuse_layer_norm __lowerCamelCase = ln_before_adapter __lowerCamelCase = list(SCREAMING_SNAKE_CASE__ ) __lowerCamelCase = default_language class lowerCAmelCase__ ( __lowercase ): @property def __A ( self : Optional[Any] ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": __lowerCamelCase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: __lowerCamelCase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
339
1
'''simple docstring''' import requests def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ) -> None: __lowerCamelCase = {'''Content-Type''': '''application/json'''} __lowerCamelCase = requests.post(UpperCamelCase__ , json={'''text''': message_body} , headers=UpperCamelCase__ ) if response.status_code != 2_00: __lowerCamelCase = ( '''Request to slack returned an error ''' f"""{response.status_code}, the response is:\n{response.text}""" ) raise ValueError(UpperCamelCase__ ) if __name__ == "__main__": # Set the slack url to the one provided by Slack when you create the webhook at # https://my.slack.com/services/new/incoming-webhook/ send_slack_message("<YOUR MESSAGE BODY>", "<SLACK CHANNEL URL>")
67
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { "google/vivit-b-16x2-kinetics400": ( "https://huggingface.co/google/vivit-b-16x2-kinetics400/resolve/main/config.json" ), # See all Vivit models at https://huggingface.co/models?filter=vivit } class _SCREAMING_SNAKE_CASE( A ): SCREAMING_SNAKE_CASE_ : List[str] = '''vivit''' def __init__( self ,SCREAMING_SNAKE_CASE__=2_24 ,SCREAMING_SNAKE_CASE__=32 ,SCREAMING_SNAKE_CASE__=[2, 16, 16] ,SCREAMING_SNAKE_CASE__=3 ,SCREAMING_SNAKE_CASE__=7_68 ,SCREAMING_SNAKE_CASE__=12 ,SCREAMING_SNAKE_CASE__=12 ,SCREAMING_SNAKE_CASE__=30_72 ,SCREAMING_SNAKE_CASE__="gelu_fast" ,SCREAMING_SNAKE_CASE__=0.0 ,SCREAMING_SNAKE_CASE__=0.0 ,SCREAMING_SNAKE_CASE__=0.0_2 ,SCREAMING_SNAKE_CASE__=1E-06 ,SCREAMING_SNAKE_CASE__=True ,**SCREAMING_SNAKE_CASE__ ,) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE :int = hidden_size __SCREAMING_SNAKE_CASE :List[Any] = num_hidden_layers __SCREAMING_SNAKE_CASE :Union[str, Any] = num_attention_heads __SCREAMING_SNAKE_CASE :Union[str, Any] = intermediate_size __SCREAMING_SNAKE_CASE :Any = hidden_act __SCREAMING_SNAKE_CASE :Optional[Any] = hidden_dropout_prob __SCREAMING_SNAKE_CASE :str = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE :Any = initializer_range __SCREAMING_SNAKE_CASE :Optional[int] = layer_norm_eps __SCREAMING_SNAKE_CASE :Optional[int] = image_size __SCREAMING_SNAKE_CASE :List[str] = num_frames __SCREAMING_SNAKE_CASE :Any = tubelet_size __SCREAMING_SNAKE_CASE :str = num_channels __SCREAMING_SNAKE_CASE :Any = qkv_bias super().__init__(**SCREAMING_SNAKE_CASE__ )
191
0
"""simple docstring""" UpperCamelCase_ ="""2.13.1""" import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse("""3.7"""): raise ImportWarning( """To use `datasets`, Python>=3.7 is required, and the current version of Python doesn't match this condition.""" ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( """To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn't match this condition.\n""" """If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.""" ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip UpperCamelCase_ =concatenate_datasets UpperCamelCase_ =DownloadConfig UpperCamelCase_ =DownloadManager UpperCamelCase_ =DownloadMode UpperCamelCase_ =DownloadConfig UpperCamelCase_ =DownloadMode UpperCamelCase_ =DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
365
"""simple docstring""" UpperCamelCase_ =[ """DownloadConfig""", """DownloadManager""", """DownloadMode""", """StreamingDownloadManager""", ] from .download_config import DownloadConfig from .download_manager import DownloadManager, DownloadMode from .streaming_download_manager import StreamingDownloadManager
128
0